4 5900066337668990814
4 5900066337668990814
4 5900066337668990814
3
Copyright © 2016 by
All rights reserved. No part of this work may be reproduced in any form except by
Printed in USA
ISBN: 978-1-62595-049-9
First Edition
4
CONTENTS
Foreword
Preface
About the Author
Introduction
About the ARRL
Part 1 Antenna Physics
1 Development of Antenna Physics
Early History
Modern History
Maxwell, the Turning Point
1901 to Present
Enter the Antenna
2 Fundamentals
What is an Antenna?
The Sphere as a Measurement Tool
Analogy Between Light and Radio Waves and
the Concept of Aperture
Defining the Critical Fundamental Terms
The 1⁄2 Wavelength Dipole Antenna
Relationship Between Aperture and Gain
Empirical Calculation of the Aperture of an
Isotropic Antenna
Free Space Path Loss
Refinements of the Aperture Concept
5
3 Radiation of Radio Waves
Modern Expression of Maxwell’s Equations
Maxwell’s First Equation, Gauss’s Law for
Electricity
Maxwell’s Second Equation, Gauss’s Law for
Magnetism
Maxwell’s Third Equation, Faraday’s Law of
Induction
Ampere’s Law
Maxwell’s Fourth Equation
The Wave Equation
Wave Equation for Electromagnetic Waves
Relationships Between the E and B fields
Impedance of Free Space
Fields, Power, and Energy
Radiation Resistance
Relationship Between Radiation Resistance
and Feed Point Impedance in Linear Antenna
Structures
Antenna Pattern Formation
PADL
Current and Voltage Distribution
Antenna Polarization
4 Transmission Lines
The Guided Wave
Radiation Loss Rr
Characteristic Impedance
6
Characteristic Impedance of Two-Wire
Transmission Line
Characteristic Impedance of Coaxial Cable
Velocity Factor
Types of Guided Waves
Transverse Modes
Losses in Transmission Lines
Reflected Waves
VSWR and Standing Waves
5 Antennas Using Multiple Sources
Driven Arrays
Mutual Impedance
Quadrature Array
Parasitic Arrays
Part 2: Applied Antenna Physics
6 Dielectric Effects Upon Radio Waves
Reflection and Absorption
Conductors and Dielectrics
Wave Propagation in Dielectrics
Snell’s Law
The Earth as a Dielectric
Effect on Wave Polarization
Horizontal Polarized Wave Reflection
Vertical Polarized Wave Reflection
Assumption of an Infinite Plane
7 Vertical Antennas
7
Ground Mounted Vertical Antennas
Radiation Resistance of a Ground-Mounted
Base-Fed Vertical Antenna
Effective Height
Example of the Effects of he on Rr
Directivity of a Ground-Mounted Vertical
Antenna
Ground Influences on Ground-Mounted
Verticals
Experimental Method to Derive Rl
Feed Point R vs Maximum Rr
Verticals Higher than 1⁄4 l
Raised Vertical Antennas
Raised Vertical Antennas Over Earth
Short Vertical Antennas
Data for Various Ground-Mounted Unloaded
Vertical Physical Heights
Ground Proximity to Current Maximums
The Near-Ground Inverted-V Antenna
Near-Ground Phased Inverted-Vs
Broadside Bobtail Curtain
Single-Element Vertical (Monopole Antenna)
Multielement Vertical Arrays
Gain Comparison of Vertical Arrays
Conclusion
EIRP Calculation for Ground-Mounted Low
Frequency Vertical Antennas
8
Ground Losses
Vertical Element Losses
Matching Losses
Summing the Losses and Determining
Required Transmitter Power Output
8 Yagi-Uda and Cubical Quad Antennas
Gain and Aperture of a Yagi-Uda Antenna
Stacking Yagi-Uda Antennas
Design and Optimization of a Yagi-Uda Array
The W8JK Array
Multiband Yagi-Uda Arrays
The Cubical Quad
Quad Parasitic Elements in Vertical Arrays
9 Specialized Antenna Configurations
Small Loop Antennas
Larger Small Loops
Helix Antennas
Small Helix Antenna
Aperture Antennas
Small Antennas
Traveling Wave Antennas
Antennas for Navigation and Location
Determination
Patch Antennas
Phased Array
Aperture Distribution
Polarization Diversity
9
MIMO Antennas
Other Configurations
10 Noise, Temperature, and Signals
Boltzmann’s and Plank’s Constants
Boltzmann Noise in Circuits
Random Processes
Radio Receiver Sensitivity
Calculating a Minimum Antenna Gain
Specification
Noise Temperature
Antenna Temperature
A Nobel Prize
Atmospheric and Galactic Noise
Photons in Radio Waves
Atmospheric Noise Considerations at MF and
Low HF
Receive Antenna Parameters for MF and Low
HF
Shannon’s Limit
Extreme System Design
Physics Defines the Ultimate Limits to Radio
Communication
10
Appendix D — Symbols and Abbreviations
Appendix E — Comparison of Elevation Plots of
Vertical-Polarized Ground-Mounted 160 Meter
Antenna Arrays
11
FOREWORD
It’s probably fair to say that the topic of
antennas has generated more interest and
discussion among radio amateurs than any other.
Antenna installation or construction is often the
first do-it-yourself project an amateur
undertakes, and most of us have tried several —
or many! — different versions during our
amateur careers. Countless books and articles
have been written describing basic antenna
theory and “how-to” construction details, and
this is the largest category in the ARRL
Bookstore.
Antenna Physics: An Introduction is not another
one of those books. In this work, Bob Zavrel,
W7SX, strives to bridge the gap between basic
theory and graduate-level engineering texts. He
explains many of the underlying principles of
antennas and antenna physics and introduces the
reader to the mathematics behind these
principles. Although some competence in
mathematics is required, readers may follow
along and understand the concepts without
needing to solve the complex equations
presented. In later chapters, in the Applied
Antenna Physics section, Bob presents examples
12
to tie the concepts learned in earlier chapters to a
number of antenna types familiar to radio
amateurs.
We hope that you enjoy this approach and gain
a deeper understanding of how antennas work. It
might even inspire you to seek out some of the
more advanced texts listed in the References
appendix to delve deeper into this fascinating
and important subject.
13
PREFACE
Why Write a Book About “Antenna Physics”?
After nearly 50 years of studying the subject
and designing, building, and using antennas for
both amateur and professional applications, I had
come to a conclusion. There are many very fine
books available on this fascinating subject, and
they usually fall within one of two categories: (1)
relatively simple treatments that provide “how-
to” material as well as basic theory to support
some level of fundamental understanding and (2)
highly advanced theoretical graduate-level
engineering texts.
There appeared to be a gap between these two
extremes. Thanks to the well-written and
comprehensive “how-to” publications of the
American Radio Relay League, I was able to
master much of the basic theory of antennas by
the time I graduated from high school. These
books provide enough theory to gain a sufficient
background to design, build, measure, and use
antennas. However these texts only instilled a
craving for deeper understanding of the
underlying physics of antennas. The leap from
these introductory texts to the masterpieces of
Kraus, Jasik, Balanis and others necessarily had
14
to wait for competence in advanced
mathematics, including the indispensable
calculus. After many years of using these books
as references and thus providing ever deeper
insights into the workings of antennas, I
concluded that many of the more advanced
treatments could be explained, at least in part,
using a simpler approach. This book is an
attempt, therefore, to “bridge the gap” between
the introductory and “how to” books and the
advanced texts. I hope some readers will use this
introductory text as a springboard to the more
advanced treatments of this fascinating subject,
indispensable to all forms of radio applications.
Although many of the antenna applications will
be most useful to Amateur Radio operators,
about two-thirds of the text should also find
more widespread use. I hope that it will serve
electrical engineering students and professional
radio engineers as well as the amateur
community.
Advanced mathematics is an essential tool for
solving complex antenna problems. However, it
is not necessary for understanding the deeper
principles. This is one of the key motivations and
challenges to writing this book: you can
understand the physics without being able to
solve a calculus problem, but it is of great
15
advantage to understand what the equations are
saying. Indeed, many of the complex-looking
equations that are found in this text are actually
only special forms of much simpler equations
which are second-nature to anyone with a
competence in electronics. In this text I often
point out the simpler forms as the calculus is
presented. For example, many of the complex
equations are merely special forms of well-
understood relationships like Ohm’s Law or the
power law.
Some competence in mathematics is required
for this text, particularly algebra. Also, being able
to deal with some trigonometry as well as
logarithms is essential to grasping the
fundamental concepts. When more advanced
equations are presented, explanations are often
offered to explain what the reader is looking at,
and what it means, particularly for equations
involving calculus and linear algebra. Of course a
mastery of calculus and other forms of advanced
mathematics will afford an even deeper insight
when combined with dedicated study of antenna
physics proper.
Physics is classified as one of the natural
sciences, together with mathematics, chemistry,
biology, and geology. Students of these
disciplines (purists) are only interested in
16
describing nature. In contrast, the engineering
sciences, concern themselves with designing and
building something. The engineering sciences are
completely dependent upon the natural sciences
for the laws and theories needed to do or build
something useful. On the other hand,
engineering that utilizes theories based in natural
science provides excellent vindication of the laws
and theories rooted in natural science. The
business of natural science is to provide
dependable, repeatable experiments with similar
results. Engineers need repeatable results to
design and build something that will be
manufactured. For example, the esoteric
theoretical subject of sub-atomic physics was
validated with the first atomic explosion. For the
pragmatic person, this is the ultimate test of a
“theory”: what are the useful results? Or, what
are the possible useful results armed with this
new theory? What can we do with this new
knowledge?
In this book we will explore how a
mathematician (Maxwell) first predicted that
radio waves exist. Then a physicist (Hertz)
actually built an experimental apparatus to prove
radio waves exist, and finally an engineer
(Marconi) actually developed radio as something
useful. Of course it wasn’t that straightforward.
17
Thousands of scientists and engineers have made
contributions, large and small to develop
“wireless” as we know it today, but the sequence
of theory, proof, and finally “let’s do something
with this” is clear. Therefore, an antenna
designer, builder, or even just a user can become
more capable if the basic underlying principles
(the physics) are understood.
After a lifetime of learning, experimenting,
mentoring and teaching antenna theory, I found
several topics to be commonly misunderstood.
Having to repeat the same ideas multiple times
over the years, extra attention is given to topics
that reflect these misunderstandings among not
only amateurs, but also professionals. These
topics, understandably are the more complex and
less intuitive. Some repetition of the basic “how
to” books is essential in writing such a text.
However, I attempted to minimize repetition
assuming the reader already understands these
basics or has access to basic texts such as the
ARRL Antenna Book.
“Applied” examples were chosen to gain
insights into common antenna types, which can
then lead to aids in engineering antennas. Also,
many of these applied examples, indeed
chapters, reflect my serious interest in Amateur
Radio and the publisher, the American Radio
18
Relay League (ARRL). Amateur Radio has been an
avocation since my childhood, and participation
in this wonderful hobby led me ultimately into a
lifelong career in electronic engineering with a
specialty in RF. The publications of the ARRL
have been and remain indispensable for the radio
amateur and many engineers and technicians
alike. In this vein I hope that this text provides
value to amateurs, students, and professionals.
Preface to the Text
The organization of this text is different from
both the more traditional introductory and
advanced texts. The beginning chapters weave
actual electricity and magnetism (E&M) material,
such as Maxwell’s equations taken from pure
classical physics, with how these concepts relate
directly to antennas. The reasoning is simple. In
order to develop a deeper understanding of
antennas, it is necessary to gain a deeper
understanding of the fields with which they
interplay and the characteristics of free space, the
medium in which they perform.
Many of the terms in common use regarding
antennas are redefined in relation to terms
typically quantified in the more advanced texts
but ignored or only briefly encountered in the
basic texts. Hopefully these discussions will shed
19
new light and deeper insights on these basic
terms used in discussions on antennas. This is
the essence of the first three chapters.
It is common practice for authors to simply
reference equations already presented in previous
chapters. I have always found this practice
awkward, having often to thumb back through
the chapters to find the applicable equation. I
would prefer readers to find the text a bit
repetitive rather than assuming equations and
other points made earlier are now embedded in
the reader’s memory. Also this book is written
somewhat in a “narrative” form. The intent is for
chapters to flow from one to the next. Hopefully,
however, the chapters will also stand by
themselves as references to the applicable topics.
Many important topics are presented in an
abbreviated form, especially when these topics
are very well covered by introductory and/or
specialized texts commonly available.
Part 1: Antenna Physics
Chapter 1: Development of Antenna Physics:
We first present an introduction of the
theoretical development of radio and then show
how antenna theory was co-developed as an
essential element to “radio.” Taken in a historical
context, it also offers an acknowledgement of
20
some of the great indispensable contributors to
the art.
Chapter 2: Fundamentals: Building upon the
introductions in Chapter 1, we begin by
introducing some of the basic physical
parameters of antennas. The approach quickly
diverges from the “how- to” books and
concentrates on developing standards of
measurements, the concept of aperture, analogy
to light, relationships among directivity, gain and
aperture, the concept of an isotropic antenna,
path loss, and radiation resistance.
Chapter 3: Radiation of Radio Waves: Again,
building from the previous chapter, we develop,
in detail, how radio waves “radiate” and how
antennas perform this task. An introduction is
made to Maxwell’s equations, the wave equation,
relationship between the magnetic and electric
fields, the impedance of free space, power,
energy, and fields, detailed definition of radiation
resistance, and how fields and antenna patterns
are formed. In order to clarify some concepts, the
dipole antenna is also introduced. The concept of
“PADL” is also introduced. This chapter, in
particular, diverges from most antenna texts in
that it roots antenna performance and
characteristics in electromagnetic physics.
Chapter 4: Transmission Lines: The thrust of
21
this chapter is to distinguish the “guided wave”
from the “radiated wave.” Rather than repeat
excellent texts available on transmission lines, we
introduce transmission lines in the context of
principles developed in the earlier chapters. For
example, the creation of a transmission line from
a 1⁄2 wavelength dipole is discussed in detail. Then
the two basic forms of transmission line are
discussed and developed from the physical
properties defined in earlier chapters.
Chapter 5: Antennas Using Multiple Sources:
We explain how directive antennas can be
constructed from multiple elements, including
multielement driven and parasitic arrays.
Part 2: Applied Antenna Physics
Chapter 6: Dielectric Effects Upon Radio
Waves: Most antennas we encounter are located
close to the Earth’s surface. Therefore, no
antenna text (even a physics book) is complete
without a treatment of the physics of the Earth
— “geophysics” — as relating to radio waves.
Since the Earth is a dielectric, dielectrics are
treated first in a general discussion. Then the
Earth is discussed as a special example of
dielectrics.
Chapter 7: Vertical Antennas: Much of the
material presented in Chapter 6 becomes
22
particularly relevant to ground-mounted vertical
antennas. Amateur Radio experimenters may
find this chapter particularly interesting
stemming from the intense interest in the
amateur community regarding vertical antennas.
A mathematical format (scalar matrix) is
introduced for the quantification of vertical
antenna losses. Also, a simple comparison of
vertical configurations is also offered with their
relative major advantages and disadvantages.
Chapter 8: Yagi-Uda and Cubical Quad
Antennas: The Yagi-Uda configuration has
become one of the most commonly used antenna
forms. Although volumes have been written on
this antenna type, the purpose of this chapter is
to relate its performance to the physical
parameters already developed in this text.
Chapter 9: Specialized Antenna
Configurations: This chapter presents some of
the more important, but less-known antenna
configurations. Loop and helix antennas are
presented first, followed by “aperture” antennas,
traveling wave antennas, and finally a detailed
discussion on small antennas (relative to
wavelength)
Chapter 10: Noise, Temperature, and Signals:
This final chapter attempts several goals. The
primary importance of noise and noise
23
temperature are presented and are shown to be
intrinsically related. Then “noise temperature” is
shown to be an important antenna specification,
especially for radio astronomy and deep space
applications. Radio link budgets are presented,
showing how noise sets fundamental limits. We
also return to include some historical narrative
culminating in great scientific achievements. The
epic discovery of Penzias and Wilson is
presented as an outstanding research application
of an antenna.
The work of Claude Shannon and information
theory are introduced. Although not related
directly to antennas, the discussion of
information theory is presented since it links
noise to fundamental physical limiting factors for
both the radio link and on the limits to
communication of information. Thus the reader
is left with all the known physical limits (and
opportunities) for communicating information
using radio links, for which the antenna forms a
critical function. These are the same limits for
garage door openers, cell phones, WiFi, and even
establishing a hypothetical link with an
extraterrestrial civilization.
Acknowledgements
First, I would like to acknowledge the American
24
Radio Relay League publications that provided
me with the critical first level of understanding of
radio technology and to the point of this text,
antenna theory. Also, Mark Wilson, K1RO, who
guided this text through ARRL’s publication
process and Larry Wolfgang, WR1B, with whom I
have worked as a volunteer ARRL technical
advisor for many years. It is my hope that this
text will offer yet another useful volume to the
League’s library.
Some of my “elmers” dating from the 1960s also
deserve mention: Martin Hornak, W8MAE, and
Orlando O. Okleshen, W9RX, provided needed
guidance and help through those difficult teenage
years. Later my professors at the University of
Oregon Physics Department, especially Dr. K.J.
Park deserve mention. Although I never met
him, the famous text written by Dr. John Kraus,
W8JK, remains my “bible” for antenna
engineering. My friends and fellow authors Wes
Hayward, W7ZOI, and Cornell Drentea, KW7CD,
showed me through example how hard work and
dedication can result in the production of fine
technical books. The narrative style of Cornell’s
book provided the inspiration to present this text
in a similar vein.
Gary Breed, K9AY, provided an excellent
editorial review of the entire text and offered
25
numerous useful suggestions and corrections.
Gary also wrote the Introduction. Finally Dr.
Nuri Celik of Trimble Navigation performed a
detailed check of the most difficult math,
particularly from Chapter 3, and yes, he found
some errors! However, the errors were minor and
few, and I hope this will ring true for the entire
text.
26
ABOUT THE AUTHOR
Robert J. Zavrel, Jr, W7SX, was born in Chicago
in 1951. His interest in radio developed very early,
having dissected his first radio at age four, much
to his mother’s horror seeing him on the floor
playing with an opened tube radio plugged in. He
was raised in Chicago Heights, Illinois, where he
passed his radio amateur Novice test when he
was 14 years old. He earned DXCC (an award for
contacting 100 different countries) when 17 years
old. After attending Roosevelt University in
Chicago, he worked as a radio navigator in the
Gulf of Mexico and as a radio broadcast engineer.
In 1979 he returned to school at the University of
Oregon, where he earned a BS in Physics in 1983.
After graduation he moved to Silicon Valley
where he worked for 12 years in RF
semiconductors for Siliconix, Philips, GEC
Plessey and several startups. In 1995 he moved to
RTP in North Carolina and worked with IBM’s
SiGe process and then to ATMEL in Colorado
Springs as Director of their SiGe process. In 2002
he started an RF engineering consulting company
that he ran for 10 years. In 2012 he joined Trimble
Navigation Ltd. In Corvallis, Oregon, where he
remains a Senior RF R&D engineer with a current
27
concentration in antenna design.
Bob has five patents, plus two additional recent
patent filings, and has published more than 60
technical papers in both amateur and
professional publications. He wrote the first
papers on software defined radio and using direct
digital synthesis (DDS)M for both FM and HDTV
applications, now all developed technologies and
industry standards. He patented the 8-circle
antenna array and has designed many antennas
for commercial and military applications. His
work has sent him to 35 countries on six
continents and has lived a total of three years in
Europe. He has earned Honor Roll status for
DXCC and CW DXCC using only tree-supported
wire antennas and has earned DXCC on all bands
160 – 10 meters. His amateur station consists of
an array of homebrew amplifiers, tuners, control
systems, and of course novel antenna designs. In
addition to Amateur Radio he grows Pinot noir
grapes and makes award-winning wine in his
boutique noncommercial winery.
28
INTRODUCTION
I have known Bob Zavrel, W7SX, for 30 years.
We share an intense passion to know more about
antennas and RF circuits — well, actually, just
about everything scientific, with a bit of
philosophy thrown in for good measure! We
understand that many other radio amateurs want
to know more about antennas, but do not have
the formal education in physics and mathematics
necessary to delve into the theory of
electromagnetism. This book was written with
the goal of helping those Amateur Radio
colleagues follow their curiosity, with the
intention of providing a substantial step forward,
beyond the how-to articles and basic concepts
presented in magazines and handbooks.
When Bob asked me to “look over his shoulder”
as he wrote this book, I agreed, knowing that we
would probably disagree at times on things such
as choosing the best example to illustrate a
particular concept, or finding an appropriate
analogy to explain a complex concept. I’m
pleased to say that those times were few.
A mid-level book like this is a challenge for the
author! How much technical depth is
appropriate? What topics need the greatest
29
emphasis? Readers will likely experience the
same sort of response. They will be pleased with
some explanations, but may find that others just
don’t resonate. One reader may find the math
difficult, while another wants more detail. If this
book encourages them to pursue further study to
gain the additional background and find the
alternate descriptions they are seeking, we will
consider it a great success!
30
ABOUT THE ARRL
The seed for Amateur Radio was planted in the 1890s, when
Guglielmo Marconi began his experiments in wireless
telegraphy. Soon he was joined by dozens, then hundreds, of -
others who were enthusiastic about sending and receiving
messages through the air—some with a commercial interest,
but others solely out of a love for this new communications -
medium. The United States government began licensing
Amateur Radio operators in 1912.
By 1914, there were thousands of Amateur Radio operators—
hams—in the United States. Hiram Percy Maxim, a leading
Hartford, Connecticut inventor and industrialist, saw the need
for an organization to band together this fledgling group of
radio experimenters. In May 1914 he founded the American
Radio Relay League (ARRL) to meet that need.
Today ARRL, with approximately 165,000 members, is the
largest organization of radio amateurs in the United States.
The ARRL is a not-for-profit organization that:
• promotes interest in Amateur Radio communications and
experimentation
• represents US radio amateurs in legislative matters, and
• maintains fraternalism and a high standard of conduct
among Amateur Radio operators.
At ARRL headquarters in the Hartford suburb of Newington,
the staff helps serve the needs of members. ARRL is also
International Secretariat for the International Amateur Radio
Union, which is made up of similar societies in 150 countries
around the world.
ARRL publishes the monthly journal QST, as well as
newsletters and many publications covering all aspects of
Amateur Radio. Its headquarters station, W1AW, transmits
bulletins of interest to radio amateurs and Morse code practice
sessions. The ARRL also coordinates an extensive field
organization, which includes volunteers who provide technical
31
information and other support services for radio amateurs as
well as communications for public-service activities. In
addition, ARRL represents US amateurs with the Federal
Communications Commission and other government agencies
in the US and abroad.
Membership in ARRL means much more than receiving QST
each month. In addition to the services already described,
ARRL offers membership services on a personal level, such as
the ARRL Volunteer Examiner Coordinator Program and a QSL
bureau.
Full ARRL membership (available only to licensed radio
amateurs) gives you a voice in how the affairs of the
organization are governed. ARRL policy is set by a Board of
Directors (one from each of 15 Divisions). Each year, one-third
of the ARRL Board of Directors stands for election by the full
members they represent. The day-to-day operation of ARRL
HQ is managed by a Chief Executive Officer.
No matter what aspect of Amateur Radio attracts you, ARRL
membership is relevant and important. There would be no
Amateur Radio as we know it today were it not for the ARRL.
We would be happy to welcome you as a member! (An Amateur
Radio license is not required for Associate Membership.) For
more information about ARRL and answers to any questions
you may have about Amateur Radio, write or call:
32
Voice: 860-594-0200
Fax: 860-594-0259
E-mail: [email protected]
Internet:www.arrl.org/
33
Part 1
Antenna Physics
34
CHAPTER 1
Development of Antenna
Physics
Early History
“Electricity” had probably been observed since the beginning
of the human species in the form of lightning and perhaps
static electricity raising hairs on skin clothing worn by early
cave-dwelling experimenters. From ancient Greece, Aristotle
and Thales described the effects of magnetism. Much later, in
the 11th century, the Chinese scientist Shen Kuo was the
inventor of the compass and employed it for navigation.
Detailed rational explanations of electricity and magnetism
had to wait until the 19th century. The invention of radio and
thus antennas followed quickly after the fundamental
breakthroughs.
Modern History
During the Renaissance in Europe, the scientific method
began to replace mystical explanations for a wide variety of
unexplained phenomena. Careful, measured experiments were
undertaken and documented. Following is a chronological
record of some of the more important breakthroughs related to
electromagnetics:
• 1590s: William Gilbert (English) conducted the first formal
experiments in electricity and magnetism. His observation that
the Earth is actually a magnet directly led to refinements of the
compass.
• 1746: Jean le Rond d’Alembert (French) formulated the one-
dimensional wave equation describing the motion of a
vibrating string. Although his work had nothing to do with
electricity or magnetism, it would prove indispensable a
century later.
• 1750: Benjamin Franklin (American) developed the law of
35
conservation of charges, found there were two types of charges
(positive and negative), and invented the lightning rod.
• 1755: Charles Augustin de Coulomb (French) invented a very
sensitive balance and used it to relate mechanical forces due to
electric and magnetic fields.
• 1780: Karl Friedrich Gauss (German) developed the
divergence theorem, which later would become part of
Maxwell’s equations.
• 1800: Alessandro Volta (Italian) invented the voltaic cell
and the battery, thus providing, for the first time, a continuous
moving charge — an electric current.
• 1819: Hans Christian Oersted (Danish) observed that an
electric current flowing on a wire deflected a compass, thus
showing that electricity and magnetism were somehow
intrinsically related.
• 1820: Andre Marie Ampere (French) invented the solenoid
and thus produced a significantly higher magnetic force from
an electrical current. He also quantified the relationship
between the current and magnetic flux.
• 1821: Georg Simon Ohm (German) derived his famous law
relating voltage, current, and resistance
• 1831: Michael Faraday (English) discovered and quantified
the phenomena of a changing magnetic field produced an
electrical current, which is the reciprocal of Oersted’s findings.
This implied that electricity and magnetism were more closely
related more than previously thought.
• 1831: Joseph Henry (American) independently observed the
same phenomena as Faraday.
Maxwell, the Turning Point
• 1860s/70s: James Clerk Maxwell (Scottish, Figure 1.1) saw
that Faraday’s work was the final concept needed to postulate a
unified theory of electromagnetism. In effect, Maxwell
combined the equations of Faraday, Gauss and Ampere into
one set of relationships that precisely define the behavior of
electromagnetism, and more importantly for this text, the
behavior of electromagnetic waves.
36
Maxwell also postulated (correctly) that light is an
electromagnetic wave by a brilliant observation. Coulomb and
Ampere, through static measurements, were able to determine
the permittivity (ε0) and permeability (μ0) of free space by
measuring the physical forces of magnets and electric fields
imparted on objects. Maxwell noticed that the expression
(Equation 1.1)
and
(Equation 1.2)
37
or
(Equation 1.3)
therefore
(Equation 1.4)
or 299,863,400 meters/second.
This is obviously a velocity, but of what? In other words the
basic free space coefficients of the magnetic and electric fields
were related to a velocity. Through simple algebra, Maxwell
reduced these measured values (distance, mechanical force,
static charge and current) to a velocity. Once he had performed
these calculations, he postulated that light must be an
electromagnetic wave after noticing that this velocity was close
to the then-measured speed of light. Since these static (and
current flow) terms could be measured with good accuracy
even in the 1860s/70s, Maxwell’s calculation of the speed of
light was more accurate than the direct measurements of the
speed of light possible in the 1870s. His calculated speed is very
close to today’s known measured speed of light.
Sixty years before Maxwell, Thomas Young discovered that
light waves could interfere with one another, supporting the
idea that light was composed of waves, and the distance
between the ripples in the interfering pattern demonstrated
the wavelength of light directly. If light were an electro-
magnetic wave, then certainly electromagnetic waves of other
wavelengths must also exist! Thus in the 1870s Maxwell not
only predicted the existence of radio waves, but also laid the
accurate theoretical basis for their study and — more
38
important — their application. This quantification of radio
waves, even before they were discovered, still forms the basis of
electromagnetic theory today. Maxwell’s equations define
precisely the relationships among the electric field, the
magnetic field, and the direction of propagation. They also
define precisely the ratio of the field strengths, and the
mechanisms for producing a radiated or a directed
electromagnetic wave.
The pivotal work of Maxwell was not universally accepted at
the time of its formulation. Today Maxwell’s equations remain
the basis of all theory pertaining to electromagnetic waves, and
thus also antennas. It would take decades before the validity of
Maxwell’s work would become difficult to refute.
The initial difficulty with Maxwell’s equations was their
complexity. His original formulation was 20 equations with 20
variables! In the early 1880s, Oliver Heaviside applied his
understanding of both electrical circuits and advanced
mathematics to “clean up” and simplify Maxwell’s equations to
just four equations in two variables. This work made it much
easier to grasp the concepts and fundamental relationships
developed by Maxwell, but we should remember that the actual
equations that are familiar to us are the work of Heaviside. We
will return to Maxwell’s equations with more detail in Chapter
3.
• 1887: Albert Michelson and Edward Morley (American)
performed the first experiment that negated the theoretical
existence of the “luminiferous ether.” Previously it was
accepted theory to assume that space was filled with an
invisible “ether.” Since light waves traveled through free space,
they required some type of matter to travel in. After all, ocean
waves need water, sound waves need air (or some other non-
vacuum medium), and so on. This assumption was in large
measure derived from Newton’s Laws constituting Classical
Mechanics. This experiment, together with Maxwell and a
growing body of knowledge from many scientists, eventually
led to Albert Einstein’s Theory of Relativity. Later we shall see
how Maxwell found that electromagnetic waves did not require
an “ether” as a propagation medium in effect, questioning the
39
then-assumed universality of Classical Physics.
• 1886: Heinrich Hertz (German, Figure 1.2) built the first
complete radio system consisting of a spark gap generator, a
transmit antenna, a receive antenna, and a spark gap receiver
(Figure 1.3 and Figure 1.4). Since radio waves cannot be
detected by any human sense, it was necessary to build a
complete system to actually prove radio waves exist! Thus
Hertz invented the antenna, the radio transmitter, and the
radio receiver, and in so doing “discovered” the radio waves
that Maxwell had predicted to exist.
40
• 1895: Guglielmo Maria Marconi (Italian, Figure 1.5 and
Figure 1.6) invented practical radio. Hertz was a physicist and
41
was interested in the discovery and study of this newly-
invented apparatus. Marconi was the engineer who used
Hertz’s inventions and findings to build something useful. If
radio were to be useful, it would be necessary to significantly
extend the range to provide a new and useful application —
wireless communication. Successive experiments and
refinements extended the range, and soon it became apparent
that the antenna played a crucial role in increasing the range
of the wireless devices.
42
In one of the experiments he greatly lengthened the wires
and configured both into ground mounted “verticals.” The
result was a range of 2.4 kilometers. Marconi was perhaps the
first to appreciate that the size of an effective antenna needed
to be comparable to the wavelength being used. This was
synonymous to raising the radiation resistance of an antenna, a
critical specification for antennas.
After moving to England, Marconi received needed funding.
After further refinements to his antennas and radios, in 1901 he
shocked the world by completing a wireless trans-Atlantic link
from England to Newfoundland. Modern wireless had been
born.
• 1897 – 1904: Hendrik Lorentz, George Fitzgerald, and Joseph
Larmor derived the Lorentz transformation, stating that
Maxwell’s equations were invariant.
• 1905: Albert Einstein used Maxwell’s equations as the basis
of Special Theory of Relativity.
43
1901 to Present
Since its infancy, radio technology has progressed through a
continuous stream of innovations. The basic theory was
derived in the 1860s and 1870s, but the technology is still
evolving quickly some 150 years later. This technical evolution
covers spark gap, the invention of the triode vacuum tube,
discovery of the piezoelectric effect, the transistor, integrated
circuits, super-miniature digital integrated circuits,
miniaturization of all the passive devices, and countless other
advances small and large.
The first vacuum tubes had severe speed limitations and thus
were usable only at relatively low frequencies. However,
vacuum tubes also were far more effective in building receivers
of much improved sensitivity and selectivity compared to the
coherer detector. In the early 20th century, the first radio
applications used the region of spectrum now used by the
standard AM broadcast band, hundreds of meters in
wavelength. Longer wavelengths were considered “low
frequency” and shorter wavelengths were considered “high
frequency.”
The relationship between wavelength (λ) and frequency in
Hz (f) is:
(Equation 1.5)
where c is the speed of light, λ is wavelength, and f is the
frequency in Hertz (Hz).
As technology improved, thus pushing the useable spectrum
higher in frequency, it was necessary to use progressively
stronger superlatives to define these new bands as they opened
up to practical use as shown in Table 1.1. Legacy trapped us
into defining 3 MHz as a “high frequency.” Perhaps frequencies
above 300 GHz should be defined as Unimaginably High
Frequencies!
44
Today, the opening of the microwave spectrum
(approximately >1 GHz) has made possible low-cost, low-power
portable devices. Microwave wavelengths are comparable to
the physical dimensions of the human hand. Thus by using
these wavelengths (and shorter), efficient antennas could be
built into hand-held equipment. The results have been
profound. For example, the personal cell phone and its
numerous derivatives were made possible by a combination of
two fundamental technical innovations: 1) miniaturization of
microwave radio components to permit a two-way microwave
radio fit in a hand-held package and be battery operated; and
2) digital integrated circuits that could permit the new
miniature two-way microwave radio to behave like a
consumer-friendly wide-area wireless “telephone” and/or
computer link. More recently, advanced computers could also
be miniaturized and included in tablets such as Apple iPads
and similar hand held computer/wireless devices.
Dick Tracy had his “wrist radio” in the 1940s (Figure 1.7),
complete fantasy at the time. The concept of “cellular radio”
was conceived in the late 1940s at Bell Labs, but at that time a
“portable” cellular radio would have been the size of a
telephone booth. In contrast, antenna technology for portable
applications at microwave frequencies is almost as old as
antennas themselves. After all, for higher frequencies you
simply make the antenna smaller! Antenna engineers had to
wait for radio and digital circuitry to catch up. But catch up,
they did, as shown in Figure 1.8.
45
46
We are all familiar with hand-held receivers that are quite
effective at much longer wavelengths (such as AM broadcast).
47
However, the receive antenna’s inefficiency (due to very small
size versus the wavelength) is overcome by the stationary
transmitters’ high power (kilowatts) and antennas that do
approach the physical dimensions of the applied wavelength.
Thus, poorly performing antennas are not out of the question
when designing radio systems, as we will explore in later
chapters.
The advent of personal computers such as the one shown in
Figure 1.9 was made possible by the miniaturization of the
microprocessor, memory, and I/O digital circuits. The
application that led to the home computer and its endless
applications and form factors was the result of the clever
innovators and pioneers of the personal computer. The first
microprocessor was developed to provide on-board navigation
for the Apollo Program! Although considered an off-shoot
development at the time, developers of this innovation went
on to form a new company called Intel. In turn, Intel and other
companies, academic researchers, and many more individuals
and organizations contributed to the birth of the modern
microelectronics and semiconductor industries. Although the
main focus of semiconductor development remains building
ever faster and smaller circuits for digital applications, the
implications of highly integrated, very small and very fast
components for micro-radio circuits has been equally dramatic.
In effect, the goals of fabricating ever smaller and faster active
devices (mainly bipolar and MOS transistors) were of equal
interest to digital (mainly computer) and radio designers (RF
engineers). “Smaller” implied transforming rack-mounted
equipment into hand held devices, and “faster” implied ever
higher frequencies of operation.
48
There is another advantage to applying ever higher operating
frequencies. The amount of data or information that can be
carried by a channel of finite bandwidth has definite limits (as
we will explore in Chapter 10). Thus higher frequencies
inherently permit wider operational bandwidths. For example,
the entire AM broadcast band spans from 540 to 1700 kHz, a
bandwidth of only 1.16 MHz. In contrast, a comparable
bandwidth in the UHF region is a small fraction of the total
band.
Enter the Antenna
Antenna design has followed this upward trend in frequency
and thus smaller form factors for shorter wavelengths. Unlike
the radically different technologies that permitted this trend in
the radio proper, many antennas remain remarkably similar in
form. For example, the standard AM broadcast antenna (circa
1920s) is a 1⁄4 wavelength, ground mounted vertical. At 1 MHz,
this implies a steel tower about 75 meters high (Figure 1.10). In
contrast, 2.4 GHz WiFi access points typically also use a 1⁄4
wavelength (or longer) simple vertical antenna, only inches
high! The differences between an AM broadcast system and a
WiFi for radio technology is enormous, but the antenna design
49
is practically identical, at least the first iterations of these
antennas. There is also a continuous effort to integrate the
antenna(s) into the various hand-held form factors. The
external flexible “rubber duckie” antennas (Figure 1.11) have
been gradually replaced with antennas integrated inside the
unit. These internal antennas are in a constant redesign cycle
to approach an optimum performance limited only by physics.
50
51
Almost all antennas, past and present, are configured of
metal. Even very complex antenna arrays are built upon
simpler long-known configurations but still consist of copper,
aluminum, and/or other metals and alloys. Although there has
been considerable progress in optimizing antenna designs for
various applications, basic antenna designs remain in
widespread use. More importantly for this text, they represent
the focal point for discussions on antenna physics, which forms
the basis for all antenna types.
In principle, virtually any type of antenna can be used at any
radio frequency, but their physical sizes will dramatically
change as a function of wavelength. Consequently, different
antenna designs may be more or less practical as the frequency
of interest changes. This is a key point in developing an
intuitive understanding of antenna design: The physics remains
constant, but the practical implementation and choices of
52
design typically vary widely with frequency.
This is not to say there has not been significant progress in
the development of antenna technology. Over the past several
decades, enormous progress has been made by optimizing
antenna parameters ever closer to physical limits. Computer
modeling has permitted such optimized designs not possible
only 10 or 15 years ago. Optimization of antenna design
involves changing the configurations of metallic structures and
often substrates upon which they are mounted.
In the early days of radio we began with Zepps, loops,
dipoles, and monopoles. Today many antennas are simply
manifestations of the simple dipole or monopole (as shown
above). Then multielement arrays (including the Yagi-Uda
design) using multiple dipoles or monopoles were designed to
create more directivity. The helix antenna was invented in the
1940s, and the parabolic reflector was simply borrowed from
optics exemplified by the reflector telescope. Patch antennas
are mainly two-dimensional flat structures, but 3D structures
are also in use. Diversity antennas of various forms use
multiple elements of the same simple structures to improve
information rate capacity and/or reduce path losses due to
changing propagation conditions. These innovations often
utilize sophisticated signal processing algorithms by comparing
and combining signals from two or more antennas. The market
wants products to be smaller, cheaper, and with better
performance!
We will explore in detail the reasons why antennas, in
general, require their minimum physical size to be close to the
wavelength of operation, or as close as possible. This is always
a critical factor in antenna design. However, this is only the
first “rule of thumb” to gain an intuitive understanding of the
workings of antennas. There are many fundamental terms to be
mastered, but again, the basic physics is constant for all
antennas. This is good news. By mastering the basics of
antenna physics, any antenna at any frequency is within grasp
of understanding. An antenna engineer is limited only by
his/her understanding of antenna physics.
53
CHAPTER 2
Fundamentals
What is an Antenna?
The relevant Webster’s definition of antenna states, “a usually
metallic device (as a rod or wire) for radiating or receiving
radio waves.” This is indeed the most common application for
antennas. Kraus provides a more technical definition: “…may be
defined as the structure associated with the region of transition
between a guided wave and a free-space wave, or vice-versa.” In
other words, it is the device that connects the guided wave
(usually a one-dimensional transmission line) to three-
dimensional free space and vice-versa (Figure 2.1). Some also
refer to antennas as “transducers,” converting the guided wave
— represented by currents and voltages — to radiated waves
and vice-versa.
54
tubing. We will also discuss some nonlinear configurations in
later chapters, but it is best to begin this introductory
discussion by restricting ourselves to the simpler linear
antenna configurations. In turn, more complex antennas build
upon the basic principles derived from simpler types, but the
calculations (particularly the integral calculus) become more
difficult. There is also the fictional “isotropic” antenna, which is
an essential reference point for all other antenna types,
practical and imaginary.
Receive antennas capture radio frequency power from free
space and concentrate it into a guided wave, and then, usually,
to a receiver. Transmit antennas perform the reciprocal,
receiving guided wave power from a transmitter (RF generator)
and radiating — or “coupling” — that power into free space.
Free space is defined as an empty three-dimensional area
devoid of any matter. For most of the radio spectrum, air can
also be considered free space for antenna work. While air (the
atmosphere) often has profound effects on radio propagation, it
typically has little effect on an antenna’s characteristics per se.
The essence of Antenna Physics deals with the physical
characteristics of the antenna and how it interfaces with these two
ports.
In electrical and electronic circuit theory, there are theorems
that can be of great help in solving difficult problems. For
example, if we have a two-port circuit (as in Figure 2.2)
composed of linear elements and apply a source voltage (and
resulting current) across Antenna 1, we will observe a resulting
current in Antenna 2. Now if we apply the same observed
current to Antenna 2 and remove the original source voltage
from Antenna 1, we will see the same voltage across Antenna 1
that we previously applied. This is called the Reciprocity
Theorem.
55
Since most antennas can be considered linear structures, we
can assume that the Reciprocity Theorem also holds for an
antenna (also defined as a two-port circuit described above).
Notice from Figure 2.2 that both antennas reflect some
impedance value, and there is also an intermediate impedance
value of the medium — in this case, free space. These
impedance terms are of fundamental importance to antenna
physics and will be carefully developed later.
The reciprocity theorem as applied to antennas is of critical
importance whenever the antenna is to be used for both
transmitting and receiving (Figure 2.3).
56
The Sphere as a Measurement Tool
Since free space is three-dimensional, 3D measurements are
of fundamental importance in describing antennas’
characteristics. We need some convenient construction that
can serve to represent and quantify what is happening in free
space relative to an antenna’s performance. We will use an
imaginary sphere for this measurement tool, superimposed
within free space with a test antenna at its center. For antenna
work, we are usually concerned with only the center of the
sphere and the surface of the sphere upon which we can define
power density over a portion of the entire sphere’s surface. The
volume of a sphere is usually only concerned with energy that is
not typically used for antenna characterization. It is usually
assumed that the space inside and outside the sphere is indeed
57
free space — no matter allowed except for the antenna!
It is often impractical to perform actual measurements in
actual free space. Antenna measurements are often taken on or
close to the surface of the Earth. This causes difficulties in that
matter (solids, liquids, and sometimes gases) can reflect and
absorb power, and worse, affect the performance of the
antenna proper. Reflections, in particular, are problematic in
that they can add or subtract from the direct path wave
between the antenna and a measurement device, thus
distorting the measurement. Such reflections can cause
multipath conditions, where a receiver receives a signal from a
straight line from the transmitter and also a reflected signal.
See Figure 2.4.
58
Since the sphere is often used in antenna measurements, a
review of spherical geometry is in order with emphasis on
terms used in this book. Figure 2.6 shows a section of an
imaginary sphere. The antenna is located at the center of the
sphere and is usually assumed to be a “point source” in that the
antenna’s physical dimensions are much smaller than the
radius of the sphere. We can see the traditional use of spherical
coordinates analogous to latitude (θ) and longitude (φ).
Therefore any unique point on the sphere’s surface can be
identified with a unique value of θ (0 – π) and φ (0 – 2π), where
the value is given in radians. We can convert the radian
measurement to degrees with this relationship: 2π radians =
59
360 degrees.
60
Analogy Between Light and Radio Waves and
the Concept of Aperture
Let us place a 100 W light bulb at the center of our imaginary
transparent sphere. We assume that this light source is
emitting exactly 100 W of light power and ignore any heat
generated through inefficiencies. Our light bulb also has a
special characteristic: Its pattern of radiation is isotropic, which
means that the light power emitted by the bulb is radiated in
all directions equally. Therefore, the power radiated is
distributed perfectly evenly over the surface of the sphere. If
the interior of the sphere is a vacuum, then all 100 W of light
power will reach the surface of the sphere. If we could calculate
the total power passing through the sphere’s transparent
surface, it would also be 100 W evenly distributed over the
surface of the sphere.
Furthermore, we can define the power of the light bulb (or
any electromagnetic wave) as a power density, or flux over an
area, or W/m2 (watts per square meter). Another term often
used is the Poynting Vector S (θ,φ). S is the vector defined as
Wm–2, and (θ,φ) represents the spherical area over which the
power density is constrained, or more conveniently by a
steradian value. See Figures 2.7 and 2.8.
61
On the surface of the 100 W light bulb’s surrounding sphere
we will place a solar panel. The surface area of the sphere is
taken to be 1000 m2 and our solar panel is 1 m2 and 100%
efficient. Therefore we can capture 1/1000 of the light power, or
1/10 W, or 100 mW from the light energy falling on our solar
panel. If we double the area of the solar panel we double the
power received. Here we introduce another very fundamental
relationship in antenna physics: the “area” of an antenna
(which is more appropriately defined as aperture). As we shall
soon describe, aperture and power gain are directly
proportional. Thus we can now express how much power is
received for a given distance and a given aperture of the
receiver:
62
(Equation 2.1)
where
Pr is the power received
Pt is the power transmitted in the direction of the receiver
Ar is the aperture of the receive antenna
4π r2 is the total surface area of the sphere with a radius of r.
Notice that the term
63
by a radius, which is the distance between the Sun and the
Earth. The Earth’s atmosphere absorbs about 1⁄2 of the solar
light before it reaches the Earth, and a 100% efficient 1 m2 solar
panel will provide about 750 W. Thus we can work in reverse
and calculate the total power output of the Sun!
The Sun is about 150 billion meters from the Earth, or 15 × 109
m. Therefore the surface area of this imaginary sphere (the
radius being the Earth – Sun distance) is 4π(15 × 109 m)2, or
about 28 × 1021 square meters. Thus if the Sun’s power is 1500 W
per square meter at the distance of the Earth, then the total
light power output of the Sun is about 42.5 × 1024 W! Thus we
can calculate any unknown variable in Equation 2.1 if the other
terms are known. Also note that thus far we assume an
isotropic transmitter and a known aperture of the receiver.
Reconciling these two variables is a key task for any advanced
discussion of antenna physics. The following discussion will
quantify this essential relationship.
Defining the Critical Fundamental Terms
We use the examples of a light bulb or the Sun, since they are
easy to visualize. Now we can take the leap and state that there
is a direct correlation between our brief discussion about light
and radio waves. We know this is a correct correlation because
Maxwell predicted the existence of radio waves using the then-
known properties of permittivity and permeability of free space
and their relationship to the speed of light. He then correctly
postulated that both light and the yet-undiscovered radio
waves were both electromagnetic radiation, in effect having
identical properties except for their differing wavelengths.
After the actual discovery of radio waves, endless
measurements and experiments have proven the validity of his
early postulations.
Solar cells remain inefficient because we remain restricted to
the photoelectric effect that relies upon the inefficient process
of a light photon dislodging electrons within the photovoltaic
material, thus generating an electrical current. Building an
array of receive antennas for light is possible, but converting
64
the “guided” light wave directly into a dc current would require
a futuristic diode. Again wavelength is the critical limitation. It
is possible to build a dipole antenna for light and/or infrared
(IR) wavelengths using micro-processing techniques. To
convert the light power into an electrical current, however, we
would need to rectify the light power into direct current. No
such diode exists. If it could be developed, solar cells could be
replaced with arrays of tiny dipoles and futuristic diodes. So for
a practical comparison of light and radio waves we need some
slight modifications, but remember that the physics is the
same. Fortunately, antennas for radio frequencies often do
approach 100% efficiency.
A transmit antenna can be compared to the light source (a
light bulb, LED, or a torch) and the receive antenna compared
to the solar panel. Unlike radio antennas, however, these
particular choices of light sources and light “receivers” are not
reciprocal. We do not normally think of a solar cell as a
radiator or a light bulb as a light receiver. On the other hand,
radio antennas are reciprocal, therefore transmit antennas are
also characterized as having an aperture (the same aperture as
they have when receiving).
A “shrinking” aperture due to inefficiency is analogous to a
reduction in antenna gain due to power losses (usually as heat).
Also, as above, if we double the aperture, we double the power
gain (about 3 dB difference). This is illustrated by doubling the
aperture (two rather than one solar cell) for a doubling of
power capability (Figure 2.9). The effect of cutting an
antenna’s physical aperture in half is identical to cutting the
efficiency of two solar cells in half. Therefore, the power gain of
a solar cell or antenna follows the same rule. For example, a
hypothetical 100% efficient solar cell would convert all the light
power falling on it into electrical power. In a real solar cell, the
absorbed power not converted into electrical power is
converted into heat. This is precisely what happens in an
antenna, either for transmit or receive antennas. Power
applied to the antenna, but not radiated from the transmitter
or not delivered to the receiver from a receive antenna, is lost
as heat.
65
One of the advantages of using aperture instead of gain in
radio link calculations is that we can directly calculate how
much power will be received by the antenna. Indeed, any
equation for calculating the power loss of a radio link must use
aperture for the proper calculation, or a derivation thereof. The
answer to the very important question posed in Figure 2.10
requires the detailed understanding of the aperture concept
which follows.
66
amount of power. This is the direction of maximum gain and
thus also the direction of maximum aperture. If you begin to
rotate the panel away from the Sun’s direction the effective
aperture decreases, and so the power received decreases.
Finally, with the panel facing 90 degrees off the Sun’s direction
there is no sunlight striking the panel and the received power
drops to zero. An antenna’s gain and aperture also depends
literally on the antenna’s point of “view”! Of course, those
familiar with directional antennas know that as the antenna is
rotated toward the direction of a transmitter, the signal power
received increases and peaks where the forward gain is
maximum. This is also the direction that corresponds to the
antenna’s maximum aperture.
We also know that any non-isotropic antenna that is 100%
efficient will radiate all the power applied to it, but not in all
directions equally. Therefore in some direction(s) a receiver
will “see” the equivalent of more than 100 radiated watts and in
other direction(s) a receiver will “see” less than the equivalent
of 100 radiated watts. In this case the antenna has gain greater
than isotropic in some favored direction(s).
Yet another optic analogy is the automobile headlamp
(Figure 2.11). The typical headlamp bulb radiates about 10 W of
light power. However, placing a parabolic mirror in back of the
bulb results in a very high power beam with a small steradian
value defining its narrow radiation pattern. In effect, the
parabola “robs” all other directions on the sphere of light
power and concentrates it within the narrow beam. If we can
concentrate all the 10 W of power into a steradian of 1% of the
total sphere 0.01 (4π)Ω we simply multiply the output power by
the power gain (in this case 100) to derive an effective radiated
power relative to an isotropic or 1000 W EIRP. That is why a
misaligned headlight can blind an on-coming driver.
67
Again, the power analogy between optics and radio physics is
exact. Also, here is another critical point: If we take the average
and thus total power radiated over the entire sphere, we find
that the value remains 10 W. In the case of the isotropic the
average and total power is also 10 W. However, the effective
radiated power (EIRP, compared to an isotropic) is 1000 W in
the main beam of the headlamp, but 0 W over 99% of the
sphere, thus the average is, again, 10 W.
The 1⁄2 Wavelength Dipole Antenna
Perhaps the simplest form of a practical linear radio antenna
is the 1⁄2 wavelength dipole, as seen in Figure 2.12. The dipole
is the shortest linear antenna that is resonant. Resonance, in
the case of antennas, implies that the source impedance (when
used as a receive antenna) or load impedance (when used as a
transmit antenna) is a pure resistance (no reactive part of the
impedance). In Figure 2.12 we use RL(load) since we are
assuming (for this example) a transmit antenna. Therefore an
antenna can be understood as a device that is represented by a
load to a signal generator that absorbs power by radiating it
into free space. Of course by reciprocity, the imaginary resistor
can also form a signal source with the same impedance.
68
The isotropic and dipole antenna types form building blocks
for both the theoretical and practical basis of antenna study.
We will soon provide a detailed understanding of their
behavior and their interrelationships.
Relationship Between Aperture and Gain
Aperture is a straightforward concept in that an antenna can
be defined as a simple two-dimensional area. Gain (dB) is
another issue: gain relative to what? Usually antenna gain
specifications reference the antenna to a defined gain of an
isotropic (dBi) or a 1⁄2 wave dipole (dBd). Without a reference, a
specification such as the antenna gain is “6 dB” is meaningless:
4 times the power of what? Since an isotropic antenna would
have a power gain of 1 (0 dBi), this is a logical reference to use.
The decibel measure of power is 10log10(P1/P2), where the P
values are linear values of power. It is essential for P is be a
ratio of power values, such as a ratio of some gain value to the
gain value of a dipole or an isotropic.
Assuming we are comparing lossless antennas, the isotropic
antenna has the minimum possible gain of any possible
antenna. For this reason, antenna gain references are
ultimately based on the gain of an isotropic antenna (0 dBi). As
a result it is necessary to know the aperture of an isotropic
antenna in order to relate any antenna’s gain to aperture.
Without equating the two, it is impossible to calculate how
much power will be received by an antenna for a given power
at a given distance and thus answer the critical question posed
in Figure 2.10. For example, you know your receive antenna has
a gain of 9 dBi and that the power density is 10 microwatts per
square meter at the receive location. How much power will you
expect to receive? A value of 9 dBi is meaningless without
69
equating it (or any antenna gain specification) to an equivalent
area. However, we know that the aperture of a 9 dBi antenna
will be about eight times as large as an isotropic aperture. But
the trick question is: What is the aperture of an isotropic
antenna?
Empirical Calculation of the Aperture of an
Isotropic Antenna
We can use the above introduction (basic physics and some
simple mathematics), coupled with a set of empirical
measurements, to derive the aperture of an isotropic antenna.
We will also use relatively simple instruments to measure
power. This calculation is critical to antenna physics, antenna
engineering, and radio system engineering. But how do we
calculate the aperture of the hypothetical isotropic antenna
that both radiates and receives to and from all directions
equally?
To begin, let us review what we now know:
1) An isotropic antenna radiates and receives in all 3D
directions equally.
2) A non-isotropic antenna does not radiate in all directions
equally, but radiates and receives more power in some
directions and less in others (relative to the hypothetical
isotropic)
3) The total and average radiated power of any antenna will
be the same as an isotropic antenna (assuming 100%
efficiencies). If an antenna radiates all the power applied to it,
the power must be going somewhere!
4) The inverse-square law applies also to radio waves.
Now we are prepared to derive an experiment and
subsequent calculations to determine the aperture of an
isotropic antenna. In this experiment, we begin with two
dipoles. We use dipoles because we have no idea how to design
an isotropic antenna. At present we assume we do not know
the gain of these dipoles, nor do we know their apertures. We
make sure that the dipoles are near 100% efficient by carefully
measuring the temperature of the antenna before and after we
70
apply power to them. We find negligible difference in the
temperature, so we assume near 100% efficiency. We also verify
a perfect (lossless) match to our RF generator by taking a
simple VSWR measurement — it is 1:1. To begin, we use 300
MHz for our first test since relatively high frequencies are
needed for practical measurements inside anechoic chambers.
In the anechoic chamber we use a dipole as a source antenna
and a second dipole at 10 meter distance as the receive
antenna. By rotating the receive dipole in three dimensions, we
can record the directivity of the dipole over the dipole’s entire
imaginary external sphere. We don’t need a calibrated power
source since we are only plotting the relative gain. Since our
dipole is near 100% efficient, we assume the relative gain is the
same as the relative directivity, where
Gain = Directivity × Efficiency
(Equation 2.2)
If we plot the relative gain of the 1⁄2 λ dipole in free space, we
find a “doughnut” pattern (Figure 2.13). By our statement 3
above, we know that the total and average power response of
this non-isotropic dipole antenna must be the same as an
isotropic antenna
71
We can now perform an empirical process to calculate the
gain difference between the dipole and the isotropic antennas.
We know that the gain of the dipole is constant for all points
perpendicular (broadside) to the line of the dipole because we
measured it. However, the gain difference can be plotted in two
dimensions (a plane) that bisects the doughnut tangent to the
doughnut’s “hole.” Since we only need to find an average of the
72
dipole power gain, we can ignore performing a three-
dimensional integration and use only a “slice” of the doughnut
and perform a two-dimensional integration. We find the total
area of the two sides of the slice, which we know corresponds to
the total power output of the dipole. Then we take this total
area and calculate the radius of a circle we construct with the
identical area, which is the equivalent pattern of an isotropic at
an identical power (circular). The radius of this circle is 0.61 of
the maximum length of the doughnut in its maximum gain
direction. This means the dipole-doughnut has a power gain of
about 1/0.61 = 1.637 over the theoretical isotropic, or 2.15 dBi.
Thus we have just measured the dBi gain of a 1⁄2 wave dipole.
Since the actual sphere is 3D, we can form a conceptual
understanding of the power integration in 3D by imagining
that we have a doughnut shaped 3D piece of putty. We measure
the radius of this doughnut to be 1 inch. Then we take the
doughnut and reshape it into a perfect sphere (like making a
snowball) and measure the radius of the new sphere. We find it
to have a radius of 0.61 inch which is about a 2.15 dB difference.
Assuming we were very careful in our measurement
techniques, Figure 2.14 shows a linear plot of the dipole’s
response in linear/linear coordinates as we rotate the dipole
360 degrees perpendicular to the source.
73
We also know that the aperture of an antenna is directly
proportional to its gain in the same direction from the antenna
as in our solar cell example. Therefore we can measure the
received power from our dipole example and calculate directly
the aperture of the dipole and thus also the aperture of an
isotropic. This is the relationship that we need to break into
new territory in understanding antenna physics.
We now know that the power gain of our 1⁄2 wave dipole is
1.637 over an isotropic. Therefore if we apply 1 W to the
transmit dipole we will have an EIRP of 1.637 W in a direction
broadside to the dipole. When we place our receive dipole 30
meters away from the transmit dipole (also broadside to the
transmit dipole) we record a received power of just over 23.036
µW. Since we know that we have a total dBi power gain in the
system of 2.6797 (product of the two dipoles’ linear gain
74
broadside to each other), we know the equivalent isotropic
assumption would be just about 7.036 µW. Our actual
maximum measured broadside-to-broadside received power
was 18.854 mW, or 2.6797 times the calculated isotropic
received power. Now we have normalized the dipole to an
assumed isotropic radiator radiating 1 W at the center of a 30
meter radius sphere and a calculated received isotropic power
of 7.036 µW. We can now calculate the aperture of the isotropic
antenna.
The area of the 30 meter radius sphere is about 11,310 square
meters. With 1 W of isotropic radiated power at the center, the
calculated power density is 88.417 µW per square meter. Since
we would actually receive 7.036 µW, the aperture of our
isotropic antenna is
or about 0.0796 square meters. This looks like the answer, but
we have another observation/potential problem. We know that
1⁄2 wave dipoles have different lengths for different frequencies,
such as the antennas shown in Figures 2.15 and 2.16. Intuitively
we believe that a larger antenna will have a larger aperture,
well, because it’s bigger! For example, a 7 MHz dipole is
considerably larger than a 144 MHz dipole. Intuitively we think
that a much bigger structure will couple more power into free
space, or receive more power from free space. If our intuition is
correct we should arrive at different aperture values for
different frequencies.
75
So we prepare an identical experiment, this time at 600 MHz
which is 1⁄2 the wavelength and thus also 1⁄2 the dipole length in
the 300 MHz experiment. If there is a difference in the
calculated isotropic aperture, we will know there is a
wavelength term in the isotropic aperture equation. We run
76
the experiment again and find that the isotropic value is about
0.0199 square meters, or exactly 1⁄4 the area of the 300 MHz
dipole. Additional experiments at other frequencies show that
the aperture of a dipole, and thus also an isotropic antenna,
increases at a rate equal to the square of wavelength, or the
square-root of the frequency.
However, we have some coefficient we must also solve to get
a true equation that defines the aperture of an isotropic at any
wavelength. If we return to the original 300 MHz results, we
know the wavelength is a nice normalized number of 1 meter,
and 1 squared is 1. Thus we see that the coefficient is about
0.07958. The reciprocal of this is 12.566. Because we’ve been
using this number so often with our spherical tool and with the
insight of a James Maxwell, we immediately recognize it as 4π.
We use this coefficient in our other measured results, and thus
converge upon the actual definition of the aperture of an
isotropic antenna as
(Equation 2.3)
where the antenna aperture (area) is in the same units of
wavelength (λ).
This is one of the most fundamental equations (and
relationships) in antenna theory in that it relates antenna gain
to aperture. Indeed, many misconceptions regarding antenna
performance can be resolved by understanding this
fundamental relationship. The most striking implication of this
equation is that the aperture of an isotropic antenna (and all
other antennas referenced to it) depends heavily upon the
wavelength of operation. Now we have the tools to solve the
problem illustrated in Figure 2.10.
Free Space Path Loss
Path loss simply defines the ratio of power lost between the
transmitter’s EIRP and the power received by the receiver’s
77
antenna. In free space, we use the simple spherical model as
above and assume an isotropic receive antenna with the
appropriate aperture. If we receive or “capture” 1/1,000,000 of
the power radiated, our path loss is 10log10(0.000001) or 60 dB.
Therefore the power received will be 60 dB below whatever
power is transmitted in the direction of the receiver. The
relationship, in most situations can be considered linear and
thus proportional.
Virtually all calculations for free space path loss are based
upon these simple concepts: power radiated (EIRP), distance,
and receive antenna aperture. These terms represent the
“starting point” to the operation of a garage door opener, an
interplanetary radio link, radar systems, cell phone coverage
area, satellite services, and so on. The great complexity (and
cost) of some radio coverage prediction software tools reflects
their ability to approximate the effects of material
environmental effects (such as a portion of the Earth’s surface)
upon the coverage. But the starting point always begins with
these fundamentals. It’s the refinement that is difficult. Free
space calculations are simple; it’s when “stuff” gets in between
the transmitter and receiver that estimations become difficult.
No matter how sophisticated a path loss predictor may be, the
result is always one single number (the path loss, and possibly
some statistical deviation around that number).
Returning to the basic situation of an isotropic antenna at
the center of our sphere, we know that the power will be
equally distributed over the sphere’s area, 4πr2. We also know
that the power received will be proportional to the aperture of
the receive antenna. If our receive antenna is 100% efficient,
the portion of the transmitted power received is simply the
ratio of the antenna aperture to the area of the sphere, or,
(Equation 2.4)
and the dB loss is simply:
78
(Equation 2.5)
Notice that Equation 2.4 is simply a ratio of two areas — the
aperture of the receive antenna to the total area of the sphere
on which the antenna is set.
Let’s assume that our receive antenna is isotropic. After the
above calculation and measurements, we know its aperture is
(Equation 2.6)
79
As implied above, we can adjust the aperture value if we
know the gain value of the antenna. For example, if we have a
6 dBi antenna, that represents about 4× power gain (calculated
through using the inverse-log term), so the aperture of a 6 dBi
antenna is simply
(Equation 2.7)
where G is the linear power gain coefficient related to an
isotropic gain for the transmit and receive antennas.
Since
Then
80
By substitution we can now also use the aperture of both the
transmit (Aet) and receive (Aer) antennas in a given link to get
the path loss by substituting these terms for the isotropic in
Equation 2.7, so we get
(Equation 2.8)
This is the free-space Friis Transmission Formula, where
again r is the distance between the antennas, λ is the
wavelength, and Aet and Aer are the antenna apertures all using
the same units. Here we can see another result of the
reciprocity theorem. Notice that both the transmitter and
receiver aperture (or gain) terms appear in the numerator of
the equation. This implies that you can compensate one for the
other. In other words, you will receive the same amount of
power by swapping the antennas, or by changing their values
provided that the product is the same.
Perhaps no better example of the concept of antenna -
aperture can be found in the parabolic (dish) antenna (Figure
2.17). The area of the dish is simply equated to the aperture of
the antenna (assuming 100% aperture efficiency). Therefore the
dBi gain of a parabolic antenna increases with increasing
frequency while the aperture (in principle) remains constant.
81
Equation 2.3 points to another issue relating to wavelength.
An often confused relationship is frequency (wavelength) and
path loss. If all electromagnetic waves conform to the inverse-
square law, why do lower frequencies seem to propagate
“farther”? Understanding the relationship between gain and
aperture solves this common source of confusion.
Refinements of the Aperture Concept
The above discussion provides for an intuitive understanding
of antenna aperture. However, this first approximation of this
critical antenna parameter needs some refinement for a deeper
82
understanding. Thus far we have only discussed effective
aperture (Ae) but have also hinted at the effect efficiency has in
reducing the effective aperture. As mentioned above, the gain
and thus effective aperture of an antenna, decreases when an
antenna becomes less efficient. Thus when an antenna has no
loss and is perfectly matched to a load, the aperture is the
maximum effective aperture or Aem. When power is lost to heat
due to inefficiencies, the result is a “loss” aperture or Al.
Therefore,
Ae = Aem – Al
(Equation 2.9)
Thus Al equates an equivalent area that captures power that is
lost to heat. This is comparable to a solar cell’s maximum
possible aperture being reduced by inefficiencies.
For a properly matched antenna to its load:
(Equation 2.10)
We have seen that power gain is proportional to area
(aperture) in calculating antenna parameters. Therefore we can
begin to formalize this relationship.
Let an incident wave produce a voltage (V) at the feed point
of an antenna, as in Figure 2.18. For convenience we will use a
dipole as an example. This voltage produces a current (I) in
both Rr and Rt. Rr is defined as the radiation resistance, which in
this case is also the feed point impedance value. In effect, Rr is
the source impedance and Rt the receiver’s input impedance
(usually 50 or 75 Ω). Rr, a fundamental term, will be dealt with
in detail later. In this discussion we will ignore any reactive
parts of the source and load impedances, since they are
routinely “tuned out” of the circuits. It is the real parts of these
impedances that determine real power terms. Assuming no
83
complex terms in the impedances (pure resistances) the power
delivered to the load is thus:
(Equation 2.11)
and
(Equation 2.12)
where V is the induced antenna voltage. Thus
(Equation 2.13)
Assuming a conjugate match, Rr = Rt. Therefore,
84
(Equation 2.14)
or
(Equation 2.15)
By our basic discussion of Ae above, we know that
(Equation 2.16)
where S is the Poynting vector, or power density per square
meter, and Ae is the aperture that captures power proportional
to its area.
Therefore, with a conjugate match and no loss
(Equation 2.17)
This discussion points to a possible contradiction. Figure 2.18
shows that 1⁄2 of the received power is “lost” in the antenna’s
radiation resistance. If we assume a 100% efficient antenna,
then the power is not “lost” to heat. Also, if we assume we only
deliver 50% of the received power to the antenna, then the
aperture and thus gain should only be 1⁄2 of what it “should” be.
The answer is that an antenna perfectly matched to its load
reradiates 1⁄2 of the power by virtue of yet another aperture, the
scattering aperture As.
In the case of the scattering aperture, the scattered power is
simply lost back (scattered) as re-radiation into free space. The
85
issue is comparable to a simple dc circuit consisting of a battery
with an internal resistance and a load represented by RL(load)
(Figure 2.19). Maximum power will be delivered to the load
when the source and load resistances are equal. In the battery
circuit, 1⁄2 the power is dissipated as heat. In the Figure 2.18
antenna circuit, 1⁄2 of the power is re-radiated back into space.
86
can achieve very low source impedances at their maximum
efficiency operating point, resulting in efficiencies of over 90%.
However, the antenna “sees” a 50 Ω source impedance but all
the power supplied by the line is radiated by the radiation
resistance. As remains present in the transmitting case, but
only applies to power incident on the antenna from free space.
Figure 2.20 shows the equivalent circuit of an antenna and a
transmitter. Transmitters are usually specified as “50 Ω output.”
The real definition of this specification is that the transmitter is
tuned for maximum efficiency assuming a 50 Ω load. Thus for a
given antenna source impedance, the receiver input is tuned at
or near the complex conjugate of the antenna impedance. In
contrast, for a given load impedance the transmitter is tuned to
some output impedance that will permit maximum efficiency of
the system.
87
However, the above is an adequate simplification for
conceptual understanding of these basic relationships.
As is often used to great advantage in antenna design and is
often misunderstood. In our basic discussion of the aperture
concept we found the maximum effective aperture (Aem) of an
isotropic and a dipole antenna. Now let’s revisit this analysis.
A properly terminated antenna will deliver 1⁄2 its power to
the load (usually a receiver) and re-radiates the other 1⁄2. The
“total” antenna aperture that includes both the Ae and As is
defined as an additional aperture, the collecting aperture (Ac).
The collecting aperture is simply the sum of the thus far
defined apertures:
Ac = Ae + Al + As
(Equation 2.18)
Therefore, in a 100% efficient antenna, Ae is about twice the
area of Ae since Ac = Ae + As and Ae = As.
As is an extremely useful property of antenna elements. In
particular, parasitic elements use this property to maximum
advantage. In a parasitic element, we set the load impedance to
zero (a short circuit), as in a director or reflector element.
When the terminal impedance of a resonant or near-
resonant dipole is zero (as in a parasitic element), As is by
Equation 2.13:
(Equation 2.19)
Since the terminal impedance is zero, then Ae = 0. Therefore all
Pr will be re-radiated, or Prr.
88
(Equation 2.20)
Thus the re-radiated power from a parasitic element is four
times that of a conjugate matched antenna! We will explore
this in more detail in Chapter 8.
(Equation 2.21)
where V2/Rr is the power term and S is the power flux/square
unit on the sphere, so the aperture is simply Prr/S.
Again, compared to As for a terminated antenna dipole, often
referred to as a “driven” element, Ae is four times as large for a
“parasitic” element compared to a “driven” element shown
again in Equation 2.22.
(Equation 2.22)
Thus, the scattering aperture (As) has a maximum area when
the dipole feed point is shorted, or no power delivered to the
load and all power re-radiated. On the other hand, maximum
Aem occurs when the dipole is matched into a load. The re-
radiation from the parasitic element can be adjusted in its
phase and amplitude response referenced to the driven dipole
to create very desirable antenna patterns. The most common
application of this principle is in creating a Yagi-Uda array
discussed in Chapter 8.
Now consider the opposite condition: an open center to the
dipole, where Rt = ∞. As becomes very small because the
antenna is no longer terminated. Mismatching and low Rr
prevent “capturing” incident wave power and also re-radiation
of the received power. This double-effect renders the antenna
to have negligible effect upon the incident wave. In effect, the
89
1⁄2 wave dipole is now two 1⁄4 wave dipoles closely spaced end-
to-end.
Now we are prepared to present a graph of how Ae, As, and Ac
change when the terminal resistance is changed. Note that Aem
appears only at one point, the maximum possible value for Ae.
Figure 2.21 plots Ae, As, and Ac against the normalized value
of Aem. When Rr = Rt then Ae = Aem and maximum power
transfer is accomplished.
90
matched to a load.
As Scattering Aperture: the equivalent area of the antenna
that re-radiates power back into space.
Ac Collecting Aperture: the total equivalent area of an
antenna that represents power delivered to the load, lost to
heat, and re-radiated, where Ac = Ae + Al + As.
Ap Physical Aperture: There is also an additional aperture
term, the physical aperture. This term defines the actual
physical area of an antenna (two dimensional physical
aperture). It is applied typically to “aperture” antennas, such as
horns and parabolic “dish” antennas, where the physical
aperture indeed resembles the collecting aperture. For
example, if we consider a dish antenna 10 meters in diameter,
then Ap = π52 = 75.5 m2. If the aperture efficiency were 100%, this
would also be the Ae. If the efficiency were only 50%, Ae would
be 37.75 m2. In contrast, the “area” of a wire dipole antenna is
very small compared to its Ac , therefore the physical aperture
has little meaning for single-dimension antennas.
91
CHAPTER 3
92
originate from Gauss, Faraday, and Ampere. These equations
are applicable not only to static electric and magnetic fields,
but also to electromagnetic radiated fields. First we will explain
the totality of Maxwell’s equations and then concentrate on the
equations and their forms that are most useful for a discussion
of antennas.
The modern expression of Maxwell’s equations represents
refinements made by Oliver Heaviside, who used vector
calculus and some other notation changes. However, the
essence of the relationships was initially developed by
Maxwell.
over the closed surface area (dA) is the sum of positive and
negative electrical charges contained within the surface (q) and
inversely proportional to the permittivity (ε) of the “material,”
or lack thereof, inside the surface.
In mathematics, the simplest form for “summing” is the
addition sign (+). A shortcut can be used if the summed values
take on a sequence conforming to some defined change (Σ), the
Greek letter “sigma” for “summing.” If we need to sum over
some function (no discrete values), we use the integral calculus
sign
93
An additional refinement
is simply the notation for a “smoothed sum” (in this case the
electric field fluxes) over the surface of the same imaginary
closed surface, usually a sphere as described in Chapter 2. The
little circle is written over the integral sign indicates we are
summing over a closed area. q is an electrical charge in
coulombs (can be either negative or positive; for example, an
electron has the minimum electrical negative charge q). We
begin with an electrical charge at the sphere’s center (Figure
3.1). We measure the total flux of the electric field through the
sphere by using the operator
94
isotropic antenna, which radiates equally in all directions. A
“point charge” is also isotropic in that the field flux will be
identical at all points on the sphere (assuming of course we are
observing it in free space and the charge is located at the
center of the sphere). However, the total flux will always be the
same over any size or shape of enclosure (here we are using a
sphere as the example). Also, the total flux will always be the
same no matter how we distribute the charge or charges within
the enclosure. We can intuitively accept this by noting that a
100 W light bulb will radiate a total of 100 W through the
sphere even if it is not located at the center. The flux will
simply be greater on the surface nearer to the light bulb.
We use a dot product, which is just a special form of
multiplication for vectors, in this case over the surface
surrounding the total charge(s). The dot product simply
accounts for the angle of an area that is not normal to the
source of the electrical charge (Figure 3.2). This is much like
our solar cell example from Chapter 2 being tilted somewhat
away from the Sun, where:
95
(Equation 3.1)
Thus if the aperture is perpendicular to the charge (90
degrees), then the aperture is equal to the actual area of the
surface. In this case, the dot product simplifies to a simple
multiplication.
Finally, q/ε0 represents the amount of charge (in coulombs
divided by the permittivity of the medium), in this case free
space where ε = ε0 This is the essence of Maxwell’s first
equation: The total E field flux through a surface containing a
total of ±q charge will be equal to the total charge divided by
the permittivity of the material or free space inside the surface-
defined volume.
If we place a negative and equal positive electric charge
inside the sphere and integrate the total electric field flux, we
will find the total flux will be zero, since the two charges cancel
and will result in zero charge. Thus in this first equation, q
must account for the sum of all charges (negative and positive)
within the closed surface.
96
place such a dipole magnet inside our sphere, we find that over
the entire surface of the sphere there will be zero net magnetic
field simply because the two field fluxes sum to zero (Figure
3.3).
97
Faraday’s Law states that an electric field E is created when
the magnetic flux ΦB changes with time over some length (dl).
In other words, a time-varying B field creates an E field. An
example is pushing a magnet in and out of a coil of wire
(inductor), creating an electric current (moving electrical
charge) in the inductor. It also explains how an electric current
is generated when rotating the inductor inside a magnetic field,
thus creating a generator (Figure 3.4).
98
Ampere’s Law
Ampere’s Law states that the magnetic field (in this case
generated along a loop conductor of length l) is proportional to
the current I in that loop and the permeability of the medium
inside the loop. An example of this is the electromagnet, or
solenoid, where a magnet is created by running a current
through a conductor (the effect is increased by configuring the
conductor into a solenoid). Notice that the third and fourth
equations are reciprocal: an electric field is formed by a moving
magnet and a magnetic field is formed by a moving electrical
charge, or current. A stationary magnetic field produces no
electric field and vice-versa.
It should now be apparent that there is some type of
reciprocity between the electric and magnetic fields. We can
generate an electric field by moving a magnet in and out of a
solenoid, but we can also create a magnetic field by running an
electric current through the same solenoid. An electromagnet
such as the one shown in Figure 3.5 is a practical example of
Ampere’s Law at work.
99
Although Maxwell’s equations look formidable to anyone
without knowledge of advanced calculus, their meanings are
easily understood by anyone with a basic understanding of
electricity and magnetism. The development of these equations
into forms that explain electromagnetic fields does require
advanced mathematics. Indeed, most of the time a calculus
100
student spends in advanced courses deals with learning how to
develop and solve the bread and butter equations (solving for
derivatives and anti-derivatives of functions). The concepts are
far easier to understand intuitively. Therefore, when possible, I
will skip most of the advanced mathematics and simply state
the relevant results, but also attempt a simple explanation.
Maxwell had a problem with the fourth equation. In the case
of a capacitor inserted into a wire loop, there is zero loop
current in that a current does not flow around the complete
length of the loop. However, a current is necessary to charge or
discharge the capacitor (to displace charge); therefore Maxwell
derived and added an additional term to Ampere’s Law to
account for this displacement current.
(Equation 3.2)
Thus. if we add this to Gauss’s Law we derive Maxwell’s
fourth equation.
Maxwell’s Fourth Equation
101
Usually, we refine the word “change” to “accelerate” for a
more precise description of how a field must behave in order to
create a propagated wave. Usually we assume a radio wave to
be sinusoidal, and indeed a sine wave is nearly always in a state
of acceleration. (See Appendix A for a proper definition of
acceleration.)
These four equations define exactly how fields are formed
and how they behave. In this book we are concerned only with
the formation and propagation of radio waves. In free space
(vacuum) there is no matter and thus no possibility of a
charged particle and no possibility of a magnet since both
require some type of matter present (even an electron). Also
there is no current present in free space because there is no
conductor on which a charge (electrons) can travel. (Of course
electrons can flow through a vacuum, but since electrons are
matter, the vacuum is no longer a vacuum!) Therefore, the first
two of Maxwell’s equations, in free space, equate to zero. The
fourth equation would also equate to zero, except for Maxwell’s
displacement current term, and this is the key equation to
describe the radiation of radio waves.
Therefore, everything you need to know about an
electromagnetic wave can be condensed into two equations:
(Equation 3.3)
(Equation 3.4)
On an antenna the situation is different. In order to create
the fields in the first place, we must either have an accelerating
charge (current). As we will soon discover, a moving charge
with constant velocity is equivalent to a dc current and will
thus create only a static magnetic field around the conductor.
102
However, an accelerating charge (current changing amplitude
and/or direction) over time creates waves that radiate. An
antenna must be made of matter and most are constructed
from conducting material, such as metals. But how do we know
these propagating fields constitute a wave?
The wave equation had been known since the 18th century
(see Chapter 1) when it was derived to describe the harmonic
motion of a plucked string. Any phenomenon that takes the
form of a wave must conform to the wave equation, a second-
order differential equation of the form:
103
(Equation 3.5)
In order to actually solve this equation for a given set of
physical circumstances, we would need to be able to solve a
differential equation. However, we only need a brief
explanation to understand what this equation means. Also the
derivation of this equation requires some rather sophisticated
calculus. I will omit this derivation and simply refer the reader
to most advanced texts on electricity and magnetism (E&M) for
the formal derivation.
The wave equation defines exactly what is happening in the
Y-axis (where the instantaneous wave’s amplitude exists). The
first term
104
tension on the string and the mass of the string/unit length.
Anyone who has ever played a stringed instrument (Figure 3.7)
knows this relationship: If you tighten a string, you get a higher
frequency (note) where
105
Many readers will recognize this equation as defining the
relationship between a radio wave’s wavelength and frequency
where the velocity is usually taken to be constant (the speed of
light, c). In the case of a violin, v is also constant and we adjust
the tone (frequency) by changing the λ by shortening the
vibrating string with our finger.
The other term that describes velocity is the mass of the
string (more precisely the mass per length). The relationship
between the “tightness” of the string, the mass per length of the
string, and the velocity is
(Equation 3.6)
where σ is the “tightness” of the string and µ is the mass/length
of the string.
Strings on musical instruments optimized for lower
frequencies (notes) will be bigger (higher mass per length). If
we convert our “tightness” term into “looseness” (tightness
being the reciprocal of looseness) then the equation becomes
(Equation 3.7)
Or,
(Equation 3.8)
Also if we define “looseness” as ε, we get
106
(Equation 3.9)
Now that we have a velocity term for the mechanical version
of the wave equation we can investigate the similarity (and
differences) to the electromagnetic wave equation.
(Equation 3.10)
Again, the third equation is:
(Equation 3.11)
The derivation of the wave equation from Maxwell’s
equations is quite complicated. For completeness, we present
some of the important steps in the process. However, again, the
important point is to understand that the third and fourth
equations lead directly to a wave equation.
From the third equation we derive the following:
(Equation 3.12)
Here we start with Equation 3.11 and equate it to Equation
3.10 by using the reciprocal to the closed integral, or the
107
differential. The integral (or anti-derivative) and the
differential (or derivative) represent the two main divisions of
calculus that solve problems by working in different directions.
A rough comparison would be multiplication and division that
solve for different circumstances in opposite mathematical
“directions.”
Also if we take the second derivative of
(Equation 3.13)
If we take the partial time differential
(Equation 3.14)
Therefore, from Equations 3.13 and 3.14 using simple algebra
we derive
108
(Equation 3.15)
And finally from Equation 3.9 we get
(Equation 3.16)
This equation is exactly the form in Equation 3.5 and
therefore satisfies the wave equation. Thus the E field (usually
defined as a sine function) is a wave. Also, by Maxwell’s fourth
equation we know that an accelerating E field creates a B field
and so on. Thus electromagnetic radiation must be a wave. Again,
we skipped much of the complex mathematics to arrive at this
conclusion for simplicity.
In Chapter 1 we described how Maxwell derived the speed of
light in the form:
(Equation 3.17)
This form of the velocity term has now been shown to be a
direct analogy to the wave action on a string. However, rather
than describing μ and ε as a mass/length and as “looseness” of
the string’s physical condition, we now define space itself in
terms of different physical conditions: permeability and
permittivity. As Einstein later postulated, the presence of a
field “bends” space, so an oscillating field “oscillates” space
itself, in space-time. There is no matter in free space (including
the imaginary “ether” once thought to exist); therefore space
itself is the medium for electromagnetic waves.
As Maxwell rightly credited Faraday for the breakthrough
109
that led to Maxwell’s theory of electromagnetism, Einstein
rightly credited Maxwell with the foundation of his relativity
theory. Einstein, in turn, discovered the photoelectric effect,
which is the mechanism for the solar cells we discussed in
Chapter 2. Also, he showed that light has momentum, and
therefore light is also a particle! The particle of light (or any
other wavelength of electromagnetic radiation) is called a
photon. When a photon has enough momentum (energy) it will
dislodge an electron, creating the photoelectric effect. This
effect is much like a billiard ball striking another billiard ball,
but at the sub-atomic level. This leads to the peculiar fact that
electromagnetic waves exhibit a wave-particle duality. In
Chapter 10 we will explore this duality in some more detail.
However, we know it is a wave because it conforms to the wave
equation, and it is also a particle because solar cells would be
impossible without the photon.
Thus Einstein played a pivotal role in formulating the basis
for the two great subdivisions of modern physics: relativity and
quantum mechanics. You may be pleased that for antennas
operating in the RF portion of the spectrum, Maxwell is all that
is required for even an advanced intuitive and practical
understanding. Maxwell can be thought of the “bridge”
between Newtonian classical physics and modern physics
pioneered by Einstein (and of course many others).
110
forms of only the third and fourth equations. Again after some
rather difficult calculus we can re-write the third and fourth
equations as:
Maxwell’s third equation
(Equation 3.18)
and Maxwell’s fourth equation
(Equation 3.19)
As before, we can eliminate the Ampere current term, here
represented by J. So the differential form of Maxwell’s fourth
equation simplifies to
(Equation 3.20)
This form of differential equation uses the (∇) operator also
known as the curl. A simple intuitive demonstration of the curl
is called the “right-hand rule.” If you shape your right hand as
shown in Figure 3.8, you can see clearly the curl result. The
thumb points in the direction of the accelerating charge
(function of time), for example from Equation 3.18
111
and ∇ E (the curl of E) is represented by the four fingers
encircling (curling) around the thumb line. These two
equations show that the opposite fields tend to wrap
themselves around a concentrated accelerating field. When a
field becomes a plane, as in a “plane wave,” the opposite field
cannot wrap itself around, and thus it also becomes a plane
wave.
There is another important result to the curl. Notice the
direction the thumb points relative to the direction the fingers
point. This implies, correctly, that the E and B fields are always
orthogonal, having a 90 degree difference — ie “vertical” and
“horizontal.” Again, the equations use very sophisticated
mathematics, but express a very simple concept.
For mathematical correctness, these equations are presented
using some different terms. For our purposes H is roughly
equivalent to B, D is roughly equivalent to E, and J is roughly
112
equivalent to µI. Since we are exploring only the conceptual
understanding of these integral and differential equations, we
may upset some advanced mathematicians by being sloppy!
See Figure 3.9. From Maxwell’s equations we find that
electromagnetic radiation occurs from accelerating charges. A
charge moving at a constant velocity (no acceleration) will
create a static magnetic and electric field, as from a dc current
in a solenoid: constant current, constant charge. From classical
physics we know that acceleration is the time derivative of
velocity or, A = dV/dt (Appendix A). For radiation to occur, the
velocity of the charge must change (usually at the frequency of
the RF signal). Charge that changes velocity (accelerating) at a
time-dependent value of a sine wave will produce a field
directly analogous to the same sine wave. Since current (I) is
moving charge, a changing rate of moving charge is the cause of
radiation. Thus antenna current is of fundamental importance
for determining the key parameters of how an antenna
behaves.
113
(the speed of light) in the X direction, or the direction of
propagation. As a second step to the intuitive right-hand rule,
image that the fist throws a punch. The direction of the punch
(the direction of propagation) is orthogonal to the direction of
both the thumb and the fingers.
The fields that propagate outward from the charge’s location
become detached at their source, since polarity is changing at
the rate of twice the RF frequency (Figure 3.10). The result is
that we create “waves” of electric and magnetic fields, not a
constant static field. The electric field can be compared to a
soap bubble first appearing on the surface of soapy water from
a bubbling air source underneath. The bubble expands until it
finally detaches from the water, propagating away from its
source, “up,” only to be followed by the next bubble already
being formed. Eventually a series of bubbles is formed. Of
course bubbles tend to become small spheres, while the shape
of the field “bubbles” will conform to the directivity of the
antenna. As they propagate away from the antenna, they
converge into “plane waves” as the wavelength becomes much
greater than the distance propagated as along the expanding
sphere.
114
Impedance of Free Space
Maxwell’s equations provide the relationships (direction and
115
ratio) between the magnetic and electric fields associated with
an electromagnetic wave. The ratio of the two fields’
amplitudes remains constant and the two fields are always
orthogonal, at least in the far field. This is the condition for
plane waves. As we will show later, there are exceptions to this
general rule.
Since the ratio of these two fields is constant in a free-space
electromagnetic wave, those familiar with Ohm’s Law (Z = E/I)
will correctly postulate that some type of resistance is present,
but in free space?
In order for propagation of electromagnetic waves to occur in
any medium (including free space), there must exist a finite
intrinsic impedance. If the impedance is zero, or if it is infinite,
the magnetic and electric fields would have to be either infinite
or zero. If one or the other cannot exist, there can be no
electromagnetic radiation. In order to have electromagnetic
radiation in free space, free space must have some finite real
impedance, greater than zero but less than infinity. Indeed, the
universe would be a very different place without a finite
impedance of empty space!
We have already defined two physical characteristics of free
space, permeability and permittivity. That should give us our
first clue that there is more to free space than just these two
simple terms, but we begin with these terms, and use some
algebra and equations from fundamental classical physics.
Ignoring the coefficients for simplicity, we show how we
derive ohms (a pure resistance Ω) from the permittivity and
permeability of free space:
(Equation 3.21)
and thus
116
where F is Farads, H is Henrys, and J is Joules where
J = Nm
(Equation 3.22)
where N is Newtons (a physical force) and m is meters. Thus we
can actually relate linear distance and physical force to the
electromagnetic terms that describe free space.
From Chapter 1 we found that
(Equation 3.23)
Then
117
(Equation 3.24)
This is the ratio of the values of the electric and magnetic
fields, again comparable to the ratio of voltage and current
equating to resistance.
We can formalize the relationship between the speed of light
in free space and the impedance of free space by:
(Equation 3.25)
Because
(Equation 3.26)
consequently
(Equation 3.27)
and if
(Equation 3.28)
then
118
(Equation 3.29)
must be Z0.
Plugging the coefficients of these terms back in, we can
calculate the actual impedance of free space:
(Equation 3.30)
The speed of light is related to the impedance of free space by
the simple term 1/μ0. Since 377 ≈ 120 π, Z0 often appears as
either term.
Notice the similarity to Ohm’s Law, in that H roughly
coincides with current and E roughly coincides with voltage.
Also, for any space occupied by matter rather than a vacuum,
the values of μ and ε are somewhat larger. Thus the speed of
light inside the material and a material’s intrinsic impedance
are calculated by the same simple equations. This will become
an important concept in later chapters.
For most practical purposes, air’s terms are sufficiently close
to a pure vacuum to be ignored. The exceptions are almost
always at frequencies above about 1 GHz, where the size of
raindrops approach the operating wavelength, or at
frequencies where air molecules (particularly oxygen) render
the air as a partial absorption medium. The atmosphere can
also affect the propagation of nearly all frequencies, but the
antenna itself is seldom directly affected by its operation in air.
Above, we derived the impedance of free space from a
derivation of Ohm’s Law. We now have an electromagnetic
specification for free space! Such a specification is necessary to
quantify how we couple RF power into free space and how we
capture RF power from free space. In effect, free space becomes
part of the entire radio system “circuit.” As such, this number is
critical to all aspects of antenna design and performance.
At this point we simply accept that “space” indeed has
physical characteristics, even when there is no matter present.
119
To explore this non-intuitive reality, classical physics fails and
the curious are forced into the non-intuitive and often bizarre
world of modern physics. People often feel a need to
understand fields and radiation. We cannot experience first-
hand electric, magnetic, or even gravitational fields. We can
only observe the effects they have upon matter. However,
science allows us to quantify the actual physical characteristics
of these non-sensual phenomena. The proof of our physical
understanding is using this understanding to duplicate
experiments and then actually utilize this physics by
engineering radio equipment, including antennas.
(Equation 3.31)
Above we see that the relationship between field strength
and power is really a special calculation of the power law, but
also involves a dimensional transformation. The direct
calculation of power density from a field strength requires
knowledge of the exact relationship between the field strength
and power. For this we need a combination of the power law
and Ohm’s Law. The form we base this calculation upon is
120
(Equation 3.32)
In this case Z0 is the impedance of free space (120πΩ), E is the
electric field strength, (volts/meter) and Sr is the power density
(watts/meter2) both at the point of measurement at distance r
from the source, on the sphere where the measurement takes
place. Again, this is simply a combination of the power law and
Ohm’s Law.
(Equation 3.33)
or,
(Equation 3.34)
For example, we can calculate the effective radiated power
from an isotropic (EIRP) antenna in the direction of the
measured point. If we can measure the field strength at a
distance r (Er) we find the power density by Sr = Er2/Z0, and
then multiply Sr by the total area 4πr2. Thus the total
transmitter power (EIRP) is:
(Equation 3.35)
Thus we can calculate through these simple equations the
relationships among power density, EIRP, field strength,
and/or the impedance of free space.
We have to know that the transmitter is using an isotropic
antenna to make the isotropic assumption. In any event, if we
121
know the distance from the transmitter and the field strength
at that distance we can directly calculate the EIRP at the point
(also defining the direction from the transmitter) where we
took the measurement despite the directivity of the antenna.
Furthermore, if we also know the power of the transmitter, we
can also calculate the gain of the transmit antenna, in dBi. So
this is just an extension of the intuitive reasoning from Chapter
2 to now include fields into our arsenal of terms.
Although the energy term is not often used in conjunction
with antenna problems, we can complete our explanations by
defining the term. Energy is defined as the time integral of
power. For example, you are billed for electric energy, not
electric power. Electric energy is defined in watt-hours, or
more conveniently kW-hours. Electromagnetic energy can only
exist as a radiating set of fields propagating at the speed of
light. If we have a 100 W isotropic light bulb burning for 1000
hours in free space, then the total radiated electromagnetic
energy will be 100 kWhr. This energy will be contained within
a sphere with a radius of 1000 light-hours, at least for that
particular instant in time. So electromagnetic energy is not only
a time integral, but also a space integral. Where the E and H
fields are measured in m, power is measured in m2 and energy
in m3, in this case the volume of an expanding sphere
Radiation Resistance
We will encounter several terms in antenna physics whose
very title can imply an incorrect intuitive understanding. The
term radiation resistance is often a source of confusion. It is easy
upon first encountering this term to interpret it as a resistance
to radiation. In other words, the higher the resistance the more
difficult it becomes to radiate. This is an understandable yet
122
mistaken interpretation of the definition of radiation
resistance. Unfortunately the confusion does not stop with the
title. We introduced this term in Chapter 2. Now we will
expand upon it, and specific applications will be presented in
later chapters to further improve understanding.
If we place an RF voltage source on a conductor (antenna)
and measure its impedance, we will see the real part of the
impedance (a resistance), and, likely, a measureable reactance.
This is the case even for a theoretical lossless conductor,
perhaps constructed of a super-conducting material. How can a
zero resistance conductor register some real value of
impedance? The answer, of course, is that power is being “lost”
to free space by radiation. The resistive value of an antenna at
a specific frequency that is associated with radiated power is
called the radiation resistance (Rr) of the antenna.
The formal definition of radiation resistance is very often
misunderstood, even by professional radio engineers. Rr is a
critical term for calculating antenna efficiency which, of course,
directly affects effective aperture, gain and overall antenna
efficiency.
Antenna efficiency is defined as:
(Equation 3.36)
where (Rl) is loss resistance, or resistance resulting in RF power
conversion into heat. This is a simple, straightforward
equation, but the determination of Rr can be complicated and
therefore often mistaken. It is obvious that you must use the
correct antenna Rr value to get a proper calculation of
efficiency. But how do you determine the correct Rr value? In
Equation 3.36, how can you determine the difference between
Rr and Rl?
In the most basic model, Rr, being a resistive value, may be
123
calculated through the power law where
(Equation 3.37)
where Pr is the radiated power and I is the current at a point
along the antenna. However, this only provides the Rr at that
point on the antenna, not the antenna as a whole, which is
necessary for a proper calculation of efficiency. The first step
toward a proper definition of Rr is to measure I at a current
maximum point along the antenna. Second, to calculate Pr it is
necessary to integrate (sum) all the power being radiated
through the sphere. The process is relatively easy if the
antenna and wavelength are small. In such cases the far field
Poynting vector may be integrated inside an anechoic
chamber. For antennas and wavelengths too large for practical
chamber sizes, power integration is exceedingly difficult,
especially when the antenna is mounted close to the ground or
other material that causes power loss.
On the other hand, I is a straightforward measurement at the
point of current maximum. Thus this equation has three
variables, two of which are very difficult to directly measure.
Consequently it is difficult to set up a closed form equation that
correctly defines Rr using Equation 3.37. Furthermore, this
equation does not explain other factors that shape the value of
Rr. Therefore it cannot explain why Rr can change, even if Pr
remains constant but I changes. To resolve this problem, Kraus
gives a more refined definition for the general definition
(Equation 3.38) where an antenna’s all-important 3 dB
beamwidth (gain) is included in the definition. The idea is that
the total power radiated will usually closely approximate the
solid angle that is defined by the 3 dB beamwidth. In effect, the
assumption is that the power remaining outside the 3 dB
beamwidth will “fill in” the solid angle, as shown in Figure 2.8
in Chapter 2 instead of the “actual” pattern shown in Figure
124
3.11.
(Equation 3.38)
Equations 3.38 and 3.37 are approximately equal. The (max)
term in S(θ,φ)max designates a specific area on the sphere: the
steradian value (Ωa) of the 3 dB beamwidth of the maximum
gain lobe. In other words, S(θ,φ)max is the area on the sphere
that contains the main lobe of the antenna pattern. When the
antenna pattern is almost isotropic (nulls are less than 3 dB
difference than the peak), then W is the same as an isotropic or
the 4π solid angle defining the entire sphere’s surface. If the
125
gain pattern of the antenna is a distinct lobe (like the pattern
from a Yagi-Uda antenna), W would be defined as the solid
angle containing the maximum gain point and surrounding
area within 3 dB of the maximum gain point.
On the other hand, Equation 3.38 is precisely equal to
Equation 3.37 when the steradian contains precisely all
radiated power as in Figure 2.8 (in contrast to Figure 3.11). For
example, let us assume that the main lobe has a maximum gain
of 10 dB, which is also a linear power gain of 10. However, this
maximum gain location typically defines a point on the sphere.
Around this point we must define an area that is defined by the
steradian value. Let us assume that this area corresponds with
the gain number of the steradian value, or 1⁄10 the area of the
sphere, or
126
the terms are being influenced by antenna gain and/or loss,
whereas Equation 3.38 isolates these two terms.
If we know the total power radiated from a lossless antenna
and its current maximum, we can directly calculate Rr by
Equation 3.37. However, care must be taken in the computation
of this current value. For example, if we feed a folded dipole
(Figure 3.12) at a current maximum, we are only measuring 1⁄2
the current flowing at that “point” (compared to the
wavelength the “folded element” is essentially at the same
points along the antenna as the fed element). This will give a
false Rr value of about 292 Ω rather than the correct Rr for a 1⁄2
λ dipole in free space of about 73 Ω. Therefore, again, Rr is not
necessarily equal to a power law calculation taken at the antenna’s
feed point. For any arbitrary point on the antenna, the radiation
resistance at that point is Rrp.
127
antenna. It is a very common mistake to automatically equate
the feed point impedance (real value) to the antenna radiation
resistance measured at a current maximum or worse, elsewhere
along the antenna.
Rr will coincide with the real part of the feed point
impedance at the center of an antenna that is less than or
equal to 1⁄2 electrical wavelength long, or an antenna that is an
odd number times 1⁄2 wavelength. For all other antenna types,
great care should be used when calculating or measuring
radiation resistance, since simply using power input and feed
point current does not necessarily yield Rr. Indeed, the
difference between the feed point impedance Rrp and Rr can be
very large. For example, a single wire 1⁄2 λ dipole in free space
has the same value of Rr no matter where the feed is connected.
The only point where the two values coincide is at the center of
the dipole, which is also the current maximum. Rrp = Rr only at
a current maximum on the antenna. Rrp is usually associated
with a feed point that is not at a current maximum.
The resonant center-fed 1⁄2 λ dipole has an input impedance
of about 73 Ω, which is also the Rr of the antenna. Now if we
move the feed point impedance to 25% the distance from one
end (off center), the feed point impedance at Rrp is now about
137 Ω, but the antenna Rr remains 73 Ω.
Another complexity: In antennas longer than 1⁄2 wavelength
the current distribution (and thus the radiation resistance) can
be changed by changing the feed point location. So, when
calculating and/or measuring the location(s) of current
maximum(s) along an antenna element, be careful that key
terms that define Rr are not changed by changing the feed
point. Thus if the current is to be measured at a current
maximum by moving the feed point, the current maximum may
be moved as a result of moving the feed point! See Figures 3.13
and 3.14.
128
129
There are few applications where Rr becomes so critical as it
does in ground-mounted vertical antennas and antennas that
are very small compared to the wavelength of operation. This
will be treated with great detail in later chapters. For any
experimenter working with antennas generally and these types
of antennas in particular, it is essential to master the proper
definition or radiation resistance.
Now let us return to Equation 3.38. Assume that the 3 dB
three-dimensional beamwidth (equivalent area) is
In other words, the area is 1⁄8 the total surface area of the
sphere. Since we know our steradian value, we also know we
have a power gain of about 8, or about 9 dBi. We carefully
measure our feed point current as 1 A.
(Equation 3.39)
Thus Rr is inversely proportional to the antenna gain and
therefore also approximately proportional to the inverse of the
3 dB beamwidth. Notice that we have isolated and precisely
defined these terms by including antenna gain (or aperture) in
the equation. Measured deviations from these predicted values
indicate loss and/or mistakes in the measurement. As such, it is
a far more useful general definition than Equation 3.37.
Regarding ground-mounted vertical antennas, a yet more
130
special form is presented later.
As a final thought, it is interesting to note that if we
extrapolate the characteristics of a 1⁄2 λ dipole with infinitely
thin wires in free space (gain and Rr) to a hypothetical isotropic
antenna, we would find the Rr = 119.9 Ω, or exactly Z0/π, where
Z0 is the impedance of free space. To derive this number for the
isotropic, we simply adjust the gain from 2.15 dBi to 0 dBi and
the consequence of this adjustment to the steradian value.
Once again we see an example of the beautiful symmetry of
antenna physics, particularly when we use proper terms that
lead directly to closed-form equations.
131
“real” antenna at an arbitrary length.
PADL
The phase and amplitude of these accelerating charges from
each of these short dipoles (antenna segments) contribute to
the total pattern of the entire antenna in the far field. There
are also two other critical parameters that define the resulting
far field: the direction of the accelerating charges and the 3D
location of these accelerating charges. Thus we need to know
four variables to determine all the parameters of an antenna’s
far field: Phase, Amplitude, Direction, and Location — PADL. The
relationships of the electric field, magnetic field, and direction
of propagation become deterministic depending upon the
PADL variables.
Returning to the infinitesimal dipole antenna, the phase is
simply the point along the RF wave defining that phase point
(0-2π radians) along the RF sine wave. The amplitude is the
instantaneous RF current amplitude. The instantaneous
direction of the current is always the same (defined by a point
on the antenna element) changing twice per RF cycle and thus
reversing the polarity twice per cycle. There are only two
directions the current can accelerate on a linear element: left
or right (as viewed from the side). The location of the
infinitesimal dipole is a point in free space. All calculations for
any antenna in free space simply sum the PADL terms (using
integral calculus) to form a far field pattern. But first we need
to determine the pattern of an infinitesimal dipole so we will
know what to sum in more complex antenna structures.
Ignoring the integral calculus to derive this important term,
we find that the linear (power) pattern of the infinitesimal
dipole follows a curve defined by a sin2 function. The pattern is
a doughnut pattern very similar to the 1⁄2 λ dipole, but has less
directivity. The directivity of an isotropic antenna is 1.
Therefore the dBi directivity of an antenna is simply:
(Equation 3.40)
132
where again,
Gain = Directivity × Efficiency (expressed as fraction)
(Equation 3.41)
See Table 3.1 and Figures 3.15 to 3.17.
133
So, as we build a full-size 1⁄2 λ dipole from many short dipole
segments, we find we increase the directivity by about 0.42 dB.
The amplitude of the current along a 1⁄2 λ dipole is sinusoidal,
and the amplitude of the current in the short dipole is assumed
to be constant along its length (a small portion of a sine wave).
Therefore, the amplitudes of each of the long series of short
dipoles will have to be adjusted to the sinusoidal value at the
134
same point on the 1⁄2 λ dipole in order for the definite integral
to provide the correct result. For example, the current
amplitude value will be maximum at the segment at the center
of the 1⁄2 λ dipole. As we move away from the center, the
segments’ current amplitudes will decrease as a sin function,
and at the ends of the dipole, the segments will have close to a
zero value.
The concept of the short dipole segmentation of larger
antennas is very convenient for direct calculation of patterns
by hand (using calculus) or by advanced mathematical
modeling tools such as MATLAB. Antenna modeling tools such
as EZNEC also use segmentation for equivalent calculations and
as shown in Figure 3.18.
135
Although finite-length segments used in antenna modeling,
by definition, offer only an approximation, surprisingly few
segments offer results very close to results obtained through an
exact solution from integral calculus. Thus we begin with a
formal definition using calculus, but then discover that
adequate accuracy can be derived from much simpler summing
techniques.
The total antenna pattern is simply the sum of these smaller
patterns at a far distance from the antenna (shown in Figure
3.19). The doughnuts also contain phase information that can
lead to either field reinforcement or cancellation in the far
field. Therefore, the complete antenna pattern is simply a
smoothed sum of the PADL parameters. Of course, the
resulting fields from each of the segmented currents are
assumed to propagate at the speed of light. All antenna
patterns are thus the result of the current distribution in
amplitude and phase at all points on the antenna and the
position of those points in 3D space. In the case of a 1⁄2 λ dipole,
none of the doughnuts are far enough apart to provide
complete cancellation. As the length is made longer than 10/8
wavelength, the phase shift along the antenna is enough to
rapidly increase nulling in the broadside direction.
136
changing in accordance with forward and reflected waves
traveling up and down the elements. For this discussion we will
only consider the RMS values.
Again, if we consider a simple center-fed 1⁄2 λ dipole, current
will always flow in the same direction on both sides of the
dipole. We assume that the transmission line is applying equal
but out-of-phase current to the antenna’s center. Since the
wires point from the feed point in opposite directions, the
instantaneous currents run in the same direction (Figure 3.20).
137
sinusoidal due to a finite diameter of the linear elements.
However, for our purposes, we will assume a sinusoidal
distribution.
Voltage is represented as “force” per charge (coulomb). As a
charge is accelerated down a linear antenna element, it
eventually comes to the end of the antenna element. Since the
charge cannot continue flowing, the density increases and this
creates a high voltage point. An analogy is the displacement
current, which is simply charging and discharging a capacitor.
If the “end” of the antenna is thought to be a capacitor (which
it is!) then accelerating charges stop, before being reflected at
the antenna end, and create a oscillating charge in the
capacitor where Vc = q/C, where Vc is the voltage a the antenna
element’s end, q is the total charge (large) and C is the
capacitance value (small) at the antenna element’s end. By this
equation, assuming the end of a thin dipole represents a small
capacitance and the amount of charge can be very large, it
becomes apparent that very high voltages will exist even for
modest transmit power.
When a charge exists, but there is no current, there must be a
capacitance involved. The instantaneous voltage at the end of
the dipole increases while current is flowing toward that end,
and reduces to near zero when the current flows in the
opposite direction. Thus, if we consider the ends of the dipole
to be capacitors, the voltages on these two capacitors charge
and discharge in an out-of-phase relationship. As we move
away from an end, RMS current increases and thus the voltage
decreases. The current flowing on a dipole can be considered to
be Maxwell’s displacement current as it never completes a
“loop” as in Gauss’s original equation.
Not surprisingly the center feed point resistance (in this case
also very close to Rr) is simply
138
A general assumption is that the maximum voltage appearing
on an antenna’s element appears at its end point. This
assumption is helpful in developing an intuitive understanding
of voltage and thus current distribution along a linear antenna
element. The sinusoid distribution of antenna I and V are
usually 180 degrees out-of-phase along a linear antenna
element (Figure 3.21). Thus you can estimate voltage and
current peaks and nulls by working back from and antenna’s
end point (if there is one!) following a sinusoidal distribution at
the assumed RF wavelength.
Antenna Polarization
Thus far we have assumed that our antenna is linear, an
antenna element in a straight line. Such an antenna is said to
have linear polarization. In a linear polarized antenna, the
current can only flow along that line, either one way or the
other. The resulting radiated fields’ directions, according to
Maxwell’s equations are exactly defined: The electric field (E) is
parallel to the direction of accelerating charge (parallel to the
wire). The magnetic field (H) is perpendicular to the line of the
wire. See Figure 3.22. In “outer” free space a linear polarized
wave’s direction is arbitrary because there is no reference for a
line in space. Near or on the Earth, the Earth’s surface defines a
plane. With terrestrial reference, on Earth if the E field is
vertical it is said to have vertical polarization. If the E field is
139
horizontal, it is said to have horizontal polarization.
140
The “hand” polarization is determined by viewing the CP
antenna from its rear and thus the direction of radiation. From
this vantage point if the field is rotating clockwise, it is “right
polarized,” if it is rotating counter-clockwise, it is “left
polarized.”
PADL thus also will define the wave polarization, mainly
which direction the charges are accelerating: up/down,
right/left, and or in a circle. An important antenna type called
a helix is a “natural” CP antenna and will be discussed in
Chapter 9.
141
CHAPTER 4
Transmission Lines
A comprehensive discussion of transmission line theory can
easily fill its own textbook. Rather than repeating volumes of
information that is readily available from multiple sources, we
present a short discussion of transmission lines as they relate
to physics presented in earlier chapters. The discussion then
evolves into a more traditional description. We have shown
that Maxwell’s equations form the basis of electromagnetic
physics. These equations can be used to solve not only antenna
and free space problems, but also problems involving the
transmission line.
142
excellent antenna design can be ruined by loss and pattern
distortion from a poorly designed or installed line.
Therefore the design goal of a transmission line is to
configure the PADL parameters (Phase, Amplitude, Direction,
and Location — introduced in Chapter 3) so that no radiation
appears in the far field from the line. If two conductors are
spaced << λ and the RF current is out-of-phase and equal in
amplitude, then minimum radiation will occur. In effect, the
entire B and E fields will be canceled according to Maxwell’s
first two equations at some finite distance from the line. This is
the essence of the PADL distribution in a transmission line.
The actual location (L) is irrelevant since at any location in free
space the conditions of phase/amplitude cancellation will be
the same.
Therefore, no matter which direction we “view” the
transmission line (from the surface of the sphere), we observe -
almost completely canceled fields and thus no radiation. If we
measure the input impedance of a perfect unterminated line,
there will be no real part of the impedance since Rr is zero. The
reactive values along an unterminated (or shorted) line change
between inductive and capacitive as we gradually change the
physical length of the line or the wavelength. Therefore we can
also use a length of transmission line as a capacitor, inductor,
parallel resonant circuit, or series resonant circuit.
An antenna typically radiates or receives RF power from
three-dimensional free space, but the transmission line
concentrates the power into a one-dimensional structure and
thus, in principle, does not permit the RF power flux to
attenuate due to dispersion (Figure 4.1)
143
Radiation Loss Rr
In the previous chapters we developed the concept of
radiation resistance as the term that defines the coupling of an
RF wave into space. If the goal of a transmission line is to not
radiate, then the Rr of the line should be 0 Ω. We are now
familiar with the characteristics of a 1⁄2 λ center fed dipole
antenna. The Rr is about 73 Ω when the wires on either side of
the feed point are aligned in opposite directions, or θ = 180° as
in Figure 4.2 and Table 4.1. As we rotate the wires from 180
degrees to smaller angles, Rr begins to drop. At 0 degrees, Rr
falls to nearly 0 Ω which is exactly what we want in a
transmission line: no radiation, because Rr = 0 W. This is a
simple derivation of how we can create a very effective radiator
144
from two conductors, and then by reconfiguring the same
conductors, we create a transmission line with no radiation.
(Of course all practical lines will have some small value of Rr
and thus exhibit some finite amount of radiation.) The
transmission line in this case is referred to as a “twin lead” or
“open wire line.”
Characteristic Impedance
Let us assume we have an infinite length two-wire line with
zero resistance, constant spacing between the wires, and the
wire diameters do not change for this very long line (Figure
4.3). The two wires each have an inductance/unit length, and
the two wires also form a capacitance/unit length. The line is
145
also assumed to be lossless, in that the resistance of the wire is
zero and the resistance of the dielectric is infinite. More often
the dielectric is specified as a conductance rather than a
resistance and also a dielectric constant (treated in detail in
Chapter 6) which is a function of the permittivity of the
material or free space.
146
there is some corresponding impedance to this line.
(Equation 4.1)
From Equation 3.30 we derived the equation for free space:
(Equation 4.2)
This term not only determines the characteristics of a
radiated wave, but also of a guided wave. If the E and B fields
are constrained between two conductors separated by free
space, then intuition suggests that the impedance of the line
will be a function of 120pΩ and the characteristics of the two
conductors.
There is a rather non-intuitive equation from Equations 4.2
and 3.21
(Equation 4.3)
Equation 4.3 is derived directly from Maxwell’s equations and
an important result is that the L and C distribution along a
uniform line is always reciprocal.
Note: there are two definitions for the term Z0 — the
impedance of free space and the characteristic impedance of a
transmission line. From Equation 4.4 through the rest of
Chapter 4, Z0 will represent the characteristic impedance of a
transmission line.
For transmission lines, the form for the equation:
147
(Equation 4.4)
is particularly attractive since the inductance and capacitance
per length are both well defined by the geometry of the line.
We will use this equation to find Z0 for the most common
transmission lines: two-wire open line and coaxial cable.
In Figure 4.3, if we switch off the source voltage (to 0 V), the
capacitance in the line will discharge at the same rate,
resulting in the same dc value for Z0. Behavior is also the same
for a continuously varying wave such as an RF sine wave, thus
transmission line characteristics are valid for all types of
applied voltages. This proves the line’s reciprocity.
148
(Equation 4.5)
where S is the spacing between the two wires and r is the
radius of the wires (we assume both wires to have the same
radius). And
(Equation 4.6)
Substituting these equations into Equation 4.4 we get
(Equation 4.7)
Notice that this equation for the impedance of an air core twin
line cable contains the impedance for free space,
149
(Equation 4.8)
where
(Equation 4.9)
and
(Equation 4.10)
or
(Equation 4.11)
where 377 Ω is the impedance of free space and 2.3 is the
conversion factor from ln to log10.
The log10 term is often the more familiar form. Sometimes
there is some confusion when using this equation. Rather than
using the radius of the wires, occasionally the diameter d is
used where
(Equation 4.12)
Therefore
150
(Equation 4.13)
These are the equations for the characteristic impedance of a
two-wire open line.
(Equation 4.14)
And the inductance/length is
(Equation 4.15)
Again we substitute these L and C terms into Equation 4.4 to
derive:
151
(Equation 4.16)
This can be simplified to the approximation
(Equation 4.17)
Thus we also have
(Equation 4.18)
εr represents the relative permittivity of the medium
separating the coaxial conductors, which is also the same as the
dielectric constant of the material where kε = εr / e0.
ke = 1 for free space. For any other dielectric, ke will be some
value more than 1 since er is always greater than e0. Typical
values of plastic insulators used in coaxial cable have ke values
near 2. For most practical purposes we can assume µ to be the
free space value of µ0, thus typically we only need be concerned
with ke. In a two-plate capacitor the capacitance is directly
proportional to ke. For example, if we have a two plate
capacitor of 100 pF with an air dielectric (ke = 1) and insert a
sheet of plastic with (ke = 2), then the capacitance is now 200
pF.
Also, for open wire lines that use a dielectric other than free
space, we use the same term
152
to adjust the equation for Z0.
The electromagnetic fields along a properly operating coaxial
line are constrained between the conductors by virtue of its
geometry. This is of great advantage in practice over open wire
line whose fields propagate well beyond the immediate vicinity
of the line. Only in the far field do the fields of open wire line
cancel. Open wire line requires installation some “stand-off”
distance from other objects to avoid distortion of the fields in
the line’s close proximity. Furthermore, by the reciprocity
theorem, coaxial cable also has much better rejection of
undesired RF sources, making it superior in spurious-free
receiver applications. Mainly for these reasons coaxial cable
has become the most widely used type of transmission line.
Velocity Factor
When εr increases by replacing the free space dielectric with
a material, usually a plastic dielectric insulator, the impedance
decreases. There is another effect upon the characteristics of
the line: the velocity of the wave along the line. We again
recall that:
(Equation 4.19)
We assume that µ will be very close to µ0 in free space,
therefore, the velocity along the line will be:
(Equation 4.20)
where c is the free space speed of light. The velocity factor of a
153
line is thus
(Equation 4.21)
For example, if we use a dielectric with a ke of 4, the velocity
factor will be 0.5, or 1⁄2 the free-space speed of light. A common
value for ke in coaxial cables is 2.3, and thus the result is a VF of
about 0.66.
The velocity factor of the line is related to the characteristic
impedance of the line only by
154
For open wire lines, we made the assumption that the
distance between the two wires <<λ. As the wires become more
separated, the “line” begins to radiate and gradually becomes
an antenna rather than a transmission line. Within a coaxial
cable, as the frequency increases the field lines inside the cable
begin to distort as the frequency is increased and the
separation of the conductors increases.
Transverse Modes
Transverse electric (TE) modes: the E field is perpendicular
to the direction of propagation, but at least a portion of the B
field component points in the direction of propagation
Transverse magnetic (TM) modes: the B field is
perpendicular to the direction of propagation, but at least a
portion of the E field component points in the direction of
propagation
Transverse electromagnetic (TEM) modes: neither the E or
B field lines are parallel to the direction of propagation (as
assumed in Maxwell’s equations).
Hybrid modes: non-zero electric and magnetic fields in the
direction of propagation.
When we consider a coaxial cable, as with open wire line the
assumption is that the separation between the conductors is
<<λ. We have also assumed (up to now) that the characteristic
impedance of the line is independent of frequency. This is not
quite so, especially as the separation becomes an appreciable
fraction of a wavelength. In such cases, the fields are
constrained by the boundary conditions within the waveguide
(or shield of a coaxial cable) resulting in a “bending” of the field
vectors (E, H, or both) within the dielectric or the space inside
the waveguide.
The implication is that coaxial cable for higher and higher
frequencies must become smaller and smaller to use the TEM
mode. This has three effects:
1) The attenuation of smaller cables is higher.
2) The attenuation for higher frequencies becomes higher in
a given coaxial cable.
3) Smaller cable reduces the power handling capability of the
155
line.
Therefore, as operating frequencies climb above 1 GHz or so,
waveguide becomes more and more attractive. Furthermore,
waveguide also becomes smaller (cross-sectional area) as the
frequency increases. At lower frequencies, waveguide becomes
impractically large for most applications and coaxial cable is
the usual choice.
In a coaxial cable operating in the TEM mode, the E and H
fields are perpendicular to each other and both are
perpendicular to the direction of propagation as shown in
Figure 4.7. Notice the H field “curl” around the direction of the
current.
156
Unlike waveguides, coaxial transmission lines do not have a
157
low frequency cut-off. As our first example of a battery and an
infinite line implies, a transmission line will exhibit its
characteristic impedance even at dc.
(Equation 4.22)
where R is the resistance of the conductors and G is the
conductance of the dielectric. ω is the angular frequency where
ω = 2πf (f in Hz).
Reflected Waves
Many students find the concept of reflected waves and the
resulting effects of these waves to be somewhat non-intuitive.
This confusion has been magnified by the plethora of published
misinformation on the topic. For this reason, we will look at
these very important parameters from several different
approaches.
In Figure 4.4 we showed how the characteristic impedance of
an infinite transmission line will set the V/I ratio of an applied
voltage of any frequency. Also if we terminate a line of finite
length with a load resistor with the same resistive value of the
line (characteristic impedance) we will see the same current for
the applied voltage. However, if the line is terminated by a
resistor of a different impedance, some of the accelerating
charges will be reflected back similar to the description of the
reflection of a wave on a dipole antenna in Chapter 3.
158
The result is identical to a free space wave striking an object
with a different characteristic impedance of free space. (We
introduce Snell’s Law in Chapter 6.) For a different
characteristic impedance there will be some reflection. This is
the principle of many forms of radio propagation and radar.
Radio waves “bounce” off of metallic objects such as aircraft,
disclosing their range (distance) and their direction. The
combination of these two parameters yields the aircraft’s point
location. Now imagine an aircraft coated with a very special
paint. This paint’s electrical characteristics show that it has a
resistance of exactly 377 Ω per square meter, the identical
impedance of free space. In effect the radar signal would be
completely absorbed, providing no reflection and rendering the
aircraft “stealth.” In effect, when the free space impedance of
377 Ω is “terminated” by an equivalent square resistance, there
is no reflection.
Returning to the guided wave, the same principle applies. An
impedance discontinuity results in reflection. Complete
reflection occurs at the end of a dipole since there is no
shunting impedance at its end. A similar condition applies for
two extremes at the end of a transmission line: a “short” and an
“open” termination where all power is reflected back to the
source. Any line terminated in an impedance showing some
value of resistance will absorb some of the power by that
resistance. If the load is a pure resistance, that resistance will
define the load resistance. In the case of a reactive portion of
the load impedance, the voltage and current will no longer be
in phase, and some reflection will take place. Use of the Smith
Chart and its applicability to determining reflection on
transmission lines has been published extensively, so I will not
duplicate that discussion. However, the Smith Chart is a very
convenient tool for such calculations….and more.
Generally, reflection on a radio transmission line is to be
minimized, since reflected power will eventually be wasted as
re-radiation and/or heat loss, both for transmitting and
receiving applications.
159
If a transmission line is terminated with the proper load
value, the RMS RF voltage will be equal at all places along that
line. Thus the ratio of the highest to lowest voltage is 1:1,
indicating a perfect “match” between the load and the line.
However, if the load impedance is not the conjugate match for
the line impedance, a portion of the traveling charges will be
reflected. The backward and forward-moving waves (measured
in RMS) will sum differently along the line. At some points they
will reinforce and at other points they will tend to cancel. In
this case, voltage maximums and minimums will appear along
the line, indicating a non-perfect match, or a VSWR of >1:1. If
the voltage maximum is twice the voltage minimum, the VSWR
is 2:1. If the terminating impedance is lower than the line
impedance, there will be a current maximum and thus voltage
minimum at the load and vice-versa.
Equation 2.24 in Chapter 2 is the applicable formal definition
of VSWR. However, we can form an intuitive understanding of
how the reflected wave creates a “standing wave” and also how
VSWR relates to reflected power. For example, let us assume
that the VSWR is 2:1 as is the case when a 50 Ω line is
terminated into a 100 Ω load. If the forward wave’s RMS voltage
is 50 V (50 W), then in order to achieve a 2:1 VSWR there must
exist a reflected voltage that will add to and subtract from 50 V
to form a ratio of two, but will be symmetrical around the 50 V
level, or 66.666 and 33.3333 V. The reflected voltage to create
these maximums and minimum must then be 16.66666 V.
Therefore the reflected power is
(Equation 4.23)
Therefore, with a 2:1 VSWR, 11% of the power is reflected. For a
transmit system the reflected power will be dissipated as heat
in the equivalent transmitter’s resistance, as in Chapter 2 in
Figure 2.22’s Rr.
Now we can summarize the critical terms:
160
1) A transmitter sends a 50 W incident wave of (50 V at 1 A)
down a 50 Ω line terminated by a 100 Ω antenna feed point
impedance.
2) A reflected wave of 16.6666 V RMS is sent back toward the
transmitter.
3) At points spaced 1⁄2 λ along the line the reflected 16.6666 V
adds to the forward 50 V resulting in 66.666 V at these points
(maximum).
4) At other points displaced 1⁄4 λ from the maximum V points,
the reflected 16.6666 V subtracts from the 50 V incident voltage
thus creating 33.3333 V at these points (minimum).
5) The VSWR is simply the ratio of the maximum to
minimum values, or 2:1. It is also the ratio of the Z0 of the line
to the Z of the load.
6) The power reflected back is as in Equation 4.23, or 5.56 W.
7) The power delivered to the antenna feed point is simply 50
– 5.56 = 44.44 W.
In a “real” line, one that has a finite amount of loss, the total
loss of the line increases as the VSWR increases. There are
several factors that determine this additional loss. First,
increased losses due to higher voltage and current at points
along the line will increase dielectric and ohmic losses at the
corresponding points. Second, since a portion of the power is
reflected back to the source, this reflected power will be
attenuated at the same rate as the forward wave. If the
transmitter side of the line reflects power it will again travel
back to the antenna, experiencing yet more loss. The ARRL
Antenna Book presents detailed graphs of increased line losses
as a function of VSWR for a variety of real lines.
For this reason, the VSWR is best measured at the feed point
of the load (antenna). For example, let’s assume that a 432 MHz
antenna’s feed line has shorted at the feed point. This would
indicate an infinite VSWR and the antenna would not radiate
or receive any power. Now assume a very long coaxial line
feeding the antenna with 10 dB loss in a matched condition.
The reflected wave would be attenuated by a minimum of 20 dB
(up and back). Therefore, if we drove the line with 100 W we
would see only 1 reflected watt at the source. We would
161
measure a VSWR of 1.22:1, indicating a very acceptable match,
but the antenna is shorted! If we measured the VSWR at the
shorted feed point, we would see the true condition, a near-
infinite VSWR.
The ratio of the reflected voltage to the forward voltage is
called the reflection coefficient where:
(Equation 4.24)
where Γ takes into account any complex components to the
load. The calculation of VSWR only requires the magnitude of
Γ, usually designated as “ρ.” Many texts only use the Γ
designation, however.
The formal equation for the above explanation of VSWR is
162
(Equation 4.25)
Let’s take two extreme cases and apply them to Equation
4.25: a perfect match where the reflected voltage Vr = 0 and an
“open” termination where all the voltage is reflected or Vr = 1.
(Vf) is taken to be 1. In the case where Vr = 0, by Equation 4.25,
VSWR = 1, or 1:1. However, if Vr = 1, VSWR → ∞, or ∞:1. Also in
the case of an infinite VSWR the maximum voltage along the
line is only twice the matched value, but the minimum voltage
is 0, therefore the ratio goes to 2/0 or infinity. See Figure 4.10.
163
CHAPTER 5
Driven Arrays
We will begin by using two simple examples that can form a
more intuitive understanding of how PADL parameters create
antenna patterns. Both examples use two parallel 1⁄2 λ dipoles
in free space and are separated by 1⁄2 λ (Figure 5.1). Both
dipoles are center fed from the same RF generator
(transmitter). Since both of the elements that constitute the
164
array are directly fed from the signal source, we call this a
driven array. In this first example the currents being fed to the
two dipoles are identical but we will use two phase
relationships: in phase (0 degrees) and out-of-phase (180
degrees).
165
166
The result is that we can provide switchable bidirectional
“figure 8” patterns by simply changing the phase relationship
between the two elements, again as shown in Figures 5.2 and
5.3.
The gain parameters in Figure 5.2 may appear a bit
conflicting upon careful investigation. The maximum gain is
5.94 dBi, or 3.82 dBi. This is counter-intuitive if we assume a
doubling of a 1⁄2 λ dipole aperture, the gain should be 5.15 dBi,
or 3 dB higher than that of the 1⁄2 λ dipole. Thus we must
explain: where did the extra gain and thus aperture come from?
Simple addition of apertures assumes completely
independent apertures and the added assumption that there is
no “duplication” of aperture coverage, like the shading of one
solar panel over another. If, however, antenna elements are
placed in close proximity, they become “coupled” and all their
independent characteristics change, including phase,
amplitude, and direction of the charge acceleration. Since these
terms directly relate to aperture, then aperture can change as a
result of the terms changing in a non-independent manner.
The intuitive process above serves as a general explanation of
how directivity is formed, but it ignores the critical subtlety
hinted above by inter-element “coupling” in multielement
arrays.
Mutual Impedance
Now we will consider this more complicated coupling
problem. Although multielement arrays can constitute driven
elements spaced at distances well into their far fields, we will
first consider near-field interactions. The far field and very
large aperture antennas will be considered in Chapter 9. In the
case of far-field coupling (through free space) there is no
measurable effect from one antenna to the other, only the
induced currents and voltages the transmitting antenna
imparts to the receiving antenna through free space. In effect, a
single antenna suspended in free space is immersed in a
medium defined by the intrinsic impedance of free space
(120pΩ). When we move the two antennas closer, approaching
the near fields of both, significant changes result in the new
167
“array.” In effect we have formed a new antenna with a new set
of parameters. The two elements become coupled not by a
conductor, but by virtue of their respective apertures.
In our example in Figure 5.1, some of the power radiated by
one element will be received by the other element; they are
not mutually independent. Because of the received power, the
element now has a modified voltage and current value: it is the
sum of the initial fed power and the received power from the
other element. This is called “mutual coupling” between the
two elements. Since the affected voltages and currents have
been affected in both elements, there must be some sort of
impedance involved.
From Figure 5.1 we now define the two antennas as antenna 1
and antenna 2 (it doesn’t matter which one is which, as they
are equal for this analysis). We now define the self-impedance,
which is also the feed point impedance, and also the Rr of both
the antennas independently. Therefore the self-impedance of
the two antennas is Z11 and Z22 respectively.
However, antenna 1 will induce power into antenna 2 and
vice-versa. The current, voltages and their phase relationships
define a mutual impedance as Z12 and Z21 respectively.
We can now determine the feed point impedances for both
antenna elements where
V1 = I1Z11 + I2Z21
(Equation 5.1)
In other words, the term V1 = I1Z11 is the voltage on antenna 1
as if it were the only element involved, and I2Z21 accounts for
the additional voltage term introduced by the close proximity
of fed element 2.
Since the two elements are identical and fed in-phase, we can
assume I1 = I2. Also, for the same reason, we can assume that Z11
= Z22 and Z12 = Z21.
Therefore the terminal impedance of element 1 is
Z1 = V1/I1 = Z11 + Z12
168
(Equation 5.2)
Also, because of the inherent symmetry of this simple array, Z1
= Z2.
Thus for this simple case, using two identical elements with
identical feeds (either 0 or 180 degree phase difference and
equal amplitude), we derive a simple equation for the new feed
point impedance, and, in this case, also the new Rr. We know
that the self-impedance (Z11) of a 1⁄2 λ dipole is 73 Ω. After some
rather rigorous trigonometry we derive Z12 at –13 + j14 Ω. Thus
Z1 = 73 – 13 + j14 = 60 + j14 Ω
(Equation 5.3)
Since there are two elements in parallel, the real part of the
feed point is 30 Ω where the two 60 Ω feeds are connected in
parallel. Again, for this simple case we can directly calculate Rr
as 30 Ω. We now provide a simple calculation that verifies this
value. We have already provided a general equation for Rr. In
Chapter 7 we will derive a more convenient form for
calculating the Rr of a ground-mounted vertical antenna.
(Equation 5.4)
Indeed, 2.433 ≅ 3.96/1.64 = 2.419, or 3.836 dBd, or 5.99 dBi.
169
This is very close to the modeled 5.94 dBi shown in Figure 5.2.
This is yet another indication of the reciprocal relationship of
gain (aperture) and Rr.
From this discussion there are two points which should be
made:
1) Often when modeling in EZNEC or other programs, exact
correlations do not occur with theory. This is mainly due to not
providing exact dimensions when modeling the antenna.
However,
is close enough!
2) Z12 is taken to be –13 + j14 Ω. This is not an error. Negative
resistances are commonly encountered when modeling
multielement arrays. It simply implies that more current is
flowing toward the source rather than flowing toward the
antenna due to the mutual coupling. Therefore in this case the
sum of the real part of the feed points is 60 Ω.
The array in Figure 5.3 uses a similar method of calculation.
As arrays become more complex, we can no longer make the
assumption of symmetry as in the simple case of Figure 5.1’s
array. Therefore the computation becomes more complicated,
but the principle shown here is the same. A natural progression
is to the quadrature array.
Note: The calculation of multielement arrays is usually more
complex than this simple analysis which is presented to show
conceptually how Rr relates to the other critical terms.
Quadrature Array
In the quadrature array (Figure 5.4), we use the same two
dipoles, as above. However, the element spacing is reduced to
1⁄4 λ (90 degrees).
170
The two dipoles retain center feed points, but now we will
define the phase relationship as 90 degrees rather than 0 or 180
degrees, with the fed amplitude (current) remaining identical
in both feeds. We can follow the same intuition as in the
examples using 1⁄2 λ spacing. In this case, element 2’s phase is
leading element 1’s phase by 90 degrees. By the time element 2’s
field reaches element 1, the phase of the current delivered to
element 1’s feed point has advanced by the same 90 degrees.
Thus, in the negative X direction (Figure 5.5) the fields add
completely and we derive maximum gain.
171
However, where element 1’s field propagates toward element
2, the phase of element 2’s current has advanced another 90
degrees, for a total phase shift of 180 degrees, and there is
complete cancellation in the positive X direction. Between
these directions of maximum and minimum radiation, there is
partial addition or subtraction of the fields, resulting in the
overall pattern shown in Figure 5.5.
Unlike the 1⁄2 λ spacing examples in Figures 5.1 to 5.3, this
array is non-symmetrical, in that Z12 ≠ Z21. Since the self-
impedances of elements 1 and 2 are identical at 73 Ω pure
resistive we can avoid very difficult computations by allowing
EZNEC to perform the necessary calculus by simply reading the
resulting Z1 and Z2 values, which are 120.3 + j76.95 and 52.85 –
j9.92 Ω respectively. Therefore we can now directly compute
the two different mutual impedances.
172
Z12 = (120.3 + j76.95) – 73 = 47.3 + j76.95 Ω
and
Z21 = (52.85 + j9.92) – 73 = –20.15 – j9.92 Ω
Parasitic Arrays
The usual configuration for a parasitic array (Figure 5.6) is to
employ one driven element and rely completely upon the
mutual impedance(s) between or among the driven element
and the parasitic element(s). In the case of multiple fed and
multiple parasitic elements forming one antenna, a
combination of these two analytical techniques becomes
necessary. The parasitic element usually employs the scattering
aperture, where the termination (center of a 1⁄2 λ linear
element) is a short, or 0 Ω.
173
The parasitic array analysis becomes quite different from the
driven array analysis. The first step is understanding that there
is a phase reversal from the driven to parasitic element. Since
like charges repel, then we can assume that accelerating like
charges will repel. Therefore the induced voltage and current
in element 2 will be 180 degrees out-of-phase from the inducing
field striking it. With this in mind, we can begin to quantify the
terms
V1 = I1Z11 + I2Z12
174
due to its own impedance is the negative value of the current
flowing on the driven element that is induced into the parasitic
element, so they sum to zero. Now we can describe the
fundamental equation for the action of a parasitic array, and its
important practical configurations including the Yagi-Uda.
175
To reiterate, the basic PADL principle applies to both driven
and parasitic arrays. Driven arrays are inherently more stable
since the characteristics are largely driven by the self-
impedances of the elements. Parasitic arrays, in contrast,
depend completely upon mutual impedances for their
excitation and very careful tuning of the self-impedances for
proper phase response. Finally, greater gain can be obtained
from parasitic arrays because of the much larger scattering
apertures inherent to parasitic elements. This is in contrast to
the quadrature driven array where the antenna apertures
simply add, thus providing only about 3 dBd gain. An optimized
two-element Yagi-Uda provides about 5 dBd.
This completes the first section of this book. Chapters 7 and 8
of the second (and last) section will build upon this and all the
previous chapters to apply these principles to understanding
specific array forms. However, the first “applied” chapter will
deal with the idiosyncrasies of antennas’ inter-reactions with
our own planet.
176
Part 2
Applied Antenna Physics
177
CHAPTER 6
178
happening. Although there are numerous physical mechanisms
for wave propagation due to the Earth, we will cover the two
most influential on forming the final gain pattern.
179
The good news is that these models often do provide fair to
excellent approximations of what a system may provide. They
are certainly better than using a free space model!
It is also fortunate that the actual geophysical terms relevant
to radio wave interaction can usually be reduced to two terms:
conductivity and permittivity. Therefore sufficiently accurate
models can be derived by reducing the Earth to these two
parameters. Therefore the subject of electromagnetic
geophysics (at least for radio waves and thus antenna
performance) is really the subject of how those waves interact
at the boundary between free space and the lossy dielectric of
the Earth as well as down to some depth.
180
(Equation 6.1)
The term m0I is the conduction current. However, if there is
a resistance in the current path (1/G), we can rewrite this term
as GE. Also the term
(Equation 6.2)
In this chapter we will assume the µ term in Equation 6.1 will
be m0 because we will only consider materials that have µ
values identical to free space. Under most circumstances, µ of
the Earth, from the surface down to about 10 meters, can be
assumed to be the same as free space or very close to it.
Therefore we can now write a ratio of the two terms as
(Equation 6.3)
where DF is the dielectric dissipation factor, G is the
conductivity (frequency independent), and ωkd is the frequency
dependent dielectric constant of the material. Conductors will
have very high conductivity, thus DF is >>1. Good dielectrics
(insulators) will have very poor conductivity, thus DF becomes
<<1. Some readers may correctly see a correlation with the Q of
a capacitor. Indeed,
181
(Equation 6.4)
Therefore we can determine the theoretical maximum Q of a
capacitor if we know the DF of the dielectric. A “perfect”
dielectric will exhibit a kd that is constant for all frequencies.
There are no materials that have this property. Only free space
has such a frequency-independent characteristic. The DF of soil
and rock is very complex and variable with temperature and
moisture level.
(Equation 6.5)
where ed is the permittivity of the dielectric. We can simplify
the relationship by multiplying the free space speed of light (c)
by a simple coefficient to derive the speed of light inside a
dielectric material
(Equation 6.6)
This is the most important effect upon an electromagnetic
wave inside a perfect dielectric (zero conductivity). The
frequency remains constant, but the speed slows and therefore
the wavelength decreases.
182
Snell’s Law
Let us assume that a free space plane wave strikes a dielectric
surface with a kd = 4. The angle of incidence is qi. The reflected
wave will be the “mirror image” of the incident wave. However,
the transmitted wave, propagating into the dielectric, will be
traveling at 1⁄2 c by Equation 6.6. Figure 6.2 shows the effect
upon the angle qi through the dielectric. Since the wave
velocity is now 1⁄2 c, the wavelength inside the dielectric is now
half the wavelength in free space. This has the effect of
changing the angle of the transmitted wave. If you imagine the
incident wave striking the boundary in slow motion, and the
wave slows down inside the dielectric, it is easy to envision the
cause of this transformation of θ where:
(Equation 6.7)
In other words, the ratio of the sin values of the angles is
equal to the inverse-square ratios of the dielectric constants of
the two medium. In this case the two medium are free space
183
and a material with dielectric constant of 4.
Notice that when the incident wave is striking the boundary
from a perpendicular direction, in this case, from straight up qi
approaches 0 degrees, but qi also approaches 0 degrees. How
can this be, when
(Equation 6.8)
The ratio of Pt and Pr is a function of the ratio of the dielectric
constants and the values of θ.
Inside an imperfect dielectric (such as soil), the situation
becomes more complicated. A finite value of conductance
implies some finite value of resistance, which in turn will cause
attenuation of a wave inside the material. This will be true for
directed as well as radiated waves.
184
(Equation 6.9)
In a dielectric, ε0 has changed to εd, where εd is always > ε0.
Inside a non-free space dielectric, a wave is still a wave, but
with two profound modifications: the speed of light has slowed
down and the intrinsic impedance has decreased where
(Equation 6.10)
and
(Equation 6.11)
Also from Chapter 4
(Equation 6.12)
and
(Equation 6.13)
where V is velocity.
185
dielectric. Vertical and horizontal waves interact with the
dielectric differently. This is particularly true of the
relationship between the phase of the incident and reflected
waves.
186
187
Vertical Polarized Wave Reflection
A vertical polarized wave’s E-field is perpendicular to the
Earth’s dielectric surface, or “vertical.” But upon careful review
there can be no “true” vertical polarization that is being
reflected. Figure 6.7 shows why this is true. The reflection
angle resulting in a 90 degree phase shift corresponds to the
minimum magnitude of the reflected wave. This is defined as
the Brewster or pseudo-Brewster angle. Figure 6.8 shows the
relative magnitude and phase of the reflected wave for vertical
polarization.
188
In the case of a 1⁄4 λ ground-mounted vertical, nulling due to
reflections is restricted to very low (grazing) angles. The point
source of radiation is simply too low to provide ground-
reflected waves at higher angles, so wave attenuation is
primarily due to ground attenuation and the requirement for
the values of the E-fields to remain constant along the
boundary of the dielectric and free space. Therefore, we do not
see ground reflections causing nulls and reinforcement in the
resulting pattern in Figure 6.9. Figure 6.10 illustrates vertical
polarized wave reflection.
189
Unlike a ground-mounted vertical, with the elevated 1⁄2 λ
vertical dipole, the center of radiation is located at a sufficient
height to provide very significant reflected waves, even at high
angles. Figure 6.11 shows a simulation of a 1⁄2 λ vertical dipole
at 2 λ over real ground.
190
Extrapolating this curve and using MF (medium frequency)
wavelengths, we can see the Earth’s curvature will serve as a
fundamental limit to how low the maximum gain elevation
angle can be (Figure 6.13). Figure 6.9 implies even more low-
angle attenuation over “average ground”. We will explore how
these terms relate to vertical antennas in the next chapter.
191
192
CHAPTER 7
Vertical Antennas
In Chapter 6 we discussed the physics at a boundary between
free space and a dielectric for radio waves. We also discussed
the differences between horizontal and vertical polarized
waves at these boundaries. In this chapter we will focus upon
the characteristics of vertical antennas, return to complete the
discussion of their relationships to the Earth, and then
conclude with comparisons among various vertically polarized
configurations.
Vertical antennas may be placed above the surface of the
Earth or mounted on the Earth proper. Ground-mounted
verticals actually use the Earth as part of the antenna, or as the
“image” of the antenna. Verticals that are mounted above the
Earth may also include the Earth as part of the antenna,
depending upon how far above the Earth the “lowest” part of
the antenna is placed. For example, a 1⁄2 λ vertical dipole with
one end just above the Earth actually may be considered a
ground-mounted vertical, although a direct connection to the
Earth will significantly change the antenna’s characteristics.
An antenna connected directly to the ground is a “driven
ground” array. An antenna located near the ground, but not
connected to the ground proper, uses the ground (for better or
worse) as a parasitic surface, but with very complex properties.
As will be shown, the gain response of driven and parasitic
ground antennas mirrors the difference in gain between driven
and parasitic arrays, as in Chapter 5.
193
systems for better overall antenna efficiency. For professional
AM broadcasters, very stringent ground requirements are
mandated so that minimum ground losses can be assumed. In
effect, the reality of compromised ground systems and
complete freedom of antenna design has resulted in much
more research in amateur than professional circles. In AM
broadcasting, antenna parameters have been established for
many decades and are highly accurate in practice. In AM
broadcast engineering, lossless ground is assumed and only
ground-coupled vertical elements are mandated. That makes
life much easier!
The general equation for radiation resistance, Rr, from
Chapter 3 is in an awkward form for easy use with vertical
antennas. Special attention must be given to Rr in ground-
mounted verticals, especially for amateur applications where
compromised ground systems are often used and/or the vertical
element is shorter than 1⁄4 λ.
(Equation 7.1)
There is an alternative and more useful method to determine
the equivalent of the power term in the general equations for
194
Rr in ground-mounted vertical antennas.
Effective Height
A critical term in our to-be-derived equation for Rr is he
which is the effective height of the antenna. he is another
potentially confusing term in that it is a function of the physical
height of a vertical antenna (including the ground image), but
should never be equated to it. It should also not be confused
with “electrical height.” Only in very specific circumstances
does he approach the actual physical height, hp. he is always less
than hp. The two terms can be equated using only the physical
height (without the image) in some specific case. The general
definition of he is
(Equation 7.2)
or
V = heE
(Equation 7.3)
The concept is quite simple. The voltage induced into an
antenna (of the same polarization) is proportional to the
incident E field and the “effective height.” The computation of
he is a bit non-intuitive for this simple definition. The
relationship between he and the physical height hp is simply
multiplying the average current along the antenna by hp and
then dividing by the maximum current along the same
antenna. For a linear vertical antenna whose he is greater than
about 1⁄8 λ, the current distribution will be sinusoidal.
Therefore, we need to calculate the average of a sine wave (or a
portion thereof) by integrating current over the physical
height. This integral simply computes the average value of a
195
portion of a sine wave. The portion of the sine wave to be
averaged is defined by two phase points along the sine wave.
(Equation 7.4)
In this equation we use two phase points α and β rather than
an arbitrary physical height. This is because the average
current can be different over the same physical length
depending upon which portion of the sine wave is present
along the length. We will show an example a bit later.
For computations of he it is convenient to take the maximum
current (I0) to be 1 A and the λ to be 1 meter to simplify the
equations. We then represent the physical height hp as the
appropriate fraction of λ, which we assume is 1. Therefore, the
term Iav/I0 is a dimensionless number which is less than 1. he
and Rr are independent of the power fed to the antenna, and
thus also independent of the actual current value or the actual
wavelength. We are only interested in the average current
referenced to the maximum current. This ratio is independent
of the amount of power applied to the antenna. Also he and Rr
are independent of λ as long as we scale the dimensions of the
antenna and thus the two phase values. For example, a 1⁄4 λ
vertical will show the same results no matter what the
wavelength, as long as the actual height is 1⁄4 λ at the actual
wavelength involved.
Using radian measure and integrating radian values for the
average current also permits us to use radian values as he For
example, a 1⁄4 λ vertical would be π/2 radians high.
A ground-mounted vertical antenna’s image is critical for
grasping a conceptual understanding of ground-mounted
verticals as well as setting up equations that define their
parameters. The image is considered to be an identical vertical
antenna that extends vertically underground, mirroring the
196
actual vertical element above ground. In the calculation of he
the image’s hp must be added to the hp of the vertical element,
or he of a ground-mounted vertical is twice the value of the
computed value for he of the vertical element taken alone.
There are some exceptions to this rule. In the case of
verticals higher than 1⁄4 λ, the image may actually be shorter
than the physical length. Also, for very lossy dielectrics (very
low conductivity and dielectric constant) that approach the
characteristics of free space, the image disappears and
significant radiation occurs into the dielectric. This radiation
must be taken into account when calculating Rr. The
disappearing image is due to the gradual decrease in ground
conductivity, which lowers the average ground current —
eventually rendering the image insignificant. Rl only accounts
for heat loss of RF current inside the dielectric, not radiated
power propagating inside the ground dielectric.
The power delivered to the load that is matched to a receive
antenna is:
(Equation 7.5)
From Chapter 2 we calculated that half of the power received
by an antenna is reradiated back into space and that only Ae is
significant for the transmit application. Now from Equations
7.3 and 7.5 we derive:
(Equation 7.6)
And from Equation 2.16
197
(Equation 7.7)
where
So,
(Equation 7.8)
Therefore
(Equation 7.9)
To confirm that this equation results in a resistance, we can
look at the dimensions within the equation, ignoring the
coefficients. Recalling Equation 7.4
therefore
(Equation 7.10)
198
The m2 and I2 terms cancel, leaving R. Therefore, this special
form for the general equation for Rr provides a pure resistance.
It is critical to understand that in a ground-mounted vertical,
the physical height (hp) also includes the image. Notice also that
the current squared term in the denominator is the maximum
current term, identical to the other two equations thus
presented for Rr. All three equations are equal or very close to
it.
199
We can now use Equation 7.9 to actually calculate the values
of Rr for these two cases.
200
loss is the same.
201
202
Unfortunately, these very encouraging graphs become
somewhat disappointing when real ground is substituted for
perfect ground, as we shall see.
203
(Equation 7.11)
where Pr is the licensed power output. Rrp is used since the
actual feed point may be a bit below the actual current
maximum. Often the broadcast tower is specified as physically
1⁄4 λ high (hp), which is about 5% higher than a resonant
vertical. If no losses can be assumed, however, the radiated
power can be calculated using the Rrp at any point and feeding
the appropriate current.
In contrast, for the average amateur installation, both hp and
the ground system will be limited by cost and/or the
availability of suitable space. Extensive empirical testing has
provided insightful approximations, especially when solidly
based upon physical principles. The usual intent of all this
work centers on the improvement of antenna efficiency.
Directivity can usually be assumed by standard models as there
is generally insignificant modification of directivity due to
antenna ground losses in the near field. However, gain
reduction directly results from these antenna ground losses. In
contrast, ground losses in the far field can have a dramatic
effect upon the final directivity and thus gain of a vertical.
These losses are a propagation effect in that they are not part of
the antenna proper. Therefore the directivity of the system is
determined in the far field, and thus little if anything can be
done to affect it, except move the antenna to a better location.
Thus far we have derived the specific equation for Rr of a
ground-mounted vertical antenna. This value is of primary
importance for understanding and designing vertical antennas.
Since such an antenna uses the Earth as part of the antenna,
the characteristics of the Earth in the vicinity of the vertical
have a profound effect upon the antenna’s performance. Of
particular interest is the ground’s contribution to the efficiency
of the vertical. Rr is usually taken to be an independent term in
the calculation of antenna parameters. The general equation
for determining the efficiency of a vertical antenna is
204
(Equation 7.12)
and
(Equation 7.13)
In Equation 7.12, Rl is the loss resistance of both the vertical
element and losses due to ground resistance (Figure 7.4). Some
authors also include Rl in matching losses. Although these
matching losses are not part of the antenna per se, they do
become important for system-level efficiency calculations.
205
converges upon an equivalent lumped resistance shown as part
of the feed point impedance. This equivalent resistance must be
normalized to Rr to satisfy Equation 7.12. Also the resistance is
distributed over and under the ground. Simply stated, resistive
losses depend upon the current applied through them. With a
distributed resistance, the current distribution over that
distributed resistance will determine the conducted dielectric
losses. Therefore the solution demands a partial differential
that will be developed later. Fortunately, however, all
contributions to the distributed ground loss can be simplified
to a single resistor and applied to Equation 7.12.
The Rl of the vertical element proper can usually be ignored
except when the hp of the element is much less than 1⁄4 λ. In
such cases, Rr can fall to less than 1 Ω or so, and thus Rl can
become the dominant part of the power dissipation, and thus
efficiency will fall to very low values. Ground losses are a
concern in all cases, but it is important to consider both Rl and
Rr in the all-important efficiency equation. We will treat the
special case of short vertical antennas in later sections.
The directivity of the vertical will be mostly influenced by
the vertical’s height and the current distribution along that
physical height, as with any antenna. Lowering ground losses
increases the efficiency, and thus increases the gain of the
vertical. Lowering ground losses in the near-field ground plane
has an insignificant effect upon the directivity of the antenna,
but does increase the efficiency and thus the gain by Equation
7.13.
The ground-mounted vertical can be broken into three
fundamental basic components that directly affect the
directivity, efficiency, and gain of the overall system:
1) The vertical radiator
2) The near-field ground
3) The far-field ground
The height of the vertical radiator and the distribution of
current along that height largely determine the maximum gain
(assuming no ground losses), he, and thus also the Rr of the
206
vertical. The near-field ground plane is the main contributor to
Rl and thus, along with Rr, determines the antenna efficiency by
the simple Equation 7.13. The limited conductivity of the far-
field ground will attenuate the power in the far field
particularly at low elevation angles and thus further reduces
the system efficiency. The usual method of reducing near-field
losses is the use of ground radials in an attempt to provide a
low resistance return path for the antenna ground current.
The electrical characteristics of the soil, rock, water, organic
material, salts, and so on, have a profound effect upon the
operation of a ground-mounted vertical antenna. Since the
ground is part of such an antenna, it contributes directly to the
efficiency, directivity, and gain of the antenna.
There is also a secondary effect that becomes somewhat
ambiguous to the performance of the antenna proper. In the far
field, ground attenuation is usually significant and thus shapes
the final transmit (and receive) directivity of the antenna
(Figure 7.5). Because this is in the far field, we usually consider
these effects to be propagation effects rather than effects due to
the antenna proper.
207
The attenuation mechanism of “real” ground can be
simplified into the one single term Rl assuming that the vertical
section’s resistance << Rl. If Rl could be determined easily, then
the basic characteristics of the antenna would be easily
modeled. Unfortunately, determining accurate ground
characteristics and how they actually combine with a radial
system to create Rl depends upon multiple variables that are
difficult to quantify.
Quantitatively, the difficulty of the problem can be
summarized by a differential equation. We will use this
equation as a step to formulating a matrix which will more
accurately quantify the interrelationships among the terms.
Each of these terms is difficult to quantify:
208
(Equation 7.14)
If these terms were mutually independent, we would have
this simple equation to define Rl. Unfortunately, these terms
are mutually dependent, thus Equation 7.14 can only be used as
a conceptual step.
This term defines the effect the feed point impedance (Zfeed) has
209
upon the overall Rl. Rl will not vary by changing Zfeed. However,
feeding a radial system with a higher voltage will result in
lower power losses, especially losses close-in to the vertical
element. An example is the use of a 1⁄2 λ vertical which
significantly reduces ground losses when compared to a 1⁄4 λ
vertical over the same ground system. Power loss is Ploss = I2R
and current will be lower for a higher feed point impedance.
Equation 7.14 is provided as a first approximation to the
actual quantification of the problem. The solution for Rl is
actually more complex than solving a simple differential
equation. All the variables listed above are mutually dependent
upon each other. When each term is partially dependent upon
all the other terms, that is, they are all mutually
interdependent, we can use a matrix to quantify this particular
type of problem. For example, if we want to know the
interrelationship between the number of radials and the length
of the radials as it affects Rl, we look at the terms
210
Notice that all values of the variables are mutually dependent
upon each other’s values. The solution is R = 1 and X = –2 and Y
= –2. “1” could easily be an equivalent resistive loss. I will not
present the method of solution. When a system of equations
becomes very complicated, suffice it to say that use of a matrix
becomes a very convenient “shortcut” to a solution. If we use
partial differentials as the elements, the form is as follows:
(Equation 7.15)
Here we use X to identify the matrix. We use four variables
which we defined above as x1-4 where
x1 is the ground conductivity (g)
x2 is the number of radials (radial#)
x3 is the length of the radials (radiall)
x4 is the base feed point impedance (Rfeed)
The 16 partial differential terms that comprise this matrix
represent all the important interrelationships that define losses
associated with a ground-mounted, base-fed vertical. For
example, in free space a ground plane antenna will have
infinite “ground” resistance, but
211
becomes significant and Rl becomes higher. Finally, when the
radials come in contact with the dielectric, Rl is maximum for
the given antenna. Ground plane antennas whose radial
systems are very close to the Earth(<< 1⁄4 λ) have been a subject
of intense study in the amateur community, mostly to
empirically determine these differential terms.
Thus far there are two key variables that have not been
treated in the ground-mounted vertical discussion: the
frequency of operation and the height of the vertical above
ground. The main difference realized when the frequency is
changed is the characteristics of the ground and the ground’s
interaction with the other variables. Therefore, the ground
coefficients of the matrix will differ for different frequencies.
The only situation where the outcome is independent of the
ground characteristics is when the ground is assumed to be an
infinite perfect-conducting plane. The height of the verticals in
the above discussion have been zero, since they are all ground-
mounted. A following section will deal with “raised” vertical
antennas, or ground plane antennas.
212
resonant vertical over the ground plane. The base feed point
impedance should show as little reactive value as possible (a
pure resistance, but some small value is acceptable). We are
only interested in the real portion of the impedance. The value
for Rr will be very close to 36 Ω. Therefore the efficiency will
simply be
Rl = Rfeed – 36 Ω
(Equation 7.16)
As you add radials, change their lengths, and so on, the feed
point should change accordingly. There are many empirical
studies in print. These can form the basis of an approximation
for a general case. However, the multiple differentials involved
will be different for any given location. There is no substitute
for empirical testing at a given location. As you experiment
simply use Equations 7.16 and 7.13 to maximize efficiency. If you
attempt to use a vertical shorter than 1⁄4 λ, then you must
calculate the Rr and adjust Equation 7.16 accordingly. For
verticals higher than 1⁄4 λ, Equation 7.16 will not provide an
accurate answer but can provide relative data.
The results of this exercise will provide a very good
approximation for efficiency of a 1⁄4 λ vertical at a given
location. It will also provide a first approximation for efficiency
of verticals of different heights and loading. Equation 7.15 is a
method for quantifying all the important variables involved.
There is also a possibility of another term that implies Rr may
itself be influenced by Rl. This is certainly the case for very low
conductive dielectrics. For example, if the “ground” under a
vertical/radial system is gradually made more lossy, it will
begin to approach the characteristics of free space. Thus the
“ground-mounted” vertical has morphed into a ground plane
with significant radiation into the ground. This will clearly
affect both he and Ae and thus modify Rr (see Appendix B).
213
We have assumed that maximizing Rr will, by definition
maximize efficiency. This remains a correct assumption
provided that the term
214
structures is unwanted radiation. As the loading conductors
increase in length they begin to radiate. This radiation is
usually highly undesirable since it will be horizontally
polarized and at a high elevation angle maximum. The usual
desired directivity from a vertical antenna is a low elevation
angle maximum.
215
Raised Vertical Antennas
As we raise the vertical antenna together with its radial
system above ground, we gradually leave the characteristics of a
ground-mounted vertical and begin to see characteristics of a
ground plane antenna (Figure 7.8). Since we have explored the
ground-mounted vertical with some depth, we will now
consider the opposite extreme, a ground plane antenna in free
space. The ground plane antenna in its simplest form is a 1⁄4 λ
vertical section fed against two or more radials. The usual
configuration is three or four radials.
216
A ground-mounted vertical antenna can be assumed to have
a complete image that mirrors the vertical element below
ground. Since there is no “ground” under a free space ground
plane, it is interesting to investigate what effect the radials
have upon substituting for “ground.”
We have already quantified he, Rr, and directivity for ground-
mounted verticals with an assumed “full” image. Since the
image directly effects he and thus Rr, we are in a position to
explore the free space ground plane.
The word “counterpoise” can reflect several definitions. In
this text we will define it as a raised ground system for a
vertical antenna. A counterpoise may consist of a single wire,
multiple wires, a flat conductive plane, or other configurations.
Unlike a large dielectric (such as the Earth) under the vertical,
a counterpoise does not create an image under the vertical
unless the size of the counterpoise becomes very large
compared the wavelength.
Step 1: 1⁄4 λ vertical with no radials in free space, simulating
no counterpoise (Figure 7.9). In both cases, the directivity is
about 1.82 dBi but Rr is about 25 Ω for the end-fed and about 12
Ω for the center fed. This result is very close to the expected
ratio using 0.64 he for the end fed and about 0.5 for the center
fed, everything else being about equal. Also notice that the
217
directivity is just a bit better than the infinitesimal dipole from
Chapter 3, which was 1.76 dBi. The directivity pattern, as might
be expected is a doughnut.
218
The feed point has risen from 22 Ω to about 42 Ω. In the case
of a single wire counterpoise, the counterpoise becomes part of
the antenna. With a multiple wire radial counterpoise, the
fields cancel and do not contribute to he, and therefore gain.
Step 3: 1⁄4 λ vertical with a four-radial configuration, also in
free space. Now we return to the typical ground radial shown
in Figure 7.8. The results (Figure 7.11) are counter-intuitive: the
directivity has dropped again slightly to 1.25 dBi, the directivity
pattern is nearly identical to the doughnut pattern in Step 1
(symmetrical in the upper and lower hemispheres), and Rr is
also very close to Step 1, about 22 Ω.
219
The only advantage the radials provide with a 1⁄4 λ ground
plane is to create a counterpoise for the vertical element. In a
practical sense, complications of end-feeding a 1⁄4 λ element are
greatly reduced by including a counterpoise. This is a very good
reason to use radials. The radials accomplish little if anything
else; again, from a practical sense, there is no effect on gain or
Rr
220
the vertical is lowered to ground level, the image gradually
returns — thus increasing antenna gain, he and Rr. In effect, as
the vertical is lowered closer to the ground, the ground itself
gradually becomes part of the antenna.
Much empirical work has been performed, mainly by
amateurs, to determine the trade-offs between ground-
mounted verticals and slightly elevated radials. In effect, this
work attempts to resolve several more partial differentials
involving the scalar Rl. These include the same terms that
appear in the Rl matrix above, but the addition of radial height
as an additional partial variable. The trade-off is the decreased
image in the ground vs. an effective decrease in ground losses
as described in the above paragraph.
Step 5: 1⁄2 λ dipole lowered from a high elevation to ground-
level. In free space the vertical dipole will have an antenna gain
of 2.15 dBi (power gain of about 1.64) and Rr of about 73 Ω. With
one end nearly touching the ground, the idealized gain
becomes 6.94 dBi (power gain of 4.94). In effect, the dipole
operating just above or connected to the ground behaves like a
2-element collinear antenna with the image providing the
other 1⁄2 λ. When lowered close to the Earth, there appears an
image, also 1⁄2 λ. Therefore, he doubles and he2 quadruples.
Aperture increases by 4.94/2.64 = 2.97, and therefore Rr
increases to 4/2.97 (73 Ω) or about 97 Ω.
221
Data for Various Ground-Mounted Unloaded
Vertical Physical Heights
Table 7.1 shows calculated data for various vertical physical
heights with no top loading. For the three shortest lengths, the
current distribution is assumed to be linear, therefore the
average current is simply 50% of maximum (we don’t need to
solve the integral in Equation 7.4). For about 1⁄8 λ and higher
the sinusoid distribution is assumed and Equation 7.4 must be
used.
222
In Table 7.1 are the physical heights in wavelengths of
various vertical elements. heλ is the effective height stated as a
fraction of the wavelength; Giso is simply the numeric gain
value; and Aeλ is Gisol2/4π, where λ is normalized to 1. And
finally, Rr is from Equation 7.9.
As also shown in Figure 7.7, notice that the apertures (gain)
for very short verticals (less than about 1⁄8 λ) remain nearly
constant as we change hp. For very short verticals (from 0 to
about 1⁄50 λ), the dBi value derived from EZNEC modeling was
exactly 4.76 dBi (a power gain of exactly 3). Also, since the
gains and thus apertures do not change with antenna height
for very short heights, we can rearrange terms by substituting
3l2/4π for Ae in Equation 7.14, and simply multiply the isotropic
aperture by 3:
(Equation 7.17)
This is the simplified equation commonly published for short
vertical antennas. With no top loading, the average current is
about 50%. If appropriate top loading is installed, then the
223
current average approaches 100% and he is now nearly twice hp.
For an optimally top loaded short vertical, the average current
is 100%, thus Rr quadruples. This equation is simpler than a first
glance implies. For short verticals with no loading (a simple
vertical whip antenna), he actually equals hp since the average
current is 0.5, and we double this including the image, which
results in a coefficient of 1 between the two terms. Thus you can
simply substitute the actual physical length (without the
image) of the whip for he. A specific example of very short
verticals is presented at the end of this chapter.
Table 7.1 summarizes the three critical issues regarding very
short vertical antennas:
1) The directivity, and therefore potential gain, does not
significantly decrease with shrinking height.
2) The major problem is a rapid decrease in Rr, which
requires very careful design of the antenna and matching unit
to maximize efficiency. For example, a 1⁄32 λ vertical would
require only 0.4 Ω of loss in order to achieve only 50%
efficiency!
3) With optimum top loading, Rr quadruples, thus offering
significant efficiency improvement.
224
over a larger area of the ground surface. (The current is not
confined to the surface, but does penetrate into the ground. For
the sake of this argument, the ground is assumed to be a two-
dimensional plane.) In the case of a 1⁄4 λ vertical, the ground
current is maximum at and very near to the antenna’s base, and
therefore all the ground current is forced to traverse through a
small portion of ground.
The ground can be considered to be a two-dimensional
“square resistance.” However, the feed point becomes a point
source for current on the square resistance. We can
approximate this idea with a simple equivalent schematic
diagram.
Figure 7.13 shows how the current values decrease through
each equivalent resistor as we move away from the vertical’s
base. Assuming a constant total current along a radial angle, we
can calculate the total power dissipated as a function of
distance from the feed point. Let us assume the value of all
resistors is 1 Ω. Therefore, the power dissipated by the first
ground ring (closest to the feed point) is 1 W (12 R). The second
ring is 0.5 W (2(0.52 R)), and the third ring is 0.25 W (4(0.0252
R)).
225
Of course the current is not constant as a function of distance
from the feed point, but this simple model does serve to
illustrate the point: for any current distribution ground
attenuation decreases as the distance from the feed point
increases. When the voltage maximum is at the feed point, it is
easy to see why 1⁄2 λ base-fed verticals have significantly lower
ground loss than 1⁄4 λ verticals for similar ground systems.
The actual current distribution as a function of distance
results from yet another partial differential equation involving
the sinusoidal function, dissipation due to ground attenuation,
and dielectric effects upon the wave velocity in the ground and
on the ground’s surface.
This simple model also shows why wire radial ground
systems are most effective. They inherently provide for the
highest density of conductivity where it is most needed: closest
to the feed point. They also provide this conductivity in the
most optimum direction(s): normal to the feed point. It also
points to the importance of the
226
term (the number of radials being used).
It should now be clear why avoiding maximum current at
ground level inherently increases efficiency in a compromised
ground system. The 1⁄2 λ vertical is one example, but other
vertically polarized configurations can accomplish this same
goal. We will show later that efficiency is only part of the
system performance. We will now explore particular vertical
polarized configurations and then compare their performance.
227
right and left of the figure) is predominantly vertically
polarized. If we compare the azimuth patterns of a single
vertical element and that of a near-ground inverted-V, we can
see that the low-angle gains are nearly identical (Figures 7.15
and 7.16).
228
coupling. The capacitive values are very low, although the
voltages are very high. Again, for total losses in such cases, a
partial differential equation is required to solve the problem.
The ground losses due to capacitive coupling from the high
voltage ends is usually far less than the advantage of getting
the current maximum(s) as far away from the ground as
possible.
229
Broadside Bobtail Curtain
The Bobtail Curtain (Figure 7.20) is an example of a
230
broadside, voltage-fed vertical array that automatically
provides for voltage maximums at the base of all three
elements. The array is base-fed at the center vertical element.
A minimum of radials are required only at the center-fed
element, in practice to help stabilize the high feed point
impedance. The three vertical elements are 1⁄4 λ and the two
horizontal wires are 1⁄2 λ making the array 1 λ long. These are
the dimensions for optimum gain. However, “mini Bobtails”
can also provide excellent results using shorter vertical
elements and/or narrower width as long as the phases and
amplitudes along the three elements are re-optimized for best
results. Figures 7.21 and 7.22 show the patterns.
231
Single-Element Vertical (Monopole Antenna)
A single ground-mounted vertical element has become the
“standard” vertical antenna for a very wide range of
applications. The 1⁄2 λ dipole is the only antenna that is more
232
common. In this text, the first mention of these simplest of
antenna configurations was in Chapter 1. The monopole’s main
advantages are its mechanical simplicity and its
omnidirectional azimuth pattern (Figures 7.23 and 7.24).
Figure 7.25 shows a 5⁄8 λ vertical for comparison.
233
Multielement Vertical Arrays
As we discussed in Chapter 5, the addition of elements allows
configuration of directional arrays. In AM broadcasting, nulls
are created in antenna patterns to “protect” other AM station
coverage areas, particularly at night. For amateur applications,
the usual goal is to maximize gain in a particular direction,
hopefully with many directions easily selectable.
In this section we will explore three multi-element
234
configurations: 2, 4, and 8 element configurations. The plots
shown in Figures 7.26 through 7.31 are taken at 1.8 MHz, but
the gain can be corrected for other frequencies with a single
coefficient for a good approximation. For 75/80 meter
applications, subtract 0.9 dB from these numbers. For 40
meters, subtract 1.5 dB (based on EZNEC modeling results).
235
236
237
In 1994 while I was working up the first known 8-circle
design, I noticed that with 1⁄4 λ spacing of the elements around
the circle, the resulting spacing of the two cardioid pairs would
be very close to 5⁄8 λ, which is optimum for a broadside array!
This results in a very effective configuration by virtue of the
array’s geometry. See Figures 7.32 to 7.34.
238
239
Gain Comparison of Vertical Arrays
We are now prepared to summarize the relative advantages
and disadvantages of the various configuration presented in the
following tables. The arrays detailed in Table 7.2 are arranged
240
by increasing gain.
241
These EZNEC models show some interesting differences
between the perfect and real ground systems. Although the 1⁄4 λ
vertical has a small gain advantage over the infinitesimal
vertical in the idealized perfect ground circumstance, it
actually shows less gain over real ground. The 1⁄4 λ vertical
begins to be affected by ground reflection cancellations,
although it remains minimal. Unlike the 1⁄4 λ vertical, the
infinitesimal vertical far field ground effects are limited to only
tangential dielectric absorption since there can be no ray
cancelations. However, from a practical viewpoint, the Rr of the
infinitesimal (or very short) vertical will be very low and thus
the efficiency will also likely be very low. Also, from a practical
sense, losses in the matching network will also add to
inefficiencies. Thus the slight advantage of a very short vertical
over a 1⁄4 λ vertical will in most circumstances will be more
than offset by efficiency concerns. Appendix E shows plots for
various vertical arrays, providing a graphic comparison of
expected gain responses.
The top-loaded (capacitive hat) 1⁄4 λ vertical shows a very
slight compromise in gain due to the raising of the current
maximum up the vertical height. However, the Rr has increased
significantly thus increasing efficiency significantly —
especially over a compromised ground radial system.
With the 5⁄8 λ vertical, the effects from reflection nulling
242
become dramatic. The current maximum on the 5⁄8 λ vertical is
3⁄8 λ above ground, thus providing enough height to generate
multiple nulling reflections in the far field. The inverted-V
configurations suffer comparatively little over real ground
because the current maximum is at the antenna’s apex, and the
vertical polarized source is distributed over both the X- and Z-
axis (sloping elements), thus “fuzzing” the reflection nulling.
The 5⁄8 λ gain maximum is at a very low 12 degree elevation
angle. Shorter verticals will have significantly less gain at these
angles. Therefore, the take-off angle desired for a given
frequency and target area should be considered when
comparing antennas. Usually for long-range F2 paths, higher
frequencies typically require lower take-off angles. Thus 13
degrees may be too low for many 160 meter paths, but may be
excellent at 40 meters. Again, from a practical sense, the slight
advantages of the 5⁄8 λ ground-mounted vertical are usually
offset by the fact it is more than twice as high as the 1⁄4 λ
option.
Also notice that the 2, 4 and 8 element phased arrays have
nearly identical maximum gain elevation angles, near 21
degrees for 160 meters. This is also nearly identical to the
monopole. Thus for a given vertical height, operating
frequency, and ground conditions, there is a conservation of
maximum elevation angle gain irrespective of the number of
elements. This also holds true for the higher vertical
configuration. Thus, for example, an array of 5⁄8 λ verticals
would exhibit similar azimuth patterns to the 1⁄4 λ arrays, but
the maximum elevation gain angles would be about half, or
about 11 degrees.
These real ground models in Table 7.2 and Table 7.3 assume a
100% efficiency ground radial system. Therefore this
comparison must be tempered with compromised ground
systems. For example, it is very unlikely that a practical
comparison between the infinitesimal and 1⁄4 λ vertical will
show an advantage to the infinitesimal antenna, especially
over compromised ground systems. Also a 1⁄2 λ or 5⁄8 λ vertical
will have higher Rr than the 1⁄4 λ vertical and higher feed point
impedance values, thus increasing the overall efficiency in
243
compromised radial systems.
The 2, 4, and 8 element directive arrays provide a coinciding
possible number of switchable azimuth directions. For
example, the 2 element array inherently provides two
switchable direction. The 4 element array provides 4
directions, and the 8 element provides 8 directions. It is also
fortunate that in all cases these directional settings provide
sufficient over-lap of azimuth patterns that result in only about
2 dB nulls for full azimuth coverage.
Conclusion
When weighing the relative advantages and disadvantages of
vertical arrays, Table 7.3 can represent a good “starting point”
in that it compares:
1) Physical dimensions and necessary area for the antenna.
2) Gain that can be expected using an “adequate” radial
system under the antenna when needed.
3) Comparative optimum take-off angles for maximum gain.
4) Relative radial requirements.
5) Feed point location.
In the case of a compromised radial system (when needed)
the quoted gain will diminish as Equation 7.15. Since this
equation is not closed form, for practical information the
volumes of empirical data from amateur experiments can
provide good approximations for real implications of Equation
7.15. However, Equation 7.15 is an attempt to actually quantify
all the important terms into one unifying equation.
244
EIRP = Pi (Gdbi)
(Equation 7.18)
where Pi is the power input to the antenna.
For example, if the input to an antenna is 5 W and the
antenna has a gain of 1.78 dBi, the EIRP in the direction of the
4.78 dBi gain is 15 W. A short vertical antenna (<1⁄8 wavelength)
has a maximum gain of 4.78 dBi (linear gain of 3) with a perfect
ground system and a perfectly conducting infinite ground. This
gain figure is independent of the vertical’s height for heights
below about 1⁄8 wavelength. Unless ground losses in the far
field are taken into account for EIRP calculations, this will be
the coefficient in calculating EIRP values. For example if we are
limited to 5 W EIRP and use a vertical shorter than 1⁄8
wavelength, our maximum power input to the antenna will be
5/3 or 1.6666 W.
At these frequencies, the typical amateur transmitting
antenna will likely be very inefficient. A full size 1⁄4 wave
vertical is about 500 feet high! The efficiency will also be
complicated by an almost certain compromised ground system.
AM broadcast stations use 120 ground radials each 1⁄4
wavelength long. The equivalent system at 475 kHz would
require a circle of radial wires about 1000 feet in diameter
(almost 20 acres).
The three primary efficiency concerns are as following for
very short ground-mounted verticals:
1) Ground losses.
2) Vertical element losses.
3) Matching losses.
Ground Losses
The short height and compromised radial system results in a
very inefficient transmit antenna system. However, the FCC
will be limiting the maximum EIRP to very low values, ≤5 W.
Assuming 1.6666 W antenna input, this implies that an antenna
only 1.6% efficient would require an antenna input power of 100
W. Wasting 98.3333 of 100 W is much more cost effective than
245
the above extravagant solution. This is a huge advantage of
regulating power in terms of radiated power rather than
transmitter output power.
If we want to calculate the required antenna power input Pi,
we use:
(Equation 7.19)
where EIRP is the desired effective radiated power.
Now we need to find the proper values of Rr and Rl.
(Equation 7.20)
where hp is the physical height of the vertical (without the
image) and λ is the wavelength, in this case about 2052 feet, or
632 meters. For the special case of verticals less than about 1⁄8 λ
high, he ≅ hp making the calculation easier. For example, a 50
foot high vertical ground-mounted antenna for 632 meters will
have an antenna Rr of 0.23 Ω. There is little advantage in
attempting top loading since the loading structure would likely
be very large, especially for verticals under about 100 feet.
The calculation of Rl requires a direct measurement of the
real portion of the feed point impedance.
Rl = Rfeedpoint – Rr
(Equation 7.21)
where again, Rr is calculated by Equation 7.20. Now we have all
the terms to calculate the needed antenna input power for a
given EIRP (Equation 7.19). It is very likely that the measured
Rfeedpoint will be much higher than the calculated Rr indicating
246
the losses will be very high and thus efficiency very low.
Matching Losses
Thus far we have limited the discussion to the antenna
proper. However, in this case, short lossy verticals will exhibit
relatively low real values at the feed point, perhaps tens of
ohms. However, the imaginary part of the feed point
247
impedance will be a high negative value (high capacitive
reactance). This calls for the use of high Q matching,
particularly the inductor. Typical inductive values for the tuner
will be about 1 millihenry. An air core inductor with 1⁄4 inch
copper tubing would be about 6 feet long, 18 inches in diameter
and 100 turns. It may be possible to build a comparable
inductor out of ferrite cores, but this very high inductor value
must also handle about 100 W.
(Equation 7.22)
where 0 dBW= 1 W and Pi (dBW) is the power output from the
transmitter. A detailed description of actual power levels using
logarithmic notation is presented in Chapter 10.
This represents the basic equation for determining the
proper transmit output power for a desired EIRP. The system
design will be challenging if the FCC limits transmitter output
power. For example, if the FCC restricts transmitter output
power to 100 W and the maximum EIRP permitted is 5 W,
implying an antenna input of 1.66666 W, a total of only 17.8 dB
loss will be required to achieve maximum EIRP output.
Many of these complications will hopefully be settled with
clear power rules. Aside from the EIRP calculation, will the
calculation for Rr require computer modeling? Will transmitter
output power consider matching losses? Answers to these and
other questions may change the coefficient values in Equation
7.22, or even change the equation itself.
248
CHAPTER 8
249
The Yagi-Uda has been the subject of intense investigation
for many decades, resulting in highly optimized designs. There
are many fine texts that detail Yagi-Uda and quad type arrays.
This chapter will only review the basic computational methods
used for optimized designs and link previous discussions in this
book to those methods. Therefore, as one of the final chapters,
the parasitic antenna can be considered as a specific
culmination of the previous chapters.
250
antenna: increase the length, and/or create a multielement
array (either by driving multiple elements, or by adding
parasitic elements, or a combination of all these).
First we will explore the increased length dipole. We have
already derived the broadside power gain of a 1⁄2 λ dipole as 1.64
(2.15 dBi). We have also defined broadside gain as a function of
length (Figure 3.17 in Chapter 3).
Again, the aperture of a 1⁄2 λ dipole is
(Equation 8.1)
If we double the aperture, the gain will also double and we
achieve about 3 dBd gain. A 2-element collinear antenna is
twice the length of a 1⁄2 λ dipole, but only achieves about 2 dBd
gain. If we extend the length to 10/8 λ, we create an extended
double Zepp. This is the length of a center-fed linear element
that achieves the maximum broadside gain (about 3 dBd).
The boundaries shown in Figure 8.3 are approximate. The
actual boundary lines are a bit “fuzzy” but this figure serves as a
good starting point to conceptualize what the aperture areas
actually look like around linear antennas.
251
The second method of increasing the gain of a single dipole is
to add elements. In Chapter 7 we showed many multielement
vertical driven arrays and their relative gains and patterns. The
same basic techniques can be used for free space dipoles with
similar results in gain. Thus free space driven arrays will not be
covered here in detail to avoid repetition.
252
We begin the discussion with a simple two-element Yagi-Uda
parasitic array (Figure 8.4). By making the parasitic element
slightly longer than a resonant dipole and placed in back of it,
the backward pattern can be nearly completely eliminated and
aperture and gain increases in the opposite (forward) direction.
By making the element slightly shorter than resonant, a
different phase response is achieved, in that additional gain is
realized in the forward direction when the director is placed in
front of the driven element.
253
aperture. On the other hand, a 2-element quadrature driven
array has a maximum gain of about 3 dBd, which is what would
be expected by a simple doubling of a single dipole’s aperture.
See Figure 8.5.
254
ratio. The addition of more and more directors is about
proportional to the log of the gain. Therefore, adding more and
more directors provides less and less gain addition per element.
This is because, again, the additional directors begin to
duplicate the available aperture around the antenna. On the
other hand, the inherent advantages (electrical and
mechanical) often outweigh this diminishing return.
Now let us assume a 3-element Yagi-Uda array. The addition
of a single director is taken at a typical value of 2 dB more than
a 2-element array. Therefore, we assume the Yagi-Uda has a
gain of 7 dBd, or about 9 dBi, or a numerical power gain of 7.94
(in free space). Therefore the aperture of our 3-element Yagi-
Uda is
(Equation 8.2)
In Figure 8.3 the shape of the apertures is elliptical. This is
due to the horizontal current distribution along the wires. As
the gain increases with a Yagi-Uda, the array proper begins to
resemble a point source, and thus the shape of the aperture
becomes progressively more “circular” as shown in Figure 8.6.
In turn we can define the radius of this aperture circle by
255
(Equation 8.3)
Stacking Yagi-Uda Antennas
We may increase the gain of a single Yagi-Uda by simply
adding a second identical antenna. In principle this should
double the aperture and thus double the gain (+3 dB). If yet
more gain is needed, doubling the aperture again will yield (+3
dB) more (using four antennas). Another 3 dB can be realized
by doubling the aperture again (now eight antennas), and so
forth. In Figure 8.3 we showed that if two dipoles are spaced
very closely (end-to-end), some of the potential aperture is
covered by both antennas. This is analogous to our discussion
256
of solar cells in Chapter 2. If one solar cell partially covers
another (shading), the amount of aperture (gain) is reduced.
The correct spacing among “stacked” Yagi-Uda antennas can be
determined with fair accuracy by calculating the equivalent
aperture (as in Figure 8.6) and then make sure the antennas
are sufficiently separated to provide minimum “shading” of the
other as shown in Figure 8.7 for antennas stacked side-by-side.
257
258
We can also “stack” the two antennas vertically. The
difference is how this will affect the directivity of the array.
Vertical stacking of Yagi-Udas over real ground provides
diversity of the maximum elevation angle of radiation. Low
horizontal antennas will radiate at higher elevation angles
than higher horizontal antennas (assuming the maximum
gain).
Figure 8.11 shows an array of eight Yagi-Uda antennas for
432 MHz using both horizontal and vertical stacking.
259
Design and Optimization of a Yagi-Uda Array
The fundamental theory of parasitic arrays was presented in
Chapter 5. We discussed the fundamental importance of the
mutual impedances of the array. However, we only dealt with a
2-element array. Now we will present numerical methods for
multielement arrays. In Chapter 7 we presented the concept of
using the matrix as a method of linking multiple
interrelationships of multiple functions. The modeling of a
multielement Yagi-Uda antenna is most easily performed using
260
such a matrix. Unlike the Chapter 7 discussion where we
resorted to differential equations for the terms, with a Yagi-
Uda we are only dealing with (in principle) relatively simple
algebraic functions, in particular solving for Ohm’s Law with
complex impedances.
Consequently, matrix algebra is used to calculate the relative
amplitudes and phases on each element that, in turn,
determine the final directivity of the antenna. We begin by
expanding the 2-element discussion in Chapter 5 to a 3-
element Yagi-Uda. The first step is to compute the voltages that
appear on each element as a result of all the self and mutual
impedances:
(Equation 8.4)
The resulting voltages on each element are the sum of their
corresponding currents and the complex impedances
determined by the self and mutual complex impedances. It is
these complex impedances that alter the relative phases of the
currents on each element. Once again the phases, amplitudes,
directions, and locations (PADL) of these currents, determine
the final directivity of the antenna.
The matrix in Equation 8.4 actually represents three linear
equations, with the currents being unknown. The mutual
impedances can be computed analytically by knowing the
spacings and electrical lengths of the elements. Thus if we
apply a voltage to the driven element (element 2), we can
derive the resulting voltages by this set of linear equations
contained within the matrix.
The final step is to derive the currents by inverting the
complex impedance matrix to yield
261
(Equation 8.5)
which, of course is simply a special form of Ohm’s Law
(Equation 8.6)
Once we know the complex currents in each element
(amplitude and phase) and the spacing between the various
elements, we can directly calculate the far-field of the array,
again by the PADL method. This is the basic technique used by
antenna modeling software such as EZNEC.
As with most antenna configurations, today computers are
used to solve for these matrices and then calculate the resulting
pattern. Furthermore, optimization for forward gain, front-to-
back ratios, and other parameters involve hundreds if not
thousands of calculations through trial-and-error algorithms to
derive an optimized design. Again, today these computations
require only seconds on a laptop computer.
262
flexibility of element length (both as a function of wavelength).
Thus, the W8JK has the inherent advantage of providing
broadside gain over one octave or more of usable spectrum (not
mentioning the required matching for different frequencies!). If
the element lengths are made 5⁄8 λ long, and use 1⁄2 λ spacing
(let’s say for 28 MHz), then appreciable gain and tolerable feed
point impedances result as low as about 10 MHz (about 3⁄32 λ
spacing). At the highest frequency, the two dipoles are
extended double Zepps, and the bidirectional gain is comparable
to a 3-element Yagi-Uda (in both directions simultaneously).
The pattern is shown in Figure 8.13.
263
For all its simplicity, as the operating frequency changes, the
feed point impedance moves “all over the Smith chart.” Thus a
practical broad-band W8JK should use open wire or ladder line
to a tuner and balun for matching.
264
frequencies the trap appears as a series inductor, thus making
the element electrically shorter at lower frequencies. Good trap
design can result in low Rl loss, but there is always some loss.
The shorter element lengths also reduce gain and decrease the
VSWR bandwidth.
2) “Interlacing” elements simply places elements for different
frequencies along the same boom. The main problem is inter-
band interaction through mutual impedances of the various
elements. Again, very careful design is needed to minimize
these effects while maintaining acceptable performance. An
interlaced Yagi-Uda array may have separate driven elements
for each band, each with its own feed line, or it may use a
broadband or multiband driven element with a single feed line.
These may be a small log periodic cell or a set of closely-spaced
elements that exhibits resonance at multiple frequencies
(typically referred to as “coupled-resonator” or “open-sleeve”).
3) Remote control of element length. This is an excellent
technique used to change the length of the elements. Designs
have included using air pressure to set the lengths against two
stops for a dual-band configuration. Currently the most popular
method utilizes stepper motors that can set the element
lengths with high precision. The one disadvantage is shared
with trap antennas: the element separation is fixed. However
the design is inherently broad-banded in that it can be
absolutely optimized for any frequency given the spacing
constraint. Unlike the W8JK matching problem, the length of
the driven element can be adjusted to resonance and a simple
broadband balun-transformer can easily match to the popular
50 Ω coaxial cable feed line at any of the frequencies covered
by the antenna.
4) Linear loading can involve many different configurations
of multiple conductors running parallel to a “main” conductor
along portion(s) of the element’s length. Parallel conductors
can provide parallel resonant dipoles or sections of open or
shorted transmission lines that in effect, create multiple
resonances at desired frequencies without the need for LC
traps. This technique eliminates trap losses, but complicates
the mechanical design.
265
The Cubical Quad
The Yagi-Uda is a derivative of the linear antenna form: the
dipole. Large loops (≅ 1 λ circumference) can be configured in a
very similar manner to the Yagi-Uda by feeding only one loop
(the driven element) and using parasitic elements slightly
longer (reflector) and slightly shorter (director) to create more
gain and/or front-to-back ratios. This is the basis of the cubical
quad antenna such as the one shown in Figure 8.14. The
calculations are identical to the Yagi-Uda.
266
element Yagi-Uda (for forward gain) by about 1.8 dB, but as the
boom length is extended for more elements, the Yagi-Uda
shows at least a parity of gain with the quad. This advantage,
coupled with the simpler mechanical structure of the Yagi-Uda,
has made the Yagi-Uda far more popular. However, the quad
has two big advantages in addition to the slight gain advantage
of 2-element performance. First, it is much easier to build
multifrequency quads. Smaller loops for higher frequencies are
simply added to the same spreaders. Inter-band mutual
coupling is also less with a quad, and there is no need for traps
or very careful design of “interlacing” elements. Second, the
quad loops inherently are lower in Q, thus providing a good
match over a greater bandwidth than a dipole/Yagi Uda design.
267
even build. High quality high voltage “doorknob” capacitors are
usually required. Also, quality aluminum tubing is expensive.
After the support, rotator system and boom, the quad requires
only wire and bamboo poles, versus telescoping aluminum,
expensive traps, and very low error tolerance in cutting and
setting the aluminum. I built my first quad when I was 16 years
old, I could not beat the performance for the required
investment. Today I use Yagi-Udas, but would still consider a
quad if my requirements changed.
268
269
There are several important notes to this type of
configuration.
1) The actual shape of the loop is not critical. We want a
response to vertical polarization, so the vertical wires should
be kept as long and “vertical” as possible.
2) We can change the loop from director to reflector by
simply adding inductor(s) for the reflector. However, the
inductor cannot be placed at the high voltage point (centers of
the horizontal wires). They will have no effect. Rather, for a
square loop, two inductors should be used, one each at the two
bottom corners.
3) The lowest wire (wire 2 in Figure 8.16) will exhibit very
high voltages and therefore should be kept well out of reach of
people and animals (as should all antennas used for
transmission).
4) The optimum length (circumference) for the parasitic
element will be greatly affected by the overall height and the
shape of the loop. Modeling will be necessary, but fine tuning
can be accomplished by breaking the loop at a corner and
resonating to the appropriate frequency above (for a director)
270
or below (for a reflector). However, despite this sensitivity it is
more inherently stable and efficient than a ground mounted
vertical used as a parasitic element.
5) An inherent problem with any parasitic element is that it
remains a parasitic element even when its associated antenna
is not being used. For example, the presence of a loop reflector
will also dramatically affect a receive antenna located nearby.
This undesirable affect can be eliminated by making the loop’s
total wire length shorter than a full size reflector, about 0.9 λ
or less (Figure 8.18). The loop is activated by switching in an
inductance and closing the loop. The switch that breaks the
loop is located adjacent to an inductor, so undesirable loading
is avoided. Figure 8.19 shows one method of creating a
variable inductor that can be used to tune the loop.
271
272
CHAPTER 9
Specialized Antenna
Configurations
Antenna design has evolved into many different
configurations since the inception of radio technology. Thus far
we have considered center-fed linear elements and derivatives,
of which the 1⁄2 λ dipole is the simplest form. In this chapter we
will investigate some of the other important configurations for
professional and amateur applications alike.
273
diameter is reduced from a large loop to a small loop. Here we
will treat the small loop in some detail, but will compare it to
the large loop to gain insight into how the two configurations
create their respective patterns.
Let us begin with a very small loop. If we reduce the loop size
to a very small size compared to the wavelength, we create an
antenna that is “almost” a shorted transmission line. As
expected, this shorting section will exhibit very high current
and radiate very little power (very low Rr). As we make the loop
larger, but still very small compared to the wavelength, the
current begins to fall, and more power is radiated. However,
since the circumference of the loop remains much shorter than
the wavelength, the current value along the loop’s conductor
will remain essentially uniform as will the phase of the current.
We can once again employ the PADL principle introduced in
Chapter 3 to gain an intuitive understanding of how a small
loop’s pattern is formed. In Figure 9.2, since the circumference
of the loop is much smaller than the wavelength, both the
phase and the amplitude can be assumed to be constant around
the loop. However, the direction of the current (accelerating
charges) is equal in all directions. If we view the loop broadside
to the plane of the loop as in Figure 9.2, the distance to all
sections of the loop is equal from any point in line with the
loop’s axis. Therefore, there is perfect cancellation in the
broadside direction.
274
In contrast, if we view the loop in Figure 9.2 from its side
(from the right, left, up or down from the loop), one critical
PADL parameter has changed. Again if we look at Figure 9.2,
but this time from a distant point to the “left” or “right” of the
loop, there is a slight difference from the distant point source
to the two sides of the loop (the distance of the loop diameter).
This difference is enough to provide a slight difference in the
phase, and thus provides a slight non-cancellation of the fields,
and thus some radiation maximum around the radius of the
loop — but again, complete cancellation broadside to the loop.
The very slight phase response normal to the loop’s axis results
in small loop efficiencies that are very low. Figure 9.3 shows
the current direction on small loop antennas versus large loop
antennas.
275
Figure 9.4 shows the resulting figure 8 pattern with the null
broadside to the loop. This is completely opposite the pattern
of a large loop 1 λ in circumference, where the nulls are in the
plane of the loop and maximum directivity is broadside to the
loop. Figure 9.5 compares the E and B fields of a small loop and
a small dipole.
276
As we might expect from any antenna that is small relative to
the operating wavelength, small loops tend to be rather
inefficient. Indeed, the common expression for antenna
efficiency also applies to loop antennas where
(Equation 9.1)
Loop antenna Rr is calculated from the general form, Equation
3.37 in Chapter 3:
(Equation 9.2)
Again where S(θ,φ)maxr2ΩA represents power concentrated in
277
the beamwidth defined as ΩA, and I2 is the current. So, again,
this is only a special form of the power law linking resistance to
power and current. By integrating the power over the pattern
(identical to the small dipole) over the sphere surrounding the
loop, we find
(Equation 9.3)
where A is the cross-sectional area of the single-turn loop.
Often the circumference of the loop is a more convenient
measurement. If we use the circumference rather than the
circular area of the loop, then the term C/λ becomes a
coefficient that is a simple ratio applicable at any frequency,
where C is the circumference. Thus
(Equation 9.4)
It is strikingly apparent that as the circumference of the loop
is made smaller, Rr drops at the fourth power of the ratio! For
example, a loop with a circumference of 0.1 λ will have an Rr of
about 20 milliohms. Cλ4 is often used rather than C/λ to make
clear that this is now a coefficient. By increasing the
circumference only to 0.15 λ, Rr jumps to about 100 mΩ, a five-
fold increase.
Furthermore, the dc resistance of the loop conductor (as in
any application) is not adequate for computing Rl in the
efficiency equation. Skin effect increases the effective RF
resistance of the conductor and must be taken into account
when calculating loop efficiency, even with 20 milliohms of
loop Rr.
Aside from using larger diameter wire, there are two
278
techniques for raising the Rr of a loop antenna: use more turns
in the loop (as in Figure 9.6) and/or wind the loop on a
material with higher permeability. These are precisely the
same techniques used to increase the inductance of a coil. Rr
will increase as the square of the number of turns, so we simply
add that coefficient to Equation 9.4.
(Equation 9.5)
The other technique is to use a high permeability material,
usually ferrite or iron core rods on which the coil is wound.
Thus we add one last term to the equation
(Equation 9.6)
Consequently, for small loop antennas, we can increase Rr and
279
thus efficiency by winding many turns on a ferrite rod core, as
shown in Figure 9.7.
280
undesirable signals and/or interfering noise.
(Equation 9.7)
We are now prepared to calculate the efficiency and thus the
gain and effective aperture of a small loop antenna. Rr is a
relatively straightforward calculation using Equation 9.6. On
the other hand, Rl is a function of several variables, all of which
are difficult to predict. However, Rr + Rl is a relatively simple
measurement. If we tune the loop to series resonance with a
capacitor we will measure a nonreactive total impedance equal
to Rr + Rl. Then
(Equation 9.8)
The directivity of a small loop is 1.5 or 1.76 dBi. If the loop had
no losses, its aperture would be
281
(Equation 9.9)
For example, we have a loop with 10 turns and Rrn = 0.0001 Ω
and Rln = 1 Ω. The efficiency of this antenna will be about
0.0001%, or 40 dB below 100% efficient (–38.2 dBi). Now we
want to know the efficiency if we replace the 10 turns with 100
turns, thus Dn = 10. The result is Rr is now 0.01 Ω and the new
Rl is 10 Ω. Thus the new efficiency is about 0.001%, or 30 dB
below 100% efficient (–28.2 dBi). Therefore, in the case of very
low efficiencies where Rl >> Rr, efficiency increases nearly
linearly with the number of windings.
282
The directivities of the Beverage as well as a short monopole
are shown here for easy comparison to the K9AY loop (Figures
9.9 to 9.11). Together with the plots in Chapter 7, a
comprehensive set of patterns for MF and low HF are
presented. The interpretation of these plots requires some
knowledge of noise in systems, and we will explore noise in
Chapter 10. In addition, we will present a discussion on
alternative methods of defining antenna parameters at a
system level. These techniques will be shown to be critical for
understanding the applicability of small loop antennas to
frequencies about ≤ 5 MHz.
283
284
Helix Antennas
As a small multiturn loop antenna’s size is made larger
(relative to the wavelength), other directivity characteristics
emerge, creating a helix antenna. In 1946, John Kraus was the
first to discover the remarkable characteristics of the helix
(Figure 9.12). The gain of the helix increased as the length
(number of turns) increased. It is inherently broadbanded and
is relatively easy to match to common 50 or 75 Ω feed lines. It
is also circularly polarized, making it preferable for a wide
range of application. It’s particularly useful for space
applications, where “horizontal” and “vertical” polarizations
become meaningless.
285
circular polarization depends on the direction of the
“threading.”
For example, when we look down on a standard thread
screw, we rotate it clockwise to tighten and counterclockwise
to loosen. If we have a threaded rod, we can turn the rod to the
other end and the same directions apply to tightening and
loosening. Helix antennas are similar in that no matter which
direction you view them from, they are either “left” or “right”
handed. A reverse-threaded screw will tighten when turned in
a counterclockwise direction, and thus has reverse threading. It
is a common mistake to assume a right-hand helix will become
a left-hand helix if it is turned over. The “threading” does not
change by flipping it over.
286
As Figure 9.14 shows, the gain increases and the beamwidth
decreases with more helix turns. There is also a decrease in
bandwidth as the array becomes longer.
287
When a circular polarized (CP) wave is reflected, its “hand”
changes. For example, if you use a right-handed helix to drive a
parabolic surface, the transmitted wave off the parabola will be
left-handed. Similarly, a left hand CP wave reflected off the
Moon’s surface will return as a right hand polarized wave.
Sometimes this is a desirable effect. For example, GNSS
navigation signals use right hand polarization. A major
problem with GNSS accuracy is ground reflections that distort
the time of arrival to the receiver from the satellite. CP
antenna inherently reject opposite polarized signals, so they
will attenuate ground reflections greatly. However, for ultra-
high performance GNSS, even the slightest ground reflected
signal can cause location errors. No “real” antenna is perfect,
and therefore a right-hand CP antenna will never have
“infinite” rejection of a left-hand signal.
288
Aperture Antennas
The parabolic antenna and corner reflector antennas are
often termed “aperture antennas.” The reflecting surfaces of
such antennas are not strictly “part” of the antenna, but rather
form passive reflection rather than taking advantage of mutual
coupling as in a true parasitic element.
The parabolic antenna derives its name from the fact that the
surface is shaped into a parabolic curve. Parabolic surfaces will
focus a plane wave onto a “focal point” where a true “antenna”
is mounted to receive or transmit power from or toward the
parabolic surface (Figure 9.15).
289
The aperture of a parabolic antenna is perhaps the simplest
aperture calculation of all antenna forms. For a perfect
parabolic antenna Ae = cross-sectional area. For example, if the
diameter of a circular “dish” antenna is 10 meters, then Ae =
78.5 m2. At 1 meter wavelength, such an antenna will provide a
gain of 986.5 or about 30 dBi. A unique characteristic of the
parabolic “dish” is that the aperture is constant for all
wavelengths. However, since gain is a function of both aperture
and wavelength, the gain (dBi) of a parabolic antenna changes
as the square of the wavelength. Figure 9.17 shows the
Cassegrain antenna, a variation of the parabolic dish.
290
The corner reflector configuration (Figure 9.18) has found
wide applications, particularly at VHF and UHF. Construction
is easier than a parabolic structure in that it is easier to build a
planar reflecting surface rather than one formed as a parabola.
Its advantages are also significant for linear polarization as the
reflector need not be “solid” but can be constructed of linear
elements in the plane of desired polarization.
291
The feed antenna of a corner reflector is not as critical as the
feed antenna for a parabolic antenna. The corner reflector’s
driven element is typically located “inside” the reflecting
surface, thus a simple dipole can be used without the concern
illustrated in Figure 9.16.
On the other hand, the maximum gain available from a
corner reflector is limited to about 13 dBd, as shown in Figure
9.19. This is due to the fact that there is no “focal point” in a
corner reflector antenna, setting a limit on the potential
aperture (Figure 9.20). Therefore if more gain is required, then
a parabola antenna is necessary where the aperture, in
principle, can be expanded indefinitely
292
293
The corner reflector can be modeled using the mutual
impedance methodology introduced in Chapter 8.
Small Antennas
In Chapter 1 we discussed the general trend in wireless
communication toward higher frequencies and smaller form
factors (size of device). Design problems arise when the
physical dimensions of the form factor become small with
respect to the wavelength of the wireless device(s). We also
discussed how antenna efficiency relates to antenna gain and
how both relate to the physical size of the antenna. Therefore,
a fundamental question arises: How small can an antenna be
294
made and still be effective?
In Chapters 2 and 3, some points we made hint at some
physical limitations on antenna size. In particular, as an
antenna size shrinks relative to the wavelength, Rr also begins
to drop. One may recall from ac circuit theory that efficient LC
matching circuits require greater Q to match higher ratios of
impedances. This also holds for matching from the typical 50 or
75 Ω transmission lines to the resistive portion of a small
antenna’s feed point impedance which is typically very low. A
simplified equation for the necessary Q of an LC matching
circuit is:
(Equation 9.10)
where R is the real component of the two impedances to be
matched.
In this equation, either the source or load real value can be
either the higher or lower value. If the load or source has a
reactive value, the equation becomes more complicated.
However, Equation 9.10 shows how the bandwidth of an
antenna becomes narrower as its load resistive value becomes
lower.
As the reactive-to-resistive portion of the antenna proper is
made larger, the Q of the antenna also rises. This shrinks the
bandwidth but also increases the antenna’s stored power versus
the radiated power. In Chapter 4 we described the basic modes
of electromagnetic waves as they relate to transmission lines.
These modes also begin to play a significant role in the
behavior of small antennas, particularly in the near field.
Therefore, one effect of making an antenna small is that
higher Qs become necessary and thus the bandwidth of the
antenna is reduced. The second major problem is that very low
values of Rr tend to drop the antenna efficiency as in Equation
9.1. See Figure 9.21.
295
A “small” antenna can be defined as an antenna whose
maximum dimension(s) are confined to a spherical volume
which has a radius of
kr < 1
(Equation 9.11)
where
(Equation 9.12)
and r is the radius of the sphere, where
(Equation 9.13)
McLean quantified the minimum antenna Q (assuming zero
matching loss) for both linear and circular polarized antennas
as:
296
(Equation 9.14)
for a linear polarized antenna, and
(Equation 9.15)
for circular polarized antennas.
For example, when λ = 2 m, then k = π. If we set kr = 1
(maximum size of a “small” antenna) then r = 1/π. Therefore a
maximum size small antenna for the 2 meter band must be
confined within a sphere of radius 1/π meters or less. The
maximum size dipole that could fit inside this sphere size
would be 2/π meter (the diameter of the sphere), or 0.64 meter
long, while a full-size 1⁄2 λ dipole would be 1 meter long. Since
kr = 1, then by Equation 9.15 the minimum Q of a dipole of this
length is 2.
These limitations are derived from the assumption that equal
power is excited into the TM and TE modes (Chapter 4) in the
immediate vicinity of the antenna. A detailed analysis of these
“higher modes” requires an advanced mathematical treatment
which can be found in some of the texts referenced in
Appendix C.
As the size of the antenna decreases, maintaining efficiency
and a reasonable bandwidth becomes increasingly more
difficult. As a practical matter, impedance matching also
becomes more problematic, which adds to the overall system
efficiency. Figure 9.22 dramatizes the challenges facing
antenna designers with ever-smaller devices. However, as the
operating frequencies have increased (λ decreases) hand-held
devices can accommodate reasonably efficient antennas.
An intuitive understanding of the small antenna problem can
be taken from our discussion on vertical antennas (Chapter 7).
As the size of the antenna is reduced (compared to the
297
wavelength), the effective height he of the antenna decreases.
Again we can see this effect in
(Equation 9.16)
which is the same as Equation 7.9.
In effect, a longer length (and therefore the size) of an
antenna distributes the current over a larger distance and thus
requires less current to create the required field. As he shrinks,
larger current values (lower Rr) are required to generate the
same field strength. The same principle applies for a linear
vertical, a planar antenna, and a 3D structure.
Finally the Q of the antenna is a function of its self-
impedance. As antennas are made smaller, Rr decreases and
thus decreases the damping factor on the increasing capacitive
reactance. See Figure 9.22. Similar to circuit theory this results
in a higher antenna Q, including an increase in stored energy
around the antenna.
298
Traveling Wave Antennas
As we discussed in Chapters 6 and 7, as a vertically polarized
wave travels along the Earth’s surface, the velocity of the wave
nearer the ground is slower. The source of these waves can be a
ground-mounted vertical antenna with its wave propagating
over the ground (ground wave), or it can be a sky wave arriving
from a distant transmitter. This effect results in a tilting of the
wave “forward.” This forward tilting effect can be used to
advantage in creating a traveling wave antenna very close to
the Earth. A wire (about 1 – 4 λ in length) placed near the
ground surface and laid out in a straight line will receive waves
from the direction of the wire. Such an antenna is called a
Beverage, after its inventor Harold Beverage, who developed it
299
in the 1920s. See Figure 9.23. There are cases where a
combination of directivity, along with better efficiency is
needed, especially for very weak signal detection in the MF and
low shortwave bands. The Beverage will exhibit better
efficiency than small loops, but at the expense of laying out a
very long wire in the desired direction(s).
300
double Zepp, where broadside gain is maximum. As the wire is
made longer, the azimuth pattern quickly changes from a
“figure 8” to a “4-leaf clover” pattern (Figure 9.24 and Table
9.1).
301
Therefore, if used as a transmitting antenna, the resistor must
be able to dissipate half the transmitter’s power output. This is
a major disadvantage of the transmitting rhombic.
302
Prior to the advent of GNSS (global navigation satellite
system), medium wave (1.6 – 2.0 MHz) and LF wavelengths
were used. These wavelengths have the advantages of excellent
ground-wave propagation and also small portable loop
antennas can provide the sharp nulls necessary to provide a
high precision azimuth.
More advanced navigation systems such as LORAN utilized
multiple beacons with time (and thus phase) synchronization.
These types of systems provided excellent 3D location for the
users. These older MF wave systems have been largely replaced
by GNSS (global navigation satellite system) which includes
GPS (US), GLONASS (Russia), Galileo (European), and others
that use microwave signals from satellites. The instantaneous
satellite locations are known with high precision (despite their
high velocity and altitude). Very precise location determination
303
(within centimeters) is possible using sophisticated algorithms
for error correction. However, even these very sophisticated
satellite systems rely upon simple position and timing for these
calculations.
Patch Antennas
With the nearly universal use of PC boards in radio
construction and the general trend toward higher operating
frequencies, it was only inevitable that antennas would soon be
fabricated onto the PC board. Also, low profile antennas
suitable for working off a ground plane were obvious choices
for aircraft and space applications. In response, a wide variety
of antenna forms have evolved. The simplest PC board uses two
sides of thin copper separated by a dielectric of fiberglass. Since
PC boards use at least two metal layers, transmission lines,
matching networks as well as the antenna proper can be
combined into one planar structure.
At frequencies above about 1 GHz, fiberglass (fire-retardant
FR4 material used in PC board construction) becomes a
comparatively high loss material. Therefore, for higher
performance RF circuits, other substrate materials are often
used. However, some compromised performance is often
accepted to keep manufacturing prices down.
At first, simple dipoles were etched onto the PC board,
followed by more complex geometric forms including spirals,
planar “patches,” and even fractal shapes. The various form
factors may either be placed above a ground plane or emulate
“free space” by keeping other metal traces away from the
antenna proper, or some combination of both.
An alternative approach has been developed that permits the
use of fiberglass PC board by creating a separate antenna
structure using much higher performance ceramic as the
antenna’s substrate (Figure 9.27). The ground plane is simply
included on the PC board’s top layer, thus making the
shortcomings of fiberglass mute. A rectangular, circular or
other shape metal plane with small spacing to the ground
plane forms most of configurations in use. The above discussion
on small antennas is directly applicable to patch antennas
304
since they are often quite small relative to the wavelength in
use.
Phased Array
In its simplest form, a phased array consists of multiple
305
antenna elements (sometime hundreds of individual antennas)
distributed over a plane (Figure 9.28). The resulting array may
be steered to any direction above the plane by changing the
phase and amplitude relationships among the elements.
306
Aperture Distribution
No matter what the antenna configuration, sidelobes begin
to appear in the main pattern as the aperture becomes greater
than about 1.5 λ. See Figures 9.30 to 9.32 for the following
discussion.
307
308
309
The simplest example is to plot the resulting pattern along
one axis as a result of the aperture distribution along that axis.
The aperture can consist of multiple individual antennas (as
with a phased array) or a large single aperture (as with a
parabolic antenna). Again, the calculation uses PADL for its
solution, but with a rather sophisticated pair of integral
functions.
The calculation of the sidelobe response from the aperture
distribution utilizes a Fourier transform. The Fourier transform
is a special form of integral that is extremely useful for a wide
variety of engineering problems. In electronics it can provide
the spectrum (frequency domain) of a signal captured in the
time domain. Time domain displays are most often shown on
oscilloscopes, while frequency domain displays are shown on
spectrum analyzers. If you have time domain data (for example
310
a voltage versus time) the spectrum of that signal (frequency
versus voltage or power) can be calculated using the Fourier
transform. On the other hand, if you have spectrum data you
can use the inverse Fourier transform to derive the time
domain signal. The simplest example is the Fourier transform
of a pure sine wave to a single spectrum line, and vice-versa.
Readers who are familiar with transitions between time and
frequency domains may recognize the similarity of the
relationships shown in Figure 9.30. If the uniform aperture
distribution shown in Figure 9.30A, were a “square wave” on an
oscilloscope, we would expect to see Fourier series of the
fundamental frequency and a series of odd-order harmonics
(3rd, 5th, and so on) decreasing at a
Polarization Diversity
In principle, an infinitely thin perfectly horizontal antenna
will have zero response from an infinitely thin, perfectly
vertical wire in free space. Since linear wire or tubular antenna
elements have non-zero diameters, there is some response
311
from one such antenna to the other. Furthermore, reflections
from the Earth and/or from the ionosphere often change the
polarization characteristics of the radiated wave.
This difference in response between polarizations can be put
to use in transmitting one channel of information using
vertical polarization and a second channel using horizontal
polarization but using the same frequency band
simultaneously. This technique has the potential of doubling
the capacity of the frequency channel. Left hand circular
combined with right hand circular can provide the same effect.
In such systems additional nulling can take place in the
receiver to offer better polarization isolation since both
polarization signals are received as shown in Figure 9.33. Since
the error signals bleed-over from the other polarization, we
can simply sample the undesired signal, correct for amplitude
and phase and provide additional nulling.
MIMO Antennas
“MIMO” is an abbreviation for “multiple input, multiple
output.” Consider the system is Figure 9.34. The transmit
power is divided between two antennas of some distance apart.
The receiver receives these two signals on two receive
antennas also separated by some distance. Therefore there are
now four possible paths for the signal to travel thus greatly
312
increasing the probability that the system can avoid multipath
problems. A second advantage is that since there are different
time delays among the four separate paths, it is possible to
transmit different information along the separate links, or a
combination of the two.
Other Configurations
The above examples of specific antenna configurations are
but a small sample of current development work in antennas.
The last few examples point to a progressively closer
relationship between the antenna and the receiver. The
development of digital signal processing and software defined
radios (SDR) has opened new frontiers for improving antenna
performance. In particular the difficulty of dealing with
expensive passive components for phase and amplitude
adjustments can now be performed using inexpensive yet very
powerful digital processors. In effect, critical adjustments to
antenna parameters are now being performed in the digital
domain. However, the physics of the antenna proper remains
constant.
313
CHAPTER 10
314
discussion by looking at thermal or “Boltzmann” noise. We will
see that thermal “radio noise” is closely related to temperature,
as its name implies, but this relationship is often confusing,
particularly when its effects upon communication systems are
considered.
Thermal noise originates from the fact that any atom or
molecule with a temperature above 0 Kelvins or 0 K (absolute
zero) will vibrate. The classical definition of absolute zero (0 K)
is the point where all atomic vibration stops. Some
experiments as of this writing (2014) have achieved
temperatures only a small fraction of a Kelvin higher than
absolute zero. (Note that unlike the Fahrenheit and Celsius
scales, Kelvins is the unit of measure for this scale — without
“degrees.” So while we would normally write 50 °F or 50 °C,
when using the Kelvin scale, 50 K or 50 Kelvins is the correct
notation.)
The absolute position of a vibrating atom is impossible to
predict at any given instant. However, the distance from a
“norm” position is a matter of probability. This physical
vibration of atoms also has another effect: it produces a
random electrical voltage. Since the voltage is random, we call
it “noise.” If a voltage is “random” it must therefore also be
“alternating.” Since this is a random alternating voltage, we can
also deduce that it is broadbanded and (to the point of this text)
extends into radio frequencies. Furthermore, we recall from
Chapter 3 that radiation can only occur when an electrical
charge is accelerated. Atoms contain electrons (charges) that
therefore also vibrate, which in turn is a random acceleration
of charges, which creates a random electromagnetic radiation.
A random RF voltage is produced within a resistor at some
temperature above absolute zero. This random voltage will be
conducted and thus appear across the terminals of the resistor.
The source of this RF voltage will also radiate, thus causing
thermal radiated noise according to the dynamics of the
antenna PADL (Phase, Amplitude, Direction, and Location —
introduced in Chapter 3) of the resistor.
As we might suspect, the random movement of atoms will
increase with increasing temperature. This is indeed the case
315
and was first quantified by Max Planck (Figure 10.1). In 1900
Planck introduced the concept of black body radiation. This was
follow-on work of Ludwig Boltzmann (Figure 10.2), who, 23
years earlier had linked entropy (from theory of gasses) and
probability. The mathematics dealing with probability
functions is of primary importance in the study of quantum
mechanics. In turn, black body radiation derives its theoretical
basis in quantum mechanics and is also described in terms of
probability functions.
316
As temperatures increase, the total power emitted from a
black body increases very quickly, proportional to T4! The
spectral distribution of this power is not linear across the
electromagnetic spectrum, but limited to RF and microwave
frequencies a linear relationship can be assumed. The peak
power typically occurs well into infrared wavelengths and
shorter. Even for very low temperatures near absolute zero, the
noise temperatures at RF frequencies change linearly with
temperature defined by Boltzmann’s constant. For example, the
peak spectral power of a very cold 3 Kelvin black body is
centered at about 160.2 GHz. Although this can be considered a
very short microwave wavelength, technically it is still in the
RF spectrum. So for radio noise considerations, the Planck
constant need only be considered at superlatively very high
(RF) frequencies and measuring at very low temperatures. The
general equation across the entire electromagnetic spectrum
for thermal noise is:
(Equation 10.1)
where h is Planck’s constant, 6.626 × 10-34 J-s (Joule-seconds)
and k is Boltzmann’s constant (given in Equation 10.3). It was
Planck who first calculated (within a few percent) the now-
known value of Boltzmann’s constant.
Fortunately for radio engineers, when f is made
comparatively very low (RF frequencies), this equation
simplifies to a much simpler linear equation using Boltzmann’s
constant. When an object is very hot, the peak frequency
becomes very high. For example, the Sun’s highest power
spectral output occurs in the visible light spectrum. The Sun is
about 5000 K. If it were made hotter, the peak power would
increase, and the frequency of maximum output would also
increase.
Figure 10.3 shows thermal noise power plotted at four
different temperatures. Note that Figure 10.3 plots black body
317
radiation. The Sun, for example emits higher power on the
spectral lines of hydrogen and helium, which make up the bulk
of the Sun’s mass. Few “real” objects emit a “pure” black body
spectrum. If we look closely at Figure 10.3, we see the noise
power (Y-axis) converges into a mostly linear function at lower
frequencies (RF). Therefore, for the majority of considerations
concerning RF, the complicated general equation converges to
a much simpler equation:
(Equation 10.2)
for RF frequencies, where T is temperature (Kelvins), B is
bandwidth, and k is Boltzmann’s constant:
318
(Equation 10.3)
for 1 Hz of bandwidth at a rate of 1 second, where J is energy in
Joules, and K is the Kelvin temperature. It is interesting to note
that k is an expression of energy/temperature. How do we
derive a power term from kTB?
“Frequency” was defined by “cycles per second” (cy/sec.) until
the ISU (International System of Units) redefined “frequency”
as “Hertz” where today
(Equation 10.4)
The old “cycles” are now dimensionless and therefore simply
become the coefficient for Hertz — and thus in a noise
calculation, also power. So if
(Equation 10.5)
and J = W × sec (energy = watts × seconds), T = absolute
temperature, and B (bandwidth) = X/sec, therefore,
(Equation 10.6)
Thus kTB is in watts, with temperature (T) and bandwidth
(X/sec). Also, W × sec = energy.
319
Nyquist noise. Like black body radiation, noise voltage is
directly related to the absolute temperature (Kelvin) of the
resistor and the value of the resistor, where:
(Equation 10.7)
This is the noise voltage value in 1 Hz of bandwidth produced
by a resistor (R) at an absolute temperature (T). To derive the
noise power we can use the simple equation:
(Equation 10.8)
or
(Equation 10.9)
320
or
(Equation 10.10)
for power in 1 Hz of bandwidth.
Adding the bandwidth term, we get
(Equation 10.11)
There are two important points to be made regarding this
equation for thermal noise:
1) In the noise power equation the R term cancels, indicating
that the noise power generated in a resistance is independent of
the resistor value, whereas noise voltage and current values are
dependent upon the actual R value.
2) The coefficient 4 is a common source of confusion. This is
the voltage seen across the thermal resistor source. However, if
we want to transfer the power to an external matched load as
in Figure 10.5 the power becomes:
(Equation 10.12)
By the following derivation (from Equation 10.7):
321
Then the voltage found across R2 is
(Equation 10.13)
or,
(Equation 10.14)
And, simultaneously adding the bandwidth correction
(Equation 10.15)
This is the derivation of Boltzmann’s noise.
A final point which is very important when considering noise
in circuits and antennas is that although the power generated
in a resistor is independent of the resistor value, the noise
source will have an impedance equal to the value of the source
resistance. Figure 10.5 shows a Thevenin equivalent circuit
which is a good approximation for source impedances of finite
value. However, in the case of near-zero or zero ohm source
impedances in thermal noise modeling, the signal noise source
is actually a short circuit. Thus the very low resistance noise
source provides a “short” across the input impedance of the
load, and no power is transferred. In effect, the random
thermal noise power generated in a non-connected resistor
simply re-dissipates itself as heat. Therefore we have a noise-
voltage/current-noise cycle inside all resistors. So, a simple
carbon resistor sitting on your bench contains a frenzy of
electromagnetic activity! Extracting a portion of that noise
power is another matter. As we will see, for antenna noise
analysis, this source impedance is simply equated to Rl for
322
efficiency and noise calculations. For circuit noise calculations,
including amplifiers, the mathematics becomes significantly
more complicated.
Random Processes
Equation 10.12 provides a simple definition of thermal noise,
either radiated or conducted. At first glance we might expect to
measure a constant voltage or power due to Johnson-Nyquist
noise. This is not the case since thermal noise is created by a
random process and thus the instantaneous power value must be
governed by the laws of probability. Equation 10.12 defines a
mean value. The instantaneous value of a black body radiator is
defined by a Gaussian distribution.
Note that the theoretical probability curve never quite reaches
zero even for an infinite power level! Using the Boltzmann
constant we can compute an actual mean power level at any
temperature. This is the reason that Boltzmann’s constant is
represented as an energy term. The energy term, by virtue of its
time integration, also finds an average power over time. In this
case it represents the average of the Gaussian function over
time, or the peak of the Gaussian probability curve as in Figure
10.6.
323
At room temperature (290 K), the power in 1 Hz bandwidth is
(Equation 10.16)
in 1 Hz of bandwidth.
This is about 204 dB below 1 W. We have found that antennas
are usually specified in dBi or dBd. These are related to power
gain and express ratios, not absolute power levels. When we get
into radio systems engineering, we need to specify real power
levels. In Chapter 7 we introduced such an absolute power term;
now we will formalize it relative to RF system engineering. The
usual convention in radio engineering is the use of dBm, which
uses the reference:
324
0 dBm = 1 milliwatt (mW)
(Equation 10.17)
Therefore, Boltzmann noise at 290 Kelvins, also +16.85 °C
and +62.3 °F (or “room temperature”) is –174 dBm/Hz. The
noise power will be directly proportional to the bandwidth,
therefore we can simply add the term
10 log (B)
(Equation 10.18)
to derive
Pnoise = –174 dBm + 10 log (B)
(Equation 10.19)
at 290 K, where B is the bandwidth in Hertz.
(Equation 10.20)
325
for actual power.
Noise figure is defined in dB. Therefore, for the NF expressed
as power,
NF (dB) = 10 log (F)
(Equation 10.21)
For example, a “perfect amplifier” will not degrade the SNR of
the amplified signal, therefore, NF = 0 while the F is 1. When we
refer to a “device” we are referring to amplifiers, mixers, filters,
and other components of a radio. Other sources also must be
considered, including noise power that is received by the
antenna. Consequently, we can divide “noise” into two basic
sources: external to the receiver, and internal to the receiver.
Both must be considered in the calculation of the quality of a
communication link. However, the first step is defining the
receiver noise figure.
Finally, the minimum performance level of a receiver requires
some level of SNR for audio, video, or digital bit error ratio
performance. In effect, the quality of a link is directly related to
the ratio of signal power (or energy) to noise power (or
energy). In analog systems, we usually refer to the SNR, which
implies a power ratio. However, in digital systems it is more
appropriate to specify an energy ratio since the symbol rate of a
digital communication system implies that the symbol value
(amplitude, frequency, and/or phase) stays constant for the
time that symbol is transmitted.
(Equation 10.27)
(Equation 10.28)
where rs is the symbol rate and rbit is the bit rate. Thus, a
326
receiver performs a time integration (Tsymbol), which results in
an energy value in contrast to an analog signal’s power value
SNR.
(Equation 10.29)
where Eb is the “energy per bit” and No is the noise energy.
Therefore, at room temperature receiver sensitivity is defined
as
(Equation 10.30)
The conversion factor between SNR and Eb/No is simply:
(Equation 10.31)
where Rb is the channel data rate (bits/second) and B is the
channel bandwidth.
To summarize, Equation 10.30 simply states that at 290 K, we
first calculate thermal noise power per Hertz, then correct for
the bandwidth of the channel, degradation from the receiver
NF, and finally the required ratio of signal to noise (power or
energy) to maintain a predefined quality of the link. This is the
definition of receiver sensitivity.
For example, let us assume that we have a system that
requires 1 MHz of bandwidth, the noise figure of the receiver is
5 dB, and the required SNR is 10 dB. What is the sensitivity of
the receiver?
Sensitivity = –174 dBm + 60 dB + 5 dB + 10 dB = –99 dBm
327
(Equation 10.32)
This receiver requires a –99 dBm minimum of signal power
assuming that only thermal noise is present at the receiver
input.
328
We also assume that we have 5 dB of transmission line loss at
both sites, for a total of 10 dB loss. How much total antenna
gain will we need to provide minimum quality to the link?
From Equation 2.6 in Chapter 2 we know that the antenna-
to-antenna link budget (assuming isotropic antennas at both
ends) to be:
(Equation 10.33)
Now we can plug in the numbers and calculate the power
received:
(Equation 10.34)
After we subtract 10 dB from the transmission line losses, we
find –114 dBm at the receiver, or 15 dB lower than we require.
We can gain this 15 dB by adding a 7.5 dBi gain antenna at each
end; or a 10 dBi antenna at one end and a 5 dBi antenna at the
other; or use lower loss transmission line with lower gain
antennas. Any of these changes will result in an increase of 15
dB at the receiver. We could also increase the power of the
transmitter, or use any combination of these approaches. This
rather complex problem simply is a matter of adding and
subtracting decibels to the actual power (dBm) levels of
Boltzmann noise, transmit power, and necessary received
power. All the other terms are simply dB loss or gain, but we
begin with dBm (transmit power) and end with dBm (signal
power) at the receiver input.
Now suppose we have transceivers at both ends of the line,
and we assume both are identical to the transmitter and
receiver specifications stated above. Since links are reciprocal
we will find the power to the receivers at both ends to be
329
identical.
Noise Temperature
In the above example, we assumed that the noise at the input
to the receiver was –174 dBm/Hz since the assumption was a
system temperature of 290 K. Therefore NF is a temperature-
dependent value related to the reference of 290 K. Suppose by
Equations 10.20 and 10.21 we measure and calculate a NF of 3
dB, the F is 2 and let’s define the gain as 10 dB. Let us drive the
amplifier with a 1 W signal with an accompanying 1⁄4 W noise.
The output will contain 10 W of signal, but 5 W of noise. We
can see the input SNR is 4/1 and the output SNR has been
reduced to 2/1, therefore F = 2 and NF = 3 dB. We have assumed
an input noise temperature of 290 K and the amplifier has
contributed an equal amount of noise (before amplification).
Therefore the noise temperature of the amplifier must also be
290 K.
For NF or F, we can calculate the equivalent noise
temperature of any device by
Te = T0 (F – 1)
(Equation 10.35)
where
Te is the equivalent noise temperature
T0 is 290 K (since F is a function of 290 K)
F is the noise factor
Good quality low-noise amplifiers can exhibit noise figures
approaching 1 dB, even operating at 290 K. This corresponds to
a noise factor of 1.26 and by Equation 10.35 an equivalent noise
temperature of about 75 K. Thus the noise temperature of an
amplifier may be considerably below the ambient temperature
at which it is operating. This very beneficial but somewhat
paradoxical feature is explained by relationships in
semiconductor physics.
Since NF is taken as a log function of F, we must use an
inverse-log function to derive the NF.
330
(Equation 10.36)
We can also lower the Te and thus also the F of an amplifier
by physically cooling it down. The relationship is linear where:
(Equation 10.37)
Here we will begin with the same amplifier as above with a NF
of 1 dB, or a Te of 75 K. If we cool the amplifier down to the
liquid helium temperature of 4.22 K, we derive a new Te for the
device of 1.09 K. In Equation 10.37, our Te290 is 75 K. (The Te of
the amplifier at 290 K.) Thus the Te4.22 is a very low 1.09 K.
The effort of cooling an amplifier to such extreme low
temperatures is worthwhile is some specific applications which
we will soon address. With this background information we are
ready to describe an antenna specification critical to many
system-level design concerns.
Antenna Temperature
A resistor can be used to measure temperature by simply
measuring its noise voltage by Equation 10.7. However, a
resistor cannot measure temperatures below its own physical
temperature, while it is possible for an antenna to do so. In a
“perfect” antenna the “resistance” is radiation resistance (Rr),
which cannot be specified by an ambient temperature. The
concept is quite simple. If we measure the power level (dBm)
being received by an antenna, then we can calculate an
equivalent temperature using Boltzmann’s constant as the
relating term. However, now the noise power source is the Rr
of the antenna instead of a “real” resistor generating Johnson-
Nyquist noise. Rr cannot have a physical temperature.
Therefore in a “perfect” antenna there can be no Johnson-
331
Nyquist noise produced by the antenna itself as with a resistor.
Antenna noise temperature is only a function of the RF noise
power received by it. And the noise power is only a function of
the equivalent temperature of the matter in the antenna’s
steradian gain response. This “temperature” is coupled to the
antenna by the antenna’s Rr. The “matter” might be the ground,
the moon, or a distant galaxy. Here again, the voltage received
in 1 Hz bandwidth will be:
(Equation 10.38)
However, in this case, T is the equivalent temperature of the
noise/signal being received by virtue of the antenna’s Rr, not
the physical temperature of Rr. Therefore,
V2 = kTRr
(Equation 10.39)
and since
V2 = PRr
(Equation 10.40)
then
(Equation 10.41)
and
(Equation 10.42)
332
Thus the noise power (dBm) received by an antenna is
directly proportional to the antenna temperature in Kelvins.
Note: this assumes that the antenna is 100% efficient (lossless).
In the case where the antenna is not 100% efficient, we must
consider the relationship between the radiation resistance and
the loss resistance Rl (actual temperature of the antenna’s loss
resistance). This is similar to the efficiency calculation of
ground-mounted vertical antennas in Chapter 7.
See Figure 10.8. We have already explained the mechanism
for a simple noise measurement of a resistor (Figure 10.8A). If
we place an antenna whose radiation resistance is the same as
the resistor in Figure 10.8A in an anechoic chamber at the same
temperature, we will measure an identical voltage and thus
power level (Figure 10.8B). Finally, in Figure 10.8C, we point the
antenna at a part of the sky with an equivalent noise
temperature and again, we read the same noise power.
333
For example, if we use a 20 dBi gain antenna, we are looking
at 1⁄100 the area of the full spherical sky and measure X Kelvins.
Now we replace the 20 dBi antenna with a 23 dBi antenna, so
we are only viewing 1/200 of the spherical sky. Therefore we
might conclude that we are only receiving 1⁄2 of the thermal
power, therefore the temperature should be 1⁄2 X. However, we
just increased the antenna gain by 3 dB so the smaller area’s
power is now increased by 2. This leaves us with a temperature
again of X Kelvins (an identical noise power)! Also, the distance
(even to the edge of the universe) does not matter by the same
cancellation. Of course the sky temperature is not constant, so
if within the beamwidth of our antenna there are different
source temperatures, we need to adjust our measurement
technique. However, in many directions, sky temperature is
constant, even for lower-gain antennas.
Now let us assume that the antenna is not 100% efficient. If
the antenna is receiving the equivalent of 50 K of thermal noise
and the antenna is 99% efficient operating at 290 K, we can
assume that this resistive loss (due to antenna inefficiency) will
contribute thermal noise, and thus raise the antenna
temperature. Thus in this example antenna noise would be:
(Equation 10.43)
or, about 53 K.
Therefore we suffer a 3 K rise in noise temperature due to the
antenna’s inefficiency. The equivalent would be to place a
resistor with a value of 1⁄2 Ω at 290 K in series with a radiation
resistance of 50 Ω at 50 K. Unfortunately, if we point the
antenna toward the Earth, we will see an antenna temperature
close to 290 K, the ambient temperature of the Earth. However,
if we point the antenna at the sky, we must also minimize all
thermal radiation from the Earth and keep our system noise
temperature very low. Only then can our antenna become an
extremely sensitive heat detector for man-made and natural
334
objects. Even a small sidelobe response to the 290 K Earth will
overwhelm an attempt to measure very weak extraterrestrial
thermal noise sources.
However, if we know Rr and Rl of the antenna (usually an
easy calculation), we can subtract the contribution from Rl and
determine the temperature of a remote object, or even space
itself. But if we point our antenna to “empty” sky, it is
reasonable to assume the antenna should record zero power
since there is no matter emitting thermal radiation from
“empty sky.” This would be true in a “static” universe, but our
universe not static, so what will empty sky provide?
Figure 10.9 shows the first radio map detailing the
temperature of the sky. This map was made by Grote Reber,
W9GFZ (Figure 10.10), who used the home built antenna
shown in Figure 10.11 for his early radio astronomy
experiments.
335
336
A Nobel Prize
Over a period between 1922 – 1923, Edmund Hubble observed
that several celestial objects were actually outside the Milky
Way, eventually overturning the previously held scientific
orthodox that the universe was constant and limited to the
Milky Way. Further observations by Hubble and others showed
that the universe consisted of countless numbers of galaxies all
moving away from each other, some at near-light speed. This
observation led to the theory of the Big Bang. If all objects are
moving away from each other, then they must have had a
common point of origin. And, if that much matter was in one
“place” at one “time,” then the universe must have begun with a
Big Bang, much like a bursting fireworks explosion expels its
fragments away from the point of explosion in three-
337
dimensional space. Observing the rate of expansion and the
size of the universe, the age of the universe was/is calculated to
be about 13.7 billion years.
One of the implications of such a violent event was that
residual radiation should remain in the universe in the form of
black body electromagnetic radiation (including RF). Through
the mid-20th century, many predictions and spirited debates
took place regarding this topic. Most of the physicists
concluded that this residual noise would have a temperature in
the range of 2 – 5 K with the most likely range being 2 – 3 K. If
this residual temperature of the universe could be found it
would all but prove the Big Bang theory to be valid.
Then in 1964, two engineers, Arno Penzias and Robert
Wilson (Figure 10.12) were working on a project at Bell Labs to
reduce noise in microwave radio links. They became puzzled
by a minimum 3 K noise temperature no matter which way
they pointed their 6.2 meter horn-reflector antenna at 4 GHz.
They would later find that they had discovered the residual
energy from the Big Bang and received the Nobel Prize for
their discovery. We now know that the actual temperature is
2.7 K and the noise is in near-perfect conformance to the
theoretical black body frequency distribution predicted by
Planck. Thus an antenna was the indispensable tool for
observing the first direct evidence of the beginning of the
universe: microwave radiation from the Big Bang.
338
Penzias and Wilson found that the absolute minimum noise
an antenna can detect is 2.7 K, which i s about –194 dBm/Hz.
This will be true of any antenna located anywhere in the
universe and pointing in any direction! It also sets a
fundamental limit to background noise.
This measurement may seem a bit puzzling. Since the horn
antenna used by Penzias and Wilson was Earth-bound, the
antenna was pointed at the sky through the atmosphere. Since
the atmosphere is considerably warmer than absolute zero, one
might expect that they would have observed a much higher
value of thermal noise. In fact, the atmosphere is a “thin gas”
that doesn’t radiate (or absorb) black body radiation in the
same manner as “true” black bodies. Once again, the answer to
this dilemma lies in the laws of quantum mechanics. For our
purposes, the atmosphere does not emit black body radiation
over the frequencies investigated by Penzias and Wilson,
allowing the above measurements to be made. In contrast the
atmosphere has an extremely high black body temperature at
much lower frequencies.
339
There are countless objects in the observable sky that emit
much higher temperatures. If the antenna is pointed at these
objects, higher noise powers will result. Figure 10.13 is
provided to illustrate Earth-bound sky noise as a function of
frequency. By using Boltzmann’s constant the actual dBm/Hz
values can be easily computed.
340
below about 1 GHz. The air molecules in the atmosphere are
constantly being bombarded by electromagnetic radiation and
particles from the Sun. The effects are twofold: ionization of
the atmosphere (particularly at higher atmospheric elevations)
and electrical charging. Furthermore, rapid vertical movement
of liquid particles (particularly water droplets) in the
atmosphere adds to electrical charging. The most dramatic
electromagnetic noise generation is lightning caused by sudden
discharging of highly charged clouds.
The Milky Way Galaxy (our home) consists of billions of stars
and a myriad of other fascinating celestial structures. Many of
these structures emit radiation at radio frequencies as well.
Detailed study of these sources of radiation began after WWII
and became known as radio astronomy, which has clearly added
very significant data to the understanding of a wide variety of
celestial objects as well as fundamental discoveries in
cosmology.
Figure 10.13 also shows a minimum of natural noise occurring
between about 1 and 10 GHz. This is the range of frequencies
where most SETI (search for extraterrestrial intelligence)
activity has been conducted, since this band provides the
minimum sky noise for Earth-based observations. Therefore if
“they” were interested in interstellar communication they
might use these same wavelengths below about 10 GHz.
341
(Equation 10.44)
where h is Planck’s constant as in Equation 10.1 and f is the
frequency in Hertz. This is the amount of energy contained by
one photon. If we should ever need to know how many
photons we are emitting per second, we simply divide our
output power by the power/photon. Keep in mind 1 Joule = 1
watt-second (1 J = 1 W-s), and to keep terms in order, remember
that h = Js and f = 1/s, therefore we end up with Joules, an
energy term.
(Equation 10.45)
For example if we are transmitting 100 W at 144 MHz, we are
transmitting about 1027 photons/second. Notice that as we raise
the frequency but keep the power constant, we transmit fewer
photons.
When we consider the momentum of a photon, classical
Newtonian physics once again breaks down. “Momentum” in
classical physics must have mass, but a photon has no mass.
However, a photon must travel at the speed of light or it loses
its energy. Any “thing” that has no mass and no energy is a
“nothing.” The value of a photon’s momentum is the same
relationship for its energy (hf). However, since momentum is
usually stated as a direction as well as a function of energy, it is
usually expressed in vector notation.
This result is also a bit peculiar in that the power of a radio
wave can be calculated from its field strength or by the number
of particles/second. The answer, in watts, will be identical. This
is the essence of Einstein’s wave-particle duality of
electromagnetic radiation. The power of electromagnetic
radiation is not “shared” between the “wave” and the “particle,”
but coexists in both simultaneously.
342
Atmospheric Noise Considerations at MF
and Low HF
Under most circumstances, RF power at these relatively long
wavelengths does not penetrate the atmosphere. This is true
for both extraterrestrial galactic noise intersecting the Earth
and terrestrial noise propagating toward the sky. In both cases
the power is either absorbed and/or reflected. On the Earth we
observe this noise as thermal noise coming from a huge
equivalent noisy resistor (the atmosphere) under a huge
reflector (the ionosphere). However, the noise problem at
these frequencies is considerably worse than the Earth’s 290 K
thermal noise.
Figure 10.13 dramatizes the noise problem at these
frequencies. Figure 10.14 provides a more detailed plot of
atmospheric noise at MF and HF frequencies with the “noise
floor” on the graph being the noise temperature at 290 K.
Table 10.1 shows the relationships among noise power levels,
equivalent noise temperature, and the S unit equivalent.
343
The solid black line in Figure 10.14 shows the maximum
natural atmospheric noise. The natural atmospheric noise level
dominates the over-all noise power when it is at its maximum
and may be taken as the total contributor to the noise level.
Lightning is the major contributor, the equivalent of hundreds
of ultra-high-power spark gap transmitters keying at hundreds
of times per second. When propagation permits, this massive
power will appear at a receiver hundreds or thousands of
kilometers away. In effect, the actual nighttime noise level at a
given location at MF will be the grand sum of lightning
occurring over the Earth’s dark hemisphere and the quality of
propagation to these areas. This is the mechanism that creates
a constant “static” from the sum of thousands of strikes/second.
The dotted line shows the minimum natural noise level.
Before the spread of electrical grids worldwide, this would have
been the lower threshold for noise. Notice the huge variation
of natural noise possible at 1 MHz, about 100 dB. Again,
lightning is the main source of atmospheric noise at these
frequencies and varies widely depending upon the season, time
of day, and geographic location.
The line with alternating dots and dashes shows the galactic
noise. In a quiet location where man-made noise is low and
344
during periods of simultaneous low natural noise, galactic noise
may predominate the total noise power at frequencies above
about 5 MHz. The reason for the low frequency roll-off is the
fact that the source is extraterrestrial and thus the ionosphere
will reflect these frequencies back into space and not affect
terrestrial receivers. The actual roll-off frequency is determined
by the MUF and D-layer absorption.
The line with dashes shows the mean man-made noise level
at a quiet rural location. When natural noise falls to low levels,
man-made noise becomes the dominating factor. Like natural
noise, nighttime noise is typically much higher than daytime
noise because of propagation effects. The usual areas producing
the most noise are highly populated regions such as China,
Japan, North America and Europe. Consequently, pointing a
directional antenna in these directions will result in a higher
noise response, if, of course the band is “open” to these areas.
Statistically we can add to this noise level to adjust for
additional local noise. For these general locations simply add
the following expected power (dB) to this line:
1) “Normal” rural area, +11 dB
2) Residential area, +18 dB
3) Business/commercial area, +24 dB
Of course these are average difference values. Actual
additional noise due to proximity to populated areas can vary
widely. Marconi certainly had a much better noise
environment when he completed the first transatlantic radio
contact! Also, man-made noise will almost certainly increase in
the future. This will result in ever increasing difficulty for
working DX on the low bands.
The line with two short dashes and a long dash alternating
shows the “best” case of man-made noise at a quiet location.
Here, “quiet” location might mean Antarctica, a remote island,
or Greenland — but only when propagation to noisy areas is
not good.
345
In the previous section we described the various noise
sources in this frequency range. The problem at these
frequencies is similar to the problem of radio astronomy in one
important way. Both applications present the challenge of
detecting a point source on the hemispheric sky with that
hemisphere filled with noise. In contrast, the problem for MF
and low HF receivers is dealing with very high noise
temperatures, while Penzias and Wilson faced measuring
temperatures near absolute zero. The general idea of point
source versus hemispheric source points to some clear
objectives for MF and low HF receive antenna design.
If we assume that the sky has a uniform noise temperature
over its entire hemisphere (which it usually does not have!)
and we increase the gain of our receive antenna, the noise
power we receive will be the same as the hemispheric pattern
antenna, as in the radio astronomy explanation. However, if
we increase the gain of our receive antenna, a desired signal
will increase as a function of the higher gain since it is a point
source. Suppose we use an isotropic receive antenna with the
resulting total received power including both noise and the
desired signal,
Pr(iso) = Pn + Ps
(Equation 10.46)
Now if we assume the noise to be constant over the
hemispheric sky, and add gain (3 dB for this example) and,
more importantly directivity, then
Pr(3 dBi) = Pn + 2Ps
(Equation 10.47)
Notice that the noise power has remained constant but the
signal power has doubled. Thus we have increased our SNR by
a factor of two. Adding more directivity will further increase
the SNR linearly as the numerical directivity increases.
SNR ∝ directivity
346
(Equation 10.48)
The receive antenna efficiency can be relatively low since we
are mainly concerned with the received SNR and the received
noise is likely to be very high. Again the relationship is
Gain = Directivity × Efficiency
(Equation 10.49)
If our receive antenna is 100% efficient, then our antenna
noise temperature will be equal to the noise temperature of the
sky. For example, by Figure 10.14, a very quiet condition for 1.8
MHz implies an antenna noise temperature of about 290 × 104
K, which places the antenna thermal noise floor at about –134
dBm/Hz or –104 dBm for a 1 kHz bandwidth. Therefore the
received desired signal strength in dBm is used to compute the
SNR against –104 dBm noise.
Now, what happens when we use a small receive antenna
such as a 20 foot vertical, or a small loop? How much
inefficiency can we tolerate? For this calculation we use the
same equation for radio astronomy equivalent noise
temperatures, Equation 10.43. For example, let us assume we
are using a K9AY loop. From Figure 9.9 in Chapter 9 we see the
maximum gain is –26.6 dBi, which is very good for such a small
loop. We will first calculate the noise powers assuming that the
loop is isotropic in the hemisphere (+3 dBi). The efficiency of
such a loop is about 0.2%.
Tant = 290 × 104 (0.002) = 6344 K
(Equation 10.50)
We will also assume a weak signal condition, where the SNR
= 1. Assume 99.8% of the noise source for this antenna is
attributed to the actual temperature of the antenna Rl, and
0.2% will originate in the antenna’s Rr. We do not care what
the actual values of Rl and Rr are, since our given efficiency
defines the ratio. Rl is simply the ohmic value of the antenna,
and thus the noise power will be –174 dB/Hz. Notice that the
347
equivalent sky noise temperature is only about 13 dB higher
than the thermal noise emanating from the antenna’s thermal
resistance. Again, this is due to the antenna’s inefficiency, thus
lowering the “received” noise power.
The sensitivity equation (Equation 10.32) further lowers the
SNR due to the receiver’s noise figure. A typical noise figure for
MF receivers is 15 dB, placing the receiver noise temperature at
9170 K. Often the receiver’s attenuator will be used in an effort
to lower intermodulation and other nonlinear responses in the
receiver proper. These nonlinear receiver responses have the
effect of raising the equivalent noise response and/or creating
spurious signals in the receiver. In-band intermodulation
products are reduced a minimum of 3 dB for every 1 dB of
attenuation added. Therefore finding the “sweet spot” is a
balance of improving the intermodulation performance versus
not affecting the sensitivity limit. The “sweet spot” will change
for different noise power (and signal) inputs.
Thus for the given efficiency, a hemispherical response, and a
very quiet atmospheric noise condition, our weak signal will
not be detected. However, the K9AY loop also has directivity,
and by Equation 10.46 we gain back some of the SNR. It is
critical that the desired point source is aligned with the
maximum directivity of the antenna, or as close as possible.
The object is to maximize directivity in the direction of the
desired point source. Sky noise can be assumed to be arriving
from all elevation angles, while some distant weak desired
signal will be arriving at a relatively low angle. A small vertical
antenna will naturally reject the noise arriving from higher
angles, where a small loop will not. A low dipole will be a very
poor receive antenna under these circumstances since its
directivity maximum is “straight up.” The loop will respond to
low-angle vertical polarized waves, but it will also respond to
noise from “straight up.” Consequently the small vertical has an
inherent SNR advantage over the loop. Small loops that also
offer azimuth gain, such as the K9AY loop, will also benefit
from some high-angle rejection but not as well as a vertical.
Starting with the vertical and then adding elements results in
the additional advantage of azimuth as well as elevation gain.
348
The Beverage also has a low angle response and gain in the
desired direction. The Beverage also exhibits higher efficiency
than smaller loops, allowing more flexibility with receiver
controls. This explains why Beverages, vertical 4-squares and 8-
circle antennas are the usual preference for low frequency
receiver antennas, not to mention a quiet antenna location.
Multiple plots of these antenna configurations are presented in
Chapters 7 and 9. When considering the all-important SNR
figure of merit, the total steradian pattern must be used to
compare expected SNRs, especially high angle responses.
Unfortunately, these very effective receive antenna arrays tend
to be very large, compared with small loops. Therefore the
small loop might be the best (or only) choice when limited by
real estate.
Shannon’s Limit
There is one more expression to define the absolute physical
limits to any communication link, including radio
communication. Although not usually considered in an
antenna text, the Shannon Limit is the fundamental limitation
of any communication link, including all radio links. Since it is
a physical limit that also involves noise, we include at least a
basic discussion in Antenna Physics. Also by reviewing the
Shannon Limit, we find that noise becomes a fundamental
theoretical limitation to any radio link regarding the rate of
information conveyed as well as a fundamental theoretical
limitation to the dynamics of the radio link. Therefore,
antennas play a crucial role in determining the fundamental
theoretical limits of any radio communication link from both
physical limitations.
The receiver must interpret the received waveform as
containing some desired information. To this end the
waveform, or signal, must contain information that is unknown
to the receiver or there is no point in transmitting the
information. The format of the signal must be known by the
receiver, but the information carried by the signal is usually
unknown. For example, when we tune an SSB receiver we
assume a certain type of waveform, but the information we
349
receive may be unpredictable (for better or worse).
Claude Shannon (Figure 10.15) was working at Bell Labs in
the late 1940s when he provided the world with a remarkable
and profound insight: Information, in whatever form, can be
directly equated with a ratio of a desired “signal” containing
power or energy to that of the noise power or energy in a
“channel.” We usually think of “information communication” as
being some abstract phenomena. Shannon actually successfully
equated it to physical terms. His work directly led to the
formation of an entirely new discipline, Information Science.
350
(Equation 10.51)
where
C is the capacity in information bits/second
B is the bandwidth of the “channel” in Hertz
S/N is the signal-to-noise ratio
In this equation the capacity (C) term needs some
explanation. We usually think of a digital communication
system specified in bits per second. This says nothing about the
information per bit. “Information,” in turn, is only conveyed
when the receiver learns something from the transmitter it did
not know before the transmission.
A simple analogy is when someone tells you something that
you already know. No new information has been sent and
received. In effect the energy and time taken to transmit the
data was wasted. Now imagine the speaker knows exactly what
you know and don’t know. The speaker then cleverly crafts what
he/she says so that each word and inflection conveys the
maximum possible amount of new information without
repeating anything that you already know. The speaker also
knows exactly how much information you are capable of
receiving, the characteristics of your hearing, and so on. This is
an example of maximizing the information contained for a
limited acoustic energy expended, constrained by the audio
bandwidth in the presence of undesirable noise, and “noise”
within the listener’s hearing, nervous system, neural receptors
and whatever else affects the reception of the desired
information.
This example is closely analogous to a digital communication
system, including radio links, which, of course, include
antennas. Now imagine a digital communication system
sending some text written in English. If the transmitter sends a
“q” then there is no information conveyed by also following the
“q” with a “u,” assuming, of course, all English words that use a
351
“q” are always followed by a “u.” In effect, the time and energy
spent sending the “u” are wasted since the receiver already
knows English.
The next leap is to imagine that some code is developed that
permits each bit of information conveyed to contain the
maximum amount of information. Both the transmitter and
receiver know English, and both also have the ability to encode
(transmitter) and decode (receiver) the 100% efficient coded
message. Therefore, one of the implications of Shannon’s Limit
is to convey the maximum amount of information per bit
possible. This has resulted is a multi-billion dollar industry
developing and implementing codes, all to overcome the effects
of noise and the resulting bit errors.
The type of modulation used in a radio system must also be
optimized to provide maximum spectral efficiency. Of course
this topic also fills volumes. However, we must consider a
theoretically perfect modulation scheme together with some
ideal code to reach the Shannon Limit.
There is another intriguing implication of the Shannon Limit
equation. Notice that the equation does not have an error
term. The “limit” is quite literal. If the information is coded
with theoretical precision, there will be zero errors if the
equation is satisfied. Below the limit, no information can be
communicated. Above the limit the error rate is zero (again
assuming a “perfect code”).
As an aside, the equations that define the amount of
information in a bit are quite similar to the equations defining
entropy in thermodynamic systems. Despite the fact that there
is no relationship between information theory and classical
“entropy” (except the equations look similar), the amount of
information in a bit is termed its “entropy”! See Figure 10.16.
352
Finally, let us assume we could build a communication
system with zero noise. Using Equation 10.51 we find that the
channel capacity becomes infinite even for an arbitrarily narrow
bandwidth. With zero noise, we could run error-free Gb/sec
links through 1 Hz bandwidth channels. Also, any radio link
could have infinite range if all noise (and interference) could be
eliminated. Indeed the universe would be a very different place
without the existence of random radio noise being present
everywhere.
Even today many people, including many engineers, are not
aware of this incredible and counterintuitive understanding of
information and noise. The reader is encouraged to pursue
Shannon and information science further. However, from a
physical standpoint, Shannon’s Limit represents the starting-
point for defining limitations on any communication system. If
you know the noise level and bandwidth, you can quickly
calculate a physical limitation to the channel’s information
capacity. Finally, because this limit is a physical limit, it applies
to all information systems, radio, fiber, and even biological
systems, including human beings. Shannon’s Limit has found
direct applications in diverse areas as neuroscience and
353
electronic communication systems.
(Equation 10.52)
354
We recall from Equation 2.5 the isotropic loss is
In this case
(Equation 10.53)
or
(Equation 10.54)
or
(Equation 10.55)
This is the amount of EIRP and receiver antenna gain to
satisfy the minimum link budget. Now we can derive the
necessary antenna gains and transmit power.
355
At 6 GHz the aperture of an isotropic antenna is
(Equation 10.56)
Now let us assume that at each end are 200 meter diameter
parabolic antennas that are 100% efficient (or very close to it).
Therefore the antenna aperture at both ends is 31.4 × 103 m2 =
Aparabola. This represents a power gain of 15.7 × 106 or about 72
dBi for each antenna. Therefore from Equation 10.52 we derive
(Equation 10.57)
The value of the linear power gains of the two antennas
(here equal) in the link budget analysis are multiplied. This is
identical to adding the dB gain numbers. The necessary
transmitter output power is 1.45 million watts assuming zero
transmission line losses to communicate a maximum of 0.69
bits/second of information a distance of 100 light years. Thus
the equations defined by the physics of the antenna, radio
performance, and information theory apply to wireless garage
door openers the same way they apply to interstellar
communication links.
356
information we can send per unit time since we will have some
finite noise from the Big Bang. Third, we are forced to conform
to the physical realities of link budgets, the inverse-square law,
antenna aperture, receiver sensitivity, and transmit power.
These limitations are due to the fundamental physical
characteristics of free space itself.
In addition to these limitations defined by this physics, there
are further engineering limitations mainly constrained by the
state-of-the-art in all relevant areas: transmission lines,
amplifiers, mixers, data converters, and, of course, even
antennas are never perfect. They all have losses and a host of
other undesirable effects. Also, the myriad of matter’s physical
characteristics and the resulting effects upon antennas and
radio links must be considered.
From these daunting realities, we actually described a system
that is limited only by physics, not current engineering state-
of-the-art. This text attempts to touch on all the basic variables
pertaining to the antenna and the antenna’s place in a system.
Further advances in theoretical limits will have to wait for the
inevitable evolution of human knowledge and the resulting
upgrading of our understanding of the limits of physics.
357
APPENDIX A
A Simple Example
See Figures A.1 and A.2 for the following discussion. Imagine
a motionless car at a starting line. The car is not accelerating,
has no velocity and is at point 0 (starting line, X = 0) along a
long straight road. Now let us consider these three variables:
358
1) Acceleration
2) Velocity
3) Position (along the road (X-axis))
These three variables will be plotted against time. At time = 0
seconds, the car begins to move, and therefore has
359
acceleration. Acceleration is defined with two functions of time,
for example 5 miles/hour/second. In other words, the car has a
velocity of 5 miles/hour after 1 second, 10 miles/hour after 2
seconds, and so on. Also notice that velocity is defined with
only one term of time (miles/hour or mi/h). Position, however is
defined with no function of time. A mile is simply a mile along
the X-axis; it does not require time to define it.
Now, if we accelerate 5 miles/hour/second at a constant rate,
it is a simple matter to multiply this acceleration rate to
determine the velocity. For example, if we accelerate at this
rate for 10 seconds, our velocity will be 50 miles/hour. At any
point in time, we know exactly how fast we are going.
In this simple example, all we need to do to calculate velocity
is multiply acceleration by time where
(Equation A.1)
The seconds cancel and we are left with mi/h (or mph).
But what about distance traveled (X)? If our velocity is a
constant 50 mi/h, then the calculation is similarly trivial, we
multiply velocity by time to get miles.
(Equation A.2)
However, in our above example, velocity is changing for the
first 10 seconds, so we cannot simply multiply by time to get
distance traveled. Arithmetic breaks down and we need
something more sophisticated to “add up” the velocities over
time to get distance traveled. For this we need to “integrate”
the velocity over time.
For this operation it is more convenient to use feet/second as
the velocity, since miles will be a small fraction after only 10
seconds of acceleration. So let’s redefine acceleration as 5
360
feet/second/second.
We can now define the velocity as 5t. That is, velocity
increases at a rate of 5 feet/second/second; in this case we’ll use
time (t). So after, say, 4 seconds, our velocity will be 20 mi/h.
If we want to find how far we have traveled during the first 10
seconds, we need to set up an integral from t = 0 to t = 10. The
equation we use is a definite integral with the start and end
points, and we define the velocity as a function of time (5t)
since the velocity is simply 5 feet/second/second (the
acceleration) multiplied by time. Now to determine the
distance traveled in the same 10 seconds, we cannot simply
multiply the velocity × time because the velocity is gradually
changing with time. We need to use integral calculus to
calculate the function of velocity over time. Again the distance
traveled (X) is equal to the function of velocity integrated over
the proper amount of time (10 seconds).
(Equation A.3)
The actual calculation of the time to solve this problem does
involve knowledge of integral calculus. However,
understanding that it is the method of determining the
distance traveled at the end of 10 seconds is a giant leap to
understanding what integral calculus is. We will cheat here and
present the actual solution:
(Equation A.4)
Now we take two points, where t = 0 and t = 10. Finding the
difference will provide the answer. When t = 10, the result is
250 feet. When t = 0 the result is 0 feet. Therefore
X = 250 – 0 = 250 feet
361
(Equation A.5)
Therefore, starting from zero velocity, with an acceleration
of 5 feet/second/second, after 10 seconds the car will have
traveled 250 feet. Similarly, we can calculate the total distance
traveled for any time. After 5 seconds, the car has traveled 62.5
feet. Without calculus an exact calculation is impossible.
362
meaning we are finding the rate of change of x as t changes.
Again we will cheat and present the answer for the
derivative. This is simply finding the rate for change in velocity
over time. In this case, as above, the velocity is increasing for
the duration of the 10 seconds (as in a drag race). Thus the
velocity is simply 5 mi/h/sec (t) or
363
It’s really that simple!
Now we can explore other examples. Energy is the time
integral of power. For example, if we run a 1 kW machine for 1
hour, we have used 1 kW-hour of energy, another simple
calculation. However, if we are constantly changing the power
being used according to “smooth” functions, we will need
calculus. For example, suppose we have a transistor amplifier
that exponentially increases its power until there is a large
surge of power, the transistor is destroyed and all power
consumption suddenly stops. If you know the function that
defines the destruction sequence you can calculate the energy
consumed by the sequence. In this way, your power meter is
performing a power integration to tell the power company how
much your bill should be. It has integrated all the power you
have used, no matter how nonlinear the consumption has
been, to yield the integral of the power over time: the energy
used.
364
This is known as a partial differential equation. This equation
informs us how the dielectric constant of a material changes
with frequency and with temperature. Of course we need some
starting point value here represented by a reference function.
By plugging in the actual numbers, we can perform an actual
real calculation. The above equation only illustrates the form.
Again, solving for an actual solution usually requires several
years of training in advanced mathematics, but understanding
what it means does not.
The applications are endless: the rate of growth of trees over
different temperatures, humidity, soil types, sunlight, and other
factors. Or we could analyze the rate of weight gain in humans
as a function of exercise, carbohydrate intake, genetics, toxic
exposures, and so on.
In Chapter 3 we presented Maxwell’s equations using a
special form of an integral
365
variables all interdependent with respect to a final desired
term, in this case loss resistance.
We live calculus functions all the time. They define almost
everything that changes as a result of something else changing.
We solve them all the time without knowing that we are doing
so. Many relationships (such as driving a car) are second nature
and become instinctive. Understanding calculus permits
insights into a very wide breath of problems impossible
without it.
To conclude, many if not most of the complicated looking
equations presented in advanced science and engineering texts
are just a more sophisticated way of defining something that
simpler techniques cannot. That’s what calculus is all about,
but you already knew that.
366
APPENDIX B
Summary of Rr Terms -
Described by Kraus
Most authors of advanced antenna texts discuss radiation
resistance. Of all the authors, Kraus offers the most
comprehensive treatment and for the most part is in agreement
with the other authors. Reference: John D. Kraus, Antennas,
2nd ed, McGraw-Hill, 1988.
In particular, the discussion of Rr for ground-mounted
vertical antennas is a common source of confusion. Kraus
provides several approaches to solve for Rr under different
circumstances. I have introduced a new term (Rrp) to
distinguish between the Rr of the antenna versus the Rr at a
particular point along the antenna (Rrp). Following is a
summary of these terms.
1) (from Kraus, page 216)
367
heat!).
Using computer modeling techniques, power integration is
typically performed over a surface, such as a sphere. Within a
lossy dielectric power “radiated” into the dielectric and “lost” to
heat are both distributed through the three-dimensional
dielectric volume. Consequently an accurate differentiation
between radiated and absorbed power is nearly impossible. We
rightly consider underground radiation to be “lost,” but a
calculation of Rr must account for all radiated power, “desired”
or not.
The portion of power that can be directly attributed to
antenna loss and not propagation loss is best calculated by a
direct measurement of Rr at a point of maximum current and
then compared with a modeled antenna over lossless ground.
Only loss in the return current can provide the proper term to
calculate losses in ground-mounted verticals. This value simply
appears as an equivalent resistance at the feed point.
Theoretically it is possible to actually perform a power
integration that will correctly differentiate radiated and lost
power, but this would likely be an exceedingly difficult model.
For direct measurements we must somehow know that the
antenna is lossless, since there is no compensation for loss in
this equation. This equation also does not provide a term for
the Ae or the he of the antenna, which additionally limits its
usefulness. Although this equation does automatically
compensate for a change in Rr for Ae, it does not provide the
necessary terms to tell us why.
The reader should be wary of any attempt to directly
calculate Rr using this equation, particularly for ground-
mounted verticals.
2) (from Kraus, page 847)
368
equivalent of the power input term from the simpler equation.
However, the radiated power term is defined in terms of gain,
which in turn, can also be easily converted to aperture, or in
this case, to a power density concentrated into a steradian
value. The actual definition of this general equation is the 3 dB
beamwidth, again stated in steradian value. A 3 dB beamwidth
corresponding to a steradian value approximately contains all of
the power radiated. However, if we hypothetically create an
area on the sphere that contains all the power (defined by a
“solid” steradian angle) normalized to the maximum gain, this
equation becomes precise.
From a practical view, the 3 dB beamwidth is usually easier to
measure than the Poynting vector over the entire sphere, so
this equation is not only closed form but is also much more
useful. Since we now know the gain, we also know what the
lossless feed impedance will be, so all the terms become
deterministic, including loss.
3) (from Kraus, page 42)
369
proportional there can be no change in Rr in the presence of
loss, unlike the implication from Equation 1. Also, this equation
provides the accurate effect gain has on Rr, an inverse
proportion. This equation represents the proper definition for
the ground mounted vertical since it includes all relevant
variables in a closed-form.
In the case of very low conductivity dielectrics, it is
exceedingly difficult to separate radiation into the dielectric
and actual heat loss in the dielectric. Eventually the radiated
power will indeed be lost to heat, but in the far field where it
can be considered as a propagation loss, not an antenna loss.
From a practical sense, any power lost in the dielectric is loss,
radiated or lost to heat in the near field since it is not radiated
into free space. However, the value of Rr “doesn’t care” it only
responds to power lost in the antenna proper (including
through mutual impedances) or power reflected back to the
antenna. Again, extreme care must be taken in the actual
calculations of “loss” and “radiation.”
Using relatively simple antenna modeling programs such as
NEC, it is possible to set up a model that can force return losses
into an equivalent single resistor and eliminate the effects of
far field loss simultaneously. The equivalent loss resistor is
placed just above the ground, with the feed point located just
above it, both very close to the ground. This simple technique
will be suitably accurate for most “real” grounds. Antenna loss
proper will converge upon a simple single resistance. The
antenna is modeled using a perfect ground to eliminate
propagation losses. A decrease in he and thus efficiency will
track perfectly with a reduction in Ae (gain) just as Equation 3
states.
4)
370
not present this special equation, but does derive a similar
expression for very short dipoles. This equation is published by
many sources, and can be seen to be a direct derivation of
Equation 3. For ground mounted verticals less than about 1⁄8 λ,
the gain versus physical height becomes nearly constant at 4.77
dBi, or a power gain of 3. Thus the aperture simplifies to
371
APPENDIX C
References
The following articles and texts were valuable in the course
of preparing this book.
372
Limits on the Radiation Q of Electrically Small Antennas,”
IEEE Transactions on Antennas and Propagation, Vol 44, No. 5,
May 1996, pp 672 – 675.
Silver, H. Ward, Ed., The ARRL Antenna Book, 23rd Edition
(ARRL, 2015).
Taub, Herbert and Schilling, Donald, Principles of
Communication Systems, 2nd Edition (McGraw-Hill, 1986).
Uda, S., “On the Wireless Beam of Short Electric Waves,”
Institute of Electrical Engineers of (JIEE), Mar 1926
Wangsness, Roald K., Electromagnetic Fields (Wiley & Sons,
1979).
Wheeler, H. A., “Fundamental Limitations of Small Antennas,”
Proceedings of the I.R.E., Vol 35, No. 12, Dec 1947, pp 1479 –
1484.
Wheeler, H. A., “The Radiansphere Around a Small Antenna,”
Proceedings of the I.R.E., Vol 47, No. 8, Aug 1959, pp 1325 – 1331.
Yagi, H., “Beam Transmission of Ultra Short Waves,”
Proceedings of the I.R.E., Vol 16, No. 6, Jun 1928, pp 715 – 740.
Zavrel Jr., Robert J., “How Antenna Aperture Relates to Gain
and Directivity,” QEX, May/Jun 2004, pp 35 – 39.
373
APPENDIX D
area
A
ampere
Ac
collecting aperture
Ae
effective aperture
Aem
374
Al
loss aperture
Ap
physical aperture
As
scattering aperture
bandwidth
B
bit-error ratio
Coulomb
C
capacity bits/second
375
C
capacitance (farads)
C
directivity
D
decibel
dBi
376
polarized isotropic antenna
dBd
energy
Eb
377
f
frequency, Hz
F
noise factor
gain
Gr
Plank’s constant
he
effective height
hp
378
physical height
current
Iave
Joules
j
Boltzmann’s constant
ke
dielectric constant
length
L
379
inductance (Henrys)
meters
Newton
NF
noise figure
N0
noise energy
power
PADL
noise power
380
Pr
radiated power
Pr
received power
Prr
re-radiated power
Pt
transmitted power
quality factor
q
electrical charge
resistance in ohms
Rb
bit rate
Rl
381
loss resistance
Rload
terminal resistance
r
radius
second
SNR
382
signal-to-noise ratio
temperature in Kelvins
Tant
290 Kelvins
voltage
V
velocity
Vn
noise voltage
VF
velocity factor
383
impedance
Zfeed
reflection coefficient
ε0
wavelength
θ
384
a degree or radian value, corresponding to
latitude in spherical coordinates
φ
ohms
Ω
beam area
385
APPENDIX E
Comparison of Elevation
Plots of Vertical-Polarized
Ground-Mounted 160 Meter
Antenna Arrays
These three plots (Figures E.1 to E.3) are presented as a
simple visual comparison among various configurations using
the same gain scales. All data assumes lossless ground systems
over average earth. The relative shapes of these plots are valid
for higher frequencies as well, so this may serve as a basic
comparison for higher frequencies. However, the absolute gain
numbers will change with different frequencies. ARRL plotting
is used to offer more detailed graphing for small differences
among the gain figures. All three graphs include a 1⁄4 λ vertical
for reference comparisons.
386
387
388
389
390