markovM阶
markovM阶
Information Signal IF
Modulator Demodulator
Source Generator Channel
Information Signal IF IF
Source Generator Transmitter Receiver
RF RF RF
Information Signal Modulator
Transmitter Channel Receiver
Source Generator Transmitter
(Media)
6g
|
| | ; Channel Layer 6 |
|
|
| | Channel Layer 5 |
|
|
| | Channel Layer 4
|
| Channel Layer 3
Channel Layer 2
Channel Layer 1
Figure 1.5 Information Transportation System (ITS) pyramid; uncoded communication system functional architecture
Signal information
Detector b | Sink
| Demodulator
Signal
Detector Information
Sink
IF Signal Information
Demodulator
Receiver Detector Sink
-» Digitized by the Internet Archive
in 2022 with funding from
Kahle/Austin Foundation
https ://archive.org/details/digitalcommunica0000simo
Digital Communication Techniques
Signal Design and Detection
Marvin K. Simon
Jet Propulsion Laboratory
Sami M. Hinedi
Jet Propulsion Laboratory
William C. Lindsey
LinCom Corporation
University of Southern California
The publisher offers discounts on this book when ordered in bulk quantities:
Corporate Sales Department
PTR Prentice Hall
113 Sylvan Avenue
Englewood Cliffs, NJ 07632
Phone: 201-592-2863
FAX: 201-592-2249
All rights reserved. No part of this book may be reproduced, in any form or by any means,
without permission in writing from the publisher.
ISBN: 0-135-200610-3
Marvin K. Simon would like to dedicate this book to his wife Anita in appreciation for her
consultation at every step of the way through the preparation of this manuscript. Indeed
this author could have never completed his portion of this book nor for that matter any
of his previous books without the support, encouragement, guidance, and sharing in the
many anxieties that accompany a project of this magnitude. This book is also dedicated to
his children Brette and Jeff who have made him very proud through their development and
accomplishments both as human beings and in their respective fields of interest.
Sami M. Hinedi would like to thank his father Mounir and his mother Hikmat for their
constant support, encouragement and guidance and for the various sacrifices they made
in order to provide their children with the best education one can hope for. He also would
like to thank his sister Fadwa, his brother Taha and last but not least, his dear uncle Abdul
Kader for their help and for being there when he needed them most. This book is dedicated
to all of them.
William C. Lindsey would like to dedicate this book to Dorothy, my wife, and to John, my
son, for the support system they have provided me throughout by professional career. Also,
to Corinne Leslie, my secretary for 27 years, for her loyal, continuous, and dedicated sup-
port; and to my sisters, Loretta and Jean for their lifelong inspiration and encouragement,
reer,
aAO eer
695@ (Ee .
a oe Pret inves
7 Ne :
= EY
5 awl et
* pw Sea”
= \ irc bed ie
5
;
=
Contents
Preface XV
Index 879
Tg
Preface
—_—————[— js SSS
a Venn diagram for error correcting codes is constructed; emphasis is placed on present-
ing the communciations efficiency achieved using Hamming, Golay, Bose-Chaudhuri-
Hocquenghem (BCH), Reed Solomon (RS), and convolutional codes.
There are a vast number of textbooks on the market today that deal with the subject
of digital communications. In fact, in a quick survey of the textbook literature, we were
able to come up with at least ten books that include the words “digital communications” in
their titles. At least three bear these words as their entire title implying, somewhat, that they
are an all-inclusive treatment of the subject. Many of these books are quite broad in scope,
but also quite shallow in detail. Striking a proper balance between these two attributes,
yet maintaining a high level of readability, is no simple task. We believe that our book
accomplishes this important goal and sets it apart from all other digital communication texts
currently on the market. Several key features that distinguish our book from the others are
as follows:
There are many specific features that make this book unique and beneficial to its read-
ers. With the advent of today’s advances in the solid state microelectronic technologies, a
variety of novel digital communication systems are appearing on the market and are serv-
ing as motivation for the introduction of new telecommunication and information services.
Chapter 1 of this book provides the reader with examples of such services and top level
system architectures thereby indicating the highly complex nature of these systems. We
believe that Chapter 2, which discusses the computation of power spectral density of dig-
ital modulations, is the best treatment of spectrum efficiency evaluation found anywhere.
This computation is essential to assessing the bandwidth (spectral) occupancy requirement
of a digital modulation, yet it is ignored in many books. Another key feature of this book
is the organization and order of presentation of the material in Chapters 3 through 7. By
first describing coherent detection and then successively following with noncoherent, par-
tially coherent, and differentially coherent detection, the reader is provided with a logical
flow starting with the conceptually simplest technique and proceeding top down to the more
complex techniques. The discussion of double differentially coherent detection in Chapter 8
is unique to our book.! Here the reader will learn how to design differentially coherent com-
munication systems that are robust to frequency offsets due, for example, to Doppler and
_ oscillator instabilities. Chapter 9 treats the voluminous subject of bandlimited communica-
tions in a condensed and unified way. Included here are the important subjects of Nyquist
1. There are a few Russian-authored textbooks that discuss double differentially coherent detection, but as yet,
they have not been translated into English.
Preface XVii
eet
with permission from problems created and used by Professors Robert M. Gagliardi, Vijay
Kumar, William C. Lindsey, Andreas Polydoros, Charles L. Weber, and Robert A. Scholtz
xviii Preface
all of the University of Southern California’s Communication Sciences Institute. The au-
thors wish to thank these individuals for permission to use these challenging problems as
homework exercises.
The authors are grateful for the help of many colleagues and students during the
preparation of this manuscript, and are particularly indebted to: Dr. Michael Mandel, for-
merly a graduate student at the California Institute of Technology; Dr. Debbie Van Alphen,
Professor of Electrical Engineering at California State University, Northridge; and Dr.
Jorge M.N. Pereira of GTE; for their comments and criticisms. Furthermore, we thank our
colleagues at the Jet Propulsion Laboratory, Pasadena, CA, namely: Dr. Tien Nguyen, Dr.
Haiping Tsou, Dr. Daniel Raphaeli, Mr. Biren Shah, Mr. Sampson Million, Mr. Alexander
Mileant, Mr. Mazen Shihabi, Dr. Ramin Sadr, Mr. Roland Bevan, and Dr. Victor Vilnrotter
for reading and reviewing various parts of the manuscript.
Marvin K. Simon
Sami M. Hinedi
William C. Lindsey
Pasadena, CA
Chapter 1
Introduction
to Telecommunications
Today’s advances in solid state electronic technologies are impacting the design and imple-
mentation of various digital communication systems, and specifically, terrestrial and space-
related telecommunication networks and services. Many systems/networks that once were
considered impractical due to their complexities are now well within the reach of current
technology and are presently being designed, implemented, and operated. Many microelec-
tronic technologies are maturing so fast that what is considered impractical today may well
be implementable tomorrow. As a result, novel telecommunication and information services
are continuously being introduced into the market.
At the system component level, microprocessors are operating faster and faster, front-
end amplifiers are achieving gains in frequency regions that were once considered unattain-
able with noise figures that approach the theoretical limit, antenna surfaces are becoming
“smoother,” thus enabling higher frequency transmission and reception. Various other ad-
vanced technologies, especially wireless, are affecting and sometimes dictating the system
design architecture and strategy. At the systems level, satellites are moving into an era
of multiple beams and high-gain spot beams allowing for frequency reuse, supplementing
global communications coverage and even allowing the use of “very” small antenna aper-
tures. Cellular terrestrial and proposed space radio systems around the world are testing
digital systems to replace their analog FM equipment, while cordless telephones are moving
outside the boundaries of the home and may merge with cellular systems, enabling users to
carry phones with them. Mobile satellite experiments [1] are currently being carried out to
test the next generation system in which a mobile user, located in a car, on a plane or a ship,
can directly communicate with a satellite, bypassing the need for a relay station. Existing
telephone networks are being upgraded using fiber-optic technology [2] to allow for higher
data rates and better transmission quality enabling novel information services. These ad-
vances and others are resulting in improved system performances, higher reliability, easier
operability and maintainability. Consequently, systems life cycles are increased dramatically
and life cycle costs decreased. Global mobile and personal communication systems are the
wave of the future, and their applications are endless.
Telecommunication engineering and more specifically, coding, modulation, ‘demod-
ulation, decoding, detection, and synchronization techniques are the technical topics to be
covered in the chapters to follow. Fundamentals of signal detection, being coherent, non-
coherent, partially and differentially coherent, are the key methodologies to the design of
2 Introduction to Telecommunications Chap. 1
optimum receivers. The choice of the modulation scheme from the “simple” binary phase
shift keying (BPSK) to the more “elaborate” quadrature amplitude modulations (QAM) is
essential in trading channel bandwidth and achievable bit error rate for a given and fixed
transmitted signal power. The synchronization function is fundamental in operating any
communication link. At the receiving site, various timing and phase references are needed
and are derived from the incoming noisy signal; these include carrier phase and frequency
estimation, possibly subcarrier phase and frequency, and definitely symbol (bit) timing to
recover the transmitted information bits. Synchronization precludes communication and in-
cludes these two functions: acquisition and tracking. Each of the carrier, subcarrier, and
symbol acquisition processes consists of first frequency, then phase acquisition, and the total
lock-up (or acquisition time) depends on the specific structures or algorithms employed. The
transition from acquisition to tracking function is nonuniquely defined and is typically said
to have occurred when the instantaneous phase error decreases and remains below a pre-
determined threshold. Automatic gain control (AGC) circuitry is essential in maintaining
reasonable received power levels in the receiver and in providing some protection against
“Jarge” intentional or accidental interferences. Finally, signal reference generators at both
the transmitter and receiver require some degree of time and frequency stability to maintain
a fixed reference throughout the system.
This chapter introduces the reader to the architecture of a digital communication
system and introduces the various terms and key parameters used throughout the book.
Section 1.2 presents practical “telecommunication networks” and discusses future networks.
The various elements and key functions of an end-to-end communication link are discussed
in Section 1.3. The key performance parameters such as signal-to-noise ratios, bandwidth,
and so on, are defined along with the various losses that need to be accounted for in a
link budget analysis. The “information” capacity of a communication channel as defined
by Shannon is discussed in Section 1.4 and in Chapter 11. Communication with subcarriers
and data formatting of various signals is the topic of Section 1.5.
transmitter. Simultaneously, a transducer is also needed at the channel output of level one,
and this is accomplished in level 2 by adding a communication receiver to interface with the
information user. Thus, we conclude that the basic building blocks of a digital communi-
cation system are a transmitter, a physical channel (communications pipe), and a receiver.
These blocks will be further partitioned and discussed.
As we shall see, the system engineering and design of a digital communication sys-
tem best proceeds top-down, while implementation and system integration proceeds best
bottom-up, that is, top-down design driven by system requirements followed by bottom-up
integration of the system building blocks yields a system that meets the system user require-
ments. From the system requirements, system engineers generate design specifications and
from these specifications, implementations (software and hardware) are perceived. Execut-
ing this top-down design process, which is driven by the system requirements and bottom-up
integration, constitutes what is frequently referred to in practice as Systems Engineering and
Integration (SE&]I).
The interfaces between the functional blocks are control signals and are specified by
interface control documents (ICDs). In what follows, a further breakdown of the functional
architecture of the information transportation system will be presented in the form of a
hierarchy. All systems are hierarchical in nature—one engineer’s part is another’s system.
The transmitter, channel, and receiver can easily be identified at each level in the hierarchy.
Figure 1.2 serves to demonstrate the building blocks for the communications transmitter
and the communication receiver in level 2 of a digital communication system architecture
in Fig. 1.1. This figure shows that each system building block is a distinct architecture within
a larger architecture in the hierarchy. These building blocks will be modeled and elaborated
upon below, as well as in Chapters 3 through 13.
We observe from Fig. 1.1 that communication system architecture centers or pivots around
the communications channel. The architecture of this channel can best be demonstrated
by introducing the channel hierarchy. The architecture of this channel hierarchy can be
conveniently represented using six levels; these levels are identified in Fig. 1.3. In this
chapter and Chapter 2, we are primarily concerned with modeling and analysis of levels
4, 5, and 6. In the remaining chapters, we will be incorporating channel levels 4, 5, and 6
as a part of the IF channel in level 3. In other words, the mathematical characterization of
the IF channel must serve to mimic the features of the RF channel, the Antenna-to-Antenna
Channel, and the Physical Channel. In a sense, the IF channel model will serve as a “virtual
channel” model that mimics the physics of the hardware and the physical channel at levels
4, 5, and 6. The concept of a virtual channel is key to understanding the modeling of the
communications process. A virtual channel is a set of resources necessary to complete the
communication link between information source and user. In this book, we elect to model
their effects, for the most part, below level 3 in the Information Transportation System (ITS)
hierarchy. This model will consider the effects of power (energy) reduction as a function
of the building blocks and source-user separation primarily below level 3 of the system
hierarchy. In the following section, we further identify the building blocks and features at
each channel level in the hierarchy demonstrated in Fig. 1.3. In essence, the building blocks
of the communication channel at any level are always defined by those building blocks in
the architectural levels below.
Figure 1.4 demonstrates the communication systems engineering education pyramid.
UOITELUIO}U|
oneuodsues)
wa}shs
eo ee ee ee oe ae ee
4UIS
20/N0S
UOITELUJO}U|
UOITEWUJOJUY
UONTEWIOJU]
ug
UOyeTUNUTIUIO?7)
wiaysAS oInps}IYIV
eansiq
*] poylduig uoNeorunWIUIOD
wia}shs sinjoajtyore pojutidoy)
wor ([g¢]
Joyiuusuel]
UOIVEUOJU|
MO|4
JOAIB0eY
-10po0eq
pueqeseg -UMOG dy Bulaisoay
JOSS89014 BARa|JA]UIBG JeyeAUOD JEAlIe0ey euuajuy
aanSIy
7] JoyIUIsuURLL
pue JaAtooa1 Surpyingsysojq payutday)
wou ([g¢]
jauueyo
puegeseg
|eUUBYDO
jeoisAyd jeuueyy
CuUUsUY-0O}-eULA}UY
OWA.
= ejel0siq ssejAuowey
jouueyD
VANSI
=F" JouULYD sINJOOIIYOIe
SNSIOA UONROTUNUIUIOD
Wa}SKs ainqoaytyose poyutidoy)
Wor ([g¢]
Sec. 1.1 Digital Communication System Functional Architecture 7
This pyramid summarizes the relationships between the theory needed to perform analysis
and/or system simulation at various channel levels in the overall architecture. This figure
demonstrates that the lower the level to be worked, the larger the knowledge base required
from a system engineering viewpoint.
In considering the more detailed breakdown of the architecture at the lower levels in the
hierarchy, it is convenient to break this into two components, viz., (1) the functional archi-
tecture of an uncoded digital communication system and (2) the functional architecture of
a coded digital communication system. Chapters 3 through 10 will be concerned with the
theory and techniques associated with the design and optimization of an Uncoded Digital
Communication System; the performance criterion will be “bit error probability.” Chap-
ters 10 through 13 will provide the necessary interface that will lead to presenting error
control coding techniques associated with the design and optimization of a Coded Digital
Communication System; the performance criteria will be the “Ro criterion” and bit error
probability.
Figure 1.5 demonstrates the Information Transportation System (ITS) Pyramid for the
architecture of an uncoded digital communications system. At each level in the architecture,
there exists a channel that serves to characterize and model the physical features associated
with the signal flow path below it. This channel hierarchy serves to define (one-to-one)
the elements associated with the channel layers illustrated in Fig. 1.3. The frequency up
converter present in practice between the IF and RF transmitter and the down converter
between the RF receiver and IF receiver (see Fig. 1.2) have been included respectively in
the RF transmitter and receiver building blocks. In this chapter, we are primarily concerned
with modeling and characterizing the power transfer and channel effects in channel layers 5
and 6.
When information source and error control coding techniques are used to mitigate the dele-
terious effects of channel interference and memory on communication transmission, the
functional architecture changes. Figure 1.6 demonstrates the /7S Pyramid when source and
error control coding techniques are assumed. Comparing the pyramids in Figures 1.5 and
1.6, we see that a source encoder and a channel encoder are required on the transmit side
and a channel decoder and a source decoder are required on the receive side. In practice,
these elements are frequently referred to as the source codec and the channel codec. The
informed reader will note that the interleaver and deinterleaver functions demonstrated in
Fig. 1.2 have been omitted in Fig. 1.6; these functions, designed to alleviate channel mem-
ory effects, are assumed to be incorporated in the channel coder and decoder of Fig. 1.6,
respectively. The details of these functions are covered in Chapters 11 through 13 from a
system design and engineering perspective. As before, the channel at any level of the coded
system architecture serves to incorporate the features of the channel below it with regard to
any deleterious effects on signal (information) flow.
The ITS Pyramids serve to demonstrate the feature of the signal (information) flow
path (the communication channel) that constitute hardware versus those that constitute the
communication medium (that is, cable, waveguide, atmospheric, and so on.) In essence,
the ITS Pyramids serve to answer the question: “What functional elements constitute the
([g¢] wor pojutiday) pruresAd uonvonpe Bulisaursue suiajsks UONeOIUNWIWOD = pT dAINSI
uolediunwWwoD
seubewospaja|y
Aioay|
voyesiunwwog
Aoay,
Buipog
Buipog Koay
‘uonewojul
‘UOoITeWO}U}
WUOjU}
BR
‘uolyewsoju|
Buipoa
pue
pue
pue
|uonesiunwwog
Aioay pue
|seuBewose|3
Aioay
‘UOIeWJOJU|‘UO!TeWJO}U}
‘UONEIUNWWOD
‘BuUBlUY
‘uoNeduNWWOD
Buuajuy
uoneBedoig
soiskyd
aeauBewos9e/3
Aioay|
pue
euvey)he]
g
BUUEUDv9A27
p
puueussafe)
|
ans
S*T
= UONULIOJUT uoneyodsuesy,
waysks(Sif) ‘prureshd papooun[eu8Ip
UONeoTUNUIUIOD
Waishs [euoNUNy oinjooytyose pajutiday)
wor ([¢]
uy) J0]09
uo yur pus (1oded
(saded pus yoeq uo 10[09 uy)
([g¢] wos poyutidoy) ainjooy1yose [euONoUN] W9\sAs UONROTUNUIWIOD
[eusIp pepoo ‘prureshd (SJ) waiIsks uoyeyodsuvsy] UONeUIOJUT QT ans
aatpadsiag Joautsuq suiayshS
Sec. 1.1 Digital Communication System Functional Architecture 11
Source Information
ee
Ee ee
ane ieriveneend| Se
ite pyencrypll ln) unas
(Seer
(RC
eeoe
ey ae
Ret |
Pr pate 5 paces |\inih
{ eneMansmion oped 4)
ay he eae pps Pa ee
|
|
|
i a [REmitTranemit
Antenna
J} Chann
Moderel
| Trans
Antenna-to- _
| RF Channel
Antenna Coded
IF Channel | Channel! Channel
|
aE pReceNE ee
ence en
|
[IF
Receive |
[—Demodulate
| __ —
Deinterleave
| J
4
\
ieaarer
en
fa
espe
a = eee
User Information
communication channel from the theoretical perspective?” The digital communication sys-
tem building blocks elements which define the channel, the signal processing functions, the
building blocks and hardware/software interfaces are integrated in the communications pipe
illustrated in Fig. 1.7. It is suggested that the reader carefully examines Fig. 1.7 to under-
stand this important communications pipe.
In digital communication networks, data and voice can be sent from one point to another in
one of two fashions. A circuit connection can be set up in the network between the source
and the user. Then the message is transmitted, and when it finishes, the circuit connection
is deleted. Another method for communication is to break up the message into smaller
packets and transmit the packets through the network (possibly through different paths) to
the destination, and then reconstruct the message at the user site. The latter technique results
in packet-switched data networks while the former technique results in circuit-switched
networks.
Figure 1.8 demonstrates the seven layers that make up the ISO-OSI Network Archi-
tecture for packet-switched networks and shows how the architecture demonstrated by the
ITS Pyramids fits into the Physical Layer (Layer 1) of the ISO-OSI Architecture: Open Sys-
tem Interconnection (OSI) network architecture is an approach to interprocessor commu-
nications developed by the International Organization for Standardization (ISO) and is the
focus of that research effort. For details of this important and detailed standard, the reader
is referred to the basic material contained in [4—10]. The functions of each of the layers are
summarized in the following list.
[pour sNJOaITYOIe YIOMJOU [TSO-OS] g’T ans
amypa}yIY YIOMION
sake
eyeg
YIOMIONyur
Jake
4eAe7 yodsuel]
12
Sec. 1.1 Digital Communication System Functional Architecture 13
1. The Physical Layer: The architecture of this layer is presented in Figs. 1.5 and 1.6. It
is further detailed in Fig. 1.7. This layer is the physical communication medium that
supports the transmission of bits and is concerned with the modulation and demodu-
lation process and functions such as synchronization and framing.
2. The Data-Link Layer: This layer provides the basic link frame, including informa-
tion packet, start and end headers, and an error check field. It ensures that packets are
reliably transmitted through the network and it shields higher layers from any con-
cerns about the physical transmission medium. The service provided by the data link
layer is the reliable transfer of a packet over a single link in the network. This is ac-
complished through positive and negative acknowledgments for error checking in the
link.
3. The Network Layer: This layer provides the networkwide (or internetwork) services
required to transmit, route, and relay data between computers (or nodes) without
regard for the communications medium. This layer generates (at the source) and reads
(at the destination) the address information and control fields for purposes of flow
control. It is responsible for establishing, maintaining, and terminating the network
connection between two users and for transferring data along that connection.
4. The Transport Layer: This layer is responsible for providing data transfer between
two users at an agreed on quality of service (QOS). When a connection is established
between two users, the layer is responsible for selecting a particular class of service,
to monitor transmissions to ensure the appropriate service quality is maintained, and
to notify the users if it is not.
5. The Session Layer: This layer provides dialogue services for those functions that re-
quire the establishment of connection and synchronization between any two machines
prior to the exchange of information. This layer provides the “Are you there”—“Yes
I am” exchange prior to the exchange of application data. A primary concern of the
session layer is controlling when users can send and receive, based on whether they
can send concurrently or alternately.
6. The Presentation Layer: This layer provides a common representation of application
data that is communicated between application entities. Common representation refers
to the encoding of data and the order of bits and bytes exchanged between processors.
This may include character code translation, data conversion, or data compression and
expansion. For example, the exchange of data between a processor using American
Standard Code for Information Interchange (ASCII) encoded characters and a proc-
essor using Extended Binary Coded Decimal Interchange Code (EBCDIC) encoded
characters requires a data translation before the information can be utilized.
7. The Application Layer: This layer provides the user program with an interface to an
OSI system. In this case, the user is any computer program or application program that
requires interprocessor communications. The layer includes services used to establish
and terminate the connections between users and to monitor and manage the systems
being interconnected and the various resources they employ. A number of common
communication functions have been identified and grouped into so-called application
entities with standard services and protocols. The application entities are the key to
open systems.
14 Introduction to Telecommunications Chap. 1
Telecommunication networks are used extensively in the modern world to interconnect ge-
ographically separated nodes or “terminals” for the purpose of providing a specific service.
The elements of a network can be spread over a relatively small area, such as a building or
a university campus, or over cities or even continents. The elements or nodes themselves
can be stationary or dynamic depending on the application. The number of nodes typically
ranges from a few elements to a few thousand or more. Each node can consist of a simple
telephone or a computer terminal to a more elaborate “receiver” or signal processing cen-
ter. The network performs a function or provides a service from which its name is typically
derived. Basically, any interconnected set of elements that communicate with each other in
some manner qualify as a telecommunication network. A very simple network can consist
of a set of two personal computers that communicate with each other to share a common
data base. It is worthwhile pointing out at this time the technology forecast made in 1945
by Arthur C. Clarke [11] about using rocket stations to provide worldwide radio coverage.
It ranks among the most accurate and history-changing of all technology forecasts in mod-
ern times. The majority of current networks rely on satellites or “rocket stations” to com-
municate the information necessary to provide their respective services. Examples of such
networks are the National Aeronautics Space Administration’s (NASA’s) Deep Space Net-
work, the Telecommunication and Data Relay Satellite System, and the Global Positioning
System.
The Deep Space Network (DSN) [12, 13] is the largest, most sensitive scientific telecom-
munications and radio navigation network in the world. Its principal responsibilities are to
support automated interplanetary spacecraft missions and radio and radar astronomy ob-
servations in the exploration of the solar system and the universe. The DSN also supports
high-Earth orbiters, lunar and Space Shuttle missions.
The DSN is managed, technically directed, and operated for the National Aeronautics
and Space Administration (NASA) by the Jet Propulsion Laboratory (JPL), California Insti-
tute of Technology, Pasadena, California. The DSN is a collection of three communication
complexes located in California (USA), Madrid (Spain), and Canberra (Australia) that sup-
port deep space missions. Each complex consists of basically three or more “large” antennas
that receive and transmit commands or data to the space probes. The location of the com-
plexes was chosen to provide global support and maximize coverage availability. Figure 1.9
depicts the DSN’s largest antenna, which is 70 meters in diameter. Each node of this net-
work consists of either a complete communication complex or a deep space spacecraft. The
network is nonstationary in the sense that the relative distance between some of the nodes is
changing as a function of time. The use of large antennas in addition to supercooled front-
end receivers has enabled the DSN to receive some of the weakest signals (—185 dBm) used
in communication systems.
segment elements, TDRS-East and TDRS-West, are shown in Fig. 1.10a and are located
respectively at about 41° and 171° west longitude in geosynchronous orbits. A spare satellite
(TDRS-Spare), located at 61° west longitude, can be repositioned at either location in case
of a satellite malfunction. The SN provides tracking and data acquisition with more than
85% coverage for orbit altitudes of 200 to 1,200 km, and 100% coverage for orbit altitudes
of 1,200 to 12,100 km. Return data rates of up to 300 Mb/sec from user spacecraft are
supported, as are a number of different services. In TDRSS-II depicted in Fig. 1.10b, the
Zone of Exclusion (ZOE) is removed using the Out-Of-View (OOV) TDRS-II. Return data
rates of up to 600 Mbits/sec can be supported. TDRSS supports the primary communication
links to the space shuttle, as illustrated in Fig. 1.11.
The third network to be discussed is NAVSTAR’s Global Positioning System (GPS) [15,
16]. Its purpose is to provide three-dimensional position and velocity information as well
as time transfer to users anywhere in the world at any time. The position determination
is based on measuring the delay of radio frequency (RF) signals from four satellites of
a total constellation of 24. A minimum of four measurements are necessary to solve for
the four unknowns—three-dimensional position in addition to the user’s clock offset. The
24 satellites are shown in Fig. 1.12 and operate in 12-hour orbits at an altitude of 20,183
km. The constellation provides visibility of six to eleven satellites at five degrees or more
above the horizon to users located anywhere in the world at any time. The four satellites
offering the best geometry can be selected automatically by the receivers using ephemeris
16 Introduction to Telecommunications Chap. 1
ANTENNA TDRS-WEST
TDRS-EAST TDRSS GROUND STEERING LONG. 171° WEST
LONG.41° WEST TERMINAL ANGLE
LAT. 32° NORTH (225° EW A,
LONG. 107° WEST 31° N/S)
Se SEs
S. SPACECRAFT
=
ZONE OF EXCLUSION
(FOR USER
SPACECRAFT HEIGHT
OF 1200 km OR LESS)
The Integrated Services Digital Network (ISDN) and its broadband version have been un-
der development for several years and are currently undergoing field tests [18,19]. The three
fundamental principles embodied in ISDN and broadband ISDN are digitization with high
bandwidth, worldwide standards, and integrated functionality and services. ISDN will fi-
nally integrate voice and data services enabling novel applications limited only by the hu-
man imagination [20]; for example, personalized newspapers in which an individual, using
his home “terminal,” can retrieve the sports section from the Los Angeles Times, the food
section from the New York Times, the world news from the Christian Science Monitor, and
so on, and print all sections in the form of one “personal” newspaper. Another example is
the cost-effective home office in which an employee can communicate with his employers
through voice, facsimile, and video simultaneously, thus enabling him or her to hold meet-
ings with fellow employees also located at home or in different company locations.
In the military domain, the next generation satellite communications (MILSATCOM) sys-
tem, designed to serve both the nation’s strategic and tactical forces, is currently in the
Sec. 1.2 Telecommunication Networks and Services 17
46°W
rr Spare TORS RS | Il 171°W
.
“<n
174°W
Nominal
In-View Relay (IVR) In-View
TDRS Il Non-Relay (IVNR)
TDRSII
7443°E
Out-of-View (OOV)
TDRS Il
deployment phase. The new system, called MILSTAR [21] and depicted in Fig. 1.13, will
provide a worldwide, highly jam-resistant, survivable and enduring MILSATCOM capa-
bility under conditions of both electronic warfare and physical attack. Through the use of
extremely high frequency (EHF) and other advanced techniques, the system will recover
very quickly from the propagation degradation caused by a high-altitude nuclear detona-
tion. MILSTAR satellite crosslinks will operate in the W-band (60 GHz) to prevent eaves-
dropping and interference from enemy ground stations. They will also enable message relay
around the world above the atmosphere without reliance on intermediate downlinks to vul-
nerabie ground stations. This capability would ensure, for instance, that a presidential jet
flying over the Pacific would be able to communicate with a tactical nuclear force in Europe
even if all ground stations had been destroyed.
Satellite systems for mobile and personal communications and their interconnection with
the Public Switched Telephone Network (PSTN) will no doubt be the next layer of telecom-
munication infrastructure developed for purposes of global communications and navigation.
Mobile satellite communication systems (MSCS) are capable of delivering a range of ser-
vices to a wide variety of terminal types. The architecture of a future MSCS is illustrated in
Fig. 1.14a with objectives depicted in Fig. 1.14b; namely, communication from one hand-
held terminal to another hand-held terminal anywhere in the world. Examples of mobile
satellite terminal platforms include land vehicles, aircraft, marine vessels, remote data col-
lection and control sites. Additionally, services can be provided to portable terminals, which
are currently the size of a briefcase and no doubt will be reduced to hand-held sizes for fu-
ture systems. Many future services will include position reporting to determine the position
Introduction to Telecommunications Chap. 1
18
SHUTTLE
GROUND STATION
WHITE SANDS, N.M.
of the mobile user, with transmission of information via the mobile satellite communications
system. In the future, Mobile Satellite Services and Radio Determination Satellite Services
(RDSS) will provide the ultimate in Global Personal Communications when interoperated
with the Public Telephone Switched Network. Ultimately, mobile satellite systems should be
capable of providing personal communications services (PCS) such as voice and data (low-
rate) to very small terminals including hand-held units. Applications requiring much higher
data rates should be possible with somewhat larger mobile and portable terminals. Mobility
and high capacity are the major new requirements that these future systems must satisfy.
At least three approaches have been proposed to the Federal Communication Com-
mission (FCC). These include different network architectures consisting of satellite constel-
lations in low Earth orbits (LEO), medium Earth orbits (MEO), and geosynchronous orbits
(GEO). In addition, Europe is considering satellite constellations in highly eccentric (ellip-
tic) orbits (HEO). These various orbits are depicted in Fig. 1.15. Presently, the International
Maritime Satellite (Inmarsat) [22] Organization and the American Mobile Satellite Corpo-
ration (AMSC) systems utilize GEO. 7
Currently, there are several proposed systems: (1) Motorola’s proposed/ridium [23]
and Loral/Qualcomm’s proposed GlobalStar [24] consisting of a satellite constellation in
LEO, (2) TRW’s proposed Odyssey [25] consisting of a satellite constellation in MEO, and
Sec. 1.2 Telecommunication Networks and Services 19
(3) Europe’s proposed Archimedes [26] consisting of satellites in HEO. It is clear that the
multinational development of global wireless communications will be a ubiquitous reality
in the twenty-first century. Table 1.1 summarizes the characteristics of the proposed systems
in terms of number of satellites, number of orbit planes, orbit altitude, and so on.
The final system/network architecture associated with this wireless world is difficult
to speculate on at this time; however, it seems safe to predict that before the year 2000, small
shirt-pocket size, low-weight, full-featured personal communication hand-held terminals
will be available with long talk time and high quality voice, data and position location
capabilities that can be used anywhere in the world—either at home, in the office, on
the street, or while on travel with either terrestrial and/or satellite facilities. This personal
communications network (PCN) will be implemented such that any of the network facilities
will be totally transparent to the end user. In the next few sections, we will give a brief
overview of several systems proposed as the infrastructure for future generation wireless,
mobile, personal communications.
Inmarsat. Inmarsat [22] is currently the world’s largest mobile satellite commu-
nications organization, backed by more than 60 national telecommunications enterprises,
as well as a worldwide network of equipment manufacturers, distributors, and service
providers. Established in 1979, Inmarsat has its headquarters in London and has 64 member
countries as of May 1992.
The worldwide services that are currently available through Inmarsat include: direct-
dial telephone, facsimile, telex, electronic mail, data transmission, flight deck voice and
data services, position reporting, and fleet management. In the near future, global paging,
briefcase telephones, and voice and data for the business traveler will be available. Inmarsat
currently uses a total of 10 satellites in its global system. While some of these satellites are
carrying commercial communications traffic, others are in orbit as backup spares to ensure
continuity of service. The satellites, as shown in Fig. 1.16, are arranged in geostationary
orbit to provide communications coverage of all areas of the globe with the exception
20 Introduction to Telecommunications Chap. 1
og ey
a
of the extreme polar regions. To provide enough capacity to meet the huge expansion in
demand, Inmarsat has acquired Inmarsat-2 satellites for the 1990s. A new generation of
Inmarsat-3 satellites with dedicated spot beams is in production for launch in 1994—95, and
its characteristics are shown in Fig. 1.17.
|. Radio determination service and two-way messaging: All subscriber units, except
some pagers, will contain this capability, which includes automatic location reporting.
Units with GPS or GLONASS services will also be available, which will provide
greatly increased geolocation speed and accuracy.
NM Digital voice communications service: The system will provide duplex high-quality
4.8 kb/s voice communications.
3. Paging: An alphanumeric pager for instantaneous direct-satellite paging in any region
of the world is being designed in a package only slightly larger than today’s pagers.
4. Facsimile: Two types of mobile facsimile units are planned, a stand-alone facsimile
unit and a unit to be used with an /ridium telephone.
5. Data: A 2.4 kb/s modem is being developed for use within the Jridium network.
Figurell4a Mobile saciluc communication
sysiem achaccme
The backbone of the Iridium system is the constellation of 77 small satellites in po-
jar Low-Earth-Orbit (LEO) at an altitude of 420 nautical miles (778 km). This relatively
icate
further
low altitude assures that 2 user will never have to commun than about 1.250
nautical miles (2,315 km) to a satellite. This allows the use of lightweight low-power radio-
telepho nes
with small low-profile antennas compared to those required toutilize geostation-
ary saicMines at ranges of more than 19,000 nautical miles (35.188 km). The constellation
comprises seven planes of 11 satellites, each plane equally spaced in longitude. The orbital
period of each saseMite is approximately 100 minutes.
The satellites are interconnected via microwave crosslinks, thus forming a global
network in space, with linkage to the ground via gateways. Each gateway interfaces with
the local Public Switched Telephone Network (PSTN), providing local area user/customer
recon keeping and interconnection capability between Iridium and non-Iridium users. The
decision to locate a gateway in any given area can be made by local governments and/or
servers wilely out of a desire tooffer local service, rather than to satisfy a functiona need
l
by the system for thegatew to ay
be a node in the networ to k
assure global coverage. Satellite
telemetry, tacking and control, and overall system control will be accompanied by one or
more centrally located System Control Facilities utilizing earth terminals and microwave
finks.
The radio links between users and satellites will operate somewhere in the 1.600 MHz
“1 1,700 Mz region of L-Band, while the gateway feeder links and satellite crosslinks
rat
will opein e ofthe 18 GHz to 30 GHz region of K-Band. L-Band is best
small portions
22 Introduction to Telecommunications Chap. 1
suited for direct satellite-to-portable-user links because it offers a good mix of relatively
short wavelengths and low propagation losses, and it includes the upper frequency limits
of economical commercial hardware using available technologies. Each satellite employs
an L-Band phased array antenna complex that forms 37 interlaced spot beams (cells). Each
beam operates independently. Each carrier of each beam has its downlink transmit power
electronically controlled to maintain link quality for each user, allowing for variations in
range and blockage/foliage. Further design details can be found in [23].
Number of satellites 12 48 66
Number of orbit planes 3 8 6
Orbit inclination (deg) 55) By) 87
Orbit altitude (km) 10354 1389 765
Intersatellite links no no yes
Repeater type bent-pipe bent-pipe processing
Frequency Bands (GHz)
¢ User to satellite 1.6 1.6 1.6
* Satellite to user Ds pie 1.6
Multiple-access mode CDMA CDMA TDMA
High-elevation angle operation, that is, greater than 30°, reduces the possibility of
signal blockage for a mobile user. The use of steerable antennas permits the coverage of
selected areas of the Earth for extended periods of time, as the satellite passes overhead.
This significantly reduces the amount of beam-to-beam switching that must be done and
simplifies the overall system design. For RDSS purposes, the system employs a combination
of ranging measurements and Doppler measurements to determine the location of a user.
For the user/satellite links, the 16.5 MHz frequency bands are divided into three sub-
bands of 5.5 MHz each. Each beam, of the 19 beam user/satellite antennas, transmits or
receives in only one of the three sub-bands. The antennas are designed to have similar spot
beam sizes and beam positions in both the 1.6 GHz and 2.5 GHz bands. For satellite/Earth
station link, each of the 19 received or transmitted sub-bands are frequency shifted to a dif-
ferent frequency requiring 16.5 MHz total bandwidth. This value includes data channels,
24 Introduction to Telecommunications Chap. 1
order-wire channel, and guard bands. Because of the relatively high values of rain atten-
uation in the 20 and 30 GHz frequency bands, diversity Earth station techniques will be
employed. Further detailed discussion of the design and performance can be found in [25].
Marecs A&B 2
Launch weight 1,006kg
Height 2.56m
Solar array span 13.8m
Type 3-axis stabilized
Capacity Approx. 75 two-way
voice circuits
Launch vehicle Ariane
First launch 1982
Manufacturer British Aerospace/Marconi
Inmarsat-2
Launch weight 1,230-1,320kg
Height 3.36m
Solar array span 15.23m
Type 3-axis stabilized
Capacity 250 two-way voice circuits
Launch vehicles Ariane/Delta ‘“
First launch 1990
Manufacturer British Aerospace consortium
intelsat-V
Launch weight 1,970kg
Height 6.58m
Solar array span 15.59m
Type 3-axis stabilized
Mobile Capacity 50 two-way voice circuits
Launch vehicles Atlas Centaur/Ariane
First launch 1982
Manufacturer Ford Aerospace and
Communications Corp.
Marisat
Launch weight 655kg
Height 3.81m
Diameter 2.61m
Type Spin stabilized
Capacity 10 two-way voice circuits
Launch vehicle Thor Detta
First launch 1976
Manufacturer Hughes Aircraft Co.
inmarsat-3
Scheduled for launch In 1995
Launch weight c.1,900kg
Height c.2m
Solar array span c.21m
Type 3-axis stabilized
Capacity 2,000 two-way voice circuits ¢ sa,
Launch vehicles = Atlas Centaur/Ariane/Proton a iy
Manufacturer Martin Marietta Astro Space
Figure 1.17 Characteristics of satellites used by the Inmarsat organization (Courtesy of Inmarsat)
the fact that the Archimedes satellites move with respect to a point on the Earth’s surface.
These differences are: the requirement to regularly hand over from one satellite to the next,
differential path length, and differential Doppler shift. These factors add to the complexity
of both fixed and mobile terminals as circuitry has to be built in to mitigate these effects.
Services are provided in much the same way as for a GEO satellite system. Broadcast-
ers uplink their channels via high frequency fixed feeder links to the satellite as shown in
26 Introduction to Telecommunications Chap. 1
Subscriber ;
Calls Subscriber
Calls
L-Band
Control/Status
K-Band
Fig. 1.21. The satellite then transmits the broadcast signal at the lower mobile communica-
tions frequencies to enable reception by mobiles throughout the coverage area. Figure 1.21
illustrates the communications system architecture showing various potential user types.
The overall coverage of the satellite system is Europe, but to reduce RF power per chan-
nel, it is advantageous to use high-gain spot beams. The formation of these spot beams is
intrinsically linked to the choice of frequency due to the laws of antenna geometry.
A critical parameter in determining the cost per channel of the Archimedes system is
the capacity of the system, in terms of the number of channels the satellite communication
system can simultaneously support. Satellites are limited in the amount of power that they
can supply to their communications payloads, and therefore, the capacity of the system is
directly proportional to the power requirement of each channel onboard the satellite. The
first generation Archimedes system is intended to serve Europe. Europe has a relatively high
latitude and therefore, the elevation angle to a GEO satellite is relatively low. The propa-
gation environment is very important for mobile communications and can be significantly
improved if higher elevation angles are given. This drives the consideration of alternative
orbits, particularly highly eccentric (elliptic) orbits (HEO). The increased elevation angles
given for HEO can be converted into decreased RF power requirements by considering prop-
agation margins. This is the most severe propagation environment Archimedes is intended to
be served by terrestrial retransmission. These measurements show the minimum advantage
of HEO over GEO to be 7 dB.
The mobile terminals are intended to be very simple and hence have nonsteerable
antennas. This means the satellite must be within the antenna beam at all times. For HEO,
Sec. 1.3 Performance Criterion and Link Budgets 27
@® TDMA/FDMA Combination
77 Processing Satellites
@ L-Band for Both Up- & Down-Link from User
488 KG; 1429 W
®@ Complex Networking
the satellite is always within a cone directly overhead and therefore a highly directional
antenna can be used, whereas for GEO, the satellite can be anywhere within a fairly large
toroid and therefore, a low-gain toroidal coverage antenna is needed. For further details of
the proposed system, see [26].
There are too many telecommunication networks proposed or in existence to enumer-
ate or discuss, but the examples described above are representative of the majority. In con-
clusion, a network is an interconnected set of “elements” or “terminals” that communicate
with each other to provide a particular service.
Both power and bandwidth play a key role in system design and analysis. Often, they
must be traded off against each other. The average power of a wide-sense stationary (WSS)
random process x(t) is given by
+00
where R,xx(t) is the autocorrelation function! of the random process and S;x(f) its re-
spective Power Spectral Density (PSD) [27] in units of Watts/Hz. By replacing statistical
1. The autocorrelation function R,,(t) and its corresponding power spectral density S,,.(f) are often denoted
using one subscript, such as R,(t) and S,(f).
28 Introduction to Telecommunications Chap. 1
averaging with time averaging, the above definition can be extended to include ergodic pro-
cesses. Although the power of a waveform can be defined precisely, the notion of waveform
bandwidth is difficult to define and is sometimes vague and misleading. The basic objective
for specifying bandwidth is to indicate the frequency range over which most of the wave-
form power is distributed. For random processes,” bandwidth can be derived from the PSD
Syx(f) which is the Fourier transform of the autocorrelation function R,,(t). The most
commonly used bandwidths (in Hertz) are the “3-dB bandwidth,” the “p% fractional power
bandwidth,” the “Gabor or root-mean-square bandwidth,” and the “equivalent noise band-
width,” defined as follows (one-sided definitions):*
1. 3-dB Bandwidth (B,_ qp): Most useful for unipodal PSDs, it is the width along the
positive frequency axis between the frequencies where S,,(f) has dropped to half its
maximum value, that is,
2. When discussing bandwidth of a filter H(f), replace the PSD S,,(f) with the magnitude squared of the filter
transfer function, that is, |H(f)|*, and all subsequent equations are valid.
3. We assume that the PSD is symmetric, that is, S..(f) = Sxx(—/f).
Sec. 1.3 Performance Criterion and Link Budgets 29
uy
Sixx (B3_ dB) == 5 Sxx max (1.2)
This definition has the advantage that it can be computed directly from the spectral
density function. It has the disadvantage that it may be misleading for spectral densi-
ties with slowly decreasing tails and it may become ambiguous and nonunique for
spectral densities with multiple peaks, that is, there may be several frequencies at
which the spectral density is half its maximum value.
. p% fractional Bandwidth (By): It is the frequency width along the positive frequency
axis within which p% of the one-sided spectrum lies, that 1s,
(Pe Sxx(f)df we
(1.3)
Spare 100
where f2 — fi;= Bp. This definition has the advantage of being unique. However, it
can be misleading for spectral densities with somehow slowly decaying tails. Also,
it has the disadvantage that it must be evaluated by integration of the power spectral
density.
. Gabor Bandwidth (B;ms): It is defined as the root-mean-square (rms) bandwidth of
the PSD, that is,
(oe) 1/2
Brms = {| PSec Dat
| (1.4)
0
This bandwidth is mostly used in radar systems and it also has the disadvantage of
being evaluated by integration.
. Equivalent Noise Bandwidth (B,,): It is the most common definition used in studies of
systems with stochastic inputs. It is defined as the width along the positive frequency
axis within which a flat spectral density gives a value of power equal to that of the
given power spectral density where the value of the flat spectral density corresponds
to the maximum of the given spectral density (see Fig. 1.22). In mathematical terms
30 Introduction to Telecommunications Chap. —
Radio
Station Voice and Data
Telephone
Radio Station
Radio Only
Satellite Control Centre
: No )
Sxx(f) = = |H(f) | (1:7)
Pee i SUC
(o.@)
/ SH) Pag
CO N in
ee)
(1.8a)
If we were to pass the white noise through another brick-wall filter with gain
Sec. 1.3 Performance Criterion and Link Budgets 31
Sxx (f)
ERR
Figure 1.22 The noise bandwidth
|H(f)|max and two-sided bandwidth W,,, then the output process y(t) has power
N
P= Was HDs max (1.8b)
For the two processes x(t) and y(t) to have identical powers, we require P, = Py
resulting in
ae
ae oc me 2B, (1.9)
he ae TG A Pe
In this context, W,, is the total (positive and negative) spectral width of an ideal (brick-
wall) filter whose amplitude characteristic is equal to the maximum of |H(f)| and
which yields the same power at its output as does H(f) when white noise is present
at the input. For any of these bandwidths, the fractional out-of-band power,. Pop, is
given by
i i
A(f)= a+j2xf | Ap a
a
It is clear that
A (f )max = H(0) = 1
fo HAP af _
Ig? Hf) Pdf
which gives
Boo
and
_, (27 Boo 1 6.3la
tan Ij“
( F ))}=0.9=;
05 Boo=
90 an = 6.31B 3- dB
5, Drape a
22]
Mi 2
| H(f) | df =atan eae ———
0 a 0 4
Therefore,
a= a
= ie eerily
Example 1.2
For an N-pole Butterworth filter, we have
Pek ON ee ry
1+ (24)
anf
and the 3-dB bandwidth is still given by a/27. However, the noise bandwidth becomes
o0 I
By =f) Fae
i aaa
a
and simplifies to
__@/2N)
7 sin(t/2N) 3-4B
As N gets larger, the noise bandwidth approaches the 3-dB bandwidth, as expected.
#
o) 3
TRANSMISSION
i
a DESIGNATION
a
5) APPLICATIONS
MEDIA > 9
= i
10° ULTRAVIOLET Ks
10
OPTICAL FIBERS z VISIBLE LIGHT OPTICAL COMMUNICATION
10°
é INFRARED 14]
es (1 micron)| a 10 1c
30-300 GHz
WAVEGUIDES, LINE-OF-SIGHT EXTREMELY HIGH fs 10!! RETR E HRe Gre ee
+ cm FREQUENCY (EHF) hed SYSTEMS
LINE-OF-SIGHT RELAYING, 3-30 GHz SATELLITE AND SPACE COMMUNICATION,
LINE-OF-SIGHT IONOSPHERE SUPERHIGH = 10!°| MICROWAVE RELAY, RADAR (AIRBORNE,
PENETRATION, WAVEGUIDES 10 em PREQUENCY (SHF) F APPROACH, SURVEILLANCE, AND WEATHER)
0.3-3 GHz TV (UHF), SPACE TELEMETRY
Peace ULTRAHIGH — 109 |RADAR, MILITARY SATELLITE
+ MELEREQUENCY (UHF) F3 |
COMMUNICATION
COAXIAL CABLES, SKYWAVE 30-300 MHz TV (VHF) AND FM, LAND
(IONOSPHERIC AND VERY HIGH — 108 |TRANSPORTATION (TAXIS, BUSES,
TROPOSPHERIC SCATTER) 10 ME FREQUENCY (VHF) RAILROADS), AIR TRAFFIC CONTROL
3-30 MHz , |BUSINESS, AMATEUR AND CITIZENS
HIGH FREQUENCY fy 10” | BAND, MILITARY COMMUNICATION,
COAXIAL CABLES, (HF) = MOBILE RADIO TELEPHONE
IONOSPHERIC REFLECTION 100 M Aas a
(SKY WAVE) MEDIUM 106 |AM BROADCASTING, AMATEUR,
FREQUENCY (MF) = MOBILE, PUBLIC SAFETY
eeee ee
30-300 kHz NAVIGATIONAL AIDS, RADIO BEACONS,
LOW FREQUENCY 10° |INDUSTRIAL (POWER LINE)
(LF) Ss COMMUNICATIONS
10 km
3-30 kH
VERY Low L104. |NAVIGATION, TELEPHONY, TELEGRAPHY,
WIRE PAIRS, SURFACE DUCTING FREQUENCY (VLF) FREQUENCY ANDEEE
EES
TIMINGEEEESTANDARDS
EEE ee eee
(GROUND WAVE) 100"km 0.3-3 kHz
VOICE FREQUENCY TELEPHONY, DATA TERMINALS
1000 km
(VF)
30-300 Hz
EXTREMELY LOW MACROWAVE SUBMARINE
yee
FREQUENCY (ELF) COMMUNICATION
Figure 1.23 Various frequency bands with typical uses and transmission media
transmitting antenna transforms the electric signal into an electromagnetic wave for prop-
agation while the receiving antenna performs the reciprocal process. Depending on its fre-
quency, the wave can take on various forms ranging from a star light, to a television broad-
cast or even a laser beam. The electromagnetic spectrum has been divided into different
bands which are assigned to serve specific types of communication as shown in Fig. 1.23.
As an example, television broadcast occupies either the 30-300 MHz range (VHF band) or
the 0.3-3 GHz range (UHF band).
In certain applications, such as terrestrial mobile communications, the radiated en-
ergy from one transmitting antenna may reach the receiving antenna by direct propagation
or by reflections by the ground or the sky as depicted in Fig. 1.24. This phenomenon is re-
ferred to as multipath, which indicates that the total received electromagnetic energy is the
weighted sum of “many” delayed versions of the transmitted signal where each component
is randomly attenuated and delayed according to the propagation path it took to the receiver.
Depending on the carrier frequency, sky reflections can result from ionospheric or tropo-
spheric scattering (reflection and refraction). The ionosphere is a collection of gases, atoms,
water droplets, pollutants, and so on, that absorb and reflect large quantities of radiated en-
ergy. It is located in the upper part of the atmosphere at an altitude of about 100-300 km
34 Introduction to Telecommunications Chap. 1
IONOSPHERE
IONOSPHERIC
“| WAVES
TROPOSPHERIC WAVE
100-300 km
TROPOSPHERE
REFLECTED WAVE
DIRECT WAVE
TRANSMITTER
ANTENNA
RECEIVING
ANTENNA
and is known to reflect signals in the VLF band. On the other hand, the troposphere is about
10 km from the earth’s surface and is known to reflect carriers in the SHF band.
The design of any telecommunication link makes use of the so-called communication
equation which contains all of the basic elements and design parameters inherent in the
communication system and makes possible the specification of the received signal power-
to-noise spectral density ratio (units of Hz) with a high degree of accuracy, viz.,
2
FE=PG, (=) LL pbalrs GEM)
Here, P, is the total received signal power (Watts), No the one-sided power spectral noise
density (Watts/Hz), P; the transmitted total power (Watts), G; the peak transmitting an-
tenna gain (unitless), 4 = c/f the carrier wavelength (meters) and c is the speed of light
(meters/sec), d the distance between transmitter and receiver (meters), Lg the pointing loss
(unitless), Lp the polarization loss (unitless), Lq the atmospheric loss (unitless), L, the an-
tenna surface tolerance loss (unitless), G, the peak receiving antenna gain (unitless), k Bolz-
mann’s constant (1.38 10-23 Watts/Hz-° K)and finally, 7° the receiver temperature (Kelvin,
°K). Each element of the “power budget equation” given in (1.11) is defined and discussed
here as to its effect on the system design.
A typical communication link is shown in Fig. 1.25 where various “virtual” channels
are identified. At the transmitter, each information bit or block of bits is transformed into
a unique baseband waveform by the baseband modulator. This signal is then upconverted
twice to IF and then again to RF. Following power amplification, the signal is fed to the
antenna for propagation. The physical channel is the propagation media between the trans-
mitting and the receiving antenna. It can be a waveguide or propagation through rain as
discussed earlier. Modeling the effects of the physical channel on the propagating electro-
magnetic field in different frequency bands and under various scenarios is a continuous and
Sec. 1.3 Performance Criterion and Link Budgets 35
ongoing task [29]. At the receiving antenna output, the RF signal undergoes filtering and
amplification using a low noise amplifier (LNA). The physics of these operations and their
equivalent set at the transmitter are referred to as the RF channel or subsystem. Some of
these operations are highly nonlinear and might introduce signal amplitude and phase dis-
tortion. Examples of such include Travelling Wave Tube (TWT) amplifiers [30] used prior
to transmission which can introduce severe AM/PM effects, depending on their operating
point. At the receiver, the output of the LNA is then downconverted to an appropriate IF for
further filtering and processing. Communication system engineers deal mainly with model-
ing the IF signal in both the transmitter and receiver, or commonly designated the JF channel
or subsystem. At the receiver, the signal model in the IF subsystem incorporates the various
effects of both the physical and the RF channel such as fading due to shadowing, cross-
polarization effects, AM/AM and AM/PM conversion in TWT amplifiers, amplitude and
phase scintillation due to plasmas, and so on. Typically, the IF signal is again downconverted
to baseband using inphase and quadrature references for further processing including bit de-
tection. In modern receivers, the latter stage is typically implemented digitally as long as the
data rates are well within the technology capabilities. The baseband signals or equivalently
the baseband channel or subsystem are equivalent to their IF counterparts and are dealt with
by communication system engineers interchangeably as the IF conversion to baseband con-
stitutes a mere frequency shift with no additional effects.
In this section, we concentrate on the link budget for the additive white Gaussian
noise (AWGN) channel. All effects in the physical channel are frequently modeled as power
degradation in the received signal. This assumption is valid for the majority of the commu-
nication links encountered in practice. Let P, denote the maximum power at the amplifier
output. Then, the actual transmitted power P; is equal to P,L. where L, denotes the circuit
coupling loss and is given by
4AZaZLr
Le (1.12)
Za Zi)?
Here, Z, and Z, denote the output impedance of the amplifier and the input impedance
of the antenna respectively. Only when the impedances are matched can maximum power
transfer occur (that is, P; = P,). The transmitted electromagnetic signal is characterized by
the Effective Isotropic Radiated Power (EIRP), defined by
EIRP = P,G;LoLy (PS)
where G; is the transmitting antenna gain in the receiver direction and Lo, L, the point-
ing and antenna surface tolerance losses, respectively. A rough surface can cause energy
scattering and loss of gain in the desired direction. The antenna surface tolerance loss is
approximately given by
| =] peea ee Satl
ye
enaptn
eee
IWOISAHd TANNVHO |
| |
ONVd3Sva
| |
ONVsasve AWNoOIS|
OUL: ‘OURS ¥yOLVINGOW
ONvsasva
YOLVINGOW30
aansigq
sz°]T y poylfdus
[eIISIP UONROTUNUIWIOD
sks wo
Sec. 1.3 Performance Criterion and Link Budgets 37
Peak Gain, G
(a)
46 beamwidth
6 is the antenna 3 dB beamwidth -3 dB from
peak
Second First
sidelobe sidelobe
Peak Gain, G
Transmitting x)
antenna be
Antenna
mainlobe
Antenna (b)
Sidelobes
angle ¢ and G, is the maximum gain taken over all angles (see Fig. 1.26a). Assuming
perfect pointing, the gain of the transmitting antenna in the direction of the receiver would
be exactly G,. However, in more realistic scenarios, the antenna is not pointed perfectly
and as a result, the gain in the receiver direction is slightly less than the peak gain G,. That
degradation is accounted for in the pointing loss Lg. Another important parameter of an
antenna gain pattern is its “beamwidth,” which reflects the width of the gain in space as
illustrated in Fig. 1.26b. For a circular reflector of diameter d,, the maximum (peak) gain
G, and the half power beamwidth yp (radians) are given by
38 Introduction to Telecommunications Chap. 1
a\e
G; ~ pa (=) (1.16)
Xr
~ ela)
Me da./Pa
where fg is the antenna efficiency factor which is a function of the aperture field intensity
distribution over the dish. This pg is typically on the order of 55%. Note from (1.16) and
(1.17) that the larger the antenna or the higher the frequency, the higher is the gain and
the narrower the beamwidth. For a parabolic antenna with pointing error @-, the antenna
pointing loss Lg is approximately given by
Lg © 072 6(Ge/ 0)” (1.18)
which indicates an exponential decrease in gain as a function of pointing error $e. Typical
pointing errors are about 0.1° — 0.3° for geostationary satellites depending on the tracking
capabilities of the antenna. Spatial pointing is maintained in the course of a track by using
several techniques, one of which is conical scan in which the antenna forms small circles
around the target while measuring the received power to determine the “best” spatial coor-
dinates.
After propagating a distance d, the field flux density (fF D—in units of Watts/meter?)
at the receiver becomes
(ET RP)LE
F D(d) = (1.19)
4rd?
where L, is the atmospheric loss due to rainfall, ionized atoms in the ionosphere, and so
on. The atmospheric losses are caused by absorption and scattering of the electromagnetic
field by the various molecules in the atmosphere. The problem becomes more severe as
the wavelength of the carrier approaches the size of these molecules. For example, severe
absorption due to water vapor molecules occur at 22 GHz while oxygen absorption occurs
at about 60 GHz. Other effects include degradation due to rainfall which becomes severe
when the wavelength approaches the water drop size, which is dependent on the type of
rainfall. For example, at a rainfall rate of 10 mm/hr, the attenuation is about 0.04 dB/km at
6 GHz but increases to 0.25 dB/km at 11 GHz. That degradation is also a function of the
antenna elevation angle as the latter affects the path length of the propagation through the
rain [29].
Assuming a receiving antenna with area A, perpendicular to the direction of the
transmitter, the collected signal power becomes [32]
(EITRP)LgA;Lp
P, = FD(d)A;Lp = (1.20)
4nd?
where L, accounts for any polarization loss due to misalignment in the received field polar-
ization. Polarization is the orientation in space of the electromagnetic field and it is typically
linear or circular. In the first, the electromagnetic field is aligned in one planar direction,
whereas in the latter, it is radiated with two orthogonal components that when combined,
-appear to rotate on a circle or on an ellipse. In both cases, the receiving antenna must have
a matching orientation. Circular polarization has the advantage of not being affected by an
erroneous phase rotation in the electromagnetic field during propagation, unlike linear po-
larization, which suffers a degradation.
Sec. 1.3 Performance Criterion and Link Budgets 39
Using the reciprocity principle [32], the effective receiving antenna area can be ex-
pressed in terms of the antenna gain as
Ar = 7 8r($.,o})
Ae / /
(1.21)
where g,(¢,, ¢;) denotes the receiving antenna gain in the direction of the transmitter. Here
again, the receiving antenna is typically not pointed perfectly and the gain in the transmitter
direction might not be the peak gain G,. Using (1.21) in (1.20), we have
P, = P,G,;LsLoLpLal,G, (1222)
where L, denotes the space loss and is given by (A/4zrd)*. Both coupling as well as pointing
losses can also occur in the receiver and need to be accounted for in (1.22).
The parameter No in (1.11) represents the normalized noise power spectral density in
Watts per Hertz (W/Hz). This factor is usually referenced to the input stage of the receiver
since the signal-to-noise ratio is established at that point. It is given by
No =kT° (1.23)
where k is Boltzmann’s constant and T° the system noise temperature in Kelvin (°K). This
equivalent system noise temperature can be expressed as
The standard technique employed to measure the noise temperature of a system is to com-
pare, in a fixed bandwidth B, a measurement of system noise power P, with the noise figure
from a resistor at a known temperature. Since P, and B are related by P, = kT°B, the
unknown system noise temperature can be determined quite simply. The quantity B is the
single-sided noise bandwidth, and noise power is always measured with respect to it.
The link budget of (1.11) is often rewritten as
P, [EIRP G,
No
ee
k Jen 15]
ye EXel li |e
ce)
r
where the transmitter parameters are lumped in the first bracket, propagation parameters in
the second, and the receiver figure of merit in the third.
Example 1.3
Table 1.2 depicts a link budget for an RF communication link between Mars and Earth as
shown in Fig. 1.27. Two different link options are considered: a direct link and an indirect
link. The first option consists of communicating from Mars to a Mars relay satellite (MRS)
and then to Earth directly at a frequency of 8.4 GHz (X-band) with a data rate of 20 Mbits/sec
[Table 1.2(a)]. The second option is similar except that the link to Earth will be relayed through
a satellite (Earth relay satellite—ERS) and the carrier frequency is at 32.4 GHz (Ka-band) and
the data rate is reduced to 10 Mbits/sec [Table 1.2(b)]. Since Ka-band suffers large degradations
in bad weather, the ERS is essential in increasing link availability.
Introduction to Telecommunications Chap. 1
40
EARTH RELAY
SATELLITE (ERS) MARS RELAY
SATELLITE (MRS)
Ka/W-BAND
X/Ka-BAND
Any performance criterion should assess the capability of a communication system to pro-
duce at its output a desired signal, that is, it should relate the desired operation to the actual
operation in the presence of the channel disturbances.
One of the most widely used measures of performance, particularly in systems with
stochastic additive disturbances, is the signal-to-noise ratio (SNR). The scenario consists of
a desired waveform (baseband or modulated carrier) contaminated by an additive interfering
waveform. When this combined waveform appears at any point in the system, the SNR at
that point (typically at the output of the receiver) is defined as
SNR = Power in desired signal waveform
— : 2t,
Power in interfering waveform ( )
SNR is an indication of how much stronger the desired signal is than the interfering signal
at the point in question. It is common to express SNR in dB, that is,
4. From here on in the book, P denotes the received signal power and the subscript “r” will be omitted.
- Sec. 1.3 Performance Criterion and Link Budgets
Table 1.2(a) Mars Return Link Direct to Earth (20 Mbps at f = 8.4 GHz)
Often, we require that a system be designed to guarantee that the SNR performance
at a given point is above a specified level, often called threshold SNR. Then the ratio of the
actual SNR at that point to the threshold SNR is called the system or design margin at that
point, that is,
SNR actual
Design margin = (1.29)
SNR threshold
which is commonly expressed in decibels. Design margins are extremely important since
42 Introduction to Telecommunications Chap. 1
Table 1.2(b) Mars Return Link through the Earth Relay Satellite
(10 Mbps at f = 32.4 GHz)
they indicate the amount by which a system can be tolerant to unexpected events or ad-
ditional degradation. The links of Example 1.3 have design margins of 2.9 and 3.2 dBs
respectively.
tive white Gaussian noise channel, coding schemes exist for which it is possible to commu-
nicate data error free over the channel provided the system data rate Ry, (in bits per second)
is less than or equal to what he called the channel capacity C. If Rp > C, then error-free
transmission is not possible. The channel capacity C, in bits per second, is defined by
P
Ca Bt og2(1 1 ++ ——
NoB> if
(1.30)
where P represents the received signal data power in watts, B the one-sided channel band-
width in Hz, and No the one-sided PSD level of the noise. Now P/NoB can be written as
1% es Ppp Me PT,/No ns Ep
ht)
Nob eaNgBH o. 1By Ree > No
where 7) is the time duration of a data bit in seconds, that is, the inverse of the data rate,
6p = Rp/B is the ratio of the data rate to the one-sided channel noise bandwidth in bits/Hz
and E,/No = PT)/No is the energy per bit-to-noise spectral density ratio. Frequently, the
parameter E;,/No is referred to as the relative efficiency or communication efficiency of a
particular communication scheme.
Substituting (1.31) and B = R»/d, into (1.30) and assuming communication at chan-
nel capacity (R, = C), we get
E
5p = logy(1 + 5—) (1.32)
No
Equation (1.32) is plotted in Fig. 1.28 as a function of 1/5, and E,/No. Notice that as
the value of E,/No decreases, the transmission bandwidth necessary to achieve 100% effi-
ciency, operation at C, increases. It is therefore instructive to allow B to approach infinity in
(1.30) or equivalently let 5, go to zero in (1.32). Taking the limit gives (see also Chapter 12)
Ep _
In(2).=--1.6.0dB (1.33)
0
Consequently, coding schemes exist for which error-free transmission is possible in the
unconstrained bandwidth, additive, white Gaussian noise channel provided that E,/No >
—1.6 dB. Unfortunately, Shannon’s theorem does not specify the data encoding method to
accomplish that.
The relative efficiency of any practical communication system can be compared with
the most efficient on the basis of how closely it approaches the Shannon limit, (1.33), for a
given error probability. In the following chapters, we consider various modulation, demod-
ulation, coding, decoding, and data-detection methods that attempt to approach the limit
expressed by Shannon’s theorem. Typical conclusions taken from later studies of various
methods are compared with the Shannon limit in Fig. 1.28. The Shannon capacity is made
more rigorous in Chapter 12.
NONCOHERENT
FSK (Pp = 10°)
COHERENT
PSk(Pe= 10P)
COHERENT
PSK (P2= 10%)
ne
w)2
SHANNON'S LIMIT
-1.6 dB (2n2)
1.0 10 100
grit BLS BANDWIDTH
6 %p\NFORMATION
BIT RATE
Figure 1.28 Relative efficiency versus the ratio of bandwidth to information bit rate
and
J
By I Lope (1.35)
where the parameter Lj, represents the transmitter modulation loss. Essentially, it specifies
the power division between the carrier and information sidebands resulting from a particular
modulation process. If the basic information carrier is a microwave, continuous-wave (CW)
sinusoidal voltage, the modulation can be impressed on the transmitted carrier by angle
modulation [phase (PM) or frequency (FM)], amplitude modulation (AM), or a simultane-
ous combination of the two. Analytically speaking, the instantaneous value of a modulated
RF carrier can be expressed by
where a(t) characterizes the amplitude modulation, 6(t) characterizes the angle modulation,
Ww is the carrier radian frequency, and /2P denotes the peak carrier amplitude when a(t) is
unity. The modulation signals, the information to be transmitted, can be classified as digital,
analog or a combination of the two.
One typical telemetry modulation is to phase-shift-key the data onto a square-wave
subcarrier and then phase-modulate onto a sinusoidal carrier (PCM/PSK/PM) as depicted
in Fig. 1.29. However, an alternate scheme phase-modulates the data directly on the carrier.
In this latter case, the data being phase-shift keyed is called PCM/PM. In either case, the
modulated carrier can be represented mathematically by
Sec. 1.5 Digital Communication with Subcarriers and Data Formatting 45
ANALOG
MODULATED
DATA DATA
SOURCES | ANALOGTO SYMBOLS SUBCARRIER | SUBCARRIER
DIGITAL CARRIER SurpuT
MULTIPLEXER PHASE PHASE
CONVERTER MODULATOR 8 MODULATOR
DIGITAL SOURCES
SUBCARRIER CARRIER
GENERATOR GENERATOR
POWER POWER
SPECTRUM SPECTRUM
CARRIER
f f
SSS ean ) bye 0 fo-lse fo fotlec
A B Cc
K
s(t) = V2P sin(wet + YA; Dj(t)) (1.37)
a
where A; denotes modulation index (in degrees) for the ith “data” signals Was 2s poe
K the total number of data channels and D;(t) the normalized data sequence (in the
case of PCM/PM) or a normalized modulated square-wave subcarrier (in the case of
PCM/PSK/PM), that is,
dj(t), PCM/PM
POT |di(t)Sin(se,t), _PCM/PSK/PM ue
where d;(t) = +1 is the normalized data sequence, Sin(ws¢,t) = +1 the square-wave sub-
carrier, and Ws, the i‘h subcarrier frequency. Using trigometric identities with the fact that
Did hee aed. b= 15.2,2.2,,.h, (1.3/7) may be expanded for K = | as
If 0° < A, < 90°, this phase-modulated carrier comprises a pilot tone (residual carrier) and
a Double-Sideband (DSB) modulated carrier. A system with A; < 90° is called a residual
carrier system. A system with A; = 90° is called a suppressed carrier system. A residual
carrier employs a phase-locked loop to track the pilot tone and provide a coherent carrier
phase reference for demodulating the DSB-modulated carrier. A Costas loop may be used
as a suppressed carrier receiver. For K = 2, (1.37) may be expanded as
When we have a single data channel, as in (1.39), the first term is the carrier, while
the second term is the data channel. Hence the modulation index A, has allocated the total
power P in the transmitted signal s(t) to the carrier and to the data channel, where the
carrier power and the data power are, respectively
P.= P cos”Ay (1.41)
P, = Psin*A, (1.42)
Comparing (1.41) to (1.35), it is clear that the modulation loss Lm is equal to cos A}.
When we have two data channels as in (1.40), the first term is the carrier component, the
second term is the modulated subcarrier D|(t), the third term is the modulated subcarrier
D»2(t) component, and the fourth term is the cross-modulation loss component. Hence, the
corresponding powers in the four components are allocated by both modulation indices Aj
and A> with
= V2P {cos(A] Sin @sct) SiN Wet + d,(t) sin(A, sin @sct) COS Wet}
[o.@)
CoO
where Jp (x) = (—1)"Jn(—x) and J,(x) is the n’” order Bessel function [37]. The first term
of (1.49) is the discrete carrier, the second term is other discrete components at @. + 2nas¢,
and the last term consists of suppressed carrier modulations centered at w. + (2n + l)ase.
The baseband data dj(t) can have different formats as illustrated in Fig 1.30. With
a Non-Return to Zero-Level (NRZ-L) data format, a logical “one” is represented by one
level and a logical “zero” by the other. With NRZ-M (Mark), a logical “one” is represented
by a change in level and a logical “zero” by no change. Two other formats, Biphase and
Miller, are also defined in Fig. 1.30. It is worthwhile noting that Bi-@-L is identical to NRZ-
L when the latter is modulated on a square-wave subcarrier with a cycle period equal to
Problems 47
t
0 27; 4 47, & 67, 4 87, 4107,
Tee Sle mile) Vlg V97s AT
the bit period, so that a transition occurs at midpoint. Thus, PCM/PM with Bi-f-L data
formatting is a special case of PCM/PSK/PM with NRZ-L data formatting.
PROBLEMS
1.1 Let Z, and Z> be two independent Gaussian random variables with moments
E{Z\}=A,, Var{Zi}=o7
48 Introduction to Telecommunications Chap. 1
3 1
y(n) — go" Saal aig ee 2) =x(n) + 2x(n — 1)
(a) Draw a block diagram representation of the system in terms of adders, delays, and coeffi-
cient multipliers.
(b) Let the input x(n) be a white WSS random sequence with PSD equal to «7. Compute the
PSD of the output y(n).
(c) Design a digital causal filter (specify its impulse response) which operates on y(n) (that is,
cascaded with the first filter) such that its output z(7) is white.
(d) Suppose that a third filter H(z) is cascaded at the output of the whitening filter. Is the output
of H(z) still white if
i, Be alaié
HZ) Falieeces |
re tel
What did the insertion of H(z) do?
1.4 Consider the system illustrated in Fig. P1.4. Assume that the input is a WSS random process
with autocorrelation R,(t) and corresponding PSD S,(f).
(a) Find the autocorrelation function of y(t) in terms of R,(t).
DELAY 5 sec
Figure P1.4
Problems 49
le O=lf\s fa — a)
This is referred to as a raised cosine roll-off characteristic with excess bandwidth factor a
where 0 < a@ < i.
(a) Suppose that x(t) is sampled at a rate f, = 1/T; =2f,(1 + @) producing the sampled
function
co
nif) =| If S fa
0, otherwise
producing X(t). Write a sampling expansion appropriate to the filter output x(t). Are the
samples of this expansion uncorrelated?
1.6 An autocorrelation function which frequently arises in practical problems is
Pr(n) = Ory er
n!
where A is the average number of zero crossings per unit time. Also, assume that x(t) takes on
the values E and —E with equal a priori probability.
(a) Sketch a sample waveform of the random process x(f).
(b) Compute the autocorrelation function of x(t).
(c) Compute the PSD corresponding to the result in (b).
50 Introduction to Telecommunications Chap. 1
Figure P1.19
when operating with wavelength 4. Determine the half-power beamwidth in the ¢; = 0 and in
the @, = 0 plane.
1.18 A communication satellite is to transmit to Earth at L-band (1.2 GHz).
(a) Assuming a LEO (1,400 km altitude) satellite, compute the required diameter of a
parabolic transmitting antenna so that the beam covers a 1,000 km? area. Assume an an-
tenna efficiency of 40%. What is the resulting antenna gain?
(b) Repeat (a) for a MEO (10,000 km altitude) satellite and a desired ground footprint of
2,000,000 km? {roughly the size of continental U.S. (CONUS)].
In the Iridium network, a phone call might require routing through various satellites to get it to
its destination. Consider the 2-satellite system depicted in Fig. P1.19. Assume ideal frequency-
translating satellites, white noise, and equal bandwidths.
(a) Compute the received signal power at the receiver as a function of uplink power P,,, satel-
lite gains G; and G2, and path losses L; and L.
(b) Compute the received noise power at the receiver as a function of Py}, Pn2, Pn3, Gi, Go,
Lj, and Lp.
(c) Compute the signal-to-noise ratio at the receiver.
1.20 A satellite at 1,000 km altitude is to use multiple downlink spot beams to achieve frequency
re-use, while covering a distance on Earth of 1,200 km (use one-dimensional antenna patterns
here). The beams are to have center-to-center separation of 0.1 radians with the gain pattern
depicted in Fig. P1.20. A receiving station in any beam must have 50 dB isolation from any
other beam (the received carrier power at any ground station from any other beam after front
end filtering is to be less than 50 dB from the carrier power at peak gain in its own beam). The
downlink signal is transmitted with bandwidth B, but can be located at any carrier frequency
fe. fo + B, fe + 2B, etc. Neglecting dual polarization, determine the amount a frequency can
be reused with this system.
1.21 In Table 1.2(a), assume the frequency is moved to Ku-band (12 GHz) and that the antenna
diameter is reduced from 5 to 3 m.
(a) Compute P,/No in dB-Hz.
(b) What is the link margin?
(c) Assume that a 34-m antenna is used on the ground instead of a 70-m antenna, compute
P,/No and the link margin.
52 Introduction to Telecommunications Chap. 1
GAIN PATTERN
Figure P1.20
(d) With a 6 dB link margin and a 34-m antenna on the ground, what is the maximum data rate
that can be supported?
1.22 In Table 1.2(b), assume that the Mars Relay Satellite (MRS) transmitted power is reduced from
90 to 50 watts but the Earth Relay Satellite (ERS) antenna is increased to 30 m in diameter.
(a) Compute P,/No in dB-Hz at ERS.
(b) Compute P,/No in dB-Hz on the ground.
(c) What is the link margin with 10 Mbps?
(d) Given a desirable link margin of 6 dB, what is the maximum data rate (in bps) that can be
supported?
1.23 Consider the case in which 100 kbps is to be transmitted over a 3 kHz voice grade circuit. Is
it possible to achieve error-free transmission with a SNR of 10 dB? If not, what modifications
would you suggest?
1.24 Consider a voice grade telephone circuit with a bandwidth of 3 kHz. Assume the circuit can be
modeled as an additive white Gaussian noise (AWGN) channel.
(a) What is the capacity of such a circuit if the SNR is 30 dB?
(b) What is the minimum SNR required for a data rate of 4800 bps on such a voice grade
circuit?
1.25 (a) Find the average capacity in bps that would be required to transmit a high-resolution
black-and-white TV signal at the rate of 32 pictures per sec if each picture is made up
of 2 x 10° picture elements (pixels) and 16 different brightness levels. All picture elements
are assumed to be independent and all levels have equal likelihood of occurrence.
(b ~— For color TV, this system additionally provides for 64 different shades of color. How much
more system capacity is required for a color system compared to the black and white
system?
(c) Find the required capacity if 100 of the possible brightness-color combinations occur with
probability of 0.003 each, 300 of the combinations occur with probability 0.001 each, and
624 of the combinations occur with probability 0.00064 each.
1.26 Assume E,/No = 3.4 dB at T = 28°K and the data rate is Ry = 1/7, = 115,000 bps.
(a) Find the power-to-noise energy ratio, P/ No in dB-Hz, at T = 28°K.
References 53
(b) For T = 0°K, 28°K, and 300°K, what is the noise power?
(c) What are the communication efficiencies at the temperatures given in (b)?
(d) Find the equivalent bit mass, say m,,, that satisfies Einstein’s result relating energy to mass,
that is, E = mc’, if T = 28°K.
(e) Find the equivalent energy associated with an electron and the ratio of the bit mass given
in (d) to that of an electron.
(f) If each spacecraft TV picture transmitted from Jupiter consists of 2,000,000 bits, find how
many electron masses make up one TV picture. (Assume T = 28°K.)
(g) What is the channel capacity if the single-sided bandwidth is B =200,000 Hz? (Assume
TE DEAS)
(h) If a picture is worth a thousand words, how many bits/word are there in a picture transmit-
ted from Jupiter?
(i) If each word contains 5 bits of information, how many bits are there in a “1,000-word”
picture?
1.27 Consider phase-shift keying with a carrier component which can be written as
REFERENCES
1. NASA’s Mobile Satellite Experiment, Pasadena, CA: Jet Propulsion Laboratory, MSAT-X Quar-
terly, no. 1, October 1984.
2. Special Issue on Optical Communications, Proceedings of the IEEE, vol. 58, no. 10, October
1970.
3. Lindsey, W. C. and M. K. Simon, Telecommunication Systems Engineering, Englewood Cliffs,
N.J.: Prentice Hall, 1973. Reprinted by Dover Press, New York, NY, 1991.
4. International Organization for Standardization, /nformation Processing Systems—Open Systems
Interconnection—Basic Reference Model, ISO 7498, New York: American National Standards
Institute, October 1984.
5. Henshall, J. and S. Shaw, OSI Explained, End-to-End Computer Communication Standards, West
Sussex, England: Ellis-Horwood Limited, pp. 10-20, 1988.
6. Knightson, K. G., J. Larmouth and T. Knowles, Standards for Open Systems Interconnection,
New York: McGraw-Hill, pp. 12-19, 20-54, 259-301, 1988.
7. International Organization for Standardization, Information Processing Systems—Open Systems
Interconnection—Manufacturing Message Specification, ISO 9506, New York: American Na-
tional Standards Institute, 1989.
8. International Organization for Standardization, Information Processing Systems—Open Systems
Interconnection—Basic Reference Model, Part 4: Management Framework, ISO 7498-4, New
York: American National Standards Institute, 1988.
9. International Organization for Standardization, Information Processing Systems—Open Systems
Interconnection—Systems Management Overview, ISO 10164, New York: American National
Standards Institute, 1990.
10. International Organization for Standardization, Information Processing Systems—Open Systems
54 Introduction to Telecommunications Chap. 1
In Chapter 1, several definitions of bandwidth were given as measures of the spectral occu-
pancy of a chosen signaling or modulation technique. All of these were in some way related
to the power spectral density (PSD) of the signaling scheme. Thus, the purpose of this chap-
ter is to characterize the PSD of various types of synchronous data pulse streams which
include as special cases the data formats introduced in Chapter 1.
generating sequence is cyclostationary. We begin with the derivation of the power spectral
density of a synchronous data pulse stream generated by a WSS sequence.
Consider the binary (+1), zero mean, WSS sequence {a,} for which it is known that
Gn =0; Gndm
= Ra(m — n) (2.1)
and the overbar denotes statistical expectation. From this sequence, we form the data pulse
stream
oe)
where p(t) is the elementary signal (pulse shape) and is not necessarily restricted to be
time-limited to only a single signaling interval, that is, the pulse train m(t) can contain
overlapping pulses. Irrespective of the properties of the generating sequence {a,}, the pulse
stream m(t) is itself cyclostationary since the expected value of the product m(t)m(t + T)
is, in addition to being a function of tT, a periodic function of t. Thus, to compute the PSD
of m(t), namely, S,,(f), we must first average the statistical correlation function
over ft [the average is performed over the period of R,,(t; t)] and then take the Fourier
transform of the result. Thus,
Since
where P(f) is the Fourier transform of p(t), then substituting (2.6) in (2.5) gives
Oge
njP"( P*
p1"
Bee. i/Sy RaGn =
y z m n
(Qe)
x Ba Oe fra ded ydz
18
1. Unless otherwise noted, all summations and integrals range from —oo to oo.
58 Power Spectral Density of Digital Modulations Chap. 2
Recalling that
/e JT dz — §(x) (2.8)
T
(2.9)
ar
. eae) dy
Recalling that the time average is performed over the period of Rm(t; T) , that is, T, then
T/2
(e204) a al ; jon GA pee ee: Oe) (2.10)
= Erie mY fT
Substituting (2.10) in (2.9) and letting / = m — n gives
j
=D Rate —j2nflT
Sn(f) es [roverx eee
pSMOt fT]
y (Qe)
v, y e JOT PAT ty
1 1 k in zk
Sn(f) = E~ Ra seer = dsP ( ee *)P*( a (2.13)
Since
sr ={o a (2.14)
and |P(f)|? is an even function of f, we get the desired result?
is the power spectral density of the sequence, that is, the discrete Fourier transform of its
autocorrelation function. Note that if the data sequence is purely random, that is,
Sa(f) = 1 (2.19)
and hence
Suppose now that we have a binary sequence {a,} that has the properties
Furthermore
where N denotes the period of the autocorrelation function Ra(n; m — n). Then, following
the steps leading up to (2.9), we see that the analogous result is now
ah (2.23)
x e f2n(y+f)nT (ef2xO+\ ay
Thus far, we have not made use of the cyclostationary property given in (2.22). If now we
use this property to evaluate the sum on n in (2.23), and as before let / = m — n, then it can
be easily shown that
N
bpR, (n;1) Te gad
ye Ra (n; 1) e Ja (yt fmt _
n=)
‘i (2.24)
x bpee
k
60 Power Spectral Density of Digital Modulations Chap. 2
f= ;
lt
(225)
1 k
ated 5 ae
. Ez d (v+s a)
Substituting (2.25) into (2.23) and performing the integration on y gives
1 a! sls
Srl) byRa(n; De cual
k
ie —j2xfITp (
(_ tr. Me
=z) \ip (—f){e
t= (pees
is the power spectral density of the equivalent stationary data sequence, that is, the period-
icity of the correlation function caused by the cyclostationary behavior of the sequence {ay}
must be “averaged out” before taking the discrete Fourier transform.
When the generating sequence {a,} has nonzero mean, then the spectrum of the correspond-
ing synchronous data pulse stream will have a discrete component in addition to the cus-
tomary continuous component. An example of a situation where this might occur is at the
output of a convolutional encoder whose input is random but not equiprobable binary data.
The procedure for handling this case is as follows:
Define the zero mean cyclostationary sequence {A,} by
An = Qn — Qn (2.30)
An=0; AnAm
= Ra(n; m —n) (2.31)
Then, using the results of the previous section, the continuous component of the power
spectrum (S;,(f)), for the synchronous data pulse stream generated by {ay} is given by
where
ee) 1 N
(Sn Pa = /ee
ane POP) (el ay as)
y n m
Since, for a cyclostationary sequence, a, and Gp» are both periodic with period N, then
analogous to (2.25) we have
— ,—j2aynT
N
OIE ||
1 y }
k
—————
DyGn’ a [Some NDE (> oP
(2.36)
ee ae owe
— ,—j2afnT -- Saal u k
n m=1 k
(Sm(Aa
Poe = ae F(A)NT
hee, Samer
c ema
k m=1
(2.37)
peaI | pik l
aa) p® |2.k\~< 5
—j2nnl/N ain
eae
. NT (<7) ( ai) n=1domme
Performing the time average over the period NT gives as before
m=\
(2.39)
x “ Ane
®
—_,j ‘nk/|N
dank 5
(F k )
Pe, uae
NT
Note that when {a,} is a WSS sequence (that is, N = 1), then (2.39) reduces to
i) k\|? k (2.40)
(Sm( fa = 7? d IP(=) os be T
62 Power Spectral Density of Digital Modulations Chap. 2
Example 2.1
Consider a binary, zero mean Markov source characterized by
Pr {dn+1 # an} = Pt
, (2.41)
Pr{@n41 =4n}=1— pr
ee cos6 — a?
Sea cos k@ = : a 5 (2.45)
=e 1—2acos0+a
Furthermore
ee 1—a?
1+2) a‘ cosko = a 2.46
k=1
1 —2acos6 + a2 ( )
Applying (2.46) to (2.43) with a = 1 — 2p; and g = 27 fT immediately gives the desired result,
namely
LG 2p
Sa(f) =
Lb = 2p) = 2 ap cos eet
(2.47)
ne 4p;(1 — pr)
2 (1 —2p;) (1 — cos2afT) + 4p?
Furthermore, if the data stream generated by this sequence uses rectangular pulses, that is,
Ati Oestrssull
p= |0; otherwise eae.
then, using (2.16) and (2.48) we get
ey)
: sin’ (z fT) 4p:,01 — p
Sm(f) = ACT = Pe be | (2.49)
(wfT)° 20 — 2p) —cos2xfT) + 4p7
Example 2.2
Consider the sequence formed by interleaving N independent first order Markov sources with
respective transition probabilities p,,;n = 1,2,...,.N. Then, the resulting sequence is cyclo-
stationary with correlation function
leat)
: =
(1—2p,,)"/";
“
1=0,4N,+42N,...
d j
ann : all other integer/ ey)
Sec. 2.2 Power Spectral Density of a Generalized M-ary Markov Source 63
The power spectral density 5z(f) of (2.29) is computed as (letting / = kN, k =0, +1, +2,...)
Noting the similarity between (2.43) and (2.51) (for fixed n), we can immediately write down
the result, namely
fi
Smf)= = Pil Si (A)?
i=l
¥
ns (*) ‘3(7 a =)
2
1 Co
iohous i)
(2.53)
n=—Oo ll
2 Bete abt 0h oe
a5 7 ke dns (f) SCP) pik le J2 i]
p= k=
where S;(/) is the Fourier transform of the i‘" elementary signal s;(t) and
co N
pilz) & > pizts —si(t) = si(t) — D- pest) (2.54)
n= k=]
n)
Also, S/(f) is the Fourier transform of s;(t). The quantity De is defined as the probability
that the elementary signal s,(t) is transmitted n signaling intervals after the occurrence of
s;(t). Hence, from the properties of Markov sequences, Py is the ik" element of the matrix
P”. Also, by definition, p\)’ = pix.
3. Note that this result is consistent with those in Section 2.1 when s;(t) = a; p(t).
64 Power Spectral Density of Digital Modulations Chap. 2
Notice that the second term (the discrete or line spectrum) of (2.53) vanishes when
M n =
Many special classes of signals exist for which the general PSD of (2.53) can be
simplified. The first class is generated by the source defined by the conditions:
(i) for each signal waveform s;(t) of the set of elementary signals, the negative —s;(t) is
also in the set;
(ii) the stationary probabilities on s;(t) and —s;(t) are equal, and
(iii) the transition probability p;x = p-s whenever s;(t) = +s,(t) and sx(t) = +5;(f).
Such a signaling source is said to be negative equally probable (NEP). Its overall spectrum
is characterized by the absence of a line spectrum and furthermore is independent of the
transition probabilities themselves. For this special case, the PSD reduces to the first term of
(2.53), namely
1 M
which is the weighted sum of the energy spectrum of the elementary signaling set.
Another class of signals of interest is those generated by a purely random source
whose properties were discussed in Section 2.1. Such a source, that is, one whose transition
matrix P has identical rows, has the property that P” = P for all n > 1. In this case, the PSD
of (2.53) simplifies to
M
1 2
Sm(f) =D Pill — Pid) Si (AYP
=
° (7 a =) (2.58)
2 M M
— FY pire Re [SHAS]
i=1
Obviously, if s;(t) = —s2(t) = p(t) and p = 1/2, then the signaling format becomes NEP
and (2.59) reduces to (2.57), which is then identical to (2.16).
At this point we turn to an evaluation of the PSDs of the various PCM signaling
formats discussed in Chapter 1.
Referring to Fig. 1.30 and assuming that the binary waveform levels are +A, the NRZ
signaling format falls into the class of signals whose spectrum is given by (2.59). Since the
elementary signal is a rectangular pulse of width 7, its Fourier transform is
p sin(z fT)
Si(f) = —S2(f) AT exp(—jxfT) (2.60)
Ce
This result could also be obtained from (2.32) and (2.40) by recognizing that
_ Fa 1— (1 —2p)*=4p(i — p); =
Gn =(1—2p); AnAm= |0: ( P) P( P) aie
(2.62)
P k\ _JAT; k=0
Pet ks k#0
When p = 1/2, the dc spike at the origin disappears and the NRZ signaling format falls into
the NEP class with
Here we have a situation where S|(f) =0 and S2(f) corresponds to the Fourier transform
of a half-symbol wide pulse, that is,
Since the source is again purely random, then substituting (2.64) into (2.59) gives
Seed Te
Se te tea P eile2h s(f= Ln3)
ie (2.65)
tig sin?
(1 fT/2)
Bea? | fT jay
66 Power Spectral Density of Digital Modulations Chap. 2
S2(t) =—s1(@)
Substituting the Fourier transform of (2.66) into (2.59) gives
Sm(f) 1 = oh We n
WP -Fa-2 Yo (=) a(t-F)
ea (2.67)
sin’
(2 f T /2)
ea 4p(1 - Clap) ——
GefT 2
As indicated by Hecht and Guida [3], the Miller coding scheme can be modeled as a
Markov source with four states whose stationary probabilities are all equal to 1/4 and whose
transition matrix is given by
1 r
vik & BE, Ih si(t)sy(t)dt i,k =1,2,3,4 (2.71)
For the Miller code, the four elementary signals are defined by
2) =—n(y
Ay ovOp Tye (2.72)
=|Ai T2S¢eT
and E; = A?T; i=1,2,3,4. Substituting (2.72) into (2.71) and arranging the results in
Sec. 2.2 Power Spectral Density of a Generalized M-ary Markov Source 67
26
ey —-—-— NRZ
22 ------ BIPHASE
a DELAY MODULATION
1.8
WHERE:
1. 6 T = SIGNAL PULSE WIDTH
ee
S A= SIGNAL AMPLITUDE
E12
10%. E=A°T
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1.0 We 1.4 1.6 1.8 2.0
de
Figure 2.1 Power spectral densities: NRZ, biphase, and delay modulation (Reprinted from [15])
1 0 0 -1
Sd (ae aeh tao
Our te) (2.73)
Alt OF 0, 1
Finally, using (2.69), (2.70), and (2.73) in the general PSD result of (2.53) yields the
result for the Miller code:
Sm(f) = ——_—_——
———
I _(23 — 2080 — 22 cos20 — 12 cos 30
E 262 (17 + 8 cos 86) (2.74)
+ 5cos
46 + 12cos 50 + 2cos60 — 8cos 76 + 2 cos 80)
where 0 Aa fT.
The power spectral densities of the NRZ, biphase, and delay modulation signaling
schemes as given by (2.63), (2.68), and (2.74), respectively, are plotted in Fig. 2.1.
Spectral properties of the Miller code that make it attractive for magnetic tape record-
ing as well as phase-shift-keyed signaling include:
1. The majority of the signaling energy lies in frequencies less than one-half of the data
fate, Rah
2. The spectrum is small in the vicinity of f =0. This spectral minimum facilitates
carrier tracking, which because of property | can also be more efficiently achieved
than Manchester coding. A reduced spectral density in the vicinity of f = 0 is an
important consideration in tape recording, primarily because of the poor de response
of the channel.
3. As a result of property 2, a lower magnetic tape recording speed (higher packing
density) can be used.
68 Power Spectral Density of Digital Modulations Chap. 2
4, The Miller code is insensitive to the 180° phase ambiguity common to NRZ-L and
Manchester coding.
5. Bandwidth requirements are approximately one-half those needed by Manchester
coding.
Other well-known techniques for reducing spectral bandwidth rely on ternary rather than
binary alphabets [4]; however, this is achieved at the expense of an increase in error proba-
bility performance and a complication of implementation.
2.3 POWER SPECTRAL DENSITY FOR DATA PULSE STREAMS WITH DATA
ASYMMETRY
In many applications [5S—7], the data pulse stream experiences data asymmetry in that the
rising and falling transitions do not occur at the nominal data transition time instants. Such
asymmetry in the data waveform causes the presence of a line spectrum in the PSD as
well as distortion of the continuous component of the PSD. Both of these ultimately cause
degradation of the error probability of the receiver. In order to assess this degradation, it is
desirable to evaluate the PSD of NRZ and Manchester streams with data asymmetry.
In this section, we use the data asymmetry model proposed in [5] which assumes that +1
NRZ symbols are elongated by AT/2 (relative to their nominal value of T sec) when a
negative-going data transition occurs and —1 symbols are shortened by the same amount
when a positive-going data transition occurs. When no data transition occurs, the symbols
maintain their nominal T-sec value. As such, we can use the generalized M-ary source
model of Section 2.2 where now M = 4 with
HO) Ay = [staAe
"10: — otherwise
—Ae =—f/2<7
<= T(k— A
200 =|6 / ote ( )/2
: otherwise
Ors)
AR) Totes 2
SOs |0; otherwise
By ey ae alot ea at
CS |0; otherwise
The stationary probabilities associated with these four elementary waveforms are, respec-
tively
Pt =2pC — p) (2.77)
Sec. 2.3 Power Spectral Density for Data Pulse Streams with Data Asymmetry 69
Taking the Fourier transform of the four elementary signals in (2.75) and substituting the
results in (2.58) gives after much simplification [6]
aD apd
S, 227s (x fT) i, 7,, sin® (1 f Tn)
(f) yee O12) + Ante. rl
Teepr APT Aste
sin (27 fT)
2 ee [Aa(p, Pr.) — As(p, Pr)] (2.78)
|
Aa(P, Pr.) = pr(1 — pr)(1 — 2p) Ecos(2xfTn) — p sin afm (2.79)
1
As(P, Pr) = PC =p CLs". p)
tS
Figure 2.2 is a plot of the continuous component of (2.78), (Sm(f))-, normalized by the
energy E = A°T, for various values of data asymmetry 7 in percent.
In this section, we use the data asymmetry model proposed in [5] which assumes that for
a +1 data bit, the first half of the Manchester symbol is elongated by AT /4 (relative to its
nominal value of 7/2 sec) and for a —1 symbol, the first half of the Manchester symbol
is shortened by the same amount. When a data transition follows, the second half of the
Manchester symbol retains its nominal value (i.e., 7/2). When no data transition occurs,
the second half of the Manchester symbol adjusts itself so that the total Manchester symbol
maintains its nominal T-sec value. In view of the above asymmetry model, we can use the
generalized M-ary source model of Section 2.3 where now M = 4 with
70 Power Spectral Density of Digital Modulations Chap. 2
(Sm(f))c/E
7 = Data Asymmetry
nF —T/2<t<AT/4
- AT/4<t <(T/2)0+A/2)
: otherwise
—A; —T/2<t<-—AT/4
A; —AT/4<t <(T/2)0 — A/2)
(OR otherwise
(2.80)
A; —T/2<t<AT/4
—A; AT/4<t<T/2
0; otherwise
The stationary probabilities associated with these four elementary waveforms are again
given by (2.76). Taking the Fourier transform of the four elementary signals in (2.80) and
substituting the results in (2.58) gives after much simplification
Sec. 2.3 Power Spectral Density for Data Pulse Streams with Data Asymmetry 71
9
(Snf))a = Arf)
DS (2.82)
ee=p a a) 1) ei
ti 1Gay) 2 Gnsayet
Pa] HAG
3 +7) 13 (¢ - =)
T
with
n=A/4
feeRvene sin? (man) [1 + 2h1(m, n)]?
4 (2.83)
H>(m, n) = sin? (2mzn)
H3(m,9) = 2 sin? (min) (cosmmn) [1 + 2h1(m, n)]
(2.84)
Sp =—
ey
2
MTG YC) HOP ve
ery ailcre.
bspiectigi ete)
(#)
(2.85)
S
Sa
+ A°TC6(n)
where
72 Power Spectral Density of Digital Modulations Chap. 2
Cia) = ee {sin(7)
(2.86)
+ sin (=) [1 - cosn/T cos nl}
ie 1
— =sin (=) sin (=) 58 8 [2 cos3z fTn + cos
2x fTn]
-- IEC!
Cs(n) = ;sin’peo”2]_ =sin? aoe
SaL2
=ant TUE] Va
cay eo ee |ee ea)
5
ee
eee
feces ed
|cosmer
For 7 = 0, (Sm(f))q = 0 and (Sm(f)), of (2.85) reduces to (2.68). Figure 2.3 is a plot of
(Sm(f)), of (2.85) normalized by the energy E = A°T for various values of data asymmetry
n in percent. Note that as 7 increases, the null at f = 0 disappears.
0.6
7 = Data Asymmetry
0.5
0.4
0.3
(f))o/E
(Sim
0.2
VY
0.1
Consider a convolutional encoder with constraint length K and rate b/n, as shown in
Fig. 2.4. In the m" information interval, b information symbols amp+j; j =0,1,2,...,b—
1 enter the encoder and n channel symbols xmy+); v = 1,2,...,n exit the encoder. The
structure of such a convolutional encoder can also be defined by a generator (connection)
matrix, namely
Sei oe a e SRK
where g;,; is either one or zero depending, respectively, on whether or not the i th modulo
summer is connected to the 7° shift register stage. For mathematical convenience, we
shall assume that both the input symbols {a+ ;} and the output symbols {x ,+)} take on
values plus and minus one. This allows modulo-2 summation operations to be replaced by
algebraic products. Thus, the encoder of Fig. 2.4 has the input output relation
Kb
Xmn+v = I] erie] (2.88)
v=
74 Power Spectral Density of Digital Modulations Chap. 2
Figure 2.4 A general constraint length K, rate b/n convolutional encoder (Reprinted from [8], [16] © IEEE)
Clearly, the output sequence {Xmn+.} is cyclostationary with period n as previously men-
tioned.
If, analogous to (2.2), a modulation m(t) with pulse shape p(t) is formed from the
sequence {Xmn+y}, then the spectrum of this modulation is given by (2.32) and (2.34). In
terms of the present notation, the continuous component of the spectrum (S,,(f)),. would
be
] n (ee) Rec
where
i ee
Sx(f) = — dd de Rx(v; nl +g — vje S27 ta-wWfT (2.93)
Equation (2.93) will be more useful later on since it shows that x, is the symbol generated
from the v'" interleaving tap output and Xni+q is the symbol generated from the g" inter-
leaving tap output.
From the definition of a cyclostationary sequence, we have that
Thus, using (2.94), we can rewrite (2.93) in the more compact form
1 n n CO
POD
S-(f). = — ies dd, B epRx(v;
nl + qq — (nl +q—v)fT]fT
v)cos[2m(nl+g —v) (2.95)
2.95
ile =)
o) = |2 140 (2.96)
We shall define the “memory” of the cyclostationary process as the smallest integer /* such
that X,Xni+g =0 for] > I*.
If {x;} is not zero mean, then in addition to the continuous PSD component of (2.89),
we will also have a discrete spectral component (Sj(f))q given by
OS
oi
D
wn Lye. (4)
nT Ysvc0s
Ysveos( (=
n ).
=—00
(2.97)
+|
Sosesin(4)] fo(s-4)
2.4.2 Encoder Output Spectrum for Independent Binary Symbol Input
Consider the input to the encoder of Fig. 2.4 to be a sequence of independent binary symbols
{a;} that take on values +1 with probabilities p and (1 — p). Let m = 0 in (2.88) with no
loss in generality; then since the symbols are independent, the mean of x, is given by
where
Kb
ay AD ani (2,99)
i=l
that is, the algebraic sum of the number of +1’s in the v'" row of G. Furthermore, it can be
shown that
(aye teqt2Bvq() Le (a) et: a= k= 1
Rx(v; nl +4 —v) = § Geta t2Fa(—) _ Gytea, —(K —1) <1 <0 (2.100)
0; otherwise
where
(K—lb
Bug) 2 Me 8v,i8q,i+lb (2.101)
il
that is, the cross-correlation of the v'" row* of G with the g'" row shifted by /b elements.
Finally, substituting (2.98) and (2.100) into (2.95) and (2.97) gives the desired results for
4. The cross-correlation is performed only on the parts of the v'h and shifted g'" rows whose elements overlap.
76 Power Spectral Density of Digital Modulations Chap. 2
the continuous and discrete components of the encoder output spectrum, namely
2
sin af T
(Sm( fe = (=) Se (f)
where a unit rectangular pulse has been assumed for p(t) and
EE ESTE a m vk ?
(Sm(P)a == DU (=ra > (a cos (= )
k=—6S n
(2.103)
2D
Ley Man bee ae eS
[Se n( =) a(/ =)
Several interesting points can be deduced from (2.102). First, note from (2.99) and
(2.103) that for v = q and / = 0, the exponent a, + a — 2Byg(/) = 0 independent of the
code. Thus, for these terms
(G)%t%4—2B vq) = 1 (2.104)
for all @ (including zero). It is also possible that the above exponent can be zero for other
combinations of v, g, and /. The conditions on the elements of G under which this is
possible are derived in [8] and are given by
which are graphically illustrated in Fig. 2.5. This defines a class of convolutional codes
whose significance will become apparent shortly. In the meantime, it is sufficient to remark
that if the code is such that (2.105) can be satisfied for any values of v, g, and / (other than
v = q and/ = 0), then (2.104) is satisfied for all a (again including zero).
Returning to (2.102), we further note that independent of the code, the exponent
a, +g is never equal to zero. Thus, the term (a)*’t will be equal to one only when @
is equal to one and will vanish when a is equal to zero (that is, p = 1/2). This last statement
is crucial to the results that now follow.
The case of a purely random data input (p = 1/2). Consider the special case
of a purely random data input with rectangular pulses and p = 1/2 (a = 0). If the code is
such that (2.105) is not satisfied for any combination of v, gq, and / (other than v = g and
! =0), then the only nonzero terms in (2.102) will be the n terms correponding to v = q
and / = 0 for which (2.104) applies. Thus, Sz(f) = 1 and the continuous component of the
spectrum is just given by (2.63). Furthermore, the discrete component vanishes. Thus, the
conclusion to be reached is that for all codes that do not satisfy (2.105), a random NRZ data
input results in a random NRZ data output scaled (expanded) in frequency by the reciprocal
ul
Sec. 2.4 Power Spectral Density of Convolutionally Coded Modulations 77
1 Oe e © ee eo ee 1 0 Oeeeeed
(k-2)b Lb
Figure 2.5 An illustration of the constraints of Equation (2.105) (Reprinted from [8], [16] © IEEE)
of the code rate, that is, n/b. We define this class of codes as uncorrelated convolutional
codes.>
If, on the other hand, the code is such that (2.105) can be satisfied for at least one
combination of v, q, and / (other than v = q and / =0), then S;(f) of (2.102) will be
frequency dependent and the encoder output spectrum will not be merely a frequency scaled
version of the input. In fact, let {v, g,/} denote the set of combinations of v, q, and / (other
than v = q and/ = 0) for which (2.105) can be satisfied; then (2.102) simplifies to
5. Note that all rate 1/2, noncatastrophic convolutional codes are uncorrelated codes.
78 Power Spectral Density of Digital Modulations Chap. 2
ete
2.0Cq abe
1
. [| :
TI
=
1.6 a ice
B
1.4
We2
uy
ee :
eee
€ = 23
ay secee ai ss
WY =]
. BEEEEEEEEH a t
SESeSSseeeeeeieait :
as Sooo SeSseeecueaia za
| SeeeeeeeeeeeseeetCeseesare
ae SEEGER EERE EERE
SGSEESTEECCEEESS veeTEE
0.2 ee as a. Rea wa a SS
Figure 2.6 Spectrum for best rate 1/3, constraint length 3 convolutional code; dotted
curve is spectrum of uncoded NRZ (Reprinted from [8], [16] © IEEE)
Multiplying (2.108) by (2.63) gives the encoder output spectrum illustrated in Fig. 2.6.
Another interesting result may be gleaned from the example just presented. Since
codes whose generator matrix have at least two identical rows are a subclass of correlated
convolutional codes for which / = 0, then for these codes (2.99) simplifies still further to
1
S;(f) =1+—- cos [27(q — v) fT] (2.109)
: ‘ 2 »
where the set {v, g} now represents all pairs of integers corresponding to pairs of identical
rows in G. Since (2.109) only depends on the difference v — q, and small values of v — qg
correspond to less oscillatory behavior of S(f), then it is clearly desirable from a spectral
concentration standpoint to have the identical rows packed close together in G. For example,
if the bottom row of G in (2.107) is moved to the top, then v — g = 2 and
2
Sef) =1+ 5 cosdnfT (2.110)
which is illustrated in Fig. 2.7. .
Sec. 2.4 Power Spectral Density of Convolutionally Coded Modulations 79
=
SreeFEE
(Sin
(Do
/E
Raeesa
GRRE
BABS
GARE
NZ|ABER
EO
RDIALGRD
eaeaes By|
|
SQRDIRERRISEE
LS
Figure 2.7 Spectrum for best rate 1/3, constraint length 3 convolutional code with
permuted code generator matrix (Reprinted from [8], [16] © IEEE)
The important point to realize from the above is that permuting the rows of the gen-
erator matrix of a convolutional code can be used to change the encoder output spectrum
without any effect on the bit error probability performance of the code.
Sa
See)
SS NO
KS
RS
KS fF
Nn
Ae
DN Cy
SS
Ie
SS
80 Power Spectral Density of Digital Modulations Chap. 2
i
Sef) = 1-5 (a4— a) + (@+a — 24) cos2nfT
a (@ —a°) [cos 4a fT + cos 8 fT] (2.111)
Since this code is an uncorrelated code, for a = 0 (2.111) reduces to Sx (f) = | as it should.
Figure 2.8 is an illustration of the corresponding continuous spectrum (Sm(f)), [ob-
tained by multiplying (2.111) by (2.63)] for several values of p = (1 — a)/2. We note that
the output encoder spectrum narrows as p is decreased despite the fact that the continuous
component of the input spectrum remains unaltered in shape as p is varied [see (2.61)].
The discrete component of the output spectrum is easily found from (2.103) to be
es ae k
(@-#)
(SnMa=Sse Yo (=) e(¢-35) (2.112)
|sed IO here
For the first-order Markov source, we first transform the information sequence {a,} into a
transition sequence {tj} where t; = ajaj_1; j = 1,2,... The sequence {tj} is a stationary
sequence of independent +1 random variables with transition probability p,. For any / and
n > 1, we have
n n
Defining
Kb
pu eh (2.115)
[=2
:
ale ais
Geom
ine
Bites
ABS. ne BA
1 vt -1
fT
Figure 2.8a Spectrum for best rate 1/2, constraint length 3 convolutional code; p = 0.1 (Reprinted from [16])
nl
|
Wl
cos (27 (nK +q —v) fT] —#’ cos [2m(n(K — 1) +q —v) fT]
1 —2t 7b cos2anfT +t 72b
where
Fa Ou
82 Power Spectral Density of Digital Modulations Chap. 2
3/%()s)
Figure 2.8b Spectrum for best rate 1/2, constraint length 3 convolutional code; p 0.3 (Reprinted from [16])
Equation (2.117) is used in (2.89) to get the continuous component of the encoder output
spectrum.
The discrete spectrum is obtained as
(Sm(P)a
(2.118)
Note that a discrete spectrum can potentially exist at the encoder output despite the fact
that the encoder has only a continuous spectrum at its input such as that in (2.49) with
Jp — (1 ay t) /2. Note that if the code is transparent, that is > each row of G has an odd
number of ones ’ then from (2.114)
(25119)
(Sin
()Q/E PCH
COOP
Be
fspibiay
Bie
Bo
Be
aS
be
SRE
Bh
ce a
aa
aes
|
am
an
Be)
ab
Bas
im
Ba
IZ
rely a (2.120)
The corresponding continuous component of the encoder output spectrum may be obtained
from the following:
. |
Se(f) =1- al (2121)
f Pie ry er i) cos4afh ert =i? LP cos 8arr
1 —2tcos4nfT +?
Figure 2.9 is an illustration of (Sn(f)),. [using (2.121)] for several values of transition
density p; = (1 —f) /2.
(Do
(Sin
/E
ee ||
:oH
Figure 2.9a Spectrum for best rate 1/2, constraint length 3 convolutional code;
first-order Markov source; p; = 0.1 (Reprinted from [8], [16] © IEEE)
| (2.122)
=Re {3(nei'} 3(t)
= Aym)(t) + jAgmo(t)
where w, = 27 f, is the radian carrier frequency, m;(t) and mg(t) are baseband modula-
tions of the form in (2.2), that is,
ee)
and s(t) is the equivalent complex lowpass signal. Assuming that {a,} and {b,} are £1, zero
mean WSS sequences as in (2.1) but are independent of each other, then the PSD of s(t) is
related to the PSD’s of m;(t) and mg(t) by
A2
SHGp = =a [Sm(f — fe) + Smi(f + fe)]
(2.124)
Ag
ea eA [Smrate - i) ae Smo(f ot |
One popular carrier modulation to be discussed in Chapter 4 that falls in the class
defined by (2.122) is quadrature phase-shift-keying (QPSK) in which {ay} and {by} are
Sec. 2.5 Spectral Occupancy of Various I|-O Modulation Techniques 85
@
||
a
||
=
=
=
f
||
=
Ba
f
oe
eB
=
S
i
||
||
Ba
Si
sea]
a
||
el
||
S
(Qo
(Sm
/E ||
{isa}
is
B
2
past
B
5
SB
SB
a
I
a
ace
1-3}
|
eeoeEeele
Ee
ASN
BARAT
PARA
AAA
PALI]
Beso
eeeee
BAAO2
"! BOLE
BEARS
ORCI
eAewa
Bee
ABBon
GHG
USRRISSRAORESNO
4 CERES
PORRABOAs
RARE
RS
GRE
PRPs
RI|
BRA
Bes
BRAN
Figure 2.9b Spectrum for best rate 1/2, constraint length 3 convolutional code;
first-order Markov source; p; = 0.3 (Reprinted from [16])
random data sequences, Ay = Ag = A, T; = Tg = 2Ty (Tp is the bit time), and p;(t) =
po(t) = p(t) is a unit rectangular pulse of duration 27). Using the results of Section 2.1.1,
we find that the PSD of QPSK is given by
ee 2
PT; Ee mf —— fe)Tths ‘iSie2 ge Cit Fe)
ii 4
Ss(f) = (2.125)
Catf=foty “Ox + poy’
where P = A? is the power in s(t).
Another popular carrier modulation technique to be discussed in Chapter 10 is
minimum-shift-keying (MSK) [11] which has an I-Q representation in which {a,} and {by}
are random data sequences, Ay = Ag = A, T; = Tg = 2Thp, and
S(f\=—>
8PT, cos? 2n(f — fe)Ti
ete me eee
cos* 2m(f + fe)Tb
(2.127)
Pee toy yi) (l= 16(F + f-)°7,) it 4
Here P = A*/2. Note that, asymptotically, the PSD of MSK rolls off at a rate hpi
compared with QPSK which rolls off as f a:
86 Power Spectral Density of Digital Modulations Chap. 2
“s)
3/0)
z
FE
0.60
aie
Lai
Pt
Figure 2.9¢ Spectrum for best rate 1/2, constraint length 3 convolutional code;
first-order Markov source; p; = 0.7 (Reprinted from [16])
Even faster spectral roll -offs can be achieved by further smoothing of the pulse shape.
For example, so called sinusoidal frequency-shift-keying (SFSK) |12] (see Chapter 10) has
an I-Q representation in the form of (2.122) where {a,} and {b,} are random data sequences,
Aj = Ag=A,T; =Tg = 27}, and
sin 2t/Tp
2mut/Tp y= otherwise
<t<Tp
SS
a PS ke
5IN R|> Bao)a)aS
7)ee (2.128)
ee
RIS eS, VIah BN
[co le
Ale yee Ss
Q Q ~
— a
ll ee
rey
is aM —!
SS | ls a: te
Aik
hs ~~)
Sa
==!
evil
=) so
a
ke)eee
A aeSIN
ers
sof
S
D
oO
=| ss
reVIVi
ta
For SFSK, the PSD can be obtained in the form of an infinite series, in particular [12]
Sec. 2.5 Spectral Occupancy of Various I-Q Modulation Techniques 87
1.0
0.9
0.8
0.7
0.6
(Sm
(Do
/E
Figure 2.9d Spectrum for best rate 1/2, constraint length 3 convolutional code;
first-order Markov source; p; = 0.9 (Reprinted from [8], [16] © IEEE)
PI; | — I
Ss(f) = s |
E (3)Ao(f eS Te) ee SS J2n (;)Bon(f rc fe)
all
Pe
= I ee
6 a Jon-1 (;)Bon—-1(f ax ‘|
(2,129)
| wo | : .
oF E (3)Ao(f ale Sc) spe y J2n (3)Bon(f ap te)
hn
2
= ]
oF oe J2on-1 (;)Bon—-1(f = ra|
i
na sin 27 f Tp
AS are,
Beka I’ 4% cos2xfT
Aot= 54 (i+ an) +3415 a5 )= Tae
1 2n 1 2n
A= 5a(f+Z) +54 (4-2 )3
‘ (2.130)
1 2n — I 1 2n — I
Arm-1(f) = 54 (r+ a )-34 (+- i )
1 I 1 1
Bon-1(f) = Spe (FFP a) ae 5 Aon I G= a)
a) =BG(f)d
B
ey (2.132)
jes G(f)df
i(owsann Se wan?)
S(t) =FAlRe Hire ate
(2.133)
CO
G(f)/2PT,
where h is the modulation index, and the phase shaping pulse, q(t), is given by
t
q(t) a r(t)dt (2.134)
with r(t) the frequency pulse, and T the signaling interval. Typically, h is chosen as a
rational number, that is, h = J/M in which case the data symbols {a,} take on values +1,
+3,...,2+(M — 1) and the number of states in CPM is finite. Furthermore, r(t) is assumed
to be of finite duration, that is, r(t) is nonzero only in the interval 0 < ¢ < LT where L is
the pulse length in symbols. The case L = 1 is referred to as full response CPM whereas
the more general case of L > | wherein the data symbols overlap is called partial response
CPM. Also r(t) is assumed normalized such that
ra |
/ r(t)dt = = (2.135)
5 4
The special case where L = | and the frequency pulse r(t) is rectangular, namely
Ae
raajwe
0;
Osis?
otherwise
(2.136)
is referred to as continuous phase frequency-shift-keying (CPFSK). The special case of
CPFSK for which h = 0.5 corresponds to MSK which, as we have already seen, has an
I-Q representation.
90 Power Spectral Density of Digital Modulations Chap. 2
(dB)
POWER,
OUT-OF-BAND
FRACTIONAL
n
-60 : :
0.01105 040/.151120..25 30 35. 400452 60, Perec tt Peto
of various carrier modulations (Reprinted
Bly from [17] © IEEE)
To compute the power spectrum of a full response CPM signal with rational modu-
lation index, we make use of the signal representation due to Rimoldi [13] which can be
written in the form
s(t) = ARe Pal ae) = Acos (Wet + 2a fut + W(ts un, On)) (2.137)
where
fo ==—h ance
fu aT 2.138
(2.138)
and W(t; Un, On) is referred to as the physical tilted phase which, for M even and 0 < t < T,
is given by
Here uy is a translated and scaled version of ay, that is, un = (a, + M — 1)/2 so that up,
takes on only nonnegative values, namely, 0, 1, 2,..., M — 1 and
t
w(t) =mh(M — le — 2rh(M — 1)q(t) (2.140)
hence, the notion of tilted phase. Finally, o, denotes the state of the modulator in the n™
transmission interval and the transition between states is described by
OD aa, De ee
k=0
(2.143)
2 2 Pale,
ae (f—fu) = Dyel hk “Na 1A
2 ai(f)
where
Wintec s.2
GN=z E d A(T — mn)
, | Met
ae El =j2e(f—fet rr
- Re |: 2. X e j2mhka.(
o( f ¢T — hk ) (2.147)
1 M-1
Gees ge Oe A PT Fe Ik)
M 1=0
Finally, letting h = 1/2 (ie., J = 1 and M =2), T = Tp, hate. (2? = A2/2, (2.147) reduces,
after much manipulation, to
32PT, cos’ 2nfTp
G(f)= 5 ae (2.148)
1 (1 — 16f27;)
which together with (2.131) yields (2.127).
92 Power Spectral Density of Digital Modulations Chap. 2
PROBLEMS
2.1 Starting with the basic definition of the PSD for an arbitrary random process m(t), namely
|
Sm(f) = pies
lim E}—
|Ts ||Mo(f)I’
o( f)|
Figure P2.2
Figure P2.3a
Problems 93
Aw
m(t) = Acos |
(0+ an) PSE a
where a, takes on values of +1 with equal probability and the 6,’s are uniformly distributed in
(—z, m) and are independent of each other.
(a) Find the time-averaged autocorrelation R,,(T) & (mime +>) where the statistical aver-
age is taken over the 6,,’s and the data sequence {ay}.
(b) Compute the PSD corresponding to your answer in part (a).
2.6 Repeat Problem 2.5 for the case of an M-FSK signal where now a, takes on values of +1,
+3,...,+(M — 1) with equal probability.
2.7 A binary-valued (+1) periodic sequence {d,} is defined by the property dp+x = dk,k =1,2,...
where p is the number of sequence positions in a single period.
Dp
(a) Show that the autocorrelation function of {d,,}, namely, Ry(k) & 2 YS dndn+k, is given by
n=l
A(k) — D(k)
= A(k) + Dik)
where A(k) is the number of sequence positions in which the two sequences agree in
polarity and D(k) is the number of sequence positions in which they disagree.
Figure P2.4
94 Power Spectral Density of Digital Modulations Chap. 2
(b) A pseudonoise (PN) sequence is a particular type of binary-valued periodic sequence that
has the following additional properties:
1. The periodp is equal to 2% — 1 (N integer).
2. Ina single period of the sequence, the number of —1’s exceeds the number of 1’s by
one. :
3. The product of a PN sequence and a cyclic shift of itself produces another cyclic shift of
the original PN sequence. Mathematically speaking, {dx} x {dk+m} = {dx+i} where m is
in general unequal to /.
Show that the autocorrelation function of a PN sequence is given by
ile k= 0
Rak) =| 1, REST
Vee be oscrea
together with Ra(k) = Ra(k +np), n= +1, +2,... That is, the autocorrelation of a PN
sequence is a periodic (with period p) two-level discrete function of the discrete shift
variable, k, possessing one value for k = 0 and another value for all other k 4 0 within
the period.
(c) Consider a binary pulse stream d(t) generated by the PN sequence defined in part (b)
where the pulse shape p(t) is a unit rectangular pulse of width T. Find the autocorrelation
function Ry(t) of the waveform d(r).
(d) Find the PSD of the PN waveform of part (c). Is it a continuous spectrum, a discrete
spectrum, or a combination of both?
2.8 Consider a binary random pulse train, m(t), such as that characterized by Eq. (2.2) together
with (2.18). After each N data bits, a parity check bit bit, b;, is now inserted into the sequence.
Thus, the sequence that generates m(t) would appear as ...d_y, d_N+1,---, Ga{nO = oad
cop Os QUSOIN AA AONEDy 0 an CPN GDo sou JUNE jth parity check bit, b;, is related to the previous
N data bits by
N
bj =- I] aiN+k
k=1
Compute the PSD of the resulting m(t). Does it have any discrete components?
2.9 Consider a binary random pulse train, m(t), such as that characterized by Eq. (2.2) together
with (2.18). This modulation is transmitted over a multipath channel that produces a single
delayed and attenuated reflection which adds to m(t). Thus, the received modulation r(t) can
be expressed as r(t) = m(t) + am (t — Ty) where a denotes the attenuation and T, denotes the
delay. Compute the PSD of r(t).
2.10 A digital pulse stream generated by a binary, zero mean, WSS sequence is impulse sampled at
arate 1/7*. The resulting modulation, m*(t), can be written as
[oe]
m*(t)=m(ti(t)=m(t) S* 3 —iT*
—19)
i=—0o
where m(t) takes the form of (2.2) together with (2.1) and the epoch fo is a uniformly dis-
tributed random variable in the interval (—7*/2, T*/2) which is independent of the data se-
quence.
(a) Find the time-averaged statistical autocorrelation function of m*(t), namely R(t) A
(m*Om*e AF 7))where the statistical average is taken both over the data sequence and
the random epoch fo. Hint: Use the Poisson sum relation of (2.12) to express the doubly
infinite train of impulse functions as a doubly infinite train of exponentials.
Problems 95
(b) Show that the PSD corresponding to Rm (t) of part (a) can be expressed as
1 ee 1
Sin® (f) a a ) iS;ee (s= iz)
ia
1=—CO
where Sj, (f) is the PSD of the unsampled modulation as in (2.15). That is, the PSD of
a uniform impulse sampled digital pulse stream is a doubly infinite sum of translations of
the PSD of the unsampled modulation where the frequency interval between translations is
equal to the sampling frequency.
2.11 Suppose now that in Problem 2.10 the sampling pulse train is nonideal in that i(t) is now given
by
i@t)= )) f@-iT*1)
where f(f) is a narrow pulse with unit area. Show that analogous to the result in part (b) of
Problem 2.7, the PSD is now given by
Sah)
4
* ew
where 0 < f,, < 2/2 is the modulation angle, w, and ws. are the carrier and subcarrier radian
frequencies, respectively, and 6,, 9s. are the carrier and subcarrier phases which are uniformly
distributed in (—7z, 2) and independent of each other.
(a) Using simple trigonometry, express s(t) in terms of its unmodulated and modulated com-
ponents at the carrier frequency, and even and odd harmonics of the subcarrier frequency
above and below the carrier frequency.
(b) Using the form of s(t) found in part (a), find its PSD. Identify the various components of
the PSD as being discrete or continuous and the average power of each.
2.13 Suppose that in Problem 2.12 a unit square-wave subcarrier were used instead of the sinusoidal
one. As such the signal is now of the form
where Sg (Ws-t +s) is as illustrated in Figure P2.13. How do your answers to parts (a) and
(b) of Problem 2.10 change?
Sq(,,t
+ 9,.)
na-@0 se
Figure P2.13
96 Power Spectral Density of Digital Modulations Chap. 2
2.14 Consider a sinusoidal carrier that is phase modulated by two binary data modulations. One
is directly modulated onto the carrier with a modulation angle of 2/2 radians. The other is
first multiplied by the first one and then applied to a sinusoidal subcarrier which is then phase
modulated onto the carrier with a modulation angle f,,. The resulting signal is of the form
s(t) =V/2P sin (ct a Smit) + Bmmy(t)m(t) sin (Wsct + Osc) + 6.)
(a) Using simple trigonometry, express s(t) in terms of its various components at the carrier
frequency, and even and odd harmonics of the subcarrier frequency above and below the
carrier frequency. Are there any discrete (unmodulated) components in this signal? If so, at
what frequencies?
(b) Using the form of s(t) found in part (a), find its PSD. Identify the various components of
the PSD as being discrete or continuous and the average power of each.
2.15 Consider a staggered quadriphase pseudonoise (PN) modulation where identical (except for
the 1/2-chip delay between them) PN waveforms (see Problem 2.7) modulate each of the
quadrature carriers. The mathematical form of such a modulation is
where , is the radian carrier frequency, T is the chip time interval, and d(t) is a PN wave-
form that has a statistical autocorrelation function Rg(t) as determined in Problem 2.7. For
simplicity, assume that the code sequence period, p, is very much greater than unity.
(a) Find the statistical autocorrelation function R,,(t) of the cyclostationary process m(t)
defined above.
(b) Find the corresponding PSD S,,(f).
(c) Discuss how the PSD found in part (b) is different from that corresponding to a staggered
quadriphase PN modulation with different and independent PN sequences modulating the
quadrature carriers, that is,
where d}(t) and d2(t) are still assumed to have the same chip period 7.
2.16 Starting with (2.53) and the properties of the Miller Code as described by (2.69), (2.70) and
(2.73), derive its PSD, namely, (2.74).
2.17 Starting with the general form of the equivalent lowpass PSD for a CPFSK signal as given by
(2.147), derive the equivalent lowpass PSD for the special case of MSK as given by (2.148).
REFERENCES
1. Jolley, L. B. W., Summation of Series, New York: Dover Publications, 1961.
2. Titsworth, R. C. and L. R. Welch, “Power Spectra of Signals Modulated by Random and Pseu-
dorandom Sequences,” Pasadena, CA: Jet Propulsion Laboratory, Technical Report No. 32-140,
October 1961.
3. Hecht, M. and A. Guida, “Delay Modulation,” Proceedings of the IEEE, vol. 57, no. 7, July 1969,
pp. 1314-16.
4. Deffebach, H. L. and W. O. Frost, “A Survey of Digital Baseband Signaling Techniques,” NASA
No. N71-37703, June 1971.
5. Simon, M. K., K. Tu and B. H. Batson, “Effects of Data Asymmetry on Shuttle Ku-Band Com-
munications Link Performance,” JEEE Transactions on Communications, vol. COM-26, no. 11,
November 1978, pp. 1639-51.
References 97
6. Nguyen, T. M., “The Impact of NRZ Data Asymmetry on the Performance of a Space Telemetry
System,” JEEE Transactions on Electromagnetic Compatibility, vol. 33, no. 4, November 1991,
pp. 343-S0.
. Nguyen, T. M., “Space Telemetry Degradation due to Manchester Data Symmetry Induced Car-
rier Tracking Phase Error,’ JEEE Transactions on Electromagnetic €ompatibility, vol. 33, no. 3,
August 1991, pp. 262-68.
. Simon, M. K. and D. Divsalar, “Spectral Characteristics of Convolutionally Coded Digital Sig-
nals,” JEEE Transactions on Communications, vol. COM-28, no. 2, February 1980, pp. 173-86.
. Larsen, K. J., “Short Convolutional Codes with Maximal Free Distance for Rates 1/2, 1/3, and
1/4,” IEEE Transactions on Information Theory, vol. IT-19, no. 3, May 1973, pp. 371-72.
. Odenwalder, J. P., “Optimal Decoding of Convolutional Codes,” Ph.D. dissertation, University of
California, Los Angeles, 1970.
. Doelz, M. L. and E. H. Heald, “Minimum-Shift Data Communication System,’ U. S. Patent no.
2,977,417, March 28, 1961 (assigned to Collins Radio Company).
. Amoroso, F., “Pulse and Spectrum Manipulation in the MSK Format,” JEEE Transactions on
Communications, vol. COM-24, no. 3, March 1976, pp. 381-84.
. Rimoldi, B. E., “A Decomposition Approach to CPM,” JEEE Transactions on Information The-
ory, vol. 34, no. 3, March 1988, pp. 260-70.
. Biglieri, E. and M. Visintin, “A Simple Derivation of the Power Spectrum of Full-Response CPM
and Some of Its Properties,’ JEEE Transactions on Communications, vol. 38, no. 3, March 1990,
pp. 267-69.
. Lindsey, W. C. and M. K. Simon, Telecommunication Systems Engineering, Englewood Cliffs,
NJ: Prentice Hall, 1973. Reprinted by Dover Press, New York, NY, 1991.
. Simon, M. K. and D. Divsalar, “Spectral Characteristics of Convolutionally Coded Digital Sig-
nals,” Pasadena, CA: Jet Propulsion Laboratory, JPL Publication 79-93, August 1, 1979.
. Simon, M. K., “A Generalization of Minimum-Shift-Keying (MSK)-Type Signaling Based Upon
Input Data Symbol Pulse Shaping,” JEEE Transactions on Communieations, vol. COM-24, no. 8,
August 1976, pp. 845-56.
Chapter 3
Scalar and Vector
Communications Over
the Memoryless Channel
In the previous chapters, a system level description of a communication link was presented
and the power spectral densities of various modulation techniques were computed. Modula-
tion and demodulation were discussed without much detail into their respective operations.
The various tasks performed in an IF demodulator can be grouped into two main functions,
namely, synchronization and detection. For the purpose of this chapter, we assume perfect
synchronization (or timing) and concentrate on the basics of detection and hypothesis test-
ing. The degradation due to imperfect synchronization on the detection performance will be
the topic of a future chapter.
Information about the world is acquired by observation and physical measurement.
All measurements are subject to error. All efforts to completely eliminate this error finally
reach a bound set by nature’s underlying chaos, which infects all observations of physical
phenomenon with an innate uncertainty (error). Acquiring information about any physical
system involves making decisions. Generally, we must decide which of a set of hypothe-
ses (statements) best describe the system insofar as observations permit one to judge. Such
hypotheses exist and, in the context of some theory, signify the state of the system. Such
a theory requires modeling the sources of uncertainty (error) that corrupt the observations.
The irreducible minimum component of error in making decisions is characterized by ana-
lyzing the decision rules that minimize some convenient measure of the average amount of
error.
The transportation of information requires a communication system. Such a system
corrupts the information by the introduction of random background noise. The need to fer-
ret weak signals out of such noise has called for the development of a Theory of Signal
Detection. This need arose during World War II in the context of determining the presence
or absence of an enemy emitter. At that time, it was realized that decisions made in the face
of uncertainty could be treated by the Theory of Signal Decisions, commonly known as Hy-
pothesis Testing. Such an elegant theory had been fully developed in the 1930s by Neyman
and Pearson [1,2], Wald [3] and others. The motivation of this theory came from “An Essay
Toward Solving a Problem in the Doctrine of Chances” which appeared in the Philosophical
Transactions of The Royal Society of London in 1763 [4]. Therein, the insightful Reverend
Thomas Bayes treated the problem of making decisions under conditions of uncertainty, and
Scalar and Vector Communications Over the Memoryless Channel 99
he recommended selecting that hypothesis for which the a posteriori probability is great-
est. This is called Bayes strategy, and the general formulation of hypothesis testing is often
called “Bayesian.”
In the context of communication of information, the abstract set of hypotheses
{Hm; m=0,...,M — 1} takes on the physical meaning of messages. For convenience,
we denote this message set as {mm, m=0,1,..., M — 1} such that hypothesis H,, has a
one-to-one correspondence with message m,,. We will apply the theory of hypothesis test-
ing to the problem of communication of information with minimum probability of error.
Our top-down approach to demonstrating the theory of signal design (message represen-
tation) and signal detection (message decision) is to apply Statistical Decision Theory to
solve this communication problem. As we shall see, it is rich in results of great practical
interest.
In the communication of information, the message-bearing signal is observed as val-
ues of the voltage r(t) appearing at the terminals of an antenna during a known interval
[0, 7] of observation. This voltage waveform is a random process described by probabil-
ity density functions that embody what is known about the message-bearing signals to be
detected and the statistical properties of the noise that corrupt them. Statistical decision the-
ory, as applied to the communication problem, shows how best to process that input voltage
r(t) in order to decide among several hypotheses (messages) about what r(t) contains. As
we shall see, it prescribes certain kinds of processing, manipulations, and appropriate log-
ical operations that end in decisions. The probabilities of error that these decisions incur
reveal the ultimate limits on the detectability (Probability of Message Error) of the message
embedded in r(f).
Using the top-down approach to decision making depicted in Fig. 3.1, we first con-
sider the scalar communication problem in which the observation is a discrete random vari-
able taking on values r = ro, 71,...,77—1. At first, this may seem to be a gross oversimpli-
fication of the problem; ironically enough, we shall see, in Chapters 4, 11 through 14, that
it is indeed the model that one must ultimately characterize in order to evaluate the perfor-
mance of an uncoded (quantized) and of a coded digital communication system. Next, we
model our observation as a continuous random variable taking on values r = peR (R is the
set of real numbers). A generalization of this model allows for our observation to be a ran-
dom vector taking on values r = pe R™ (R% is the set of N-tuples of real numbers). This
model allows for evaluation of the performance of most uncoded communication systems,
as will be discussed in future chapters. Moreover, the vector model offers great insight into
the detection problem wherein the transmitted messages are represented by waveforms. It
will also be shown that the waveform communication problem can be reduced to a corre-
sponding vector problem with the appropriate transformation, and hence the vector problem
serves as a platform from which the waveform problem can be approached. Figure 3.2 il-
lustrates the detection problem for all cases of interest. The information source outputs one
of M messages, which is transmitted using a scalar, a vector, or a waveform. The channel
is represented by the probabilistic transition mechanism which outputs an element of the
observation space. Finally, the decision rule maps the observed element into one of M hy-
potheses (messages), according to a certain prescribed criterion. A communication link from
a systems engineer perspective is shown in Fig. 3.3. This perspective has been discussed in
great detail in Chapter 1. For our purpose, it suffices to note that depending on the appli-
cation, an appropriate architectural level can be chosen to model the channel as a discrete
memoryless channel (DMC) or a continuous memoryless channel (CMC).
100 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
Top Down
CASE 2:
{s,;} KNOWN VECTORS
n lS A RANDOM VECTOR
CASE 3:
{s,(t)} KNOWN WAVEFORMS
n(t) IS A RANDOM PROCESS
PROBABILISTIC OBSERVATION
INFORMATION
TRANSITION SPACE
SOURCE DECISION
MECHANISM
RULE
mo — Ho:r=so+n
m, — M:r=s;
+n
GaD)
mMuM—1 — Hy-1:r=SsSm—-1+n
OjU|WW UOITe
doy umogq
UOIJEWO}U|
HUIS
UOIeWJ
Oyu]gounos
UONeWOJU jeubis JeuBis
gounos Joyesauayg 10198}9Q
YUIS
DANSI SWIa}SAG-¢"E
Ja9uIsUd datdadsiod
101
102 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
for the additive noise channel where s; is a known deterministic voltage representing the i’
message and n is in general a zero-mean noise with a known probability density function.
The function of the receiver is to operate a decision rule that maps the observed random
variable r into one of the M hypotheses. Typically, the optimization criterion will be the
minimization of the message error probability, but another criteria such as minimizing the
average cost, as in Bayesian detection, can be chosen and is discussed in more detail in
Appendix 3A in the case of vector communications.
Let’s consider first the discrete memoryless channel (DMC) which corrupts the signal by
adding a discrete random variable n. In this case, the output r is constrained to assume one
of a set of J discrete values. We assume that the M a priori (before reception) message
probabilities P(m;) are known. These are the probabilities with which each message gets
transmitted over the channel. In this case, the channel is described by MJ conditional
(transition) probabilities p;,; = P(rj/m;) where M is the number of messages and J the
number of channel observable outputs, as depicted in Fig. 3.4. Often, these probabilities are
grouped into a single M by J matrix P given by
Examples of such channels are depicted in Fig. 3.5. In the binary channel (BC), M = J =2
and the two messages mo and m, experience in general different transition probabilities. On
the other hand, they experience identical transition probabilities in the binary symmetric
channel (BSC). In other channels, the message either is received correctly or is erased.
Such is the case in the binary erasure channel (BEC). Other examples include the noiseless
channel in which p;,; = 1 fori = j and 0 fori ¥ j, the useless channel in which pj, ;= p;
for alli and j, and finally, the M-ary symmetric channel defined by pj, ;= pj,i-
For any discrete memoryless channel (DMC), we would like to derive the optimum re-
ceiver that minimizes the probability of error. That is, upon observing the received symbol
r;, the receiver should decide which message m; was transmitted in such a way as to max-
imize the probability of correct decision, P(C). We therefore expect the optimum receiver
to outperform any other receiver after a “long” series of transmissions. Using the method of
conditional probabilities, the joint probability of transmitting m; and receiving rj is given
by
P(m;, rj) = P(mj)P(rj/mi) (3.3)
where P(mj) depends solely on the message source. The total probability of receiving r j AS
the sum of all joint probabilities P(m;, r;) over all possible transmitted messages, that is
M-\
P(rj)= » P(mi, rj) (3.4)
i=0
Using Bayes rule [5,6], the a posteriori (after reception) probability that m; was transmitted
Sec. 3.1 Scalar Communications Over the Memoryless Channel 103
M9
m4
DISCRETE
MEMORYLESS |
CHANNEL
(DMC)
U2
™M-1 4
M MESSAGE J OBSERVABLE
INPUTS OUTPUTS
(a)
ah PL t9/mo |
m4
ney
ba PL y.4/ my]
Figure 3.4 The discrete memoryless
(b) channel (DMC)
P(rj/mj)P (mj)
P(mj/rj) = P(r;)
oe
(
m9
Poo 1-p
Mo uy Mo uf)
P10 p
B
Poy P i,
m4 4 m4 "4
Py a 1-p i
Figure 3.5 Examples of discrete memoryless channels: (a) Binary channel (BC), (b)
Binary symmetric channel (BSC), (c) Binary erasure channel (BEC)
m(r;) denote the transmitted symbol attributed to observing r;; for example, the receiver
decides that m; was transmitted if r3 is received, that is, m(r3) =m. Then, the conditional
probability of being correct given rj is observed is just the probability that m(rj) was
actually transmitted, that is
P(m
x) P(rj/my) = P(mj;)P(rj/m;) for alli (3.9)
and is referred to as the maximum a posteriori (MAP) decision rule. In the particular case
where the a priori probabilities are equal, the decision rule simplifies further to: set m(r p=
my, if and only if
Example 3.1
Consider the mathematical model of a discrete communication channel having 3 possible input
messages {a, b,c} and 3 possible output symbols {1, 2, 3} as shown in Fig. 3.6. The channel
model is completely described by the set of nine conditional probabilities
1. Given that the received symbol is 1, what is the decision of the optimum receiver regarding
which message is transmitted?
Using Bayes rule, we have
_ P(1/a) PG)
P(a/l)= mak Veh W
where
P(1) = P(1/a)P(a) + P(\/b)P(b) + P(/c)P(c)
Therefore
(0.6)(0.3)
P(a/1) = ———___—
(0.25) = 0.7
().72
Similarly
P(1/b)P(b) — (0.1)(0.5)
ese) 1X) (0.25)
P(1/c)P(c) (0.1)(0.2)
; aes = = 0.08
ey P(1) (0.25)
Since P(a/1) is greater than both P(b/1) and P(c/1), the optimum receiver will decide that
message a was transmitted given that the received symbol is 1.
2. Repeat (1) for the received symbol 2.
Similarly
P(2/a)P(a)
P(a/2)= P(2)
106 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
where
P(2) = P(2/a)P(a) + P(2/b)P(b) + P(2/c)P(c)
Therefore :
(0.3)(0.3) _= 0.25
P(@j2)= 0.36
and
P(2/b)P(b) (0.5)(0.5)
= = = ()'6944 =.
Fey? P(2) (0.36)
P(2/c)P(c) (0.1)(0.2)
= = —__
— _=()'0555...;
Rte) P(2) (0.36)
If the received symbol is 2, the receiver decides that message b was transmitted.
3. Repeat (1) for the received symbol 3.
EGS)
_i= —
P@/a)P(a)
Pa)
where
Therefore
P(3/b)P(b) (0.4)(0.5)
P(b/3) PG)
= —___— 0.39) _ =
__= ——— 0.5128
0.5128...
The receiver will decide that message b is transmitted if the received symbol is 3. Note from
(2) that message b is also decided when symbol 2 is received. Therefore, message oe 9? is never
chosen by the MAP receiver. This is acceptable since each received symbol is assigned a unique
transmitted message, but not necessarily a distinct message. This problem illustrates that in
some DMCs, the MAP decision rule, which minimizes the average probability of error, does
not always decide on all messages in the transmitted set.
4. Given that the received symbol is 2 and the receiver decides that message b was sent, what
is the probability that the decision is correct, that is, that message b was actually transmitted?
0.25
P (correct decision/2) = P(b/2) = 0.36 = 0.6944...
(a)
Ss
{m;} — {si}
Figure 3.7 Continuous memoryless channel (CMC) model for scalar communications
In most channels, the noise n is a continuous random variable with probability density
function (pdf) fy, (a).! As a result, the received random variable r is also of the continuous
type and the continuous memoryless channel (CMC) in the scalar case is modeled as shown
in Fig. 3.7. For a particular observed value of r, say r = ~, assume that the receiver sets
m(p) =m;. Then, the conditional probability of correct detection is just the probability that
m; was actually transmitted given that r = p is observed, that is
1. Strictly speaking, the subscript n in f,(a@) denotes the random variable while the argument is an independent
variable (in this case, w). Sometimes, we will denote the independent variable by the same symbol as the random
variable, i.e., fn(n).
108 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
where f;(o) is the probability density function of the received voltage r. Here again,
P(C) is maximized when P(C/r = p) is maximized since f,(p) > 0. But, P(C/r = p) =
P(m;/r = p) is maximized when m; is actually transmitted. As in the discrete memory-
less channel, the optimum receiver maps p into m; for which the a posteriori probability
P(m;/r = p) is maximum. Therefore, the decision rule is: set m(o) = m; if and only if
Thus, in cases of both discrete and continuous noise, the optimum receiver is a “prob-
ability computer” as depicted in Fig. 3.8. The optimum receiver computes the a posteriori
probability, given that r = p, for all messages in the set and decides on the message with the
largest computed a posteriori probability.
So far, we have shown that the optimum receiver is a probability computer that computes
all the a posteriori probabilities. Indeed, this is the case for any channel with arbitrary
signal and noise statistics. We would like to know if the receiver can be simplified for some
specific channels and how it can be mechanized in a relatively “simple” manner. We will
first consider the case of binary communications to understand the basic fundamentals, and
then generalize the results to M-ary communications.
Binary scalar receivers. In the binary case (M = 2), the optimum detector sets
m(p) = mo if and only if
P(mo/r
= p) > P(m/r =p) (3.14)
Using Bayes rule in mixed form, we obtain
fr (p/m) P(m;)
P(m;/r=p)= (3-05)
fr(p)
where f,(o/m;) is the conditional probability density of r given that m; was transmitted.
Inserting (3.15) into (3.14), we have the following decision rule: decide that mo was trans-
mitted if and only if
fr(p/mo)P(mo) _ fr(e/m)P(m))
fr(p) ‘i fr(e) ee
Cancelling out f;(o) from both sides, we have
mo
f-(e/mo)P(mo) = f-(e/m)P (m1) (3.17)
m\
Sec. 3.1 Scalar Communications Over the Memoryless Channel 109
f - (p/ my ) P(m4)
fr (p/mg ) Pim )
So | S{
The above notation indicates that the receiver decides that message mg was transmit-
ted if the left-hand side of (3.17) is greater than its right-hand side. However, message m is
chosen if the reverse is true. In the case of equality, either message can be chosen randomly.
Figure 3.9 depicts a typical division of the observation space into two disjoint decision re-
gions, say Jp and J). If pel), then m(o) = mj, that is, m, is decided. On the other hand, if
pelo, then m(p) = mo. For p = p* (that is, p¢/p and pe/; simultaneously), both messages
are equally likely to have been transmitted and the receiver can choose either of the two
randomly. Sometimes, (3.17) is expressed as a ratio, that is
A,(p) 4
fr(p/mi) ZI POmo) _ (3.18)
f(p/mo) mo Pom)
This is referred to as the likelihood ratio and it measures which of the two messages is more
likely (in terms of larger a posteriori probability) to have been transmitted. Note that the
threshold 2* depends only on the a priori probabilities of the messages.
To further evaluate the rule, we need to make additional assumptions on the channel
and the statistics of the noise and the signal. Assuming that the signal is disturbed by
additive noise and that the latter is independent of the signal source, then if s; is transmitted,
r = :¢ is received if and only if the noise takes on the value n = p — sj, that is
where the latter equality is due to the independence of the noise from the signal. Therefore,
the decision rule of (3.17) simplifies to: set m(p) = mo if and only if
fn(e — 80)
P(mo) > fn(e — 51)
P(m1) (3.20)
Assuming further that the additive noise is Gaussian with probability density function (pdf)
exp(—x?/207)
fnr(x%) = (3.21)
V2102
where o? is the noise variance (and is assumed to be known at the receiver), (3.20) reduces
110 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
s4
Decide m (P) = my
iff
fi(p)
Aq.>a 2
Otherwise decide m,
So
to
eee,
exp [-| exp [-os) |
P(mo) > P(m\) (3.22)
V2102 V2n02
Note that when the noise has a nonzero mean, the latter can be added to the signals sg and
s, and (3.22) still holds for the modified signals. Taking the natural logarithm of both sides,
(3.22) can be simplified to yield
Hj): mj—r=n
Ay: m,;—r=s-+n
Sec. 3.1 Scalar Communications Over the Memoryless Channel 111
where s and n are independent random variables with respective pdf’s (assume a > 0, b > 0
and a 4 b)
—aS
ON Cem eS ()
fs(8) Fa S<0O
and
be°" N>O
n(N = 4 <
Joh) fe N <0
Find the decision rule that minimizes the error probability if mo and m, are equally likely.
Using the likelihood ratio test of (3.18), the decision rule becomes
2) fr(p/m}) es
A(p)
fr(p/mo) in
as the a priori probabilities are equal. Since r is the sum of two independent random variables
when message mj, is transmitted, the conditional pdf f,(o/m},) is the convolution of the indi-
vidual pdf’s, that is
= /* geale-N be NaN
0
b
Se axb
a ml
A(p)= (ise er ae
ap mo
Rearranging and then taking the natural logarithm of both sides yields
Baz Inf—], ab
M-ary scalar receivers. We have shown that the optimum M-ary receiver sets
m(p) =m; if and only if the a posteriori probability P(m;/r = p) is the largest. Using
Bayes rule in mixed form as given in (3.15) and cancelling f,-(p), the decision rule becomes:
set m(p) = m; if and only if
Simplifying, we have
4) bo
Select m (P)=m,
iff
max iGd= 95
(CN ei (3.32)
or equivalently, the receiver decides m(o) = m; if and only if
In other words, the receiver decides on the message m; whose voltage s; is “ closest” to the
received voltage p. This rule is referred to as minimum distance decoding and it will be used
again when considering vector communications.
Example 3.3
A system is used to transmit M equally likely messages using the voltages s; =iA, i =
0, 1,..., M — 1. The received voltage, when the i'" message is transmitted, is r = s; +n where
Sec. 3.1 Scalar Communications Over the Memoryless Channel 113
m 0
| l | l
m m m m
| lal Taal Bs M-1
| | l | |
ALS lo Shs ee | laa i
0 l A 2A | 3A | Og 6 | (m-1) A ;
| | | |
| | | |
1 1 i] i} i}
A 3A 5A 7A
2 2 22 2
n is a zero-mean Gaussian noise with variance o*. Draw the decision regions of the optimum
receiver.
Since the noise is Gaussian, the optimum rule is minimum distance decoding as shown in Fig.
3.12. Hence, if r< A/2, decide mo. If A/2 <r < 3A/2, decide m, and so on.
In this section, the error probability is used to assess the performance of the optimum
receivers for both binary and M-ary scalar signaling. In other fields such as radar, signal
acquisition, antenna pointing, and so on, other performance criteria are sometimes used in
detection and these are discussed in Appendix 3A.
« exp(—B*/2)
(m})
+ P(m 1 -|i: re
————d B
© exp(—y*/2)
0
O@) (3.40)
56 /2
which results in
|s| Se
Pmin(E) = Q =] — Q =) (3.44)
oO
Thus, the error performance depends only on the signal-to-noise power ratio (SNR) given
by s*/o”. The larger this ratio, the smaller is the probability of message error.
M-ary scalar receiver performance. The optimum receiver in the M-ary case
divides the observation space (a line in the scalar communication problem) into disjoint
regions denoted Jo, 1), ..., —1 respectively, as shown in Fig. 3.13. If the received signal
pel, then m(p) = m; and the probability of being correct is given by
M-1
PC)= S> Pom) P(C/m) = Y* Pom) i}fr(o/midp
i=0 i=
(3.45)
Example 3.4
Compute the probability of error of the receiver in Example 3.3.
Sec. 3.2 Vector Communications Over the Continuous Memoryless Channel 115
So | $1 | eee | SM-1 Pp
— |
Ip ee ly cEgit FC (al emit IM-1
The probability of error is the weighted sum of the conditional probabilities of error, that is
, Mol
P(E) = — E P(E/m;)
© exp(—r?/207) (=
P(E/mo)
NS = P(E
lS ==Uae|| ama re ad ie
———_—_——— dr = =)
and
rera(3)0(8)+2(4?)0(8)=2(“)e(8)
Hence
VECTOR
A (p)
VECTOR VECTOR
SOURCE TRANSMITTER CHANNEL RECEIVER
Select m(p)=m;
iff
PROBABILITY
COMPUTER Nees {P(m,/ r=p)}=P(mj/ r=p)
that the a priori probabilities {P(m;)}, the signal set {s;}, and the statistical characterization
of the channel {f;(p/s = s;)} are all known. The receiver observes the vector p and decides
which message m; was transmitted.
Let P(C/r = p) denote the conditional probability of correct decision given that r= p
is observed and assume that the receiver sets m(p) = mx. Then the probability of correct
decision is the probability that m, was indeed transmitted, that is
where N is the dimensionality of the observation space. Since f-(@) > 0, then P(C) is
maximized when the conditional probability P(C/r = p) is indeed maximum. Hence, the
optimum receiver sets m(p) = mx if and only if
P(m,/r
= p) = P(m;/r
= p) for alli (3.48)
Here again, the optimum receiver is an a posteriori probability computer as depicted in Fig.
3.15. Using Bayes rule in mixed form, we have
Note that the maximum likelihood receiver does not depend on the knowledge of the a priori
probabilities {P(m;)} and that if all messages are equally probable, then the maximum like-
lihood vector receiver is equivalent to the maximum a posteriori probability vector receiver.
When the channel transition probability densities f,(p/s = $;) are known, the maximum a
posteriori probability vector receiver can be simplified further. We will first consider the
optimum receivers in the presence of additive white Gaussian noise (AWGN) and then
investigate the receiver structures when the noise is colored.
When the signal s is disturbed by an additive N-dimensional noise vector n with zero
mean, equal variance and uncorrelated components, the noise vector is defined as “white.”
In the special case when it is Gaussian in addition to being white, the noise components
are independent and we define the channel to be additive white Gaussian noise (AWGN).
Letting n = [719, 145.0'.. nn—1]!, we have
fnj(0j) = Foams
I ox [-«2/20
Dae (3.52)
resulting in
N-1 1 N/2 eee
fn(@) = I fnj(@)j) = (=) exp |-s30 a (3.53)
Binary vector receivers in the AWGN channel. In the case of binary signal-
ing, the optimum receiver sets m(p) = mo if and only if
P(mo) exp |e
lo =i] 2°
rs P(m}) exp
lo — eysil?
oe he (3.55)
P(m})
le —— sol’ $7 < le —§,|?
—Sil — 2071
o nf ee (3.57)
For equally probable messages, the rule becomes: set m(p) = mo if and only if
DECISION @?2
4 BOUNDARY
Ho
i (P)
This rule decides on the message whose corresponding signal s is “closest” to g. This rule
leads to the notion of Minimum Distance Decoding encountered earlier. Coding theorists
work from this perspective.
Example 3.5
One of two equally likely messages is to be communicated over the AWGN channel. Assuming
that the transmitter uses the signal vectors
(i) (3)
draw the decision space and show a mechanization for the optimum receiver.
Since the messages are equally likely and the channel is AWGN, the optimum receiver employs
minimum distance decision and the observation space can be divided into two distinct regions
as shown in Fig. 3.16(a). Let on = (| £2), then the decision rule becomes
mo
i 2 > v3 a
2p So —|Sol” = 2p-s: —'|s3|
m|
or
mo
2(p1 + p2) —2 = 2(—2p; — 22) — 8
my
or
mo
pit+m= —1
m
Hence, the receiver can be mechanized as shown in Fig. 3.16(b). The vector operation p' so iS
referred to as cross-correlation in practice.
Select m @) =m,
iff
max {q,} = 4;
Figure 3.17 M-ary vector cross-correlation receiver for the AWGN channel
where
The vector cross-correlation receiver can be mechanized as shown in Fig. 3.17, where
Gip= p's; + B;. The received signal p is correlated with each possible signal s; (that is, form
the inner product p’s;) and a bias B; is added to compensate for different signal powers and
a priori probabilities. When all the signals have equal lengths and a priori probabilities, the
bias can be ignored.
Kilso wei
DONS
—s Kos + In P(mg )
0
max {q.}= 4,
a KAS a tin P(t)
Figure 3.18 Optimum cross-correlation receiver mechanization in the presence of colored Gaussian noise
fn(p — §;) and taking the natural logarithm, the decision rule in the binary case becomes
I T y-! ee I Ty-l
In P(mo) — 50 — So) Ky (@ — So) = In P(m) — 5 (0 —spiky (@—s1) -G-65)
my)
To arrive at the above equation, we assumed that the noise has zero-mean. The
nonzero mean case can be handled by incorporating the means into new modified vectors,
given by s; = s; + my. Expanding the products and simplifying, the rule reduces to
mo
pK,
so siee 8 (3.64)
mM
where
1
Tiss oa a P(m})
A 5 {80DK n 180—S,K,
a iG 1
Si} |
+1n P(mo) ;
(3.65)
Note that in arriving at (3.64), we used the fact that p’ K, 'so = s)kK, |p since both quanti-
ties are scalars and are the transpose of each other (recall that lis symmetric since Ky is
symmetric and hence K,! = (K,!)"). In the M-ary case, the decision rule can be general-
ized to give: select m(p) = m; if and only if
ae bas
Dp Kee sae 581 Ky Is; + In P(m;) (3.66)
is maximum and the resulting cross-correlation receiver mechanization is depicted in Fig.
3.18. In the case of AWGN channel, K,! = (1/07) and (3.66) reduces to (1/07)(p"s; +
B;) where B; 1s given by (3.60).
Let’s consider an alternative derivation of the optimum receiver in the presence of
colored Gaussian noise. Knowing how to perform signal detection in the presence of white
Gaussian noise, a natural extension would be to perform an operation on the received vector
p, which contains colored noise, such that the resulting observation is corrupted by white
noise only. This is referred to as a whitening operation and assuming that it is linear, the
Sec. 3.2 Vector Communications Over the Continuous Memoryless Channel 121
output noise vector, in addition to being white, would remain Gaussian, since it is a by-
product of a linear operation on a Gaussian vector. Following the whitening process, the
optimum detection rule (assuming equally likely messages) is minimum distance decoding
as we have seen before. However, the detection has to be performed on the modified set of
transmitted vectors at the output of the whitening operation. This two-stage approach to the
detection problem is very intuitive, but it is not clear if it is equivalent to the optimum MAP
receiver.
Let’s start by drawing on a key result from the field of linear algebra on the factor-
ization of matrices, namely that any nonnegative definite symmetric matrix of dimension
NXN has N associated orthogonal eigenvectors e;, i = 1,..., N [8,9]. As a result of this,
the covariance matrix of any random vector, being nonnegative definite symmetric, can be
decomposed as
Ka = Q2AQ7 (3.67)
where &2 is the matrix of orthonormal eigenvectors given by
A=AlPal/2 (3.71)
where
aa wae 0
Ka = 2A1/2(A/4)T
QT = (QA4)(QA'/)F = QOT (3.73)
where the matrix Q is defined as
0=2A!/2 (3.74)
Now in order to whiten the zero-mean vector n with a covariance matrix Ky, suppose we
form the linear transformation
w=Pn (3.75)
122 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
DECISION
Uf WHITENING WHITE NOISE
| TRANSFORMATION SIGNAL
OBSERVATION NEW OBSERVATION DETECTION
IN COLORED IN WHITE NOISE
NOISE
where P is an NxN matrix chosen such that the resulting vector w is white, that is, Ky = Z.
From (3.75), we have
where the new observation y = Q-'p. From (3.58), the optimal decision rule for white
noise is
my
ly — O7'so/? z ly — Q7!s;/ (3.79)
0
or equivalently
i =i
LHS = (2"'p) Q7' (s; — s0) =p" (Q7) Q7' (si — 50)
re (3.81)
=p" (20°) (si — So) = p’ Ky! (S1 — so)
pKa
Tiel
80) a2 5Depa)
Gie §) SsKe $0)
Pre
(3.83)
mo
which is identical to (3.64) for equally likely signals. Hence, we have shown that the op-
timum MAP receiver, in the presence of colored noise, does indeed whiten the observa-
tion and then perform signal detection in the presence of white noise. Again, the cross-
correlation process is involved in signal detection.
Example 3.6
Consider transmitting the equally likely messages mo, m, using the signals
0 0
so= 10 and s;= 10
over the additive vector Gaussian noise channel with covariance matrix
m= (2 5)
4 2
Derive the optimum receiver structure and draw the decision space.
Using (3.64), we have
cates(3 4)(20)
4
=3 (2)
Ren eee ey NEEGS
5 ue
q(-p1 +202) < 0
3 m\
In this case, the factor 5/3 can be dropped without changing the rule. Also, note that the first
component of the observation, 1, is being used in the detection process, even though the
transmitted signals have identical first components. This is because the noise is not white and
the observation of p; offers information on the noise component of 2. If the noise vector were
white, the decision rule would have been
and p; would not be used in the detection process. The decision spaces of both cases of white
and colored noises are shown in Fig. 3.20.
In all derivations so far, we have assumed that the covariance matrix Ky is nonsingular
(or equivalently, that the inverse does exist). What if K, | does not exist? In this case, the
noise vector lies in a space of dimension less than NV, even though it has N components
(similar to three vectors lying in a plane even though each vector can have three components
that are nonzero). Expanding the noise vector in terms of the eigenvectors of Kp, it can be
124 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
Py
CHANNEL
1
CHANNEL
2
MULTIPLE
VECTOR VECTOR fi (p)
TRANSMITTER RECEIVER
shown that n has at least one noise-free dimension or direction. Perfect detection is then
possible in this case, provided that the projections of the transmitted vectors in these noise-
free dimensions are distinct. This is referred to as singular detection. If the projections
are not distinct, then the detection should be performed in the reduced dimension space,
spanned by the noise vector [8].
where L denotes the number of received vectors (or diversity channels) and
rf)(12,9. PN (3.85)
Given that r’ = p’ = (p,", py", ..., p,,") is observed, the MAP receiver sets m(p) = m;
Sec. 3.2 Vector Communications Over the Continuous Memoryless Channel 125
SCALAR
TRANSMITTER
MULTIPLE
SCALAR
RECEIVER
tg=Ny*No
No
if and only if
Sr( py, Pr,---,Pp)= fry /r7 es r,(P|/Pr; tees PL) Sr Hee ry,(Pr; =. Dy) (3.88)
(3.87) simplifies to
P(mj) fry /r,....7,,5(P 1/P2> 1+; Pz,S=Si) fr Rites r,/s(P2, ++, Py_/S=S§;) (3.89)
is maximum. In this case, the knowledge that r); = p, is ignored by the MAP receiver
as it does not enter into the maximization of (3.87) with respect to 7. Hence, r} = p, is
irrelevant to the decision process and the performance of the receiver is not degraded [10].
This reasoning can be applied iteratively to (3.90) to ignore all irrelevant observations and
simplify the MAP multivector receiver to the extent possible.
Example 3.7
In the multiscalar communication system depicted in Fig. 3.22, the transmitted signals (s; =
—so =) and the noise nj, 2 are all statistically independent random voltages. Assuming that
the messages are equally likely and that the noises are Gaussian with zero-mean and variance
o”, derive the structure of the optimum multiscalar receiver.
The decision rule is given by (since L = 2)
m(
Fersrr(P1, P2/Mo) = fri,r2(01, 22/1)
m|
But
exp \-333
I
((o1 +)? 2 + (p2 = 1-9)53 )|a2 0 |—357l ((o1 — 5)?
\2
+ (p2 a1 +)’)2
Taking the logarithm and rearranging, we get (assuming s > 0)
my\
pi-=—2 = 0
2° mo
The optimum decision rule partitions the N-dimensional observation space into M disjoint
decision regions Ro, Rj, R2,..., Ry—1 corresponding to the M messages. The decision
regions are defined by the decision rule which sets m = mx if the decision function |r —
s;|2 — 207In P(m;) is minimum for i= k.
To gain geometric insight into this decision rule, consider first the case of equiprob-
able messages. In this case, the second term in the above decision function can be dropped
and the receiver decides m = m, if the reduced decision function |r — §; | is minimum for
i =k. Noting that |r — s;| is the Euclidean distance between r and s;, then the decision rule
has a very simple geometric interpretation, namely, the decision is made in favor of that
signal (message) that is closer to r. This is called minimum distance decoding which is a
familiar concept in coding theory.
When the messages are not equiprobable, then a particular message m; will be most
probable. In this case, a decision is made more often in favor of m; and the decision regions
will be biased in favor of m;. This point is illustrated by the appearance of the term In P(m;)
in the decision function.
(b)
Figure 3.23 Decision regions and decision boundaries for binary communication
not the bisector of) the line that joins the vectors $9 and $s; at a distance jz from So (see Fig.
3.23b). Here, pp = (c* + d*)/2d = (07 /d) In{ P(mo)/
P(m )] + d/2 where d is the distance
between So and $1.
To prove the above, we make use of the geometry in Fig. 3.23c. It is clear from this
figure that ae =a? +? and d? = a* + (d — 1)”, so that dp — d? = 2du — d* =c?. This
is the desired result since along the decision boundary, d? — d; is constant and equal to c?.
Note that for equal a priori probabilities, c = 0, and along the boundary, dp = d).
Figure 3.24 Decision regions and decision boundaries for ternary communication
In this section, we derive the message error probability of the optimum MAP vector receiver
in AWGN channels. First, we consider the case of binary signaling and then generalize
to M-ary communications. The performance in the presence of colored noise can also be
derived following similar steps.
| IC
p™ (si — 80) < 5 (Isi? = [sol?) — 07 In ( on (3.93)
P(mo)
Sec. 3.2 Vector Communications Over the Continuous Memoryless Channel 129
1 P(m})
So +n)? +m)’ (Ss;
(So — so)
(Ss; — So) <=< =5 ([s1|7
(isi = Iso| |sol?)
) — o* 07In
In (|——
P(mo) (3.94)
Performing the inner products and simplifying, we have
1 P(m,)
n’ is(S1
(Sj— So)
So) < =|S;
2 —o~12
5 iar — So|“
50g n(5) (3.95)
19>
i inner product n’ (s; — so) is a zero-mean Gaussian random variable z(u), with Saran
oa? = E{z*(u)} == (S} — So)! Kn(S1 — So). Since Ky = 07Z, the variance reduces to o?
ae so|?. Using the transformation z(u)= o|s; — So|x(u), where x(u) is a nonmalited
(zero-mean, unit-variance) Gaussian random variable, (3.95) can be rewritten as
dio oO
x(u) < — — —In (Fe) (3.96)
20 dio P(mo)
where do = |S; — So| denotes the distance between vectors sp and s;. Using the Gaussian
tail integral, the probability of error, conditioned on mo, is then given by
Aging? P(m})
P(E = — +— |] 3.98
Shee be dip ster She
Hence, the total probability of error becomes
P(E) =P(mo)Q
mr \26 dio
a) )
Ses \-P(mo)
(3.99)
+ P(m)Q
ics 2z in (Fe)
han ARG
When the signals are equally likely, the expression reduces to
d
P(E) = o(3) (3.100)
20
which is a function of the ratio of the distance between the signals to the noise standard
deviation. This is a fundamental result which suggests selecting the binary modulation
technique that maximizes the “distance” between the vectors that represent the messages
in the channel.
M-ary vector receiver performance. In the M-ary case, the receiver partitions
the observation space into M disjoint decision regions R; as depicted in Fig. 3.13 [since the
figure deals with scalar communications, R; (a multidimensional region) is replaced with /;
(a one-dimensional region)] and the conditional probability of correct decision given that m;
was transmitted is given by
and
M-1
P(C)= S47 P(m;) P(C/mi) (3.102)
i=0
For some specific signal constellations, the multidimensional integral can be reduced to the
product of N one-dimensional integrals when the noise is white and Gaussian. However, in
the general case, the integral cannot be simplified any further.
Example 3.8
A ternary communication system employs the vector signal set illustrated in Fig. 3.25 to trans-
mit equally probable messages mo, mj, and m2 over the additive white Gaussian noise channel
with zero-mean and covariance matrix Ky = o°Z. Write a mathematical expression for the ex-
act probability of error P(E).
Noting that cos 60° = 1/2 and sin 60° = /3/2, the signals vectors are given by
= =e,
4(5) = Yuk
i=
elle
A( 7H) and n=
nah elle
a( /f, )
Since the noise is white and Gaussian, minimum distance decoding is the optimal rule, as shown
in Fig. 3.25. Since the signals are symmetric and equally probable
P(E) = 1 — P(C/mo)
But
or
P(C/mo)
0 =f i
Fin SN exp | 5: ((01 - A)’
) +0} d p2dp
and
P(E) = 1 — P(C/mpo)
The above integral cannot be simplified any further and needs to be evaluated numerically for
performance prediction purposes.
To facilitate analysis and/or to minimize the average required transmitted energy, it is often
advantageous to translate and rotate the signal constellation {sg, s1,..., Sy—1}.
Consider the conditional probability of correct decision given that m; is transmitted,
that is, P(C/m;). This conditional probability is the probability that the received vector
r=s; +n lies in region R;, as depicted in Fig. 3.26a. This is also the probability that
the noise vector drawn from the point s; lies within R;. Since this probability does not
depend on the origin of the coordinate system, we can translate the coordinate system to
facilitate the analysis. Such a translation is equivalent to translating the signal point and
the corresponding decision region. Consequently, P(C/m;) = P(C/s;) for the translated
system shown in Fig. 3.26b is identical to that of the system in Fig. 3.26a. We conclude
that translations of the entire signal constellation and the corresponding decision regions do
not alter message error probability. This in turn suggests that “different” signal sets yield
equivalent performances.
In the case of spherically symmetric additive white Gaussian noise, we make another
important observation, namely, the rotation of the coordinate system does not affect the
message error probability because the pdf fn(@) possesses spherical symmetry. To prove
this, consider Fig. 3.26c which is generated from Fig. 3.26a by translation and rotation.
Notice that rotation of the coordinate system is equivalent to rotating the signal constellation
in the opposite sense. The probability that the noise vector n (drawn from s;) lies in R; is the
same in Fig. 3.26a as the probability that the noise vector (drawn from s;) belongs to R; in
Fig. 3.26c. This is because the probability P(C/m;) is characterized by the integral over
the region R; of the noise pdf fnp(@ — s;) in Fig. 3.26a. If one translates s; to the origin, the
probability in region Rj is not affected by any rotation of the region R; because symmetric
Gaussian noise (same variance in all directions) possesses spherical (circular) symmetry.
Example 3.9
One of sixteen equally likely messages is to be communicated over a vector channel which adds
a Statistically independent zero-mean Gaussian random variable with variance No/2 to each
transmitted vector component. Assuming that the transmitter uses the signal vectors shown in
Fig. 3.27(a), express the P(E) produced by an optimum receiver in terms of the Gaussian tail
integral.
Since the channel is additive white Gaussian, the decision rule is the minimum distance de-
coding rule as shown in Fig. 3.27(b). Dividing the observation space into three major types as
depicted in Fig. 3.27(c), there are 4 regions of type ©, 4 regions of type @, and 8 regions of
type @. Since translation and rotation does not affect P(E) and hence P(C), we have
132 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
Figure 3.26 Translation and rotation of signal vectors and decision regions
ee ares I Ee
= - =——exp |= (x +9)
y?)[axa
bs
4 J_aj2 J—a/2 ™No No :
d/2 lo) 1
1 1
ou >/ —— exp a (x 35 )|dxd
2 J—aj2 J—aj2 No No | :
1 ee ee i 1 A ;
A ae ND) ls (x ae )|
dxd
4 J—aj2 J—aj2 No No 3 ?
Sec. 3.2 Vector Communications Over the Continuous Memoryless Channel 133
I 1
P(C)
(C) = —-(1—
4! P) p)” eae p)( 2p) +7 2p) 2
Based upon our arguments presented in the previous section, there exists infinitely many
signal constellations which yield the same message error probability. From a system de-
sign prespective, it makes sense to choose the set of signal vectors that minimizes the mean
energy required to yield a specified message error probability. This suggests choosing the
signals which are closest to the origin. By translating and rotating the signal set, one can
always find an arrangement of signals which require minimum energy. For example, trans-
lating a set of orthogonal vectors to minimize the required energy yields the transorthogonal
(simplex) set.
Consider the mean energy
vector Ss; to create a new vector s; = s; —a for all i. In other words, we wish to find the
vector a which minimizes the new mean energy
M-1
E'= )> P(m)lsil « C100)
i=0
One can easily show (see Problem 3.42) that the vector a specifies the center of gravity of
the M vectors So, $1, ..., Sy—1, that is
M-1
a= Dy P(m;)s; = E {s;} (3.105)
1=0
This result readily follows from the fact that the energy E’ is equivalent to the moment
of inertia about a point a of the “M particles of probability masses” P(mo), P(m}),.--,
P(my-1), respectively.
Example 3.10
Consider the equilikely binary orthogonal signal vectors s) = (W/E, 0) ands? = (0, VE). Cre-
ating the mean signal vector a and subtracting it from So and s; yields the new signal vectors
sy’ = (—/E/2, 0) and s,7 = (/E/2, 0) = —sp. The message error probability for both sig-
nal sets is Q(./E/No); however, the latter requires a smaller mean energy. This is one of the
reasons why BPSK (antipodal vectors) is selected in practice over BFSK (orthogonal vectors).
Example 3.11
Transorthogonal (simplex) signal vectors are the minimum energy equivalent set of equilikely
orthogonal signal vectors. This is illustrated in Fig. 3.28 for the case of two- and three-
dimensional signal space. Note that energy minimization effectively reduces the number of
dimensions (bandwidth) of the signal space by one. For two-dimensional orthogonal signal vec-
tors (BFSK), the corresponding transorthogonal signal vectors lie on the same straight line, that
is, the one-dimensional antipodal signals (BPSK) as shown in Fig. 3.28a. For three-dimensional
orthogonal vectors, the plane formed by the three signals is constructed where the center of
gravity of the signal vectors on this plane becomes the new origin. It is obvious that the equiv-
alent transorthogonal signal vectors are reduced to two dimensions as depicted in Fig. 3.28b.
Similarly for the four-dimensional case, the minimum energy set lies at the vertices of a tetra-
hedron (as shown in Fig. 4.21 in Chapter 4) reducing it to a three-dimensional space. It can
be shown (in Chapter 12) that for a given mean energy, the equilikely transorthogonal signal
vectors are optimum in the sense of minimum error probability in white Gaussian noise.
For the orthogonal signal vectors s; = VE@;, the mean signal vector a is given by
and
so that
1
B= 15)?=6 (1~ x)
M
Hence for the same message error probability, the mean energy of the transorthogonal set is
(1 — 1/M) times that of the orthogonal set. For large M, orthogonal and transorthogonal signal
sets have the same message error probability.
Sec. 3.2 Vector Communications Over the Continuous Memoryless Channel 135
For most signaling schemes, the computation of the exact probability of error is tedious and
an upper bound on the performance might be sufficient. In this section, we will derive such
a bound for vector communications over the additive white Gaussian noise channel. The
bound is general and can be applied to any set of vectors.
We know from the previous discussion that the optimum a posteriori probability re-
ceiver sets m(p) = mz, if and only if
or if any three events Ej, Em, En (j #k, mAk andn#k) occur jointly, ..., or if all
M — events Eo,..., Ex—1, Ex41,..-, Em—i occur together. Mathematically, this can be
denoted using a union of events as
This bound is referred to as the Union bound and is very useful in communication prob-
lems. Consider the event that {p/sx < p's; /mx}. Assuming that the noise is white Gauss-
ian, the probability of this event is given in (3.100) as
dk j
P(E ;/m) =| — (3.112)
20
where dx; is the distance between the vectors sx and s;. Hence, the Union bound simplifies
to
M-\
dk
P(E/m) < 9) Q @ (3.113)
j=0
i#k
The probability of error P(E) can then be bounded using (3.113) with the appropriate a
priori probabilities. This is a key result that will be used in the chapters on coded communi-
cations.
Example 3.12
Assuming that a communication system uses M orthogonal signals to communicate over an
AWGN channel with Ky, = 027, derive a bound on P(E) when |s; |? 4 Dei m—a Ole eS oe M —-1.
The signal can be modeled by
7 =VEO) = 01M
Sec. 3.2 Vector Communications Over the Continuous Memoryless Channel 137
where
for all i and j in the set {0, 1,..., M — 1}. The distance between any two vectors is given by
d€d=VE+E=V2E
and is independent of which pair we choose due to the orthogonality. Using (3.113), we get
jh;
PREYS 1)0 (z
where we have assumed that 0? = No/2 =kT°/2 Joules. At room temperature, T° = 300K =
28°C and kT° = k300K = 1/40 eV (electron-volt).
In Section 3.2.4 we described the traditional approach for evaluating the probability of
correct decision for N-dimensional, M-ary vector receivers wherein the observation space is
partitioned into M disjoint N-dimensional regions Rj; i = 0, 1,2,..., M — 1 surrounding
the signal vectors. The origin of coordinates for defining these decision regions is assumed
to be the same for each of them and is located in the N-dimensional space defined by the
received vector r. In particular, for the i signal transmitted, the conditional probability of
correct decision is described by the N-dimensional integral in (3.101) which, in the general
case, cannot be simplified any further. Even for the case of a two-dimensional signal vector
set and an AWGN channel (such as in Example 3.8), we saw that the expression for the
conditional probability of correct decision still took the form of a double integral. Thus, in
general, the average probability of correct decision, or equivalently, the average probability
of error, would be expressed as the sum of M such double integrals.
Recently, Craig [16] cleverly showed that the evaluation of average probability of
error for the two-dimensional, AWGN channel case could be considerably simplified by
choosing the origin of coordinates for each decision region in the two-dimensional space de-
fined by the corresponding signal vector. This shift of the coordinate system origin from the
received vector space to the signal vector space allows the integrand of the two-dimensional
integral describing the conditional probability of error to be independent of the transmitted
signal. In particular, this integrand is simply the pdf of the two-dimensional noise vector
n which is characterized by a two-dimensional zero mean Gaussian pdf with independent
components. When expressed in polar coordinates, that is, (r, @), this pdf can be integrated
on r to yield a closed form result, that is, a single exponential, regardless of the complexity
of the integration limits or their dependence on @, the latter being dependent on the shape
of the decision boundaries. Thus, at worst, the average probability of error would now be
expressible as the sum of M single integrals. In short, this new approach of computing av-
erage error probability has shifted the complexity of the computation from the integrand to
the integration limits.
To illustrate the above approach, we consider a generalization of Example 3.8 cor-
responding to an M-ary communication system with a vector signal set having M points
138 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
uniformly distributed around a circle of radius A. The geometry for the decision region as-
sociated with the signal point
—1
So=A ( ‘)
is illustrated in Fig. 3.29. An error is made if the noise vector n lies outside the wedge-
shaped decision region bounded by the semi-infinite lines OA and OC, as for example, the
point Z. In particular, the probability of an error given that signal So is transmitted is twice
the probability that Z lies above the boundary AOD, namely
(M—1)2/M re
/2KGS, imo) =2 f ao | fa (r,
9) dr (3.114)
0 R
where R is the distance from the signal point to the boundary point E (in general, a func-
tion of #), and fn (r, @) is the above-mentioned bivariate Gaussian pdf of the noise vector
expressed in polar coordinates, that is,
Te r2
far, 0) ae exp (-z) (Selilis))
which is clearly independent of 6. Substituting (3.115) into (3.114) and carrying out the
integration on r gives
1 (M—1)x/M AZ sin2 4
P (Elmo) = — | ex dé 3.116
mt Jo P 202 sin? (0+ a) ( )
Making the change of variables ® 4 [(M — 1)/M] — 0, the final form for P (E |mo) is
J fdM@-Dx/M A? sin? 4
P (Elmo) =— f ep) eo 3.118
Tt Jo : 202 sin? © ( )
Finally, since from the symmetry of the vector signal set, we see that P (E |m;) is indepen-
dent of 7, then the average error probability is also given by (3.118), that is,
1 p(@-Dx/M Avene ES
Pe@)=— f exp | d® (3.119)
P(E) = af a - | d® (3.120)
m Jo 20? sin? ®
which is a considerable simplification of the result previously obtained for this case in the
form of one minus a double integral. Furthermore, for small values of P (E), the direct nu-
merical evaluation of (3.120) will yield improved accuracy relative to the previous indirect
result where a number close to unity (the probability of correct decision) is first evaluated
and then subtracted from unity.
For the binary case (M = 2), (3.119) evaluates to
1 m/2 AZ
P(E) = - exp -ss| do .121)
Tw JO 202 sin* ©
However, for this case, the result as obtained from the traditional evaluation approach in
Section 3.2.4 is expressible as a Gaussian tail integral, namely
A
p(e)=0(=) (3.122)
Thus, we see that the Gaussian tail integral, which is normally defined as the integral over a
semi-infinite interval of a Gaussian pdf, can be expressed as the finite integral
1 m/2 x2
= — exp |-——~— |d® (Gall23))
24) =} | 75
which is useful for numerical evaluation purposes.
It is straightforward to generalize the above results to signal sets whose decision
region boundaries are polygons surrounding each of the signal points. A typical example
of such is described by the geometry in Fig. 3.30 where the nth signal point s,, is illustrated
inside its correct decision region described by the four-sided polygon ABCDA. If signal s,
(corresponding to message m,) is transmitted, then an error occurs if the tip of the noise
vector lies outside of this polygon. The region outside of this polygon is made up of the four
disjoint regions ©, @, @, and @. The probability that the tip of the noise vector falls in region
@ is described by an integral analogous to (3.116), namely
6 ew)
Pi (Elm)=— f | alba h | (3.124)
1 0 202 sin? (0 + W1)
140 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
The probability that the tip of the noise vector falls in any of the other three regions is
given by an expression similar to (3.124). Since the four regions are disjoint and comprise
the entire error space, then the conditional probability of error given that message m, was
transmitted is
G; __ Apsin’
A? sin? wi
EACEMin;, = (CEMA) = dé Brl25
i he ey Chas aefe o|-
S52 Ga (0
ase
+ Wi)
The conditional probability of error for any other point in the signal constellation is
given by an expression analogous to (3.125) except that the number of terms in the sum
might be different depending on the number of sides in the polygon that defines its decision
region. Finally, the average symbol error probability would be obtained by averaging all
of these conditional error probabilities over the a priori probabilities of the message set.
As such, the average symbol error probability is expressible as a weighted sum of single
integrals with finite limits and an exponential integrand.
Appendix 3A
Other Detection Criteria
So far, we have considered the probability of error detection as the only performance mea-
sure for the receivers. Depending on the application (for example, in radar system design),
other criteria [11] might be more suitable and these make use of differing amounts of in-
formation and problem specification. The two criteria for the decision rule that we have en-
countered are the maximum likelihood (ML) criterion and the maximum a posteriori (MAP)
criterion. In this appendix, we will consider three more criteria: the Bayes decision rule,
the minimum-maximum rule, and finally, the Neyman-Pearson criterion. Only the basics of
each rule are discussed assuming a binary hypothesis and the reader is referred to appropri-
ate references for a more detailed treatment.
and
Since correct decisions are less costly than erroneous ones, we can safely assume that
so that Co} — Cj; > 0 and Cio — Coo > 0. Bayes decision criterion selects the decision re-
gion Ro in order to minimize the risk. Since the first two terms of (3A.5) are independent of
Ro, they are not involved in the minimization. The integral, on the other hand, is minimized
by selecting Rg to contain all the observations p that satisfy
Ho
P(Ho)(Ci10 — Coo) fr(@/Ho) = P(A1)(Co1 — Cit) fr(@/A1) (3A.9)
Ay
or equivalently
Ar(p) A
fe(p/ Hi) 21 P(Ho)(Cio=Coo) _ , (3A.10)
fe(o/Ho) py P(H)(Cor = Cui)
which is a likelihood ratio test. Note that when Co; = Cio = 1 and Coo = Cj; = 0, the Bayes
rule reduces to the MAP decision rule, as given in (3.18) for the scalar case. All results
derived herein can be generalized to M-ary decisions as given in [12-14].
Example 3A.1
A decisioning system must decide between hypothesis Ho and H;, where Ho is the hypothesis
that an observed random variable x(u) has a pdf
l x?
fx(x/ Ho) = ex (-5)
Vir, Wee
and H is the hypothesis that x(w) has a pdf
i F
0, otherwise
The a priori probabilities of Ho and H; are 3/4 and 1/4, respectively. The costs are given by
Co =0 Ci=1
Cor =2 Cy =0
Determine and describe the optimal Bayes decision procedure.
Sec. 3A.2 The Min-Max Criterion 143
fe) 1.081 3
Since f,(x/H)) = 0 for x < 0 and x > 3, then decide Hp in this region. For 0 < x < 3, use the
ratio in (3A.10) to get
pen——.,
exp(x*/2) for0<x <3
i,0) 2/2
or
A, 9
felp/Ho) Zz
$0 PCHi)(Cor = Cus) Gen
:
fr(p/Hi) yy, Cl — P(A1))(Ci0 — Coo)
associated with P(A), or equivalently
fr(e/A1) Ua
> lbey(19 poe— P(A1))(C10 a— Coo) (3A.15b)
fr(e/Ho) Hy = PH) (Cor — Cin)
This is followed by a maximization to find the least favorable P(H)}).
Example 3A.2
Assuming that the a priori probabilities are unknown, repeat Example 3A.1 for the min-max
rule using the pdf’s
exp(—x) = 2( P(A) )
2 exp(—2x) i, 1 — P(AM;)
or
siPte (Luo
A
Mae e(€2)))
or
Ho =
rare gels (soe
a) By.
P(Ao)
The decision rule is valid as long as 7T* is positive (that is, P( Ho) < 4/5). In this case, the risk
is given by [using (3A.1)]
P(H\/Ho) = / exp(—z)dz
0
and
foe)
P(Ho/H\) = / 2 exp(—2z)dz
ti!
P(Ho) = 2/3 maximizes the risk and results in 7* = In 2. Therefore, the min-max decision rule
reduces to
A
Rie In?)
Ay
The first event is called “target detection” and the probability associated with this event
is called the probability of detection, Pp. The second event always occurs if event (1)
does not occur; therefore, the target is missed and the miss event occurs with probability
Py = 1 -— Pp. Event (3) is called a false alarm since a target presence is declared even
though it is not actually present; its probability is denoted by Pr4. Using the same reasoning
as in events (1) and (2), the probability of event (4) must be the complement of event (3),
or Pc = 1 — Pra, and is sometimes called the probability of correct dismissal. Since the
decision process just described is binary, only two probabilities are needed to fully specify
the radar (signal presence indicator) performance. Notice that the error probability is given
by P(E) = P(Ho) Pra + P(A1)(1 — Pp), when the prior probabilities are known. In terms
of the channel transition densities, we have
The Neyman-Pearson criterion maximizes the probability of detection, Pp, under the con-
straint that the probability of false alarm, Py, is equal to B. The solution is obtained by
using Lagrange multipliers to reformulate the problem as an unconstrained one. Construct
T= Pp— A(Pra
— B)
2 / {fe(o/H1)
~Afe(p/Ho)} do+28 3A.20
ae
Ry
146 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
Clearly, if Py4 = 6, then maximizing [ maximizes Pp. As before, the integral is maxi-
mized if R; is set to contain all observed vectors p, for which
is satisfied. Note the similarity of (3A.22) to the maximum likelihood criterion, in which the
threshold A is unity.
Example 3A.3
Consider the binary problem with the following conditional pdf’s
1
fx (x/ Ho) a Jin exp(—x"/2)
IU
1
iy) = exp(—(x — m)?/2)
iaG
We require that Pr, = B. Derive the Neyman-Pearson test.
From (3A.22), we have
Ay InXt m
nie rey
Ho mM
The false-alarm probability is therefore
[e,2)
1
Prea=P(Hi/Ho) = | =
|
o(m +3)
Ind+-=)]=0.1
Q(1.282) = 0.1
and hence, Ind + 1/2 = 1.282 resulting in A = 2.18. The decision rule reduces to
A
We 1280
Ho
Appendix 3B
The Gaussian Integral Q(x)
In this appendix, we present a numerical table and a series approximation for Q(x), which
are useful in decision-theory problems. The Gaussian tail integral is defined by
foe) 1 sie
Q(x) aftae ex lamers
(-5) dy=1-—
dice Q(-x
Q(-x) 3B:1
(3B.1)
a)
ace) 8 =u") ddu
f exp (—u?) (3B.2 )
where
(1)(3)G)...(2n
Pr waa
+ 1) >0 (3B.7)
% error <
x?
x K=O) ca 1—O(%)
0.50000 00000 00000 . 0.84134 47460 68543
. 0.50797 83137 16902 0.84613 57696 27265
0.51595 34368 52831 ° 0.85083 00456 69019
0.52392 21826 54107 0.85542 77003 36091
eecess
aorflno 0.53188
ooo°o°o 13720 13988 ococo
Oolhno 0.85992 89099 11231
0.53982 78372 77029 0.86433 39390 53618
e 0.54775 84260 20584 Ha 0.86864 31189 57270
. 0.55567 00048 05907 0.87285 68494 37202
0.56355 94628 91433 0.87697 55969 48657
Onrrno 0.57142
Pree
eY 37159 00901 0.88099 98925 44800
and get
Le ail x ale Zz
O= aS {©exp (-5) =| oe (-5) ac} (3B.9)
1 ae
Q(x) < ——— ex (-5) (3B.10)
J 20x ae
Also integrating by parts the second term in (3B.9) with
1 3
= ee du = me
we obtain
| 22 1 x2 oO 3 22
; noe ar az =~ SAD ore ~| AOD ro dz (3B.11)
Thus
Ol) ett (x l Fe 8 a i
x) = ——j-exp(—-—]-—
ae p 5 = &XP -—5 #f —ox (—_—=) cf (3B.12)
Ignoring the third term which is always positive, we have for x > 0
«)> =l (1-5)
1
ex (-5)
x2
3B.
Rilke Se \ ape ee erie
= (-8)=o(-2 anaes
Combining (3B.10) and (3B.13), we have
This is an upper and lower bound on Q(x) for x > 0. This bound can be improved by
rewriting (3B.12) as
1 oe "00 3 i
1
Q(x) = one {(1— =)exp (-5) +x] a exp (-5) a:| (3B.15)
Sec. 3B.2 Series Approximation for Q(x) 151
ee A SOG).
we have
| ae x2 Pr(G)G6) 22
Oa) =
Fa (int w)or(-F) fo SP ew (-5)ac}caro
Again, integrating the second term by parts with
~ B&O) _ (3)(5)(7)
u= eg yp —— es eee
4 ke
z 2
dv =zexp (-5) dz, v=—exp (-5)
we have
2
O(x)=pees
anepie b 2aa3)4G)
{1 ier =
(3B.17)
2 (ove) 2)
+X exp (=) / ae exp (-5) a:|
1 xe 2 (1)(3)(5)...
(2k — 1)
Q(x) = c= t= le 2 a1) eR
p( ) )| xk n (3B.18)
3 (3B.21)
jouO Red) fe BENG a
=p OO ey : (1+ 5) exp(—t)dt
or
...(2n-1
Rn = fey ee |(x) (3B.22)
152 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
where
pel
D(x) =i ( + =) ; exp(—t)dt (3B.23)
0 x
=
i: (:+- =)
ary} exp(—t)dt < i exp(—t)dt = 1
~
0 x 0
Let
l ¥2 n
(1)(3)(S)...
2k — 1)
Qn(x) = exp (-5) fne NN ee ener (3B.24)
J 2X k=1
be the n'” order approximation to Q(x), then the percentage error is given by
exp (-5) Rn |
Q(x)
— Qn(x)
% error = |—————_—
OG) _| = SOs (3B.25 )
exp ae
% error < J2nx
(-5) OO
Q(x) x2(n+1)
rs (3B.26)
Example 3B.1
Assume that x = 3. Calculate a simple bound on the percentage error when Q(x) is approxi-
mated by the first n terms in the asymptotic series, for n = 2, 3. Repeat for x = 5 and n = 2, 3,
and 4.
(a) For x =3
1)3)(5
ned Ghtor Se oa eaa
36 (12 )
(1)(3)(5)(7)
n = 3, % error < = 0.018 = 1.8%
38(1- 4)
Sec. 3B.2 Series Approximation for Q(x) 153
(b) For x = 5
DDO
n= 2, % error’< eine:
56(1 seei)
(1)(3)(S)7)
n = 3, % error < = 0.0003 = 0.03%
s8(1- 4)
n= 4, % error<
(DAE) = 0.0001 = 0.01%
Hel = 4)
Appendix 3C
Gram-Schmidt Orthogonalization Procedure
The Gram-Schmidt orthogonalization procedure [15] is a useful tool in constructing a suit-
able coordinate system to define a decision space with the least possible coordinates or
equivalently, with the smallest dimensionality. Given a set S = {so(t), s1(t),--.,8u—1()}
of M signals, we wish to find a corresponding set of orthogonal signals, such that all signals
in the original set S can be represented as a linear combination of the orthogonal signals,
such that the number of the orthogonal signals is minimized. The procedure is very simple,
and it consists of choosing the first orthogonal signal f(t) as the first signal, that is
> Bi (t)si(t)dt
Bo(t) = s1(t) — (ene |Bilt) (3C.2)
Jo*°
By@at
The third orthogonal signal 63(t) is again that component of s2(t) which is linearly indepen-
dent of £;(t) and B2(f), that is
Ts d
B3(t) = so(t) — Ga
Ts 2 Bi(t)
(t)dt
: Jo Bi (C.3)
Jo* B2(t)s2(t)dt
SE
9 By(t)dt
In general, we have
— ( Jo’ Bi Osx-1at
Belt) = se-1) — )|| > —— |8) (3C.4)
j=l Jo’ By (t)dt
for allk =2,..., M. To get these signals normalized, we simply define [for B;(t) 4 0]
Bi(t) :
(O=——— t= 1M GC5)
Jo’ Beat
to obtain an orthonormal (orthogonal and normalized) set of signals. The same procedure is
applicable to vectors by noting the 1s si(t)sj(t)dt =s!;.
Given the orthonormal set of signals (or vectors), each signal s;(t) (or vector s;) can
then be written as a linear combination of the orthonormal functions (or vectors) as
N
si(t) =) sijj(2) (3C.6)
j=l
for 0 <t < T, and s;(t) = 0 elsewhere. Here
Ts
Sij= [ Sj (t)pj(t)dt (3C.7)
App. 3C Gram-Schmidt Orthogonalization Procedure 155
Note that the number of orthonormal elements N is always less than or equal to the number
of original signals M. If the set S consists of linearly independent signals, then M = N
and the number of orthonormal elements is identical to the number of signals. Once the
set of orthonormal functions {¢;(t)} are identified, each transmitted waveform given in
(3C.6) is completely determined by the vector transpose s! = (Filo HApc ooo Gon) or all =
0, 1,..., M — 1. In vector notation, each N-tuple denotes the vector
where the inner product oo j = 4ij and where @ ; denotes the unit vector along the j'*-axis,
j =1,2,..., N. In this sense, we can visualize the M vectors {s;} as defining M points
in the N-dimensional Euclidean space, the signal space. The coefficients s;; in (3C.7) are
equivalently found from the vector s; by evaluating the projection s;; = s! o ; for alli and j.
Example 3C.1
Consider the following three vectors:
Then
Bi =si = (1,1,1)
T
piPore
=s' |(SiS
Pre Jaf
ee=c.0.0 oe(1,1,
1)ae
=3(1,2 2,1)
eras - (Bi)ot(25)0
T T
I
= (07 1,0) = 3 (1 LL) alae. 1) = (0, 0, 0)
Note that B, can not be in the new set since by definition, the elements of a basis are a set of
nonzero vectors. Check orthogonality property of the new vectors:
pip, =*-2**
L—2+1
=o
Normalizing B, and B,, we have
fe
TM shied 1)
Cig a”
4i
Re Js
$2 |B>| Gen a
Expressing the original signals in terms of the new orthonormal vectors, we have
156 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
ik T ae? if 1 2
So = (J3, 0), s; = ( ’ Ve Ss = ( a aes )
2 v3 A/6 NS ans
We have therefore defined a new observation space @, and @, such that the signals can be
represented in two dimensions. Note that in terms of the standard base o, = (50:0); o3 =
(0, 1, 0), and o; = (0, 0, 1), the signals were represented using three coordinates even though
they lied in two dimensions (that is, a plane). As can be seen from this example, choosing
the appropriate base can reduce the dimensionality of the observation space and could greatly
simplify the detection problem and reduce the receiver’s operations.
Appendix 3D
Connecting the Linear Vector Space to the Physical Waveform Space
The relationship between signals and vectors will be needed; this suggests the possibility
of a geometrical representation of signals and an application of analytic geometry. Consider
an N-dimensional vector space with unit vectors @,, @>,..., @y along the coordinate axis.
These vectors are called the basis vectors and form a complete set. The set of N basis
vectors is not unique. One can choose any N independent vectors; they may or may not
be mutually orthogonal. In our work, we shall assume the set of basis vectors to be mutually
orthogonal and each normalized to a unit length. Such a set is called an orthonormal set.
Once the N basis vectors are chosen, any vector r in this space can be specified by
N numbers r? = (r},72,..., rN), which represent the magnitude of components of r along
the N basis vectors, respectively. The vector r can be expressed as
N
ee aCe (3D.1)
n=
where, for normalized basis vectors, the projection of r onto @,, is the scalar product
Lage t @,
reat: (3D.2)
for all n. The vector r can be represented geometrically by a point (71, r2,...,7N) in the
coordinate system formed by the basis vectors.
We have an analogous situation for signals which evolve with the passage of time.
Analogous to the base vectors, suppose we have N waveforms ¢$1(t), $2(t),..., @n(t)
which form an orthonormal set of basis signals, N may be finite or infinite. Then , an ar-
bitrary signal r(t) in this N-dimensional Euclidean space can be expressed as
N
r(t}=) raga@), O<t<T (3D.3)
n=l
where, for orthonormal basis signals, the projections of r(t) onto the n!” basis signal is
7
tar i r(t)dn(t)dt (3D.4)
0
for all n. The set of basis signals {f,(t)} must form a complete set of signals for the space.
In practice, the basis signals are usually generated using a local oscillator to generate the
inphase and quadrature components 2/7 cos wct and /2/T sin wet.
Once the basic signals {#,(t)} are specified, one can also represent the signal r(t)
by an N-tuple (r1,7r2,..., 7). Alternatively, we may represent this signal geometrically
by a point (r1,7r2,...,7N) in an N-dimensional space. We can now associate a vector r
with signal r(t). Note that the basis signal $)(t) is represented by the corresponding basis
vector @, and ¢2(t) is represented by @, and so on. The vector r is called the projection
of r(t) onto the vector space. In the applications that follow, the linear vector space needed
by the optimum vector receiver is created from the physically observable waveform r(t) =
s(t) +n(t) by projecting the waveform onto that space. Such projections are equivalent to
filtering and sampling the received waveform, namely, the operation of cross-correlation.
158 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
The scalar product. In a certain signal space, let s(t) and r(t) be two signals
represented by their projection vectors s? = (51, 52,..., SN) and r? =(r,1r2,..., rn). If
{dn (t)} are the orthonormal basis signals, then
N
s(t) =) )snGn(t) (3D.5a)
n=1
Hence,
iD N
since
[‘ Pn(t)bm(t)dt dee
= (i a
eee (3D.7)
However, the right-hand side of the above equation is by definition the scalar product or
inner product of vectors r and s
if
r’s= i}r(t)s(t)dt (3D.8)
0
We conclude that the integral of the product of two signals is equal to the scalar product of
the corresponding vectors.
E=s's=|s/? (3D.10)
where |s|? is the square of the length of the vector s and |s| is called the norm of s. Hence,
the signal energy is given by the square of the length of the corresponding vector.
Example 3D.1
Consider the two-dimensional signal vectors depicted in Fig. 3D.1. For Binary On-Off Keying
(OOK), we use the inphase component generated by a local oscillator to write
so = 0 > so(t) =0
$4
So 34 So S} So
(a) On-off keying (OOK) (b) Binary frequency shift keying (BFSK) (c) Binary phase shift keying (BPSK
for 0 < t < T. For Binary Phase Shift Keying (BPSK), we have
One of the most powerful analytical tools used in the development of communication theory
is that of linear vector spaces, needed in the representation of signals (codes). In vector com-
munications, the Pairwise Probability of Error (PPE) depends on the Euclidean distance (a
true mathematical distance measure induced by the inner product or cross-correlation) and
the signal space with this distance forms a Hilbert space. Analysis and synthesis of vector
communication system techniques can be reduced, in many practical cases, to consideration
of the geometrical features of this Hilbert space.
A basic building block in the development of M-ary vector communications is the
so-called PPE. It also serves as the basic building block in the development of the Union
bound, which is used to characterize the symbol error probability (system performance) and
is composed of M — | terms, each of which is a PPE. This bound is used to develop the
ensemble symbol error probability arising in the Shannon theory, which demonstrates that a
code exists that exhibits a zero asymptotic error probability in the presence of noise, under
some conditions.
The pairwise error event is defined as
where dj, denotes the distance between the vectors. Defining the unit vector ej,A (s;—
sx)/|S; — S,| and introducing the random variable z(u) = ne jk» One can write the pairwise
error event as
The probability of this event is just one minus the value of the probability distribution
function of the random variable z(u) evaluated at d jk/2, that is
ae
PAIRWISE ERROR
DECISION BOUNDARY
«© d ik 2
dik‘ /4No
= exp [-22\o Jaz = a| |<@
I J™No V ‘0
Gaussian random variable with variance 0* = No/2. Using this fact in (3E.4), we obtain
1 CO
exact error probability performance of all binary communication systems transmitting equal
energy, equal probable signals, (2) the symbol error probability performance of an M-ary
system is Union bounded by a linear combination of PPE’s, that is
M-1
P(E/my)< 2 Pris}. (3E.7)
ink
This suggests that the performance of an M-ary vector communication system is bounded
by a linear combination of the performances of M — | binary systems employing signal
vectors $; and Sx, that is
M-1M-1
P(E) < DD) Ply) Pa(Sj, St) (3E.8)
k=0 j=0
i#k
Finally, there is no loss in generality in allowing s; and sx to be any two arbitrary
encoded data sequences (vectors) so that (3E.6) can be used in evaluating the performance of
coding-decoding algorithms of the sequential type, for example, decoding of convolutional
codes using the Viterbi algorithm. This CORY, building block will be applied in Chapters 4,
12, and 13. If one identifies the vectors Ss) = (Sm, ---,Smn) With the coded message m of N
symbols, then the PPE that message m ’ (consisting of another coded message of N symbols)
represented by si SeNS ras! Tone 6 ae Sen ES USE
d2,
Pim->m)=0( 4 (3E.9)
V2No
where Ci A |Sm —Sm’|’ /|2 denotes the squared Euclidean distance between the two vectors
Sm and S,,’. This result will be used numerous times in later chapters.
PROBLEMS
3.1 An observation x is defined as follows:
Ah: x=mb+n
lava eNO
CLG b and n are two independent, zero-mean Gaussian random variables with variances op
and Cp respectively and m is a constant. Assuming equiprobable hypotheses,
(a) Find and sketch the optimum (MAP) receiver.
(b) Calculate the minimum probability of error for the receiver found in (a).
(c) Repeat (a) and (b) when the random variable b has mean value my.
(d) Using (c), find the error probability when 0, — 0 and discuss your result.
3.2 An observation x is characterized under two hypotheses according to the following pdf’s:
Hie!) ie cai
0,
i
we ())
The probabilities of the two hypotheses Ho and H; are 5/8 and 3/8, respectively.
Problems 163
(a) Find the minimum error probability test to choose between Ho and A).
(b) Find the minimum error probability that results from the test in (a).
3.3 Consider the following scalar communication problem:
where so = 1, s} = —1 and
1 1 : o| ae
f—a)- fe 1 . o| (n +a)?
ee
fr(n) = 5
g 256 Digg 256
SSS ee <n<c7
fn(n) _
m(1 +n?)
(a) Determine the likelihood ratio test that specifies the maximum-likelihood receiver.
(b) Evaluate the false alarm probability Pr4 = Pr{H is chosen/ Hp is chosen} and the prob-
ability of detection Pp = Pr{H, is chosen/H is chosen} if the threshold voltage is set to
0.5 volts. Also calculate the probability of error.
3.6 (a) Using just a single observation, what is the optimum receiver to choose between the hy-
potheses
“i . 5 r D)
Ho: the sample is zero-mean Gaussian, with variance 09
r é i : D) ) 9
H,: the sample is zero-mean Gaussian, with variance oy (oy > 99)
LA at
Ate Sa len
Hy: r=-l+n”
where n is a zero-mean Gaussian random variable with variance 0”. Assume equal a priori
164 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
probabilities for each hypothesis. Show that the test statistic may be chosen to be
a ri
=|-—
i=l
Jaye (PS
Joh =a sew
2 2
p po +a ap
i felo/ ty =F exp (= Fo2 )0(%). p=0
where q@ and o are known and /(x) is the zero-order modified Bessel function of the first kind.
(a) Specify the likelihood ratio test if
]
Prob{ Ho} = Prob{ H,} = 5
3.10 Consider the following binary decision problem representing the detection of a known signal s
1. See M. S. Roden,““Digital and Data Communication Systems,’ Englewood Cliffs, N.J.: Prentice Hall, 1982.
Problems 165
Table P3.9
ce ee a 8 oe a
B
a 1 2 3 4 5 6 7 8 9 10
A i A OE ea a 0
1 0.875 0.268 0.044 0.003 0 0 0 Oe 0 0
2 ONS" = 01603, 970213" (501034. 0,002 0 0 0 0 0
3) 0991-05885 0595 —0:195° 0.031 ~~0002 0 0 0 0
4 0.999 ©0986 0:875 0.575 «420.186 ~=—:0.029._-—«- -:0002 0 0 0
5 0.999 0.999 0.984 0.865 0.564 0.180 0.027 0.002 0 0
6 0.999 0.999 0.999 0.984 0.860 0.543 0.176 0.027 0.002 0
TE FOSO9 E9999 50,999 920.999. 0.981 <0:855.. 0523 0.173 0,026... 0.002
8 0.999 0.999 0.999 0.999 0.999 0.980 0.850 0.511 0.170 0.026
9 0.999 0.999 0.999 0.999 0.999 0.999 0.977 0.845 0.497 0.167
10 0.999 0.999 0.999 0.999 0.999 0.999 0.999 0.975. 0.840 0.482
The pdf which characterizes the additive random noise voltage is the generalized pdf
let 0 Vesa)
Higa — sce
Find the decision region which minimizes the error probability if Ho and Hj are equally
probable.
166 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
r=s+n
Figure P3.15
as fr(e/ A)
A,(p)
Fr(e/ Ho)
If we replace the observation p in A,(p) by the random variable r, then A,(r) is itself a random
variable with special properties. Prove:
(a) E{A,(r)"/Hi} = E{Ap*!(r)/Ho}
(b) E{A,(r)/Ho} = 1
(c) E{A,(r)/Hi} — E{A,(r)/Ho} = Var{A;(r)/
Ho}
3.15 The random variable n in Fig. P3.15a is Gaussian, with zero-mean. If one of two equally
likely messages is transmitted, using the signals of Fig. P3.15b, an optimum receiver yields
PEs O00K
(a) Compute the variance o? of n.
Problems 167
-4 0 4
(a)
Figure P3.16
(b) Erasures are sometimes employed in receivers to allow them to avoid a decision when it
seems that a decision might contribute an error. Consider the decision rule of Fig. P3.15c:
Compute d such that the erasure probability is 10-4. Compute all necessary probabilities
P(r;/m) to reduce the continuous channel to a discrete memoryless channel (DMC) with
2 inputs (M = 2) and 3 outputs (J = 3).
(c) In some applications, it is better to quantify the quality of the received symbol. This is typ-
ically accomplished by measuring the confidence level of the observed voltage. Consider
the decision rule depicted in Fig. P3.15d. In this case, the receiver distinguishes between
a soft “1” (denoted by 1—) and a strong “1” (denoted by 1+). The new decision rule is as
follows:
Compute d such that the probability of any soft symbol is 10~*. Compute all necessary
probabilities P(r; /m ;) to reduce the continuous channel to a discrete memoryless channel
(DMC) with 2 inputs (M = 2) and 4 outputs (J = 4).
3.16 The random variable n in Fig. P3.15a is Gaussian, with zero-mean. If one of two equally
likely messages is transmitted, using the signals of Fig. P3.15b, an optimum receiver yields
P(E) = 0001.
(a) What is the minimum attainable probability of error when the channel of Fig. P3.15a is
used with the three equilikely signals of Fig. P3.16a?
(b) How about with the four equilikely signals of Fig. P3.16b?
(c) How do the answers of (b) change if it is known that the noise has mean m.
3.17 A communication system employs two channels to transmit a voltage s;, i=0,1, as shown in
Fig. P3.17. The voltages sg and s; correspond, respectively, to the messages mo and m;. Two
received voltages r; and rz are available to the decision device, that is, the receiver makes its
decisions based on a joint observation of r; and r2. Assume that the additive noises n; and n2
are zero-mean Gaussian random variables with variances o; and oF, respectively. Also, assume
that s;, 11, and np are statistically independent. Letting so = VE with probability P(mg) and
s} = —VE with probability P(m)),
(a) Determine the structure of the optimum receiver that results in the minimum probability of
error.
168 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
ue)
DECISION
TRANSMITTER DEVICE
Figure P3.17
s;° a
Figure P3.19
(b) Calculate the average probability of error for the case P (mo) = P(m,) and o? =o.
3.18 (a) Sketch the optimum vector receiver for each of the signal sets illustrated in Fig. 3D.1.
(b) Sketch the decision regions and boundaries.
(c) Compute the minimum probability of error for each case.
3.19 (a) Given two equilikely messages, mo and mj, to be transmitted (using the signals illustrated
in Fig. P3.19) over a channel that adds white Gaussian noise, find an expression for the
minimum error probability.
(b) What can one conclude regarding the energy of the two signals from the result found
in (a)?
3.20 Consider the following N-dimensional decision problem: the received vector r = s; + n where
= (ifn EN)
Assume that the noise vector is Gaussian with mean 0 and covariance matrix o?Z, where T is
the N x N identity matrix.
(a) Design a decision rule to choose between the four equally likely hypotheses.
(b) For the case N = 2, assume the received vector is r = (4, 0)’. Which hypothesis would the
receiver choose?
3.21 The two equally likely hypotheses are
1 xe xh
Hos. Fis %2/ Ho) = XD ae
2
Ne OC eG
20 or 209 2092
1 x2 x2 x? x?
Hictaas (Xiseos exp (i ee aeexe ee
41 0001 Ss 207 209 P 20 20; E
—COl< Xoo
Problems
169
Figure P3.23
So S1
(a)
Figure P3.24
Venere
d
Ss Lea S
7
ie Go
N ” wo N N.
N
Figure P3.25
Figure P3.26
disturbed by additive white Gaussian noise, an optimum receiver will achieve P(E) = q for
equilikely messages. Assuming that the seven vectors indicated by x’s in Fig. P3.24b are used
as signals over the same channel, compute the probability of error achieved by an optimum
receiver in terms of g, 0, and /, assuming equilikely messages.
3.25 Consider the PCM (pulse code modulation) signaling system where each PCM word has a
length of 3 bits, that is, there are 27 = 8 possible symbols. The signal vectors are placed on
the vertices of a “3-dimensional hypercube” as shown in Fig. P3.25. Assume all signals are
equilikely and that the noise is an AWGN with PSD of No/2 watts/Hz. Find an expression for
the probability of symbol error.
3.26 Consider the 6-QAM constellation set depicted in Fig. P3.26. Assume all signal vectors are
equilikely to be transmitted and that the channel is disturbed by AWGN with PSD No/2.
(a) Find the boundaries for the decision regions.
(b) Find the probability of symbol error.
Problems 171
Figure P3.27
135° 45°
AMPLITUDE = y2-
180° 0° ABSOLUTE
Sil Consider the 4 x 4 QAM signal constellation depicted in Fig. P3.27. Assume an AWGN chan-
nel and equilikely signals.
(a) Draw the optimum decision boundaries and illustrate the decision regions.
(b) Using an appropriate basis, write an expression for each of the 16 vector signals.
(c) Compute the average energy for this signal constellation.
(d) Draw the optimum detector.
(e) Write an expression for the probability of symbol error.
3.28 Consider the V.29 signal constellation depicted in Fig. P3.28, which transmits 3 bits/symbol.
Assume an AWGN channel and equilikely signals.
(a) Draw the optimum decision boundaries and illustrate the decision regions.
(b) Using an appropriate basis, write an expression for each of the eight vector signals.
(c) Compute the average energy for this signal constellation.
(d) Draw the optimum detector.
(e) Write an expression for the probability of symbol error.
3.29 Consider the signal constellation shown in Fig. P3.29. This signaling scheme is referred to as
8-QAM (quadrature amplitude modulation). Note that the energy of each signal is not equal.
(a) Find the average energy of the signaling set. Assume all signals are equilikely.
(b) Find an upper bound for the probability of symbol error.
3.30 Consider a 3-dimensional “biorthogonal” signaling scheme as shown in Fig. P3.30. Assume
all these signals are equilikely and that they are sent over an AWGN channel with PSD No/2.
Using the union bound, find an upper bound for the probability of error as a function of E;/No
where £, denotes signal energy.
172 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
Figure P3.29
Figure P3.30
3.31 Consider the V.29 signal constellation depicted in Fig. P3.31, which transmits 4 bits/symbol.
Assume an AWGN channel and equilikely signals.
(a) Draw the optimum decision boundaries and illustrate the decision regions.
(b) Using an appropriate basis, write an expression for each of the 16 vector signals.
(c) Compute the average energy for this signal constellation.
(d) Draw the optimum detector.
(e) Write an expression for the probability of symbol error.
3.32 Consider a 32-QAM signal set as shown in Fig. P3.32. Assume all signals are equilikeiy and
are transmitted over an AWGN channel with PSD No/2.
(a) Calculate the “average” energy of this set.
(b) Find an upper bound for the probability of symbol error.
3.33 Consider the problem in which you must decide which of two hypotheses is true, based on the
observation of x:
Hy: x=so+no
Ay Ve sq 27)
where sg and s; are known constants and no, n; are zero-mean independent Gaussian random
variables with respective variances OG and at:
Problems 173
Figure P3.31
Figure P3.32
(a) Derive the decision rule that minimizes the probability of error (assuming equilikely prob-
abilities) and write expressions to characterize P(error/ Ho) and P(error/H}).
(b) Suppose that the problem is reformulated as follows: decide which of two hypotheses is
true based on the observation of the vector x where
Ho: x=Sot+n
A: x=s,;4+n
where so = (so, 0)7, s; = (0, 51)’ and n is a zero-mean Gaussian random vector with
covariance matrix
oO2
bal cine)
0
Kn=( °
174 Scalar and Vector Communications Over the Memoryless Channel Chap. 3
if W1/2- =1/2
r= (-172 1 =")
=i) ltt
(a) Let the random variable y be formed by the inner product y = h’ x, where h is constrained
to have unity magnitude. Find the h vector which maximizes the mean-squared value of y.
(b) Determine the mean and variance of the random variable y with h chosen as in (a).
(c) Let w = Gx. Find the matrix G such that the resulting vector w has a covariance matrix
equal to the identity matrix (that is, perform a whitening transformation on x).
(d) Express x as
n
x= yy J riwiei
i=l
where the w;’s are zero-mean, uncorrelated random variables with unit variance (that is,
find the A;’s and the e;’s).
3.35 Let x be a correlation matrix with zero determinant. Prove that the matrix ee has eigenval-
ues zero or one where €,, is the matrix whose columns are the eigenvectors corresponding to
nonzero eigenvalues of 7Rx.
3.36 Based on the observation of the random vector x, determine the detector that will minimize the
probability of error between the two equilikely hypotheses:
Ho: x=(1,1,0,0)7
+n
Hy x=(0,031,-1)
ton
where the covariance matrix of the Gaussian zero-mean noise vector is given by
al 0 0
a OFeEO
Kn =
ORO leo
On0 St l
Hint: Is the covariance matrix singular? Is perfect decisions possible? How about preprocessing
the observation x with a matrix 7 such that Hx has a nonsingular covariance matrix?
O55, You are asked to design an optimum decision rule to decide between the following two hy-
potheses:
(d) Assuming that the noise s(n) is no longer white, but has PSD
Su(f) =
ga ee) with
0 < |p| < 1
1 — 2ocos(27f) + p?
derive the optimum decision rule. Check to see that when p = 0, the decision rule reduces
to that in (a).
(e) Bound the probability of error of the receiver in (d).
(f) Derive an exact expression of the receiver in (d). Check to see that the expression reduces
to that in (c) when p = 0.
3.38 A decisioning system must decide between the hypothesis Hp and H; where
1 2
Ho: frlp)= = exp(—>)
apI WES Shes)
Joie ok Setheee
fr(0) | 0, otherwise
The a priori probability of Ho is 3/4 and H, is 1/4. The cost of decisioning is
Co =0 Ci=1
Co=1 Cy =90
Determine and describe the optimal Bayes decision procedure for this problem.
3.39 Consider the hypotheses
ley Ra
j(S\=exp(—S), S20
Ho
pe y
A
Find the numerical value for y for the Bayes test of (c).
(e) Compute the risk for the Bayes test in (c).
(f) Compute the threshold for a Neyman-Pearson test with Pr, less than or equal to ime
Find Pp for this threshold.
(g) Reducing the Neyman-Pearson test of (f) to
3.40 A decisioning system must decide between two hypotheses with pdf’s
Choosing Hy when H, is true costs three times as much as choosing H; when H, is true.
Correct choices cost nothing. Formulate a decision rule using the min-max criterion.
3.42 Show that the vector
M-1
a= P(mj)s;
i=0
minimizes the required mean energy
_ Mal
E’ = )° P(m)|s; — al?
i=0
without altering the message error probability.
REFERENCES
1. Neyman, J. and E. Pearson, “The Testing of Statistical Hypotheses in Relation to Probability A
Priori,’ Proc. Comb. Phil Soc., vol. 29, (4), pp. 492-510 (1933).
2. Neyman, J. and E. Pearson, “On The Problem of the Most Efficient Tests of Statistical Hypothe-
sis,” Phil. Trans. Roy Soc. (London), Ser. A, 231(9), pp. 289-337 (1933).
3. Wald, A., Statistical Decision Functions, New York: John Wiley and Sons, 1950.
4. Bayes, T., “An Essay Toward Solving a Problem in the Doctrine of Chances,” Phil. Trans. Roy.
Soc. (London) , pp. 370-418 (1763). Reprinted in Biometrika 45, pp. 293-315 (1958).
5. Papoulis, A., Probability, Random Variables and Stochastic Processes, New York: McGraw-Hill,
1965.
6. Larson, H. J. and B. O. Shubert, Probabilistic Models in Engineering Sciences, Vol. I: Random
Variables and Stochastic Processes, New York: John Wiley and Sons, 1979.
7. Weber, C. L., Elements of Detection and Signal Design, New York: Springer-Verlag, 1987.
8. Strang, G., Linear Algebra and Its Applications, 3rd ed., San Diego: Harcourt Brace Jovanovich,
1988.
9. Friedberg, S. H., A. J. Insel and L. E. Spence, Linear Algebra, Englewood Cliffs, N.J.: Prentice-
Hall, 1979.
10. Wozencraft, J. M. and I. M. Jacobs, Principles of Communication Engineering, New York: John
Wiley and Sons, 1965.
11. Blackwell, D. and M. A. Girshick, Theory of Ganies and Statistical Decisions, New York: Dover
Publications, 1954.
References 177
12: Van Trees, H. L., Detection, Estimation and Modulation Theory, Part 1, New York: John Wiley
and Sons, 1968.
13% Helstrom, C. W., Statistical Theory of Signal Detection, 2nd ed., Oxford: Pergamon Press, 1968.
14. Melsa, J. L. and D. L. Cohn, Decision and Estimation Theory, New York: McGraw-Hill, 1978.
ID: Friedberg, S. H., A. J. Insel and L. E. Spence, Linear Algebra, Englewood Cliffs, N.J.: Prentice-
Hall, 1979.
16. Craig, J. W., “A New, Simple and Exact Result for Calculating the Probability of Error for Two-
Dimensional Signal Constellations, ” IEEE MILCOM’91 Conference Record, Boston, MA, pp.
25.5.1-25.5.5.
Chapter 4
Coherent Communications
with Waveforms
In the previous chapter, we investigated the problem of scalar and vector communications
and derived decision rules that minimize the probability of message error. Other perfor-
mance criteria were also considered to illustrate alternative parameter optimization. In the
presence of white Gaussian noise, the decision rules resulted in optimum receiver structures
that correlate the received vector with all possible transmitted vectors (with the addition of
the appropriate bias), and then decide on the message with the maximum correlation. In this
chapter, we consider the problem of waveform communications in which each message is
transmitted using a distinct voltage function. We start with a baseband signal model which
assumes perfect carrier phase coherency. We will first derive the optimum receiver structures
in the presence of white Gaussian noise and then evaluate the resulting performance for var-
ious sets of transmitted waveforms. The approach is then generalized to handle nonwhite
Gaussian noise, using various techniques, namely, the sampling approach, the Karhunen-
Loéve (K-L) expansion, a direct derivation using a sufficient statistic, and a “whitening”
operation on the observed waveform. Finally, inphase (I) and quadrature (Q) modulation and
demodulation are introduced to establish the link between the derived receiver structures and
communication signals. We will then relax our assumptions regarding the carrier phase co-
herency and the effects of imperfect carrier and symbol synchronization are investigated for
the simple model of a slow time-varying phase error process. Finally, the deleterious effects
of data asymmetry and suboptimum detection on system performance are treated.
Our approach to waveform receiver design [1] is to first approximate the waveforms
using finite samples to provide a quick but nevertheless correct derivation of the receiver
structures. The more rigorous approach, which does not require the finite sample approx-
imation, uses the K-L expansion. It will be shown that the results obtained are the same;
however, it is felt that time sampling is more appealing from a practical viewpoint. In the
current case, the transmitter maps each message m into a distinct voltage waveform s(t) for
transmission as shown in Fig. 4.1. It is assumed that each transmitted message lasts for T
seconds and that the system is perfectly synchronized in the sense that the receiver has full
knowledge of the start and end epoch of each signal. The receiver observes a sample func-
tion p(t), 0 <t <T, of the random process r(t), and maps the observation into one of the
messages in the transmitted set. Formulating the detection problem as a hypothesis testing
problem, then
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 179
where sinc(x)A (sin x)/(mx) and t, 4 k/2B. Moreover, we shall have occasion not only
to represent time functions as samples but also reconvert from sums of samples to time
integrals. Thus, we note that
represents the energy in the sample function n(t). Now the joint probability density function
of the vector n? = (n(t1), n(t2), ---,n(kT)) created from sampling a T-second strip of n(t)
is given by
as the 2BT samples are uncorrelated and Gaussian. Here, C represents a normalization con-
stant. Using (4.4) to convert from sums of samples to a time integral, we readily approximate
180 Coherent Communications with Waveforms Chap. 4
(4.5) via
ie n*(t)dt
|- i (4.6)
Fni(ty),--:n(kT)(M) ~ C exp
for the k’” time interval 7; = [(k — 1)T, kT]. Without loss in generality, we set k = 1 and
consider the time interval [0, 7]. Hence
_ fo Pat
2(t)dt
Fn(O),---.n(T)(M) = C exp N (4.7)
0
In the binary case, the receiver needs to decide between messages mg and m,. Using MAP
decision theory, the receiver sets m(p) = mo iff
P(mo/r
= p) > P(m/r=p) (4.9)
or equivalently iff
P(mo) fr(e/mo) s P(m1) fr(e/m1)
(4.10)
fr(p) fr(p)
Using (4.8) in (4.10) and assuming equilikely messages, the decision rule becomes: set
m(p(t)) = mo iff
ai T (p(t) — so(t))“dt
2
5 exp | 6T &Ot) =
— 510)
2
s,(t))*dt rey
No No
Taking natural logarithm and simplifying, the rule becomes: set m(o(t)) = mo iff
r T
i (p(t) — so(t))?dt < if(o(t) — si(t))?dt (4.12)
0
or equivalently, choose the message corresponding to the waveform s;(t) that is “closest” to
p(t) where distance is defined through
DECISION
DECISION
S ,;(t)
Figure 4.3 Alternative implementation of the binary waveform receiver (equal energy, equiprobable signals)
L 1
aga [ P(t) (so(t) — s1(t)) dt > meee (4.16)
0
For equal energy signals, Eg = FE) and the test reduces to
MW mo
ag= | pte (sot) —s)) 2 0
0 my
(4.17)
Hence, the receiver correlates p(t) with the difference signal so(t) — s,(t) over (O, T) and
compares the outcome to zero as shown in Fig. 4.2. If the correlation is positive, the receiver
selects message m,. Otherwise, it selects mo. Alternatively, the receiver can be implemented
using two correlators as depicted in Fig. 4.3 and selects the signal corresponding to maxi-
mum correlation. Note that the decision rule operates linearly on the received signal and that
no other rule (linear or nonlinear) can produce a smaller error probability. The zero thresh-
old is due to the assumption made on the binary signal. If the signals are not equal energy
or equally probable, the threshold will be biased away from zero.
To gain further insight into the structure of the optimum waveform receiver, let’s consider
the problem of detecting the presence of a known signal s(t) using the block diagram shown
182 Coherent Communications with Waveforms Chap. 4
No of
0) => / h? (Add (4.23)
—0Oo
or alternatively
with equality if and only if h(t) =ch,(t) (c being any constant), (4.25) can be upper
bounded by
SNR
PO
—
cal aie eo
nae de
: re) [Wada ae
or since E = PT = (e s*(t)dt, then
2 ® , 2E
SNR < — / s°(T —A)ddA = — =2(PG)(CNR) (4.28)
ING dies No
where PG denotes the processing gain of the matched filter and is given by PG=WT and
CNR = P/(NoW) where P is the signal power and W is the front-end one-sided bandwidth
of the receiver. Here, CNR denotes the signal-to-noise ratio in the front-end bandwidth W
where as E/No = PT /No denotes the signal-to-noise ratio in the data rate bandwidth 1/7.
The ratio of these two bandwidths is the processing gain PG of the matched filter. Exact
equality in (4.28) occurs only when (c any constant)
i CSC) (4.29)
Therefore, SNR is maximized when the filter h(t) is matched to the incoming signal s(t),
thus the terminology matched filter. Of course, such a filter with a finite impulse response is
not physically realizable. However, as a practical matter, the filter becomes both physically
realizable and essentially optimum whenever almost all of the energy in the signal s(t) is
located within 0 < t < T and —B < f < B where B is the one-sided baseband bandwidth
of the receiver and is related to the IF bandwidth of the receiver through W = 2B. When
implementing a matched filter, the output y(t) can be expressed as
which is exactly the correlation between the incoming waveform r(t) and the signal s(t).
From Fig. 4.3, the optimum binary waveform receiver can alternatively be implemented us-
ing matched filters as shown in Fig. 4.5. In this case, two filters matched to the possible
incoming waveforms are used to filter the observed signal and their outputs sampled appro-
priately at t = T to provide maximum SNR. The resulting voltages are fed into a comparator
for selection of maximum value. Note that the signal component of the matched filter output
is given by
(oe)
Recalling that the time-autocorrelation function [4] of any real function g(r) is given by
Select
message Decision
corresponding
to largest
cross-correlation
Figure 4.5 Matched filter implementation of the binary waveform receiver of Figure 4.3
Hence
Ely@l=k =) (4.34)
In the case of M-ary signals, the optimum receiver sets m(p(t)) = m; iff
Pera ae T OD) D
t) — sj(t))“dt
No
é (4.36)
canek {Pemex eo rare
9)
t) — sg(t))“dt
No
Simplifying and taking the logarithm, the receiver reduces to maximizing
i Me fish
In P(m;) — — / s?(t)dt + p(t)s;(t)dt (4.37)
No Jo No Jo
or equivalently, the optimum receiver sets m(o(t)) = m; iff
i T
qi ah p(t)sj(t)dt + B; es {| P(t)s(t)dt + a,| (4.38)
0 0
where the bias term By, is given by
N E
BA = In P(m,) — a (4.39)
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 185
Set m (p(t) = m,
iff
q;La = max{q,}
ieeo
where E;, is the signal energy given in (4.15). In the case of equally likely messages, the
bias reduces to — E;/2. In addition, if the signals had equal energy, the bias term can be ig-
nored. Note that the bias term is equivalent to that given in (3.60) when 0” = No/2. Hence,
the receiver should correlate the observed signal with all possible transmitted waveforms
(add the appropriate bias for unequal energy or unequal a priori probabilities) and select
that message corresponding to maximum correlation as shown in Fig. 4.6. This implemen-
tation requires M correlators and hence, the receiver complexity increases linearly with the
number of transmitted signals. If the signals were designed such that (see Appendix 3C)
N
HO=) sy)(0) (4.40)
= i)
then
; ig T N
[ ewsmar= [oY syojtoat
0 a
: (4.41)
N V1 N
=Yisj f pwmajnar =) sur
‘a1 0 a
186 Coherent Communications with Waveforms Chap. 4
SIGNAL
WEIGHTING
MATRIX
if m (p (t))
q; palets {q,}
SM-11 SM-12esM-IN |
Figure 4.7 Alternative implementation of the M-ary waveform receiver using N correlators
Hence, the correlation can be achieved by correlating p(t) with the N orthogonal functions
and then weighting the output with the signal coefficients to form the required signal cor-
relation. This processing is depicted in Fig. 4.7 and requires only N correlators for any
signal set. When the number of signals M is greater than the dimensionality N, the sec-
ond configuration requires fewer correlators and is more suitable for implementation. Since
the correlation operation can also be implemented using a filter and a sampler, the opti-
mum receiver can alternatively be implemented using either M or N matched filters (with
the appropriate weighting) as illustrated in Fig. 4.8. Note that when the signal waveforms
satisfy (4.40), we can map the waveform problem into an equivalent vector problem where
s! = (si1, Si2,+-+, Sin), i =0, 1,+++, M — 1 and use all the tools developed in Chapter 3.
In this section, the performances of the various coherent receivers are derived for both
binary and M-ary signaling. First, binary signals are considered to gain insight into the
problem [5], and then the results are generalized to the M-ary case. Equiprobable signals are
considered as they reflect most communication systems. But the results can also incorporate
unequiprobable signals with the appropriate weights.
S02 eeeSon
Figure 4.8 Implementation of the M-ary waveform receiver using matched filters
T
EtAg/mi= | si(t) (so(t) — s1(t)) dt (4.42)
and
T 1
Ka — a {(f n(t)(so(t) — su()ar (| n(u)(so(u) — sitw)du) |
0)
(4.43)
N L.
= ee ( (so(t) — s1()ar
2 0
Because of the equal energy assumption, ig sp(t)dt = Us sa(t)dt = E and (4.43) reduces
to
r
OKa — Nf {ze
— / suesottnae} (4.44)
0
If we denote
0 1 (Aq — Bqo)”
Pe) =f = Op d(Aq) (4.50)
Fee [ 2OK, 20Kg
E
P(E)=Q (\
Na ea vo) (4.52)
where E’/ No denotes the signal energy-to-noise ratio. Alternatively, (4.52) can be expressed
as
d
P(E)=@Q (=) (4.53)
where d;; denotes the distance between signals s;(¢) and s;(t) and is given by
ij
dj, A i (si (¢) <= sj(t))° dt
0 (4.54)
= Ej; + Ej —2/ EE jvij
For the two binary signals with equal energy, din = 2E(1 — yo) and since the most negative
value 19 can take is minus one, the corresponding P(E) is then
[2E
P(E ea (az) Antipodal (BPSK) signals (4.55)
)
which occurs when so(t) = —s}(t) and the pair is referred to as antipodal signals. Note that
any pair of antipodal signals will generate the same P(E) for a given energy E. Examples
of three pairs of antipodal signals are given in Fig. 4.9, namely: (a) pulse code modulation
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 189
So (t) n S
—_ = oo
VET
So (t)
VE/T
VE/T
T
t t
ali
- JER Vert
(b) MANCHESTER OR BI-o SIGNALS
Sq (t)
S(t)
/J2E/T J2E/T
+
a ~
Figure 4.9 Examples of antipodal signals: (a) Pulse code modulated (PCM) signals, (b)
Manchester or Bi-¢ signals, (c) Binary phase-shift-keyed (BPSK) signals
(PCM), (b) coded PCM typically referred to as Manchester or Bi-¢, and (c) binary phase-
shift-keyed (BPSK) signals. The various signals are given by
PCM:
For orthogonal signals, ie sy(t)so(t)dt = 0 (that is, yi9 = 0) and P(E) of (4.52)
reduces to
E :
P(E) =O (4 Orthogonal signals (4.59)
\ 0
which is a factor of two (or 3 dB) worse in E’/ No than the performance of antipodal signals.
Another signaling scheme is coherent frequency-shift-keying (FSK) where the messages are
transmitted using two distinct frequencies, that is
DB
Sit yaa) cp sin(@jt +0), O<t<T (4.60)
I2E 3
So(t) = ap eae) OST (4.61)
and wT, woT > 1. In this case, the correlation coefficient is given by
1 Ve
sin(@, — wo)T e sin 27 f\, oT
Y10 = =f si (t)so(t)dt = = 4.62
E Jo (@| — wo)T 2 fiol ( )
where the correlation is plotted in Fig. 4.10 as a function of the frequency separation f|.9 =
fi — fo. The minimum possible correlation is given by yj9 = —<4, resulting in
E 2
ECE) =O —j{1+— Coherent BFSK (4.63)
No 3
which is 2/(1 + ys) = 2.1 dB worse in SNR than antipodal signals. The corresponding fre-
quency separation is (f| — fo)T = f\,0T = 3/4. On the other hand, when f;,97 = 0.5, that
is , the frequency separation is equal to half the message rate, orthogonal binary signaling is
obtained and coherent FSK with this minimum frequency separation is sometimes referred
to as minimum shift keying (MSK).
In all cases so far, we have considered equiprobable, equal energy signals. For unequal
energy and unequiprobable signals, it can be shown (see Problem 4.10) that
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 191
P(E) = dio
(E) o(=) (4.65)
4.65
as expected. An example of a scheme with unequal powers is “On-Off” keying where so(t)
is sent by turning the power off and s;(t) is sent by turning the power on, that is
and so(t) = 0 for 0 <t < T. In this case, yj9 = 0 since so(t) = 0 and de =F Hence
Note that the maximum energy delivered is E, but the average energy is E/2 since both
signals are equally likely. Hence, (4.67) can be expressed in terms of the average energy
Eayg as
Eavg
P(E)=O ie (4.68)
the performances of binary and M-ary communication systems. The probability of message
(symbol) error, P;(E), can in general be expressed as
M-1
P,(E) = )) Ps(E/mi)P (mi) (4.69)
i=0
where P,(E/m;) is the probability of message or symbol error assuming that m; is transmit-
ted. In order to evaluate P,(E/m;), the outputs of the integrate-and-dump filters have to be
jointly characterized through the M-dimensional pdf. Recalling that
qk = [™ pidsaltyat (4.70)
where p(t) = s;(t) + n(t) when m; is transmitted, the output of the je? integrator becomes
reas eerie 7
YM-10 YM-11 °** YM-1M-1
Equation (4.74) assumes that Kgq is invertible. If Kg is singular, the dimensionality of the
observed vector can be reduced without any loss in generality and can result in singular
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 193
Note the order of integration where the last integral is over gj. Since P(C/m;) depends on
yji, J =0,1,...,M —1, the resulting probability of symbol error, P,(E), depends solely
on the symbol energy E,, the noise spectral level No, and the signal correlation matrix T'!
given by
Zi = (Gi — Gi)
/VSGi (4.79)
then (4.77) may be rewritten as
Co rzt,/2Es/NoU—v%1) Zita/2Es/No(l-yim-1)
Pe(C/m) = f if af fi(z/mj)dz (4.80)
—oo J—00 —0oO
where z! = (zo, z1, Z2,-*:,ZM—1) and f,(z/m;) = G(z; 0;T), independent of m;. Using a
similar procedure as in Chapter 3, a Union bound can be derived for waveform communica-
tions resulting in
M-1
ie
P(E/m) < o(5) (4.81)
h »» J2No
i#k
; a. NoEs
1. The covariance matrix Kg is given by Kg = 21.
194 Coherent Communications with Waveforms Chap. 4
a) Mo!
(4.83)
= ‘Ss (Avg. # of erroneous bits/m;)
i=0
The last equality is due to the fact that we are transmitting a “large” number of symbols and
as a result, all messages occur equally likely. Hence
1 M-l
»(E)
P,(E) Taipei
= —— Avg. g # of erroneous bits/m; /m) (4.84)
4.84
Now let n;, ; denote the event that message i is transmitted but message j is decided (that is,
vector $; is closest to observed vector p) and (i, j) the number of bits that symbol i differs
from symbol /; then
M-1
(Avg. # of erroneous bits given symbol 7) = SC Pn, jG J) (4.85)
j=0
Hence
1 M-1M-1
P(E) == Ppp GD) (4.86)
MGR ME =o j=0
The probability of bit error, P,(£), depends on the symbol-to-bit mapping, otherwise
known as the code. The computation of P(nj;,;) is very difficult if not impossible for a
general M and signal constellation. It is therefore desirable to obtain an upper bound on
P,(E). Let §;,; denote the event that message i is transmitted but signal j is closer to the
received vector, ignoring all other signals. The probability of the event &; ; is referred to as
the pairwise probability of error (PPE) and it will be used extensively in future chapters
to bound the probability of bit error. Figure 4.11 illustrates the regions where no,1 and & |
occur for an example of four signals. In all cases, the region where the event &;,; is true is
greater than the corresponding region where nj; ; is true. As a result
Equations (4.86) and (4.88) are exact and upper bound expressions for the bit error probabil-
ity for any signaling format and any code (that is, symbol-to-bit mapping). Both expressions
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 195
No.1
can be further simplified by noting that (7, 7) = (j,i) for all i and 7. However, the events
ni,j and n;,; are not always identical and depend on the signal set of interest.
Example 4.1
Consider the four signals illustrated in Fig. 4.12(a) with the corresponding bit mapping. This set
corresponds to quadrature phase-shift-keyed (QPSK) signals, which will be discussed in more
detail when considering M-PSK signals. Using (4.88), we have (M = 4)
1 e) 5
P,(E)
»(E) <<———~
Dew 22! i, j)P(&.,j)
j)P (Ei,j
and
Es
P (0,1) = P (0,3) = P(&1,0) = P (1,2) = P (2,1) = P(&2,3) = P(&3,2) = P(§30) = Q (‘|
and
2E;
P (0,2) = P(&1,3) = P (2,0) = P(&3,1) = @ ( No
196 Coherent Communications with Waveforms Chap. 4
S|
So
$3
(b) (c)
Figure 4.12 Signals and bit mapping of Example 4.1
Hence
E, 2E,
nie) =0(/%) +0 |
and E,/No =2E,/No. Note that P,(E) can be made arbitrarily small by making E,/No larger
and larger. This is not the case in binary or bit-by-bit signaling where the probability of a bit
error occurring approaches unity for any E,/No when transmitting many bits. To illustrate
that, consider the probability of a bit error in a single transmission [that is, Q(,/2E,/No)] as-
suming antipodal signals. Then, the probability of a bit error in K transmissions is 1 — [1 —
Q(./2E,/No)|* and approaches unity as K — oo (for a fixed E,/No). Note that this compar-
ison is between block transmission as the block length increases and bit-by-bit transmission as
the number of bits grows. The key advantage is obtained by varying the block length in one case
versus a fixed block (or bit) in the other.
Orthogonal Signals. Orthogonal signals are characterized by the fact that any two
distinct signals have zero correlation, that is
n= {9 ee
OMe
(4.89)
Hence, their signal correlation matrix [9 is given by
PO mes
On uk. siOe cee
To=]|. ae . | =Im (M x M identity matrix) (4.90)
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 197
In this case, the integrate-and-dump filter outputs go, g1, ---, gy—1 are Gaussian, i.i.d., with
Okey
qk/i = Bathe (4.91)
(4.92)
x 1 exp _ Gi ra Eeye ae
V 2102 ee
Since P;(E/m;) is independent of hypothesis i, the unconditional probability of symbol
error P;(E) becomes (using Appendix 3B)
pEy=1—
sh) = [- enn
ae NG (pp eed A i ale
ee 4.93
(4.93)
and is depicted in Fig. 4.13 as a function of E;/No for M = 2, 4, 8, and 16. The performance
degrades for increasing M because even though each M-ary system is transmitting at the
same symbol rate R, (symbols/sec), the corresponding bit rate 72, (bits/sec) is changing
and is given by
|
Rie a log, M (bits/sec) (4.94)
AY
Hence, as M increases, the bit rate also increases. In order to compare all M-ary systems at
the same bit rate, we need to assess P;(E£) as a function of bit SNR, Ey/No, given by
Ep = Ee 1
(4.95)
No Nolog,M
Figure 4.14 depicts P;(E) versus E,/No when M = 2”. It is clear that the bit rate adjust-
ment leads to an improvement in performance as M is increased.
To convert from symbol to bit errors, note that with orthogonal signals, the incorrectly
decided symbol is equally likely to be any of the remaining M — | symbols since each
symbol (equivalent signal vector) is equidistant from all others. Consider Fig. 4.15, which
depicts the bits for eight symbols (that is, M = 2? = 8). In any column of bits, there are
exactly 2”—! ones and 2"! zeros (four of each in Fig. 4.15). Thus, if a symbol in incorrectly
decided, then for any given bit position (or column), there are 2"! out of a possible 2” — 1
ways in which that bit can be in error. Thus, the probability of bit error, P,(£), is given by
n—|
P(E) = Ps(E) (4.96)
2" — |
and is plotted in Fig. 4.16 as a function of E,/No for several values of n. Using the union
bound of (4.81), P;(£) can be upper bounded by
iO.
10:7
Le
a”
10°
104
Example 4.2
A typical set of signals that may be generated in practice for M-ary transmissions is the set of
M nonoverlapping sinusoidal pulses
a) = ;
V2PM sinact, iT,
FF oe< G+dT
i
, otherwise
This set is orthogonal and is useful in Pulse Position Modulation (PPM) systems. Another
orthogonal set of signals consists of sinusoidal pulses of different frequencies
where the frequency separation is 1/27 Hz to ensure orthogonality. Such a set is referred to as
multiple frequency-shift-keyed (M-FSK) signals.
n i}
We
i]
n i
n=
n
n
n CORN
Ss)
Gn)
(GO):
NI)
00
mee}
Sh—-
nm
©
SIGNAL PPIN
Mo _ 0 0 0— CORRECT SYMBOL
m, 00
mM, 010
4 OUT OF 7 WAYS
THIS BIT CAN BE Figure 4.15 Symbol-to-bit mapping for
IN ERROR orthogonal signals (M = 8)
0 i#j,\i-sjl4#4
Wij 1 | i. (4.98)
-1 ji-jl=%
and as a result
Im -—Im
Ip = ( 2 2 ) (4.99)
200 Coherent Communications with Waveforms Chap. 4
| et
n=1
n=2
10"! n=3
n=4
n=5
n=6
10? ew
n= 8
n=9
x n=10
n= 15
192 n=20
104
-5
1¢ i 15 5 Figure 4.16 Bit error probability for M
orthogonal signals (M = 2”) as a function of
E., /No, dB bit SNR (Reprinted from [12])
where Zm is the u x a identity matrix. Since the signals and their negatives are in the
2
signal set, the receiver can be simplified and implemented using ¥ correlators as shown
in Fig. 4.17. In the decision rule, the signal is first chosen to correspond to the maximum
correlation in absolute value. Then, the sign of the signal is chosen to correspond to the
sign of the correlation. The joint pdf of go, g1,°--,4 M_\ is the multivariate Gaussian pdf of
(4.74) with an identity covariance matrix corresponding to the set of u orthogonal signals.
As a result, the probability of correct symbol detection, when m; is transmitted, is the
probability that |g;| > |qx| for all k Ai andi =0, 1,---, 4 — | and has the correct sign
(assumed positive without any loss in generality), that is
P;(C/m;) = Prob {4 > 0, |gil > (gol, --+, lgi-il, l@i+il, +: lay _,D/mil
fe i Ch
~ Jo —$
pert (4.100)
“_1 integrals
x fq(a/mi)dqodqi ---dqi-1dqi+1-: dqu_\dqi
Since the qj’s, i =0,1,---, 4 — 1 are i.id., then (with change of variables)
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 201
DECISION
DEVICE
. SELECTs; (t)
WHERE
a= Gn
2. SIGN OF s| (t)
CORRESPONDS
TO SGN (qi)
(4.101)
ogg? ; Es
p(E)=1— f ae —erfe (q+ [7 dq (4.102)
The symbol error probability, P;(£), is plotted in Fig. 4.18 as a function of bit SNR, E,/No.
Since for any choice of signals, one of the remaining M — 1 is antipodal (distance = 2./ E;)
202 Coherent Communications with Waveforms Chap. 4
fel
pee
(PSK)
P(E)
iT]
iT]
Wu
now
er
eh
la)
e)
eb
Sle!
el
SiS
= Moe
DY
=|
FW
ON
ON
+0ono
a 15 20
Figure 4.18 Symbol error probability
for M biorthogonal signals (M = 2”) as a
ae be function of bit SNR (Reprinted from [12])
while the rest are orthogonal (distance = /2F;), P;(E£) can be bounded using (4.81) to
produce
In order to compute the bit error probability, there are two types of symbol errors that
can occur—when the negative —s;(t) of the transmitted signal s;(t) is selected or when any
one of the M — 2 remaining orthogonal signals is decided. We will refer to the former event
as errors of the first type and to the latter event as errors of the second type. The probability
of symbol error of the first type is given by
Pyi(E) = Prob {qi<0, |gil > (Idols *++5 lgi-tls laiils+s la _D/mi}
=f ak
iL abst
iM (4.104)
wal
Honneeats
[fee EB; ie
Pa(E) = | ed —~1+erfc(q + Re dq (4.105)
The probability of error of the second type, P;2(E), is greater than P,,(E) since the or-
thogonal signals are “closer” to the transmitted signal than its negative version. Since both
error events are disjoint and form all possibilities of symbol errors, then P;;(E) + P,2(E) =
P;(E) and as a result
l
+E) Sete Lene) (4.107)
P(E)
largest achievable negative correlation [6]. The value of the signal correlation is limited as
follows:
M-1M-1
Ts M-\ Ts M—l Ts
| OD s(nyar = f a soars [ De DE silts) (hat (4.108)
0 i=0 0 j=0 ee | ay
= ME, +M(M — lEsy
where y is the correlation between any pair of distinct signals. Since Io Saran S (t))?dt >
0 must be true for any signal set, then
or
es
> — =
(4.110)
This is the maximum negative correlation achievable with M equally correlated signals. If
we choose a set with y = —1/(M — 1), the signal set is called a simplex signal set with
corresponding signal correlation matrix
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 205
ph ee ee (4.111)
en ie eT
One method to construct simplex signals is to start with M orthogonal signals, s;(t),
i =0, 1,---, M —1, each with energy E,, and translate each signal in the set by the average
of the signals, that is, let
, Mel
a(t) A 2 si(t) (4.112)
Ts Ts Ts Ts
E\ = / (s}(t))?dt = i s?(t)dt + if a*(t)dt —2 / si(t)a(t)dt
0 0 0 0
, . ; (4.114)
=E;+ yn Es) roy we =E, (1= x)
1 Ts 1 ie 1 M-1 1 M-1
Vii = ey if Oe / /
Oi ! (si) — dXSKS) — 7 »Dsi(t))dt
- a (4.115)
1 Es Es ME, ] : . .
=—— |0 +, |= ——-,._ independent
of jandj
E. M M M M-1
Hence, the new signal set achieves the minimum correlation possible for M signals and the
correlation is independent of the signal pair. The simplex signals for M = 2,3 and 4 are
depicted in Fig. 4.21. When M = 2, the simplex set reduces to two antipodal signals as
expected. Since the simplex set is obtained from an orthogonal set by translating the signals,
the symbol error rate remains the same as it is insensitive to signal translation or rotation in
AWGN channels. Using (4.93) and realizing that the signal energy is scaled by the factor
dad - a) then
M-\ 2
Fo sk Es M e?
where we let E, denote the energy of the simplex signal (that is, replaced E{ by E;). Since
all the signals are equidistant from each other, a symbol error is equally likely to be in favor
of any incorrect symbol and as a result, the bit error probability is related to the symbol error
probability in the same way it was for orthogonal signals. The Union bound for simplex
206 Coherent Communications with Waveforms Chap. 4
a) M=2
(b) M =3:
(EQUILATERAL TRIANGLE)
®o
92
c) M=4
(REGULAR TETRAHEDRON)
signals is identical to that of orthogonal signals of (4.97) except for the adjustment in energy,
l=]
so that
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 207
20 : :
ioe cos( =
Gi? I 4 :
(4.119)
: os cos(4%
7)
cos (24-5) ASE ee 1
For M = 2, the signal set reduces to antipodal signals, while for M = 4, we obtain a
biorthogonal signal set. Notice that this matrix is singular since the sum of elements in every
column is zero; hence, the matrix has zero determinant.
For multiple phase-shift-keying (M-PSK), a convenient set of signals that may be
generated is the polyphase signal set
where , is the carrier frequency and 6; = 27i/M fori =0,1,---, M — 1. Notice that for
M = 2, we get antipodal signals or binary-phase-shift-keying (BPSK), the case M = 4 pro-
duces quadriphase-shift-keying (QPSK), and M = 8 yields octaphase-shift-keying (8PSK).
The three signal sets are depicted in Fig. 4.22, along with the decision regions of an opti-
mum receiver (which employs minimum distance decoding).
Since the matrix [y—psx is singular, the dimensionality of the integration in (4.77)
must be reduced in order to evaluate P,(E). A simpler method is to transform the problem
into polar coordinates by representing the i‘” integrator output in terms of its amplitude and
phase. The optimum receiver computes for all i the quantities
Ts Ts
Gn = / r(t)s;({)dt = V3 | r(t) sin(@ct + 6;)dt
0 0
t Ts (4.121)
LIAS co6 | r(t)sin@-tdt + sin a | rn cosextdt|
0 0
= Acos[6; — n]
where defining
Ts
ne a r()||v2PP sin we t|dr (4.122)
0
Ts
A. af ri) V2P cos wet|dt (4.123)
0
we have
AA,/A2+ A2 (4.124)
and
n tan! ae (4.125)
As
On the basis of this representation of g;, it is important to note that the receiver of Fig.
4.6 requires M correlators, whereas an equivalent receiver (see Fig. 4.23) requires only two
208 Coherent Communications with Waveforms Chap. 4
[°2
(c) OCTOPHASE SIGNALS (M = 8) Figure 4.22 The signal vectors for various
polyphase signal sets: (a) Antipodal signals
THE DOTTED LINES REPRESENT DECISION (M = 2), (b) Quadriphase signals (M = 4),
THRESHOLDS IN THE RECEIVER (c) Octaphase signals (M = 8)
correlators. Since for any 7, gj represents a sample of a Gaussian process, the joint pdf of A
and n, given that s;(t) was transmitted, is
"CO
fn) = | © exp{—[r 2 — 2r,/ Ft
Es
cos(n — 61) + Es
Ht||dr, Wr 7Reg (4.127)
0, : elsewhere
Since qj is a maximum when the distance |6; — 7| is a minimum, then for 7 in the region
(6; — 7/M, 6; + 1/M), the inequality |0; — n| > |6; — n| is valid for all j #i and hence
gi > q; for all j #1. Based on this result, the probability of correctly detecting the i’” signal
is the probability that 7 is in the region [(2i — 1) /M, (2i + 1) /M}, that is
(2i+1)/M
Ps(C/mi) =|! f(mdn (4.128)
(2i—1)x/M
Letting y 4 n — 2i/M, we see that P;(C/m;) is independent of i; hence, the probability
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 209
20 COS Ot
2 E, (o-e) m/M Eg
pc) == exp (-=) | rexp (-r?) [ exp 2r,|-cosw )dy |dr (4.129)
Introducing the change of variables u =rcos yw, v=rsinw and noting that v =u tany,
(4.129) reduces to
9) ee) E, A utanz/M 5
p(c)=— | exp4 — (eh No | exp(—v av du (4.130)
The corresponding symbol error probability is derived in Appendix 4A and can be written
as
PAE
s(E)|m=4
as Ao =erfCS Es
No amie 4, rfc?
fil Es
2No
gtee. (4.132)
Ske
Es
:
P(E) ~ ere (= (4.133 )
This implies that the symbol error probability performance using four signals is about 3 dB
worse than that of orthogonal signals discussed earlier. For values of P;() less than 10m
210 Coherent Communications with Waveforms Chap. 4
—— mPSK
- - - - MDPSK
the region of great interest in practice, the error probability for M > 2 is well approximated
by
] arate | .
since when an error occurs, one of the two neighbors of the transmitted signal is chosen
with high probability. This approximation becomes extremely tight for fixed M as E;/No
increases and serves as an upper bound on the probability of symbol error for any E;/No
(see Problem 4.23). The performance for various values of M is depicted in Fig. 4.24 for
M =2,4, 8, 16, 32, and 64, as a function of bit SNR.
The computation of the bit error rate for M-PSK signals when M = 2” is considered
for Gray code mapping, as given in Table 4.1. The Gray code has the property that in going
from one symbol to an adjacent symbol, only one bit out of the log, M bits changes. As
a result, an error in an adjacent symbol is accompanied by one and only one bit error. At
large symbol SNRs, we can assume that the only significant symbol errors are those whose
corresponding signal phases are adjacent to that of the transmitted signal. Hence, the bit
error probability can be approximated by
Ps(E)
P,(E) = eee, (4.135)
2
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 211
Table 4.1 Gray Code Bit Mapping and Hamming Weight (Reprinted from [7])
OANDOaAARWNH-O
OO
—~NWN—"N
$O
HN
S—~NMNWNHWHWNHH]NHWNH
WA
AA
OC
=
—-$NMNW
WNHHANWNWAWNHH]NMWNH=|
At lower symbol SNRs where the approximation of (4.135) is not valid, it was not until
recently that the bit error rates were able to be evaluated exactly. Only the key results are
presented here and the reader is referred to [7, 8] and its references for additional detail. Let
Rm denote the decision region when message m is transmitted. Let S,, be the probability that
the received signal vector r falls into R,, when message “0,” corresponding to zero phase, is
transmitted (no loss in generality is obtained by considering message “‘0”’). Then, it can be
shown that
M
Sm = Sm—m for fara ae Net (4.137)
Let W,, be the Hamming weight of the bits assigned to symbol m (see Table 4.1). Since
the probability of a bit error is Wn/logyM when message “0” (all zero bit pattern) is
transmitted and symbol m is decided (that is, re Rm), then
M-1
P(E) = De Wn Sin (4.138)
log, M m=1
Since d) sina, = d2 sin a, the bit error rates for BPSK and QPSK are identical. For M = 8,
we have
1
Pp(E) = ~{S, + 2524+ S3 + 254 + 385 + 2S6 + S7}
3
5 (4.141)
Fis {Q(d3 sin a3) + Q(d3 sin 3a3) [1 — Q(d3 sin a3)]}
For
M = 16,
(2 5
PAE) = 5) Dd)Sm+ D|Sm + $5 +286 + Sy (4.142)
m=1 =o
The exact BERs for M-PSK are shown in Table 4.2 for M = 2,4, 8, 16,32, and 64 and
plotted in Fig. 4.25.
A more recent study [9] presented a technique for evaluating the bit error probability
with any bit mapping for M-PSK signals. A simple technique for evaluating S,, is suggested
which uses closed form expressions for the half-plane and quadrant plane probabilities in
terms of the Q-function. Bit error probabilities for M-PSK with Gray, natural binary, and
folded binary bit mapping are derived.
L-Orthogonal Signals. _L-orthogonal signal sets [10] are generalizations of biorth-
ogonal signal sets and consist of M = NL equal energy, equal time-duration signals. A
given L-orthogonal signal set has the following two properties: (1) the signal set may be
divided into N disjoint subsets, each subset containing L signals; and (2) signals from
different subsets are orthogonal. One instance of an L-orthogonal signal set is realized when
each signal subset is chosen as a set of polyphase signals. A construction of such an L-
orthogonal signal set is given in Reed and Scholtz [10]. The signal correlation coefficients
are given by”
|aA]+[4
cos (22 - j)), a)-[e
vij= 41, i J (4.143)
(E)
P,
-7
10 0 4 8 12 16 20 24 28
E,/N, dB
Figure 4.25 Bit error rate of M-PSK over AWGN channel (Reprinted from [7])
The corresponding correlation matrix 7 —o;th can be partitioned in the following form:
I'7—psk O vee O
O [ne
Pz -orth = nee (4.144)
O BA Bat) sae
where I’; psx is the matrix defined in (4.119) with M replaced by L, and © is the L x L
null matrix whose elements are all zero. Again, the special case L = 2 yields biorthogonal
signal sets. Other special cases appear of little practical importance in the implementation
of phase-coherent systems.
A set of M signals that have the correlation properties of (4.143) can be realized
by combining the orthogonal M-FSK signal discussed in Example 4.2 with the polyphase
signal representation of (4.120). Specifically, the set of M = NL signals
Table 4.2 Exact BER Values of M-ary PSK with Gray Code Mapping
(Reprinted from 2
Py (E Ph (E me (E) Pp (E Pp (E
E,/N,.dB} (M= 54) hae M=16) | (M= 32) (M = 6a)
2.132E-01 2.468E-01 2.867E-01 3.185E-01 3.376E-01
1.861E-01 2.217E-01 2.646E-01 3.003E-01 3.242E-01
1.584E-01 1.961E-01 2.420E-01 2.817E-01 3.099E-01
1.306E-01 1.708E-01 2.191E-01 2.629E-01 2.951E-01
1.038E-01 1.461E-01 1.965E-01 2.442E-01 2.800E-01
7.865E-02 1.227E-01 1.745E-01 2.256E-01 2.649E-01
5.628E-02 1.008E-01 1.535E-01 2.073E-01 2.497E-01
3.751E-02 8.061E-02 1.338E-01 1.892E-01 2.345E-01
2.288E-02 6.225E-02 1.155E-01 1.714E-01 2.194E-01
1.250E-02 4.589E-02 9.865E-02 1.538E-01 2.043E-01
5.954E-03 3.186E-02 8.292E-02 1.368E-01 1.895E-01
2.388E-03 2.048E-02 6.816E-02 1.207E-01 1.747E-01
7.727E-04 1.195E-02 5.429E-02 1.055E-01 1.599E-01
1.909E-04 6.181E-03 4.145E-02 9.147E-02 1.452E-01
3.363E-05 2.748E-03 2.998E-02 7.840E-02 1.307E-01
3.872E-06 1.011E-03 2.025E-02 6.614E-02 1.165E-01
2.613E-07 2.937E-04 1.256E-02 5.451E-02 1.030E-01
9.006E-09 6.338E-05 7.010E-03 4.349E-02 9.027E-02
9.417E-06 3.427E-03 3.325E-02 7.848E-02
8.756E-07 1.421E-03 2.406E-02 6.752E-02
4.516E-08 4.789E-04 1.627E-02 5.724E-02
1.246E-04 1.010E-02 4.747E-02
2.342E-05 5.642E-03 3.819E-02
2.925E-06 2.763E-03 2.950E-02
2.187E-07 1.147E-03 2.163E-02
8.573E-09 3.876E-04 1.486E-02
1.011E-04 9.417E-03
1.907E-05 5.394E-03
2.393E-06 2.725E-03
1.799E-07 1.177E-03
7.099E-09 4.176E-04
1.159E-04
2.361E-05
3.264E-06
pice ff welt)
(4.146)
| ffPEO Daan
N-1
dx\dx2
A(x1)
where A(x) and B are regions of integration as shown in Fig. 4.26. In particular, B is the
set of points (x1, x2) in the wedge bounded by the straight lines
3. For L = | and L = 2, we have the well-known results of orthogonal and biorthogonal signaling respectively.
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 215
a RADIANS
., REGION
Yt A(x)
5
tH Z Uy)
Xx Gite an :
x X41 an 4.
and A(x;) is the set of points within an L-sided regular polygon circumscribed on a circle
of radius x; + ./E;/No. Substituting (4.147) and (4.148) in (4.146) and making the change
of variables uj = x; + /E;/No gives (for L > 2)
P,(C) = = / exp|—(ui-
JO
/=) ||f
No 0
exp (=x})
dx
(4.149)
x
|A(uj~s/Es/No) a
Since any L-sided regular polygon can be subdivided into L equal area triangles, then from
the symmetry of the double integral, we get
2
2 ee) E uy, tan(a/L)
I(r =f exp |= (.— ‘| | exp (—x3) dx?
mt Jo No 0
N-1
uy y, tan(r/L) exp {— y2 + y2
0 —y, tan(7r/L) wu
(4.150)
Z
= — exp) —
Ta iy
haat U
LT mal hime
Es
No erf(u tan -= )
Now ths
L uy 5 *e N-1
x Fal ee —yj})
exp ( yi)e erf(yi
YI tanSF=)dn| d Uy
216 Coherent Communications with Waveforms Chap. 4
The special case of L = 2 corresponds to the biorthogonal signal set and hence (4.150) re-
duces to (4.102) (see Problem 4.26). When N = 1, we merely have a single set of polyphase
signals, and by inspection (4.150) reduces to (4.130). For other combinations of N and L,
the integrals in (4.150) must be evaluated numerically.
Upper and lower bounds on the probability of symbol error P;(E) = 1 — Ps;(C) may
be readily calculated by circular approximation to the area of integration A(u)). In particu-
lar, an upper bound on P;(£) [lower bound on P,(C)] is obtained by approximating A(u1)
by its inscribed circle of radius u;. Similarly, a lower bound on P;(£) may be calculated
by approximating A(u1) by its circumscribed circle of radius u; sec(2/L). Considering first
the upper bound on P,(£) and letting yj) =r cos w and y2 =r sin y gives
»
2 (oe) E; uy tan(ar/L) >
Pees
5 ( == = | exp Sa
(. 4 | exp (aya
x3) x2
O16 (4.151)
9) ie E; ¢ uy tan(z/L) 5
=— exp | — [ui —,/— / ex (= )dx,
m Jo No 0 ee
N-1
x [1— exp (-«7)] duy
v1 N=!
{1—exp [-«7]| =~ y-1Cx(—1) exp (—kut) (4.152)
k=0
where ,,Cx is the binomial coefficient given by
CA n!}
nk Get (4.153)
Substituting (4.152) into (4.151) and making the changes of variables u = uj./1 +k, v=
x2V1+k gives
be Web.) tsi]
PC)
(0) 2 pg CE exp|=
nC l+k (4.154)
te E;/No 2 utanz/L v2
x ~— exp} —lu— - —
Alay hee 1+k | exp a} a]
Recognizing that the k =0 term in (4.154) corresponds to P;(C)|m—psxK as given by
(4.130), we write the upper bound on P,(E) as*
4. P;(E)|1—psx is the symbol error probability of M — PSK signals with M replaced by L, hence the notation
L— PSK.
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 217
Noa
Ct) k kEs/No
5
P(E) (E) = s(E)|~-—psK
PACE) 7 = _ aS pee VtSiG, exp |
— eee (4.155)
:
a} oo yee 2 utanz/L v2
oy = ©,90) || = a du|\d
=i le VeLek | exp | ai 7 ‘i
where P;(E)|z-psk A 1 — Ps(C)|z—psx. Similarly, one can derive a lower bound on P,(E)
with the result
(E)
PEG) >> PCE)\
P(E) |r p= psx _
+D ——
Cb
eee n-1Cx exp {-
k'Es/No
(4.156)
1+k’
2
a co E;/No utanz/L v2
x = [ exp | = |u— / exp }—-—.} dv |du
a Jo 1+k’ 0 LE ke
where k’ 4 k sec? (1/L). The bounds of (4.155) and (4.156) can be simplified by reducing
the double integration involving an infinite limit to a single integration having a finite limit.
The procedure for doing this reduction parallels that given in Appendix 4A. The results are
N-1 Cay k kEs/No
pe)
PCE) == PAE
P(E) YAP.
SP.) hi psk.
ir — Ss ep = Ke cep |
=: cE
6 1 Ee INGps
[45
— + —erf (Ge sno)
0 (ANlioi )
1 / (Es/No)/(k+1) sin@ ( :
N-1 k }
(1) k'Es/No
P(E) = P; (E) & P;(E)|L—psk = Da (aaa n—-1Ck exp {~<a
k=]
yest | E;/N
Op Natit BasMa
| i Il+k (4.158)
Ge
J(Es/No)/
(+1) sin6!
veal exp (-»’) erf (y cot 6’) dy
with 0 4 tan~! [se |.We note from (4.157) and (4.158) that for N = 1, the two bounds
become equal and give the exact performance; that is P;(E)|, psx. Also for large L (small
N), the bounds are very tight since both the inscribed and circumscribed circles are good
approximations to the L-sided polygon. Figure 4.27 illustrates the upper and lower bounds
of (4.157) and (4.158) respectively as functions of E,/No in dB for M = 64 and various
combinations of N and L.
218 Coherent Communications with Waveforms Chap. 4
P(E)
\
‘
‘
‘
My
‘
- - LOWER \
10°4+— BOUNDS
‘
iY
‘
\
— UPPER ‘
iY
ir BOUNDS ty
‘
‘ x
|. XxX SUBOPTIMUM x
p RECEIVER . | | | | |
10° Figure 4.27 Symbol error probability of
5 0 Say iG 15 L-orthogonal (M = 64) signals as a function
E,/No.dB of bit SNR (Reprinted form [12])
Optimum detection of the set of signals in (4.145) requires the correlation receiver
of Fig. 4.6 and yields the performance as given by (4.150). When M becomes large, the
number of correlators required to implement the receiver of Fig. 4.6 becomes prohibitive.
Hence, we investigate a suboptimum receiver that requires only 2N correlators and whose
error probability performance is quite close to that of the optimum receiver.
The suboptimum receiver illustrated in Fig. 4.28 uses a noncoherent detector to de-
termine the most probable index n characterizing the received signal [see (4.145)] and si-
multaneously makes a maximum-likelihood decision on the phase 6), or equivalently, the
index /. We note from (4.145) that the index i is correct if and only if n and / are correctly
chosen. The probability that n is correct is the probability of the event {V2 = max, V2} or
equivalently, the probability of choosing the correct frequency from a total of N choices.
This probability P;(C)|now is evaluated in Chapter 5 and can be obtained by taking one
minus the error probability performance as given by (5.77) with M replaced by N. Having
selected the correct frequency, the conditional probability P;(C) of correctly choosing / (or
equivalently, the phase 6; from a total of L choices) is given by (4.130) with M replaced by
L. Thus, the probability of correct reception is
or equivalently, the symbol error probability for the suboptimum receiver of Fig. 4.28 is
given by
Ts / ¢. :
pe a (32) (42)dt =1 for alli and j (4.162)
Hence, the signal correlation matrix I" has unity entries and is noninvertible. The optimum
receiver is the minimum distance decision rule and since all the signals are expressed in
terms of one function #(t), the waveform problem can be reduced to an M-ary scalar
equivalent problem. When ./£; =iA, the optimum receiver is given in Example 3.3 and
its performance is evaluated in Example 3.4. Noting that o* = No/2, we obtain
Pry = M = 1 | A?
tx, seen: 4.163
(2) ( M jo 2No
If we further assume that the transmitter is subject to an average power limitation of Eayg/Ts
(Watts), it follows that
M-\1
@ 1 _ Eavg (4.164)
Fe 4 M(M—1)2M -1)
iy — 6 (4.165 )
j=0
5. Quite often, M-AM is defined by a set of amplitudes that are uniformly spaced and symmetrically located
around zero.
YHAIYYVO
NOLVZINOYHONAS
220
18s. Loa
=()s Q))'s
ONIGN
Ol Od
ONIGNOdS3YHOO
a! %,
» oi}
Yi} Ou {[ou
eat
SNIGNO
Ol d
lane [Pu
y b-Ny
NXOrFrtOrOL
ASWHd YOLVWILSS
ee ee oe ee Roo
AINSI]
P 87 y [eLNdogns
JOALID0I
JO} JUA1AYO9 UoNa!9p
Jo [eUOBOY}S[PUBIS
i0-7
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 221
eerie tayy
ee i2 4.166
VG DOM 1 ae)
Consequently, the average probability of error may be written as
P,(E)
aa =2 (=>)
NeM QO Noi —aDQM —1) 4.167
See
When M = 2, (4.167) reduces to the “on-off” keying performance of (4.68) as expected.
Given a symbol-to-bit mapping, the bit error probability can be computed exactly in closed
form using (4.86) and the probabilities of Example 3.4.
We are now in a position to relate the limiting error probability performance of block codes
as the number of codewords M approaches infinity to the notions of channel capacity,
communication efficiency, and the Shannon limit introduced in Chapter 1. We shall deal
with orthogonal signal sets, although the conclusions drawn apply to a more general class
of block codes, including biorthogonal and transorthogonal sets.
Consider the probability of correct symbol detection for orthogonal signals as given
in (4.93), that is
(ee) see)
| jer (-«- i) anleeg (4.168)
el 2
With a simple change of variables, (4.168) can be expressed as
* we: M-1
If we define 6 4 E,/No and consider the limit of the logarithm of the expression in brackets,
we obtain
Mot. In Q (—v — \/2 logy M)
jim,mn[Q(—-»—v26t0e.M)| = Jim“gy ITD
As M > oo, the limit approaches In Q(—0o)/0 = 0/0. Treating M as a continuous variable
222 Coherent Communications with Waveforms Chap. 4
(4.172)
bool v?
A,| ee a 4.173
cAv ia (4.174)
In2
and noting that
d 1 aed:
os 28 1025 M |= ——SS 4.175
dM (v+ paces ) Min MV 21n2 oe
In M
exp (—6———In2 ) = m~(6/In2) (4.176)
(4.172) can be expressed as
exp (—c2w)
exp ({ 1 — aa w?
lim (—c})
w— oo Ww
(( : ) Ihe <a
Me ©,©)
Bebe i)
B/ In2s<4
The limit is independent of the sign of cz as the exponential in w* always dominates the
exponential in w as w — oo. Since this is the limit of the logarithm in brackets, the limit of
the expression itself is
M-1
: 2Ep 1, Bee >in
lim —v —./— 1 Bead Boxe
ain. ‘ \ No cn) fc E}/No <In2 eee
6. To compute the derivative of Q(x), use Leibnitz’s rule
q
cs
aa Forvayax=ae fgeeage dq - Bayedp
Pp
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 223
1, E,/No>1n2
ee PAC)=| 4 E»/No <\n2 ce
or, in terms of error probability
0, E,/No> In2
jim rey= {9 Ep/No <In2 ae
Allowing M to approach infinity is equivalent to considering the case of unconstrained
bandwidth. Hence, an orthogonal coding scheme produces error-free transmission in the
unconstrained bandwidth, additive white Gaussian noise channel provided that E,/No >
In2. The limiting behavior is particularly significant from an information-theoretic point
of view. Shannon’s formula for the capacity of a channel of bandwidth B perturbed by
additive white Gaussian noise with one-sided PSD No is given by (1.30) and repeated here
for convenience:
P
This represents the maximum bit rate for error-free transmission. Letting the bandwidth B
approach infinity and using the expansion In(1 + x) =x — x7/2+x3/3—---, |x| < 1, then
B je B P 12
li = lim
poe gua ere—~In{ n(1+ +a)
— ]=([-—~](——]=
(=) (x) NED 4.184
TMi
Thus, from (4.182) and (4.184), it follows that orthogonal signals achieve error-free trans-
mission in the limit as M — oo, provided that the bit rate 7, is less than
In this section, we shall briefly discuss methods of generating orthogonal, biorthogonal, and ° :
transorthogonal signal sets for the binary case, that is
N
Sia) Sippj(t) POE oom AGE E= II (4.186)
j=!
where sj; = +1 for all i and j and {$j (t), j = 1, 2, ---} is any set of orthonormal functions
over the interval (0, 7;). The signals as defined by (4.186) are normalized with energy N.
For any energy E;, the signals should be scaled by /E;/N. The signals can alternatively
be represented by their respective vectors $; = (81, 5i2,+°-, siv)-. Let Hy be an N x N
matrix defined inductively by
FA Soa ik
TEE Wy. alae
Hy (4.187)
;
where
r= ( A) (4.188)
224 Coherent Communications with Waveforms Chap. 4
Such a matrix is called “Sylvester-type Hadamard” matrix and satisfies the property that
Fintan (4.189)
By definition, the construction of such matrices is known when N = 2*. It is also conjec-
tured that when N = 0 mod 4, such matrices exist [11]. As an example, 7112 is given by
1 1 1 1 1 1 | 1 1 1 1 1
1 =| 1M el 1 l 1 =f =1 =1 ||
1 = ee 1 =1 1 1 t=) Si) al 1
1 a | 1 s=! 1 1 |
i ey
1 a 1 -1. -! 1 i 1 1 PRS ee
i =| —I!I ete — 1 ea 1 1 eel
(4.190)
eres |STN Fe Ae oO ny a eee
1 1 =1 =! --! ees Na | geal 1 1
1 1 1 =—1 -1 —] | a oe | 1 =] 1
1 1 1 1 =-1 -l -! ii eal es |
if =i 1 [ =f =b L es 1
1 Po, 1 1 Pe eer (eee | a |
Since Hy HH", = NIn, the columns of these matrices can be used to construct M orthogonal
signals each with N coordinates with M < N. To construct M biorthogonal signals, use any
M/2 columns and augment the set with the negative of the columns to obtain the complete
set of M signals. To obtain M simplex signals, delete the top row (or first column) of Hy
and use the remaining columns as vectors. The number of coordinates will be one less than
the number of constructed vectors as expected.
Example 4.3
Design binary vectors representing eight orthogonal, biorthogonal, and simplex signal sets with
E; = 8. Use Sylvester-type Hadamard matrices to obtain the vectors.
Since M = 8, we need to compute 71g as
[oe pewWitk Eo Up oe
} pee SESS ia est
eis at eed 1 rs eC
Hele H4 ie ee 4 l bk ie 1
Ha —Ha 1 l i 1 ee ee Pe ees
tos aes Drs bel We Re:
| Li sk Sheep er aa 1
1S =besets “lv St Fees
= (So S$} S2 ---S7)
So S1 S2 -::87
For simplex vectors, delete first row from Hg and use remaining columns to obtain
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 225
1 —1 1 —1
] ] =] —1
1 —1 —] ]
So= [et sp ies = ja) ec ead Rn
1 —1 1 1
1 1 —1 1
1 =| —1 —|
We can check to see that
1 1
ee
1 Sy Fd
s( ee
VE ( 3)
and
Thus far in our discussion of optimum coherent detection of M-PSK signals, we have
tacitly assumed that the receiver was perfectly synchronized to (had exact knowledge of)
the frequency and phase of the received signal’s carrier. In practice, the receiver derives its
frequency and phase demodulation references from a carrier synchronization loop which,
depending on its type,’ may exhibit an M-fold phase ambiguity. That is, the receiver is
capable of establishing a phase reference only to within an angle @g of the exact phase.
Since the phase ambiguity indeed takes on values equal to those of the M-PSK data phases
themselves, then any value of ¢, other than zero will cause a correctly detected phase to be
erroneously mistaken for one of the other possible transmitted phases, even in the absence
of noise! Clearly, something must be done to resolve this phase ambiguity else the receiver
will suffer a performance degradation. For example, for a phase ambiguity corresponding
to m = M/2, the receiver will be 180° out of phase with the received signal and all data
decisions will be inverted resulting in very poor error probability performance.
One simple solution to the above is to employ differential encoding at the transmit-
ter yet still maintain coherent detection at the receiver. The theory of differential encod-
ing and how it resolves the above-mentioned phase ambiguity is described as follows. Let
5i(t) = /2P cos (wet + 6;) represent® the M-PSK signal to be transmitted during the i‘
transmission interval i7, <t < (i + 1)7; where 6; ranges over the set of allowable values
Bm = 22m/M;m=0,1,...,M — 1. Ordinarily, 6; would be exactly equal to the informa--
tion phase, say A6;, to be communicated in that same transmission interval. At the conclu-
sion of the i‘ transmission interval, that is, at time t = (i + 1)7;, the receiver makes its
M-ary decision 6; on 6; (or equivalently, A@;). Because of the ambiguity associated with the
demodulation reference phase, however, this decision consists of an unambiguous part by.i
plus the ambiguity ¢, itself, that is, 6; = By.i + oq. Suppose now that instead of assigning
0; identically equal to A6;, we instead encode A6; as the difference between two adjacent
7. A complete discussion of the various types of carrier synchronization loops for M-PSK modulation is pre-
sented in Volume II of this book.
8. In this section, it is necessary to introduce a notation to characterize discrete time. For simplicity, we shall
use a subscript, e.g., i, on a variable to denote the value of that variable in the i” transmission interval. To avoid
confusion with the previous notation of 6; to denote the i'” message phase 277i /M (i" hypothesis), we introduce a
new Greek symbol, i.e., B;, to denote the i'" message phase. Whenever it is necessary to represent a variable both
in discrete time and/or by an associated hypothesis, we shall use a parenthetical superscript for the latter.
226 Coherent Communications with Waveforms Chap. 4
transmitted phases, that is, A@; = 6; — 6;—1. If the receiver still performs decisions on the
transmitted phases 6;_; and 6; but arrives at its decision on the information phase based on
the difference decision variable 6; _ 2 = (4. + ba) — (4.11 = ta) = 6,7 —0y.7=15
then clearly the phase ambiguity has been eliminated. Inherent in the above ambiguity res-
olution is the assumption that the phase ambiguity ¢, introduced by the carrier synchro-
nization loop remains constant over the two transmission intervals of interest, that is, for
a duration of 27, seconds. If we continue to encode the information phases in the above
manner and base our decisions on the difference of adjacent decisions made on the actual
transmitted phases, then clearly a sequence of Ns + | transmitted phases is needed to con-
vey N, symbols of information. That is, the first symbol in the transmitted sequence is used
to resolve the phase ambiguity.
As a first example, consider a BPSK modulation (M = 2) and suppose that we desire
to transmit the information sequence A@; = 7, Adz = 1, A63 = 0, A64 = 7, and As = 0.
Assuming 69 = 0 as the reference phase, then the remainder of the transmitted (that is,
the differentially encoded) phases reduced modulo 27 would be 6; = 6) + A@) = 7, 62 =
6; + Adz = 0, 63 = 62 + Ab3 = 0, 04 = 63 + AO4 = 77, 05 = 64 + AOs = 7. In the absence
of noise and assuming no phase ambiguity introduced by the carrier synchronization loop,
the receiver would make perfect decisions on the received sequence, that is, 6=6; i=
0,1,2,...,5 and would then differentially decode the modulo 27 difference of these de-
cisions as 6 — 6) = 1, 6) — 6 = 7, 63 — 65 = 0, 64 —-63 =z, and 6s —64=0 which is
identical to the information sequence. If now the carrier synchronization loop introduced a
phase ambiguity of z radians thus converting the received phase symbols into the sequence
6, = 1, 0; = 0, 05 = 1, 6, = 7, 0, = 0, and 6{ = 0, then in the absence of noise we would
make perfect decisions, that is, 6/ =6', i=0,1,2,...,5 and after differential decoding
(modulo 27) obtain 6; — 6 = 7, 65 — 6; = 7, 65 — 05 = 0, 6, — 6, = 7, 65 — 8, =0, which
is once again identical to the information sequence.
As a second example, consider a QPSK modulation (M = 4) and suppose that
we desire to transmit the information sequence A@; = 21/2, A@2 = 2, AO3 =0, AOy=
32/2, AOs = 3/2. Again assuming 6) = 0 as the reference phase, then the remainder
of the transmitted (that is, the differentially encoded) phases reduced modulo 27 would
be6; = 69 + AQ) = 1/2, 02.= 0; + Ado =31/2,.03=
02 + Ads
= 37/2, 04 = 03 + AO,=
mt, 05 = 64 + A@s= 7/2. In the absence of noise and phase ambiguity, the receiver would
differentially decode (modulo 21) its perfect decisions, that is, 6; =i= mint 6) oa 6 ==
Toe) 6.-6, =n, 6; — 6) = (0), 64 — 63 = 37/2, and 65 — 64 = 3/2 which is identical to
the information sequence. In the presence of a phase ambiguity, say @g = 7/2, the re-
ceived phase symbols would be converted to the sequence 6) = 1/2, 0; = 7, 6, =0, 05 =
0, 0, = 32/2, 05 = 7 and in the absence of noise the receiver would sioke merce decisions
that is, 6/ = of i=0,1,2,...,5 and after differential decoding (modulo 27:) obtain
0, — 6) = 1/2, 0, — 0 = 1, 65 — 65 =0, 0) — 64 = 3/2, 0, — 6 = 3/2 which is once
again identical to the information sequence.
Figures 4.29a and 4.29b are illustrations of a differential encoder and a differential
decoder, respectively, where the inputs and outputs of these devices are the phase symbols
themselves. Alternately, one can represent differential encoding and decoding in terms of
the complex baseband representation of the signals which results in the configurations of
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 227
modulo 2r modulo 2x
Lee
asterisk denotes
complex conjugate
multiplication
namely, (ou; 6r). The ultimate decision on AQ; would then be obtained from A@; =
6! — oe modulo 277.
The design of an optimum MAP coherent receiver for jointly detecting 6/_, and 6;
follows the general likelihood function approach described in Section 4.1.3 and illustrated
in Fig. 4.6 and later applied to conventional (without differential encoding) M-PSK in Fig. _
4.23. Here however, for fixed @,, there are M7 hypotheses (pairs of possible values for
6!_, and 6!) and a joint decision is made based on a maximum correlation over the 27;-sec
interval (i — 1)T, < t < (i + 1)7,. In particular, analogous to (4.38), the receiver computes
(i+1)T;
x) & i) p'(t)s* (tat (4.191)
(i—1)Ts
where the prime on p(t) denotes the fact that the sample function of r(t) includes the phase
ambiguity gg and
with s(t) = /2P cos (wet + By); iT; <t < (i + 1)T,. The decision rule on the transmit-
ted phases 6;-; and 6; in the presence of the phase ambiguity is then made as follows.
Choose Oo, = B; and 6! = f; where x (él) is the maximum value of x“-. Finally, the de-
cision on A@; would be Ad; = 6! _ oS = B; — B;. Since (4.191) can be obviously written
as
iT. +07,
ed p'(t)s(t + T,)dt +f p'(t)s (dt (4.193)
(i-1)Ts iTs
a receiver that implements the above decision rule is illustrated in Fig. 4.31. Furthermore,
because of the white noise model and the assumption of uncoded data [that is, s;_(t) is
independent of s;(t)], then
iTs (i+1)Ts
max x*) — max { p'(t)s(t + roa} + max / p'(t)s(t)dt
( i
a trys Ts (4.194)
A max pat + max mo
k l
and the joint decision on 6:_, and CH can be visualized as successive decisions on 6!_, and
6‘ independently. In view of this, the receiver of Fig. 4.31 can be redrawn in simpler form
(M correlators as opposed to M? correlators) as in Fig. 4.32a. Furthermore, transforming
the problem to polar coordinates as was done in arriving at Fig. 4.23 from Fig. 4.6, we
obtain the familiar form (see [12], Fig. 5-20) illustrated in Fig. 4.32b which requires only
two correlators. We emphasize that the equivalence between Figures 4.31 and 4.32a or 4.32b
would not be valid for coded (correlated) data despite the white noise assumption.
Error Probability Performance. To evaluate the symbol error probability perfor-
mance of the receiver of Fig. 4.32, we begin by considering the possible ways of making
a correct decision on A6j. Clearly, if both 6/_, and @/ correspond to correct decisions,
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 229
s(t) |
x! 0,0 )
Delay
T,
Choose
Largest
r'(t)
Delay 6
sO 7)
Choose
Largest
Figure 4.32a Optimum receiver for differentially encoded M-PSK equivalent to Figure 4.31
that is, ge, = 6;-1 + bq and 6) = 6; + da, then AO; will be detected correctly. The prob-
ability of this event occurring is PC) ey PAC Mab where P;(C)|y—psx iS given
by (4.130). Furthermore, if both On and 6! are in error but by the same amount, that is,
Oey = 6;-1 + dq + 2k /M and 6! = 6; + da + 2k /M, then A@; will still be detected cor-
rectly. Denoting the probability of this event by POC ), then referring to Fig. 4.32, we
have
k
PL (C) = Pr {(2k — 1)t/M <ni-1 — 6-1 — ba < Qk + I)z/M,
(4.195)
(2k — 1)n/M < ni — 6 — ba < (2k + 1) /M}
i: —°¢lu
[eyuualassjiq
Japooaq
SOM 4 O[Npow
uz
}SOT[PUIS : —
ul , 9 oh
ainjosqy|
(1)4
A, ane |
9= [u—tgfuia
I)
| Aelaq
l a
ol ee eee
Py
JatIes
“yous
doo]
aaNSIq qzE"p NWINdQ Joatooo1
Joy Ajenuasaryip
Peposua YSq-W JUaTRAINbs
0} SANBIJ BZE'p
231
232 Coherent Communications with Waveforms Chap. 4
where nj—) and 7; are the receiver’s noisy estimates of the phases in the i — 1™ and i®
transmission intervals, respectively. Analogous to (4.127), the pdf’s of nj-1 and n; are
a(t") 2
Using the same change of variables as in going from (4.129) to (4.130) and following the
approach taken in Appendix 4A, we arrive at the following result:
2 ;
pes E, u tan[(2k+1)z/M]
p® — ih
exp |—v° }dudu
aur:mone See ae \ene utan{((2k—1)2/M]
No
=a
frei E
tie = sin
aed
$7
1 E
7 =|- ten] in
=
, : (4.198a)
ce a9(-)
Es /No sin{(2k-+1)2/M] 2k+1
ert
|yor SEU
a = |ay
Since the preceding events are mutually exclusive and exhaustive of the ways to make a
Sec. 4.1 Optimum Waveform Receivers in White Gaussian Noise 233
M-1
PQ eeP(C) (4.199)
k=0
M-1
s P&C) (4.200)
=2 Ps(E)Im=psk
s(E) |v —psk |1—
5
3 Ps(E)im—psk
1
Ps(E)|m—psk — 25 py
e=|
Ps(E)|m—psk
yp Es 3 E, i 4 ES
Es
Pelb
Ce) jee 2 eric’
erfc \ No — 2erfc*
erfc | —— fc?
No + erfc |No
—— — =erf
a Cc INo (4.202 )
Analogous to the formation of a QPSK signal from two independent BPSK signals,
a differentially encoded QPSK signal can also be formed from two differentially encoded
BPSK signals. The optimum receiver would then differentially decode each BPSK signal
independently and then combine these two binary decisions to form the resulting decision on
the QPSK information symbol. The receiver for this system is illustrated in Fig. 4.33 where
the symbol phase decision is made in accordance with the bit to symbol mapping of Fig.
4.12a. To demonstrate that the symbol error probability performance of Fig. 4.33 is identical
to that of Fig. 4.32, we proceed as follows. For Fig. 4.33, the probability of a correct
symbol decision is equal to the probability that both binary decisions are correct. Since each
binary decision is the result of differential decoding a BPSK signal, then the probability of a
correct binary decision is one minus the probability of error given in (4.201) and hence the
probability of a correct QPSK symbol decision is equal to [1 — PE) lees | Finally, the
probability of error for the QPSK symbol is
2
P;(E) |p x. QPSK = a [1 ~ PE)\px. BPSK |
(4.203)
=2 P,(E)\px. ppsk — (Ps(E)Ip.z. BpsK)”
Substituting (4.201) into (4.203) and simplifying gives the identical result to (4.202).
234
dm —1@)so9
(°@
joqurds
aseu
(1)4
UOISIDaq
7A—
(’
—1°@)uts
aaNS1y
¢¢"p
= wind Jantaoo1
10j ATenusayjip
papoouaASdO
Sec. 4.2 Optimum Waveform Receivers in Colored Gaussian Noise 235
The problem of detecting signals in colored Gaussian noise can be approached in several
ways, namely, using the sampling approach, the K-L expansion, by performing an intuitive
whitening operation on the received signal or by a direct derivation with a sufficient statistic.
All four methods will be explored in this section to derive the optimum receiver in colored
Gaussian noise. Unlike detection in white Gaussian noise, the detection problem in colored
Gaussian noise and the performance of the optimum receiver depend on the length of the
observation interval. In white noise, the observation of the received signal over the interval
(0, 7), that is, over the signal duration, is sufficient to perform optimum detection. In
colored noise, the observation of the received signal outside the signal duration (and hence,
the observation of the corrupting noise) is of key importance as it provides information on
the noise present inside the signal interval. Limiting the observation interval to the signal
duration will in general result in a receiver performance inferior to that of observing a larger
interval. There are two key issues that will be addressed in this section: (1) the structure
and performance of the optimum receiver assuming a finite observation interval, and (2)
the structure and performance of the optimum receiver assuming an infinite observation
interval. The problem is then to decide between M messages where r(t) = s;(t) + n(t)
under m;, i = 0, 1,---, M — 1, over some observation period (finite or infinite) and n(t)
is Gaussian, zero-mean, nonwhite with autocorrelation function Ry(t1, t2) A E{n(t})n(t2)}.
Let’s first start by investigating the time-sampling approach. In this case, the received signal
r(t), observed over the interval [0, 7], is sampled to produce the N-dimensional vector
The MAP receiver decides on that message with maximum a posteriori probability, that is
1
lim
im pURvs
Pn vs = iim
li (4tpyf) cepts
Apne 4,
(4.212)
Define
be Ey
ge Ay en Si (4.213)
or
sj = RygiAt (4.214)
By continuity of s;(t) and Ry(t, s), we write in the limit as N > oo, s;(j At) > s;(t) and
and g;(t) is determined by solving the integral equation. From (4.213), we find
= / r(s)gi(s)ds
0
Sec. 4.2 Optimum Waveform Receivers in Colored Gaussian Noise 237
Similarly
yt = in ade
: Tp-l ‘
T
ii
(4.218)
= / si(s)gi(s)ds
0
Thus, the MAP test reduces to the following: select m(o(t)) = m; iff
where
th
Sij a} Si(t)Pj(t)dt (4.223)
0
DECISION
AOmMOD>Yrr
(see note)
Figure 4.34 Waveform receiver in colored Gaussian noise (derived using sampling approach)
The Gaussian noise random process n(t) can also be expanded in terms of the basis func-
tions to yield
where
th
nj af n(t)pj(t)dt (4.226)
are random variables. For these expansions or sequences to converge, we require that
‘ 2 8,
N 2
rn) where
if
rj= | r(t)pj(t)dt (4.229)
0
and
when m; is transmitted. The random variables n;, i = 1,2,---, N are Gaussian (since they
are the result of a linear operation on a Gaussian random process) with
if
E{nj}=E {[ no, ar|=) (4.231)
0
If we set
E{ry/mi} = Sj (4.235)
where s! = (S;], Si2,°**, Sin). Hence, from (4.210), the a posteriori probability P(m;/rjy)
1S
| \Tp-! rar
Pag
b P(mi) exp {3 (oy — Si) Ry (On si)|
cecee
P(m;j/t¥n = py) =
Ne e ON , (4.986)
Ben P(mj) exp {-3(en — sj)
Ry’
(ey — s;)|
To eliminate the problem associated with Big: as N — oo, we use a special case of $j (t)’s,
which render all Ry to be invertible, that is, let A; and ¢;(t) (eigenvalue and eigenfunction)
be solutions to
T
AiPi(t) = / R(t, s)Pils)ds (4.237)
0
240 Coherent Communications with Waveforms — Chap. 4
where A; ; = A;4;;. Thus, the sample function expansion of the noise n(t) = ee njQj(t) is
the Karhunen-Loéve (K-L) expansion. This implies that the n;’s are uncorrelated and
he 708
i,j =Ai9i,j = / i dj (t) Rn(t, S)pj(s)dsdt (4.238)
0 JO
In this case, (4.233) reduces to
CO
or
(o,)
In P(m;/r(t))
e =InC +InP(m)
OT + YOa5! |EE
ryjsij eee
ag (4.243)
J=
for alli =0,1,---, M — 1. Therefore, the optimum receiver sets m(r(t)) = m; iff
CO CO
SIGNAL
DECISION
AQOMODY>-
MHOOLO
Figure 4.35 Waveform receiver in colored Gaussian noise (derived using K-L expansion)
is easier to solve. The eigenvalues are solved for simultaneously by imposing certain con-
straints on the form of the solution of the differential equation (see Problem 4.31).
ed) ee) if
See r(0)bj(Odt
=I j=l (4.246)
T
=i! r(t)gi(t)dt
0
where
oO
= 047!sj(AiGj)) (4.248)
7
Y- sijj(s)
j=l
= si(5)
Recall g;(t) in the sampling approach defined through (4.216). Thus, the K-L and the
sampling approaches are equivalent if
N
Rae
2
Sip < (4.251)
The convergence of the random sums computed in the receiver is handled in the following
way; let Sy denote
N
SrA Aten (4.252)
i=
Sec. 4.2 Optimum Waveform Receivers in Colored Gaussian Noise 243
This sequence of random variables converges in the mean-square sense as N —> oo when-
ever
This is the Cauchy convergence criterion [14] for random sequences. Suppose that m, is
transmitted. Then, the r;’s are independent Gaussian random variables with
T
rj =| r(t)oj(t)dt = sj +n; (4.254)
and
then for finite noise power in the observation interval, we require that
lim’, = 10 (4.258)
Noo
N+M 52
lim Le —/ —0 for any M (4.259)
iliac TS | j
Since the denominator elements are going to zero, we not only require that
lim — =0 (4.261)
244 Coherent Communications with Waveforms Chap. 4
If the above is satisfied for all 7, then it is guaranteed that the remaining term in (4.256) also
converges because of Schwartz’s inequality, that is
(c) let Fe/A; be the signal energy-to-noise ratio for signal i in the direction of $;(t), then
=
lim Sj < oo for alli
is oF = 1 Ay
When condition (c) is not satisfied, we arrive at the case of singular detection in which there
exists at least one direction or dimension that is noise free (that is, zero eigenvalue). In this
case, perfect detection is possible provided that the projections of the various transmitted
signals onto that coordinate are distinct, that is, unequal. If the projections are not distinct,
the dimensionality of the problem can then be reduced by the number of zero eigenvalues
and the problem reworked in a less dimensional vector space.
Sey Lisi;
wet + B = max yj LhjSkj
sp: + Byk (4.263)
=
where B;’ is given by (4.245) and the rule results in the functional block discussed pre-
viously in Fig. 4.35. Assume now that the transmitted signals are designed such that they
occupy the first NV dimensions, that is
N
s(t) = x sijj(t) (4.264)
j=l
T ? : :
then r; = fo r(t)b;(t)dt = s;; +-n;, where the last equality assumes that m; is transmitted.
Decompose the vector r into
p= (tigre) (4.265)
where
and
r= >, 174;(0)
j=l
N ee)
= DrigiO+ Di 1joj@ Oe
j=l j=N+1
—| ———4”
IAs) R= /)) =
_ P(m) fe(p/mi)
Hasan fr(p) (4.269)
= P(mj) fry,r. (P15 P2/m;)
fr(p)
P(mi)
fr,(p,/mi) = max { P(mk) fr, (91 /mx)} (4.272)
or alternatively iff
N N
| yi eB ee max EG rjSkj + mt (4.273)
j= j=
2
raat = VrjSij + B;" = max Fe
2~<
5 a B, (4.274)
No = k | No Pa
246 Coherent Communications with Waveforms Chap. 4
SIGNAL
MATRIX
mnooro
DECISION
AOMODDYr
In this case
N T
1 Seis te os / r(t)s;(t)dt for alli (4.275)
jel
and ie Side AS Si; = E;. Hence, the receiver reduces to selecting m(p(t)) = m; iff
it T
iEr(t)s;(t)dt + B; = ne {| r(t)sp(t)dt + Br (4.276)
0 0
where By = (No/2) In P(m,) — Ex/2 as given in (4.39).
Since the optimum MAP structure of Fig. 4.36 whitens the received signal by projecting it
on the appropriate eigenfunctions, let’s rederive the receiver by imposing a whitening filter
up front.
Sec. 4.2 Optimum Waveform Receivers in Colored Gaussian Noise 247
DECISION
where Ago is the mean conditioned on mo and oaq the noise variance at the output of
the matched filter. Since the noise at the output of H(f) is Gaussian in Fig. 4.37, (4.277)
is still applicable with the appropriate mean and variance. Since the output Aq(t) =
Sor@)h(t — t)dt, then
where S,(f) is the PSD of n(t) and S(f) the Fourier transform of the signaling pulse s(t).
Hence
My Jian (pe Tt df (4.280)
AA tae eS
AAMAS
CO AO) patbd)
Defining
we have
oa a) 1/2
A< (
/ mats ar) (4.285)
or equivalently
K( f\p—J2ufT
Wf) = Aelia an (4.287)
Sn(f)
This is the optimal filter for antipodal signals with colored Gaussian noise with PSD
Sn(f). The filter can be decomposed as a series of two filters, that is
where
1
H
Le Seeaaee
=
Gere
4.289
and
Silent
Hof) = (4.290)
V Sn(f)
The filter Hj(f) whitens the noise? and the filter H2( Ff) is matched to the “whitened”
signals as shown in Fig. 4.38.
Assuming optimal filtering, the bit error probability is given by P(E) = Q(Amax)
where
TSC
Nor a
i Sf) 2 d
Gare
4.29]
Hence, P(E) is now dependent on the choice of S(f), that is, it depends on the form of
the actual signaling pulse s(t), not simply its energy E as in AWGN channels. Then, what
Whee 17)
Sn(f) =c 2
I]jai t = Pj)
where z; and p; are the zeros and the poles, respectively. Since S,(f) is real and nonnegative, then the complex
roots of S,(f) have to occur in complex conjugate pairs and the real roots occur with even order. In this case, the
whitening filter can be constructed using
where Im(z;) = 0, 1 <i <n and Im(p;) > 0, 1 < j <m, that is, H\(f) is formed with the right-hand side zeros
and poles of S,(f). In this case, H;(f) is guaranteed to be a causal filter.
Sec. 4.2 Optimum Waveform Receivers in Colored Gaussian Noise 249
Hy (f) Ho(f)
sre! oT
V Sn (f)
WHITENING MATCHED
FILTER FILTER
would be the optimal signal shape to transmit for a given channel S,(f) or equivalently,
what is the choice of s(t) that maximizes A max? That is
Consider expanding the received signal r(t) using a K-L expansion as in the previous sec-
tion (remember that only the first N components are sufficient for detection), that is
N
ri(t) =) ri@i(t) and rj = sei +7: (4.293)
i=l
assuming my, is transmitted with
E{njnj}= 1
ce Rr ew,
oe (4.294)
Then, r(t) can be mapped into the vector r; and the whitening operation can be represented
by the equivalent vector transformation!?
S12
di 0) Ses @)
—1/2
2 ot eras (4.295)
nya
0 ip ees
since
hee ak 0 A 0 0
—1/2 S 0 X
; @) \ —1/2 0 Or ty N
ae (4.296)
iS 0 0
=
a e 0
(RS aaa ie ei
10. r},, stands for the whitened received vector and is given by rj» = Sw + Mw.
250 Coherent Communications with Waveforms Chap. 4
and the noise components of rj, are actually white. The signal component Of Lig. Sus
becomes
eg RO Pies? | 0
—1/2
of le ee 5 (4.297)
0 Osan
and
Nae foe) KY iz
Sa =f side = | ST af (4.298)
a Ki —0oo —co Sn(f)
where the first equality is due to the fact that a signal and its expansion have the same
energy and the second to Parseval’s theorem [15]. The optimization problem can therefore
be restated in terms of the vector components as
Ne‘Se N
max
<
) i subject to y s? = iE (4.299)
i=l “" =I)
Ns: NTS: E
Dee = (4.300)
with exact equality if we place all the signal energy EF in the dimension with the smallest
noise variance Amin. Therefore, the optimum performance is obtained by transmitting all
the signal power in that coordinate where the noise is weakest. Thus, the optimal antipodal
signal pair in nonwhite Gaussian noise is
Finite observation interval. When the noise is still WSS but the observation in-
terval is finite, the derivation of the optimum receiver is not as straightforward and requires
solving integral equations. Let (0, 7) denote the observation interval, then the output of the
whitening filter in Fig. 4.39, when message m; is transmitted, becomes (From here on, we
will let H(f) denote the whitening filter. Note that in Figures 4.37 and 4.38, H(f) denoted
the combination of whitening and matched filters)
1
Ha) = i!hG@ — t)r(@)dt=syi@) +N) (4.302)
where
t
Sw i(t) = Si(t) * h(t) =i) si(t)h(t — t)dt (4.303)
0
je
Ny(t) =n(t) *h(t) = / n(t)h(t — t)dt (4.304)
0
Sec. 4.2 Optimum Waveform Receivers in Colored Gaussian Noise 251
WHITENING
FILTER
Since the noise at the output of the filter, n,,(t), is white, the optimum receiver for equally
likely messages computes
FE Jé
InA(ry(t)/m;) = i!ry (t)Sw,i(t)dt — ;/ s2,,(t)dt (4.305)
0 0
for every transmitted signal and decides on the message with the largest In(r,,(t)/mj;) as
shown in Fig. 4.40(a).!! Substituting for sy j(t) and n»(t), we have
Tig ept i
In A(ry(t)/mi) =i) {| A(t — or(eide | A(t — r)si(e'yar| dt
0 0 0
(4.306)
1 ort et ji
oe > if {| h(t — dsi(eydr [ h(t — r)sirar'} dt
2Jo (Jo 0
Defining!”
7
OG) af h(t —t)h(t — t’)dt (4.307)
0
then (4.306) simplifies to
ToT.
In A(ro(t)/m) = f / r(t) Q(t, t’)sj(t’)dtdt’
ie
1 T af
(4.308)
_ =f | 5i(t) O(c, t')8;(t’)dtdt’
2Jo Jo
and can be implemented as shown in Fig. 4.40(b) with
1 fhe T
Cc; A -; f i 5i(t) Q(t, t’)sj(t)dtdt’ (4.309)
2Jo Jo
Defining
fi
gi(T) af Q(t, t')sj(t')dt (4.310)
0
11. Ey,; denotes the energy of the i‘? whitened signal, that is, Ey; 4 i s2 (t)dt.
12. Note that with a simple change of variables u = t — t and when the observation interval becomes (—00, 00),
then
(a)
DECISION
WHITENING
FILTER
r(t) ty (t)
MHOOLZO
HOmMODrYrr-
(b)
DECISION
mnoozro
4Homowvrr
Figure 4.40a, b Alternate structures for colored noise receiver (derived using the whitening approach)
Sec. 4.2 Optimum Waveform Receivers in Colored Gaussian Noise 253
DECISION
with
1 i
B= -; f si(t)gi(t)dt (4.312)
0
which is identical to (4.221) for equilikely signals. This suggests the alternative implemen-
tation shown in Fig. 4.40(c). The computation of the functions g;(t),i=0,1,---,M—1
require solving (4.310) which depend on Q(t, t’) which is a function of the whitening filter
h(t). The g;(t) functions can also be expressed directly in terms of the transmitted signals
and noise autocorrelation function R,,(t) as follows: since n,,(t) is white, then
ifs Vf
E{ny(t)nw(s)} =| | Rae =F AE = Th HT dt az”
0 0 (4.313)
= d(t —s)
Multiplying the right-hand side by h(t — rT) and integrating over r, then
i a ie
/ b(t —s)h(t —t)dt et / RAG — tt")
0 0 JO
(4.314)
T
x i)A(t — vt)h(t — t)dth(s — t")dt'dt”
0
or
Paint
h(s =) =a) ih Rie’ —t JOG, te)h(s —t")dt'dt” (4.315)
0 0
254 Coherent Communications with Waveforms Chap. 4
F
si(t) =i) gi(t)Rn(t — t)dt (4.319)
0
This is an integral equation in g;(t) as a function of the observation interval (0, 7), the
signal s;(t), and the noise autocorrelation function R,(t). Note that when the observation
interval is (—oo, oo), then s;(t) = g;(t) * R,(t) and G;(f) = S;(f)/S,(f), which is the
filter obtained earlier!? when discussing antipodal signals.
Example 4.4
Consider the binary communication problem over the colored Gaussian noise channel with
Sn Gi (No/2)(f2 at 1)/(f? + 4). Derive the optimum receiver assuming infinite observation
interval and antipodal signals
where
S(f)
G
Sao— =
2 3
G(f) = are |S
Since the Fourier transform of e~@!"! is es then
at) =ees
5 [si a ue —2n|t| +(0)|
13. The filter obtained earlier is denoted by H(f) and given by (4.287) as H(f) = S*(f)e/2"/T /S,(f). The
filter h(t) and the pulse g(t) are related through h(t) = g*(T — 1) since h(t) is actually the filter matched to g(r).
Sec. 4.2 Optimum Waveform Receivers in Colored Gaussian Noise 255
1
t- 3 t t +5 0 nN,
(a) ts-1
2
1 1
(b) - 3 <t £5
1 1
Ouh= 5 tee 5 r
or
2 Az!
eO= = [so 16x VES]
where f(t) = fi(t) * fo(t) and
fae rel
Leases
fat) =
Q otherwise
In order to compute f(t), we need to consider three cases as shown in Fig. 4.41, namely
For
t < —1/2
256 Coherent Communications with Waveforms Chap. 4
1/2 :
1-1/2 a
For —1/2
<t < 1/2
0 t+1/2 1 e TE
Fort = 1/2
,o= {Pe ee
fale a
Therefore
e” VE ovE
&G) =9(——)
= g( == 126 (ec
er ae No + oon
6—
- VE
0) == ((14
g(0) — 12e7”) — ~ 13.5—
ar) No No
The function g(t) is plotted in Fig. 4.42 as a function of f. Since it rolls off exponentially, the
function can be truncated when it is implemented in a receiver. A “good” engineering judgment
would be to neglect (No/WE)g(t) when it is less than 0.1 (which is less than 1% of its peak
value of 13.5), that is
1 0.1
12 sinh (x) e602" #!.<'0.1 oe = ——1 ~ 1.15
Memes "1 = 5 ln sinh Gr) ee
Therefore, the receiver for all practical purposes observes r(t) over the interval
SSeS
Recall that the optimum receiver requires solving for g;(t) which is given by the
integral equation in (4.319). No attempt is made to derive the solutions to these equations,
but the key results are given and the reader is referred to Van Trees [13] or Helstrom
[16] for additional details. When the observation interval is finite, the noise autocorrelation
Rpt, t2) = E{n(t))n(t2)} can be classified as either rational or separable. In the rational
Sec. 4.2 Optimum Waveform Receivers in Colored Gaussian Noise 257
case, Ry(ti, 12) = Rn(t) — t2) = Rn(t) and assuming that the noise n(t) can be written as
the sum of two independent noises, that is, n(t) = n,(t) +n¢(t) to separate out the white
from the colored noise contribution, the PSD becomes
No
Sn(f) = o + Sc(f) (4.320)
where S-(f) is the PSD of n,(t) and is given by the ratio of two polynomials
N (f?)
Sf)
4 —— (4.321)
ee)
The numerator is a polynomial of order 2n and the denominator a polynomial of order 2m.
We assume that m — n > | so that n-(t) (colored noise contribution) has finite power. When
No = 0, that is, no white component, (4.319) reduces to
x
5;(t) =i) R(t — t)gj(t)dt (4.322)
0
which is a Fredholm equation of the first kind. The general solution is given by
2n m—n—1 k
dk d
Bi(t) = Boo(t) + 2 aight + aE, eo + G5 - r)| (4.323)
where 2.0(t) is the solution to the integral equation assuming an infinite observation interval
(obtained using Fourier transform techniques), gp.)(t) is the /‘” homogeneous solution to!*
(let p 4 d/dt)
No uf
si(t) = 5 8) +f Re(t — t)gi(t)dt (4.325)
0
which is a Fredholm equation of the second kind. Its general solution is given by
2n
SOLUTION
OBSERVATION WHITE NOISE GIVEN BY
INTERVAL KERNEL COMPONENT TECHNIQUE EQUATION
where A; and ¢;(t) are the eigenvalues and eigenfunctions of R-(t1, tz), then it can be shown
that
2 K ahi
gi(t) = No [si@) ~ Lint ENo/2% | ms as 1 (4.328)
0 elsewhere
where s; = ie si(t)@;(t)dt. Note that if we had allowed K = oo, all kernels would be
considered separable since we can always write
mi >r(t)=s(t)+n(t) te(O, T)
my —>r(t)=n(t) te(O, T)
and n(t) is a zero-mean Gaussian noise with R,,(t1, t2) = exp(—a|t, — t2|) for all t, tre (0, T).
Derive the decision statistic A.
Since R,(t) = exp(—a@|t|), the PSD of the noise becomes
2a
Sn(f) 3 a2 + (nf)?
or
Boolt)th=—
== [ars(t) ~ -5")]
1 2 y
= i
0
exp(—a|t — t|)q(t)dt = bye! + coe”)
] t
sara i
0
en [o%s(r) ~ s(t) |dt
] T
? eee
=e" |
as2a [o*s(r) ~ s"(r)| dt
Integrating the terms containing s(t) twice by parts, we get
1
s(t) = bee™ + coe®"—)) — 5, fe“las(0) — s'O)]
a
+ eT [as(T) +.s'(T)] — 2as(t) }
In order for this equation to hold, the coefficients of e®’ and e~® must vanish. Hence, by and cg
must satisfy
We rederive the optimum receiver in the presence of colored Gaussian noise using a fourth
separate approach. !> A received waveform r(t) is observed on a finite interval (0, 7) and
is known to have either the form!® r(t) = n(t) (message mo) or r(t) = s(t) + n(t) (mes-
sage mj) and s(t) is a known real function and n(t) is a zero-mean Gaussian noise with
autocorrelation R,(t, tT). Consider those signals s(t) that can be represented as
ih
s(t) =i] Kyi, teat, O=t=<T (4.330)
0
Consider an arbitrary set of orthonormal functions &)(t), €2(t), --- on the interval (0, T).
Add s(t) to the set. The augmented set {s(t), &1(t), &2(t), ---} is still complete and spans the
same space as the original set. Orthonormalize the augmented set using the Gram-Schmidt
15. This approach is in several references [13, 16, 17]. We follow the paper presented by Brown [18] in 1987.
16. The more general binary detection problem (mo — r(t) = so(t) + n(t) and m; > r(t) =s\(t) +n(t)) can
be handled by defining r’(t) = r(t) — 1 (t) since s\(t) is a known signal and s(t) = so(t) — 51 (t).
260 Coherent Communications with Waveforms Chap. 4
Note that since the set is complete, knowledge of the r;’s allows approximation of r(t) in
the mean-squared error sense with arbitrary accuracy. For j > 2, we have
V4
E{rirj;/mo} = elf neowj(ode f n(a)e(r)dr|
. (4.334)
=f i Ra(t, e(eyj (drat = | s(t)yj()dt =
and .
E{r;/m,}
= E{r;/mo} =0,- j >2 (4.338)
Next consider the cross-covariance terms of the sequence {r2, r3, ---} when either mo orm]
is transmitted:
Tare
Etriry/mo) = [ [ Ralt, tWit)wj(t)dtdt, i,j >2 (4.339)
Sec. 4.3 Inphase and Quadrature Modulation and Demodulation 261
10. Sh
E(rirj/my}=E |[0 [JO sosmvtovjeacar
[ey th
+ i;[ n(n (oil (e)drdr | (4.340)
0 JO
ih ap
=| / Rit, WpiOwj(t)dtdt, i,j >2
0 0)
Hence
E{rjrj/mo}
= E{rirj;/mi}, i,j >2 (4.341)
Therefore, we have shown that the projections {r2, r3, - - -} are statistically independent of ry
when mo or my is transmitted and that the means and cross-correlations among the terms of
the sequence {r2, r3, -- -} are identical under mo and my.
Forming the likelihood ratio test based on the finite set {7}, 72, ---, 7}, we have
ap Miata
sos NyTA) [Aol rg WE Ort m
(4.342)
f (ri, 172,°-+,1nw/mo)
where the joint densities are Gaussian under either hypothesis. Since rj and {r2, r3,---, rv}
are independent under both mo and m, then
_ f£ri/m1)
f(72, 73, +++, 1N/m1)
N= "FOrJo) fra T3. 4 Fw/m0) one
However, the joint densities of {r2,73,---,rn} are multivariate normal with identical
means and cross-covariances independent of which message was transmitted. Hence,
f(r2,173,-++,7n/mo)
= f(r2,73,-++,7nN/m) and
pecs ada) (4.344)
f(ri/mo)
Since Ay is independent of N, the same test is obtained in the limit as N — oo. Hence,
we conclude that rj = tos r(t)g(t)dt is a sufficient statistic for deciding between mo and
m,. This technique can be expanded to M messages and results in the optimum receiver
depicted in Figures 4.34 and 4.40(c) as expected.
referred to as the inphase (I) carrier and the sinw,t as the quadrature (Q) carrier. As an
example, antipodal signals are constructed by setting Q(t) = 0 and /(t) = +p7(t) for 0 <
t < T. Alternatively, /(t) can be written as
where ay = +1 representing the binary random data and p;(t) is the baseband inphase pulse
of duration T seconds. A specific case of antipodal signals is BPSK where
il Osy ssh
= aes 4.347
Pit) ts otherwise ( )
where 6; = 0 or z radians, depending on the data polarity. Another example is QPSK where
both J(t) =+p/)(t) and Q(t) =+p/)(t) for kT <t < (k + 1)T, independent of each other.
Alternatively, QPSK can be represented by (4.345) with J (rt) given by (4.346) and
where by; = +1 representing an independent binary data stream. For kT <t < (k+1)T,
s(t) can be written as
COHERENT
PHASE
REFERENCE
Figure 4.43 Functional diagram of the I-Q modulator/demodulator using real signals:
(a) I-Q modulator/signal generator, (b) I-Q demodulator/signal detector
r(tyri(t)inp = VP/21(t)
(4.355)
r(t)ro(t)|zp = ¥ P/2Q0(t)
which for kT <t < (k + 1)T, reduce to the js signaling baseband pulse s;(¢), assuming that
m; is transmitted. The corresponding I-Q modulator/demodulator is depicted in Fig. 4.43(b)
using real signals.
I(t) +) Q()
i(@ct + 8)
e
Sometimes, it is more convenient to use complex notation to denote the I-Q signal.
Letting m(t) denote the complex baseband modulation, that is, m(t) = /(t) + j Q(t), then
s(t) can be written as
Ge ai Say ek a b w
1 1 i 0 1 -3 W- — 3wa
1 -1 1 | -1 Wc — Wd
1 -1 1 2 1 Wc + Wa
1 -1 -1 3 l 3 We + 304
-1 1 1 0 -1 -3 We — 3@¢
-1 1 -1 1 -1 -1 We — Wd
-1 -1 1 2 -] 1 Wc + @d
-1 -1 -1 3 -1 3 Wc + 3w¢
by similar noises and as a result, the receiver provides the same performance in both cases
as long as B >> 1/T.
Example 4.6
One realization of a biorthogonal set of M signals is a combined frequency/phase modula-
tion (FPM) wherein N = M/2 frequencies uniformly separated by 2 fg = k/2T, (k an integer)
are each BPSK modulated, that is, N-FSK/BPSK. Typically, N = 2” (n integer) and the fre-
quencies are symmetrically placed around a carrier, that is, they are located at fe + fu, fet
3 fa,.--, fe + (N — 1) fa where f. = w,/27 is the carrier frequency in Hertz. Such a signal-
ing set is generated as follows. Consider as a data sequence a binary (+1) random sequence
{dj} of rate 1/7) [equivalently a rectangular pulse data stream d(t)]. Blocks of A = log, M =
1 + log, N bits of duration 7; from sequence {d;} are assigned to FPM signals of duration
T; = XT). We specifically use 7; to denote the duration of a symbol so as to be consistent with
what is done later in Chapter 10 where this example is treated as a special case of a more general
form of modulation. The first bit in each of these blocks is assigned to a sequence {a;} that is
used to define the BPSK portion of the biorthogonal modulation. The remaining A — 1 = logy N
bits in each block are assigned to an N-ary [+1, +3,...,+(N — 1)] sequence {b;} which is
used to define the N-FSK portion of the biorthogonal modulation. The coding rule that maps
{d;} into the binary amplitude sequence {a;} and the N-ary frequency sequence {b;}, respec-
tively, is described as follows.
For a given block of A bits of the input sequence, say {do, dj,...,dn}, the first bit (do) is,
assigned directly to the amplitude level a , that is, a = do. The remaining A — | bits in the block
(di, ..., dy) are represented as an N-ary number, say k, in accordance with the natural mapping
rule which is then associated with the frequency number b = 2k — (N — 1) and results in the
actual transmission of the frequency f. + bfa. Note that if a block of 4 bits corresponds to b,
then its complement (negative) also corresponds to b. An illustration of the above coding rule
for N =4 (M = 8) is given in Table 4.4.
Defining the random pulse trains a(t) and b(t) from the sequences {a;} and {b;}, respectively,
foe)
for example, a(t)= >> ajp(t —iT;) with p(t) a unit rectangular pulse, then an FPM signal
i=—0co
is characterized by
Using the methods presented in Chapter 2, it is straightforward to show that the equivalent
lowpass two-sided power spectral density of FPM with minimum frequency spacing 1/27; is
given by
266 Coherent Communications with Waveforms Chap. 4
JB
S(/T,,
dB
S(f/T,,
PI Aan i N-1
(N==z X sinc” (1
S(f) = — | fT;tana
+ = — —— )
In coherent demodulation, the carrier phase should be known precisely. Assuming that the
references r;(t) and rg(t) are not ideal, that is
AG = V2 cos(wet +6)
sf (4.358)
ro(t) = —V2sin(wet + 6)
where @ is the receiver’s estimate of the incoming carrier phase 6, then (4.355) becomes in
the absence of noise
r(t)ri(@)|zp = ¥ P/2[1(t)
cos @— Q(t) sing] 4 I(t)
(4.359)
r(t)re(t)|zp = ¥ P/2[1(¢) sing + Q(t) cosg]4 Qo(t)
where ¢ = 6 — @ denotes the carrier phase error, which is the phase difference between the
received carrier and that generated locally by the carrier synchronization circuitry. Note that
a nonzero carrier phase error results in cross-talk between the I and Q arms, that is, the
observed I channel, /o(t), is a combination of the desired I and Q channels. Similarly for the
observed Q channel, Qo(t). If 6 = 1/2, then
I(t) = —v P/2Q()
(4.360)
Qo(t) = vyP/21(t)
and the I and Q channels are reversed in addition to a polarity inversion in /o(t). If on the
other hand @ = 7, then
I(t) =—JVP/21()
(4.361)
Qo(t) = —-vV P/2Q(t)
and both J and Q channels are inverted, but not reversed. Using complex notation, this
phenomenon can be easily expressed using
The true measure of degradation is the average probability of error P(E) obtained from
10°?
10°3
where f(@) is the pdf of the phase error that depends on the carrier synchronization (sync)
algorithm employed. Assuming that f(@) is of the form
exp(og00s) 14) <
f@= | 2m 1o(eg) ” (4.365)
otherwise
where pg 4 Wor then using (4.365) in (4.364), P(E) can be evaluated as a function of
E/No for various values of pg. The results are plotted in Fig. 4.46 illustrating the noisy
reference effect on system performance. For pg = 12 dB at P(E) = 10~°, the required
increase in E’//No is about 2 dB.
Assuming that perfect carrier synchronization is achieved with BPSK signaling, the base-
band received signal in the presence of noise is given by (4.357) with P replaced by 2P
[since O(t) = 0]
where 5 denotes the bit epoch which needs to be estimated by the bit synchronization
Sec. 4.3 Inphase and Quadrature Modulation and Demodulation 269
circuitry, a, = +1, and p(t) is the baseband pulse (with power P). 20 Consider the situation
where the bit sync is in error by « = 6 — O=AT sec, where 6 denotes the estimate of the
bit epoch. The detector statistic on which a decision is made for the n" transmitted bit
now depends on the received data in the time interval (n — 1)T +6 <i Sh aN Stated
mathematically
nT+6 (ea
an = sgn / r()ri@)\~p pit — (n — 1)T — djdt (4.367)
(n—1)T+6
Gn = sgn O (4.368)
and
If
We note that in general, three input bits are involved in estimating the n" transmitted bit.
Actually, if A is known to be positive (or negative), then only two bits are involved, namely
the n" and one of its adjacent bits. Denoting the signal and noise components of Q by Q,
and Q,, respectively, then
where R,(rt) is the signal pulse autocorrelation function given by (4.33) and repeated here
for convenience
ie
With regard to the input bits affecting the calculation of Q;, one of four situations can occur:
either the n" bit and its adjacent bit are alike (both positive or both negative) or they are
unlike (negative followed by positive, or vice versa). Hence, the overall error probability
conditioned on knowing A is given by
2
P(E/A) = ) [Prob {n'® and adjacent bits = (—1)*}
k=
x Prob {error/n' and adjacent bits = (—1)*} (4.372)
+ Prob {n'® bit = (—1)* and adjacent bit = (—1)**"}
x Prob {error/ n'” bit = (—1)* and adjacent bit = (—1)*+"}]
Since the random variable Q is conditionally Gaussian on a,—1, Gn, Gn+1 and fixed A,
with variance
=
See
T =
rs(T) 0 int (4.376 )
= “el
Fray es
erf(x) exp[—(x oi/2a)*\dx, |A| < 1/4
Sec. 4.3 Inphase and Quadrature Modulation and Demodulation 271
where
a A VE/2No(1 — 41a)
b A JE/2No (4.381)
c A /E/2No(1 — 2|A))
The average unconditional error probability P(E) is determined from
Amax
where, as before, f(A) is the probability distribution of the normalized sync error and
Amax is the maximum value for which A is defined; for example, Amax = 1/2 for NRZ and
Amax = 1/4 for Manchester, RZ, and Miller-coded data. The pdf, f(A), depends on the way
in which bit sync is derived at the receiver. For the purpose of comparing the average error
probability performance of the various data formats, we postulate a Tikhonov pdf for f(A)
which is characterized entirely in terms of the variance o of the sync error. Thus, for NRZ
data [let p, 4 1/(270,)7]
exp (~, cos(27 i))
fa) = » [Al<1/2 (4.383)
1o(px)
whereas for baseband signaling that employs transitions in the middle of the symbol inter-
val, f(A) is characterized by
10°2
P(E)
10°3
10°
no data transition occurs, the bits maintain their nominal T-sec value. An illustration of
this distortion for a typical sequence is given in Fig. 4.51. Assuming an integrate-and-
dump (matched) filter, then for random NRZ data (50% transitions) with asymmetry, the
integrate-and-dump output depends, in general, on the polarity of the bit over which it
is integrating and that of the preceding and succeeding bits. As such, one must compute
the integrate-and-dump output for each of the eight possible three-bit sequences and their
corresponding conditional (on the particular sequence) error probabilities. Then, averaging
these conditional error probabilities over the eight equally likely three-bit sequences gives
the error probability performance in the presence of asymmetry.
Defining the measure of data asymmetry, 7, by the difference in length between the
shortest and longest pulses in the sequence divided by their sum, that is,
T(i+4)-T(i-4) A
T(+9)+T0-3)
4
2 a4.385
then the average error probability performance described above is given by [20]
5 LE l YS Per |E
Sec. 4.3 Inphase and Quadrature Modulation and Demodulation 273
10°
10" § -|
i
0.10
ie = 0.07 #5
6, = 0.06
a J
0.05 4
10°3
\\ 0, = 0.04 |
10
: “he
1A
| P(E)= 1/2erfc\/E /N,
2 4 6 3 10 12
4
{4
Figure 4.48 Average probability of error
versus signal-to-noise ratio with standard
deviation of the symbol sync error as a
E/N dB parameter (Manchester code) (Reprinted
o’ from [12])
For a given value of n (typically measured in percent), one can evaluate the SNR degrada-
tion by computing the additional E'/ No required (A E’/No) due to asymmetry to produce the
same value of bit error probability in the absence of asymmetry (7 = 0). This degradation
(in dB) is given by
E/N
AE/No = 10 logo (=o (4.387)
where for a given error probability P(£) and data asymmetry n, E/No is computed from
(4.386) and (E/No)¢ is computed from (4.386) with 7 = 0, namely, the value corresponding
to ideal coherent BPSK performance. Table 4.5 gives these degradations for error probabil-
ity values between 10~7 and 10~®°, and asymmetries in the range of 5% to 25%.
Inherent in the above discussion was the assumption of perfect bit synchronization
which in the presence of data asymmetry is tantamount to having a clock with a misalign-
ment equal to half the asymmetry. If in addition to data asymmetry we include the effect
of imperfect bit synchronization (additional clock misalignment) discussed in Section 4.3.3,
then the conditional error probability in the presence of both imperfections is given by [20]
Coherent Communications with Waveforms Chap. 4
274
10°
10°2
P(E)
10°3
Serte
/E +terfe(/% —n-210)
+ erfc (/za pets 21A1))
P(E =} +rherte
(\/£ (1-20): O<|al<% (4.388)
berfe /£ + ferfe (/£ (1 -n 21)
+herfe (EU
—n +210); P<|al<3}
where as before AT denotes the misalignment of the bit sync clock relative to its nominal
position. Note that for 7 = 0, (4.388) reduces to (4.377) as it should.
The average probability of error, P(E), is obtained by averaging (4.388) over the pdf
of the bit sync misalignment. Postulating, as in Section 4.3.3, a Tikhonov pdf for 4, which
is entirely characterized by the single parameter ao? [see (4.383)], then P(E) is given by
(4.382) where (4.388) is now used for P(E|A). Figure 4.52 illustrates P(E) versus E/No in
dB with oy as a parameter and a fixed value of asymmetry equal to 6%.
Manchester Data. The signal model (Model | in [20]) used to characterize Manch-
ester data with asymmetry assumes that for a +] data bit, the first half is elongated by AT /4
(relative to its nominal value of 7/2 sec) and for a —1 data bit, the first half is shortened
Sec. 4.3 Inphase and Quadrature Modulation and Demodulation 275
10°
10°2
P(E)
10°3
by the same amount. When a data transition follows, the second half of the Manchester bit
retains its nominal (7 /2-sec) value. When no data transition occurs, the second half of the
Manchester bit adjusts itself so that the total bit maintains its nominal T-sec value. An illus-
tration of this distortion for a typical sequence is given in Fig. 4.53. Assuming an integrate-
and-dump (matched) filter, then for random Manchester data (50% data transitions) with
asymmetry, the integrate-and-dump output again depends, in general, on the polarity of the
bit over which it is integrating and that of the preceding and succeeding bits. Thus, as for
NRZ data, one must compute the integrate-and-dump output for each of the eight possible
three-bit sequences and their corresponding conditional (on the particular sequence) error
probabilities. Then, averaging these conditional error probabilities over the eight equally
likely three-bit sequences gives the bit error probability performance in the present of asym-
metry.
Defining the measure of data asymmetry, n, relative to the bit time 7 as was done for
NRZ data?! then analogous to (4.385) we have
21. By defining data asymmetry relative to the symbol time for both NRZ and Manchester data, equal percent
asymmetry implies equal amounts of asymmetry (in seconds) as measured by the actual time displacements of the
waveform transitions.
Coherent Communications with Waveforms Chap. 4
276
(a)
Figure 4.51 (a) Symmetric NRZ data stream; (b) Asymmetric NRZ data stream (Reprinted from [20])
AE / N,, dB
ee
T(i+4)-T(1+4)
ee ee
a 4.389
"“ T(i+4)+7 (1-4) .4 vee
and the average bit error performance becomes [20]
l E 1 sh
BAB) = 17otic (\550-20) + pene([F¢-9) (4.390)
Figure 4.54 illustrates a comparison between the SNR degradation due to data asymmetry
for NRZ and Manchester signals. We see that for analogous definitions of asymmetry as
in (4.385) and (4.389), the Manchester signal always yields a larger SNR degradation for
a given percent asymmetry as one might anticipate since, on the average, the Manchester
waveform has 3/2 as many transitions as the NRZ waveform.
When both data asymmetry and imperfect bit sync timing are present, then the analo-
gous result to (4.388) becomes
Sec. 4.3 Inphase and Quadrature Modulation and Demodulation 277
10°
1071
10-2
10°3
P(E)
10-4
10°5
10-6
gerte(\/F (1-2n))
+g erfo(,/# —n—21a0)
+4 erfe(\/£(—n +21A)) O<|al<3
PCE|X)= (4.391)
herte(\/#( a4 AD) +3 erfo(/ #1 =r? ID)
which reduces to (4.378) when n = 0. Averaging (4.391) over the pdf of the bit sync
alignment”? modeled as a Tikhonov pdf [see (4.384)], P(E) is given by (4.382) where
(4.391) is now used for P(E|A). Figure 4.55 illustrates P(E) versus E/No in dB with
0), as a parameter and a fixed value of asymmetry equal to 6%. Comparing Fig. 4.55 with
Fig. 4.52, we observe that for fixed values of n and A, the error probability performance of
Manchester data is always worse than that for NRZ.
We conclude this chapter with a discussion of antipodal signaling with rectangular pulses
but using an arbitrary filter H(f) instead of a matched filter (I&D) as in Fig. 4.56. The
ae F(t 5) aC 5) if F (1-3)
Dero! yao GY Ear A 2 2
Figure 4.53 (a) Symmetric Manchester data stream; (b) Asymmetric Manchester data stream
(Reprinted from [20])
Here the a,’s are the independent +1 data bits and p(f) is the rectangular pulse
noe ee
0,
US
otherwise
(4.393)
Zero Signal Distortion. Assume that |H(f)|max = 1 and that H(f) has a single-
sided noise bandwidth By that is sufficiently wide so as to pass m(t) “undistorted.” Typi-
cally, a By of 3 to 4 times 1/T is sufficient. Thus, after passing through the filter H(f), the
signal plus noise is
E
E* {y(T)} = * (4.396)
Sec. 4.3 Inphase and Quadrature Modulation and Demodulation 279
E/No
=1.5 dB
MANCHESTER
(n=A/4 =6)
wo
nw
AE/No,
DEGRADATION
SNR
dB
NRZ
(n= A/2 =26)
P(E)=Q (= (4.397)
Since for H(f) a matched filter, By = 1/2T and SNRmatched = 2E/No, then the SNR
penalty of using other than a matched filter (assuming undistorted signal) is
SNR
10 logy (Fe) = 10logi)2ByT (dB) (4.398)
Clearly as By get smaller, the distortion of m(t) increases and the mean of y(T)
becomes less than E/T. The value of the mean of y(7T) for any specific sample time
t =nT depends on the past sequence of a,’s, that is, the previous bits interfere with the
mean of the sample corresponding to the present bit. This phenomenon is called intersymbol
interference (ISI). In general, the effect of ISI is difficult to compute since the distortion
depends on the infinite past. Typically, we make some approximation, that is, we assume
Coherent Communications with Waveforms Chap. 4
280
10°
1071
10-2
10-3
P(E)
10°5
10-6
that the filter has finite memory, that is, the distortion of the mean of the filter output sample
for a given bit depends on a finite number of previous bits. We now present an example to
illustrate this point.
Worst Case Design—Upper Bound on Performance. Let H(f) be a single pole
Butterworth filter, that is
] 4
HU) = sept Bw = (=) fe (4.399)
20 fe
Assume now that a sequence of all —1 data bits preceding and succeeding the data bit
of interest which is a +1. This input modulation m(t) (assuming rectangular pulses) is
illustrated in Fig. 4.57. Assuming that y(t) has value —1 just prior to the bit interval of
interest, then during this interval we have
E
E{y(t)}= We {-1+2(1 —e~*")} = mets — 2e~@") (4.400)
The value of the mean of y(t) at the end of the pulse interval of interest is
Sec. 4.3 Inphase and Quadrature Modulation and Demodulation 281
n(t)
DECISION
m(t) J aR
DECISION
mit)
y(t)
Vet-—-p-------=-------
ey Ge Ae E eis:
eae ( oy )=¢ (,T NoBn (1 He )
(4.402)
a E (1 —2e744n7)
~2\V No VENT
We wish to choose By so as to maximize the argument of the Gaussian tail integral, that
is, produce the best trade-off between ISI and noise power in so far as minimizing error
probability. Differentiating the argument of the Q function in (4.402) with respect to ByT
gives
m(t)
y(t)
The corresponding worst case error probability is determined by substituting (4.404) into
(4.402), resulting in
E
P(E) =O (iis] (4.405)
\ No
which when compared with the matched filter result represents a —10 log; 9(1.118/2) = 2.5
dB SNR loss. We now reexamine the design of the same receiver structure in light of a
minimum average error probability criterion.
Average Error Probability Performance Design—Finite Memory of Past Two Sym-
bols. Instead of considering the worst case ISI sequence, we now assume a random data
sequence as shown in Fig. 4.58. We shall make the following approximations to the response
of H(f) to m(t): At any potential data transition point, there is either a transition or no tran-
sition, each occurring with equal probability 1/2.
1. Assume that after two pulse intervals, the response achieves its maximum value. Thus,
if we have ax, = ax—1, then the mean of y(T) equals /E/T.
2. If there is a transition in the waveform, that is, a, # ax—1, then the preceding bit, that
is, ax—2 will either be identical to ay,_, or opposite to ax_}:
(a) ag—2 F ag-1
This case is illustrated in Fig. 4.59(a) and we shall assume that the sequence
has alternated continuously into the infinite past. Then, to determine the mean of
y(T), we have
E
=F lct (= ev) (1esec) = E{y(T)} (4.406)
or
LEU Seed)
E{y(T)} ~( eS (4.407)
m(t)
y(t)
y(t)
This case is illustrated in Fig. 4.59(b). Here, we assume that the response starts
from —./ E/T. Thus
E{y(T)} = (a eect
yee
E —w-T
) (4.408)
Case | occurs with probability 1/2 (half the time, there is no transition). Cases 2a
and 2b each occur with probability 1/4 (half the time, there is a transition. Half
of that time, the two preceding bits are identical and half of that time, the two
preceding bits are opposite). Thus, the average error probability is
ast |
# 50 (ah (1a a)
Again, one can choose ByT to minimize P(E) by first rewriting (4.409) as
284 Coherent Communications with Waveforms Chap. 4
utanm/M
P,(E) =1— = [ exp {-(ua VEIN) || exp (-»”) audu (4A.1)
JO 0
(E)
P,(E) =1-——
I pet (VEN) E;/No, d
a (4A.2)
4A.2
where
(z+x)
tan 1 /M
f (x,z) = exp(—z”) / exp(—v7)du (4A.3)
0
Taking the partial derivative of P;(E) with respect to ./E;/No, we get
ee ie Af VESTN: 2)4,
a | J—/E,jNo 9V/Es/No
From (4A.3)
a 12
Z) _ 2 1A
Ae on
Of (VSEs/No, = exp (-z )tan uM exp |-{(zef Es/No) tan =} (4A.5)
M4 Es 4
= tan = exp|- |2sec? © +22 E, /No tan? WE wn? TI
Of (/Es/No,Z) _ | Es i |
== (EN (2.4 0) 4
0./Es/N M No M
ioe: (4A.6)
2
x exp {-sec” 7 (z+ /Es/No sin® —)
— Bio ® 7aS
=-(- )tan(= exp aha
Es/No M
2 (4A.7)
lee) E,; I
x ik expe sec? 24+ |= sin’ — dz
-/E,/No oe)
286 Coherent Communications with Waveforms Chap. 4
= ~
1 E
sin a exp \-= sin? at [P+ erf(./ Es /No cos =)
1
where m(t) denotes the complex baseband modulation function, consisting of both ampli-
tude a(t) and phase 6(t) modulations, that is
5, (0)
x(t)
Consider the processing depicted in Fig. 4B.2 where the output z(t) of the linear time-
invariant (LTI) filter consists solely of the positive frequency content of the input x(t) by
defining U(f) as
A eel
’ =kena
U(f)S 0, otherwise eae
Since the filter is LTI, the output z(t) is also WSS with PSD
S:(f) == |2U(f)Sx(f/)
|2 a (4B.8)
=4S.(f — fe)
The autocorrelation function of the modulation m(t) is defined as Rym*(t1, t2) = E{m(t))
m*(t2)}. Using m(t) = z(t) exp(—j2z fet), then
m(t)
= mR(t) + jm7(t) (4B.11)
where both mr(t) and m;(t) are real functions. In order to compute second-order statistics
of mr(t) and m;(t), we need an expression for Rz-(t), t2) 4 E{z(t;)z(t2)}, which is dif-
ferent from the autocorrelation function R,,«(t), t2) for complex processes. Consider Fig.
4B.3, which depicts a real WSS process x(t) through a complex LT] filter h(t). The output
z(t) becomes
COMPLEX
LTI FILTER
resulting in
Reha / / h(a)h(B)E
(x(t —«)x(t2 —B)\dodB
See (4B.13)
- / / h(a)h(B)Rex(th— ty — a + B)dadp
Note that R,-(t), t2) is a function of t; — tz only. Therefore
where F.T.{ f(t)} denotes the Fourier transform of f(t). Referring back to Fig. 4B.2,
H(f) =2U(f) and
Sez(f) = S.(f)2U (f)2U (—f) =0 (4B.16)
since U( f)U(—f) is nonzero only at f = 0, but S,(0) = 0 since it is a bandpass function.
Hence, R,,(t) = 0 and
Noting that
Rinm*(t1, 12) = Rig (th, t2) + Ry (th, #2) + JRmimp(t, 12) — JRmem,
(th, t2) (4B.18)
Rim(t1, t2) = Rmg (ti, 12) — Rm, (A, 12) + J Rmpmg (ts 12) + J Rmeym) (th, 22) =90 (4B.19)
we have by adding (4B.18) and (4B.19)
1
Ring(h, 2) = Ret Rmme(t1, t2)} (4B.20)
1
Rinpmg (th, 12) = 5Im{Rmm*(t1, t2)} (4B.21)
where Re(x) and Im(x) denote the real and imaginary parts of x, respectively. Subtracting
both equations, we obtain
1
Rm, (tH, t2) = 5 Ret Rms (ti ,t2)} (4B.22)
1
Rm gm;(thy #2) = — 51m{Rmm*(t1, t2)} (4B.23)
290 Coherent Communications with Waveforms Chap. 4
Hence
and
From (4B.10), Sn(f) =4Sz(f) and since Rinm*(T) = pee Sn(f) exp(j2aft)df, (4B.22)
and (4B.23) become
where mp(t) and m;(t) are zero-mean jointly WSS real random processes with second
order statistics given by (4B.26) and (4B.27). Note that if S;(f) is symmetric about f = 0,
then Rmpm,(t) = 0 for all t and mr(t), m;(t) are uncorrelated. Furthermore, if we add the
constraint that
where n-(t), ns(t) are uncorrelated random processes with PSD as depicted in Fig. 4B.5(b),
that is
1. By writing m(t) = mp(t) — jm;(t), x(t) can be expressed as x(t) = m,(t) cos(2z fet) + my(t) sin(27 fet)
with Rmgm,(T) = 2 f°>, SL(f) sin(2xft)df.
App. 4B Bandpass Random Process Representation 291
x(t)
-sin2nfot
Figure 4B.4 Baseband recovery: (a) Complex implementation, (b) Real implementation
In some instances, it is more convenient to represent the noise n(t) with an additional = 5)
factor as
Nc
Su f) = Sy (f) = {2: fiaeblalge
otherwise
(4B.34)
to produce P, = 2NoB as in (4B.32). It is often convenient to express the noise n(t) in its
complex form as
where a(t) and @(t) are the envelope and phase noise processes given by
292 Coherent Communications with Waveforms Chap. 4
5, ()
0
-f, fo f
10 = Sn
No
PROBLEMS
4.1 Consider a binary “on-off” communication system which is used to send one of two equally
probable messages in the face of AWGN (with one-sided PSD No) during the time interval
(0, 7). Assume the “on-signal” has energy E.
(a) Assume N = 2 (dimension of vector space). Construct the vector representation and illus-
trate the optimum vector receiver. Find the probability of error.
(b) What waveform design is compatible with the vector in (a)? Illustrate the optimum wave-
form receiver.
4.2 Show that in an AWGN channel, when the average signal energy is constrained not to exceed
E, that is
T T
Pimo) | so(t)dt + pom) [ s(t)dt <E
0 0
that the optimum signals are given by
Problems 293
Figure P4.3
P(m)
a= S@), OSes
P(mo)
T
an s*(t)dt
0
4.3 Consider the following binary communication system. The information is sent using one of
the two signals shown in Fig. P4.3, which are antipodal, nonreturn-to-zero (NRZ) waveforms.
Assuming equilikely signals and an AWGN channel with two-sided PSD No/2, design the
optimum receiver, using the minimum number of correlators, and compute the probability of
error for the case A = 2 volts, T = 1 sec, and No = 2 watts/Hz.
4.4 Consider the pulses depicted in Fig. P4.4.
(a) What is the impulse response for filters matched to pulses (a) and (b)?
{b) What is the output of the matched filters if they are excited by the signals to which they are
matched?
(c) Make comments on the difference in results. Suggest a practical implementation.
4.5 Consider an amplitude-weighted binary FSK system described by the pair of signals
si(t)
= A(t) cos 4 | a + (-1)
i+) 5
A@ Phen
;
Theat Wesyesve
yofj4 ae So(t)sy(t)dt
Io ss(dt fy spat
vs
(b) If the weighting is exponential, that is, A(t) = Aexp(—Wr), evaluate yo; for the case
where the frequency separation Af = Aw/2z7 is an integer multiple of the data rate 1/7.
(c) Assuming an AWGN channel and an optimum matched filter receiver, what is the error
probability of the system for the amplitude weighting of (b)?
(d) Is the error probability found in (c) better or worse than that for an orthogonal binary FSK
system with A(t) = A?
4.6 For coherent FSK, the transmitted signal in the n" transmission interval nT <t < (n + 1)T is
Coherent Communications with Waveforms Chap. 4
given by
Ao
s(t) = Acos out+ an p(t — nt) |
with a, taking on values +1 with equal probability. Conventionally, p(t) is the unit rectangular
pulse
oe i 0Se= 7
PX’) 0, otherwise
in which case the two possible signals (corresponding to a, = —1 and a, = 1, respectively), are
given by
A@
so(t) = Acos (vx= =) |
2)
Aw
s1(t) = Acos |
(vx
ate =) |
Suppose now that the pulse shape p(t) is the unit square wave
OSRSi2
PM
(t)=a L
ae)Cana fae Pace &
Figure P4.11
4.7 The received signal in a binary baseband communication system takes on values +A in each T-
sec interval. The probability of transmitting +A is 1/4 and the probability of transmitting —A
is 3/4. An integrate-and-dump filter is used for detection with the comparison threshold set to
Vo. Assume an AWGN channel with single-sided PSD No.
(a) Derive an expression for the average error probability, P(E), as a function of the detection
threshold Vo.
(b) What is the value of Vo that minimizes P(E)?
(c) What is the corresponding minimum value of P(E)?
4.8 A binary communication system (AWGN with 2-sided PSD No/2) transmits a binary one by
sending either a waveform /)(¢) or fo(t) (energy E, time T), selected by tossing a coin, and
sending nothing for a binary zero. Assuming equally likely bits, sketch the decision mechanism
that minimizes bit error probability.
4.9 (a) Obtain the maximum output signal-to-noise ratio of a matched filter for the following input
signals: A and J, are constants. Assume that the two-sided noise PSD at the input is No/2.
(1) x1(t) = A{u(t) — u(t — 5T¢)}
QO) @) Alu@) — 24 = sic) u(t — 47) — ut — 57,)}
where u(t) denotes the unit step function.
(b) Find the amplitude of an input signal pulse of duration 7, that will give the same maximum
output signal-to-noise ratio as the signals in part (a).
(c) Obtain the autocorrelation functions of the signal in parts (a) and (b). Comment on the
utility of each for resolving delay or relative time differences.
4.10 For the binary coherent communication problem, show that the performance of the optimum
receiver for the unequal energy, unequiprobable signal case is given by (4.64), that is
4.11 Either of the two signals shown in Fig. P4.11 is transmitted with equal probability in a binary
communication system.
(a) Find the matched filter impulse response h(t) that would be required to implement the
optimum receiver in an AWGN channel.
(b) Calculate the optimum threshold and the error probability performance of the receiver as a
function of fo.
(c) What is the best choice for fo such that the error probability is minimized?
(d) Sketch a correlator receiver for these signals.
296 Coherent Communications with Waveforms Chap. 4
g(t),
i=0,1
N\
r(t)| WAVEFORM | ™
DECISION
DELAY T/2
Figure P4.12
4.12 Consider the binary communication system shown in Fig. P4.12. Assume the signals are equi-
likely and that n(t) is AWGN with PSD No/2 watts/Hz.
(a) Illustrate the optimum correlator receiver and the optimum matched filter receiver.
(b) If s;(t), i =0, 1 were antipodal signals, evaluate the probability of symbol error.
4.13 A ternary communication system transmits equally likely messages mo, m;, and m2 using three
orthogonal signals with equal energy. The channel adds white Gaussian noise with one-sided
PSD No.
(a) Define vector and waveform signal models.
(b) Illustrate the optimum vector and waveform receivers.
(c) What is the message error probability in terms of E;/No?
(d) What is the minimum number of correlators required to implement the optimum receiver?
(e) Illustrate the optimum waveform receiver in (d) and define all waveforms.
4.14 Consider the following hypotheses testing problem:
fori = 1,---,
8 where
ane A . E
5;(t) =V2E sin (1oS)
This signal set is known as 8-PSK. Assume n(t) is am AWGN with PSD No/2 watts/Hz.
(a) Choose a suitable orthonormal set for the signal space. What is the dimensionality of this
space?
(b) Indicate the optimum decision regions and use the union bound to find an upper bound on
the probability of symbol error.
4.15 A waveform signal set is described by
20
Si(t) = A cos(@ct +1 ae VSS EF
for i = 1,---, 6. Assume all signals are equilikely and that the signals are transmitted over an
AWGN channel.
(a) Find a suitable set of orthonormal basis using minimum number of dimensions. Sketch the
signal vectors and indicate the decision regions for optimum reception.
(b) Use the union bound to find an upper bound for the probability of symbol error.
4.16 Consider the problem of communicating over white Gaussian noise channel (No = 1/2). One
of four possible waveforms is transmitted every second. The source messages are equilikely
and independently chosen. The transmitter has a maximum energy constraint defined by:
|Es| <3
(a) What is the optimum choice for the set of four signals in this situation? Describe the set as
vectors.
(b) Convert the set of vectors in (a) to waveforms.
(c) Draw a block diagram for the optimum waveform receiver. Use the smallest number of
correlators (matched filters).
Problems 297
90°
1352 45°
AMPLITUDE = yo
180° 0° ABSOLUTE
(d) Using the result from the union bound, what is an upper bound on symbol error probabil-
ity?
4.17 Consider the 4-ary hypotheses testing problem. Each observation is a 3-dimensional vector
Ld en — Siete ta(al aS
(GC) (4
where
The components of the noise vector are independent identically distributed Gaussian random
variables, with variance No/2.
(a) Choosing an appropriate basis, express the signal vectors as waveforms.
(b) Sketch the optimum receiver assuming equilikely signals using the least number of corre-
lators.
(c) Compute the probability of symbol error.
4.18 Consider the V.29 signal constellation depicted in Fig. P4.18, which transmits 3 bits/symbol.
Assume an AWGN channel and equilikely signals.
(a) Write an I-Q waveform model using rectangular coordinates for square and arbitrary I-Q
pulses.
(b) Write an I-Q waveform model using polar coordinates for square and arbitrary I-Q pulses.
4.19 Consider an M-FSK system that employs the signal set (k an integer)
k
sit) = V2P sin |( + aniaf)), PS Oe laake lt 1
(a) Find the minimum frequency separation spacing that produces orthogonality and the effec-
tive bandwidth occupancy of this set.
(b) Using this orthogonal set, define a biorthogonal signal set and find its effective occupancy.
(c) Repeat for a transorthogonal signal set.
4.20 Consider the V.29 signal constellation depicted in Fig. P4.20, which transmits 4 bits/symbol.
Assume an AWGN channel and equilikely signals.
(a) Write an I-Q waveform model using rectangular coordinates for square and arbitrary I-Q
pulses.
(b) Write an I-Q waveform model using polar coordinates for square and arbitrary I-Q pulses.
298 Coherent Communications with Waveforms Chap. 4
Figure P4.20
4.21 Consider a ternary communication system which transmits one of three equally likely signals
EQ TAC Sie),
Ay: r(t)=Asinw.t
+ n(t)
during 0 < t < T. Assume that n(t) is a white Gaussian noise process with PSD No/2 watts/Hz.
(a) What is the optimum receiver which minimizes the error probability? Sketch this mecha-
nization and show all operations which must be performed in making a decision.
(b) What are the conditional probabilities of correct reception when Ho, H), and HA are true?
(c) Using the results found in (b), write an expression for the average probability of error.
(d) Compare this ternary system with the binary PSK system which uses Asinqw,t and
—A sin wet to transmit data.
4.22 Establish (4.129) by substituting (4.127) into (4.128) and averaging over the equal a priori
probabilities of the transmitted signals. Show that (4.129) reduces to (4.130) and the average
symbol error probability is given by (4.131).
4.23 For large signal-to-noise conditions, Gilbert” has found an approximate formula for the average
symbol error probability of a maximum likelihood receiver in terms of the geometrical config-
uration of the transmitted signals in signal space. Applying his formula to an M-PSK system,
the M signals are represented as uniformly spaced points on the circumference of a circle of
radius R and
where rmin is the smallest of the set of M(M — 1)/2 Euclidean distances between pairs of signal
2. E. N. Gilbert, “A Comparison of Signaling Alphabets,” Bell System Technical Journal, vol. 31, May 1952, pp.
504-22.
Problems
299
points, and N, is determined as follows: For each signal s;(t), let Ns, denote the number of
signals that are at the minimum distance rmin. Then
M
Ne asi Ns;
——————
M
(a) Show that rmin = 2R sina /M.
(b) For M > 2, show that N, = 2.
(c) From the results of (a) and (b), verify (4.134).
4.24 A general representation for the L-orthogonal signal set is given by?
si(t) = VP [xn(t)
cos + ya(t) sin@] O<t<T
fof — TN Pt 2 Ltn 1)L + 1 with 6 = 2(/ — 1)z/L and
tf
Ti, Xjedd (t)XK( =a |e (Dat arate
= | oesh
pay:
1 T
a} Xj(t)
yx(t)dt = 0 for all j and k
0
If
and
rae ape
u(j +2n)t
yn(t) = /2cos n=1,2....,.N
Show that the signal set {s;(t)} is L-orthogonal. Assume that / is any integer.
4.25 (a) In terms of the L-orthogonal signal representation of Problem 4.24, let {x,(t)} be a set of
N nonoverlapping sinusoidal pulses defined over the interval 0 < t < 7/2 and {y,,(t)} bea
set of N nonoverlapping sinusoidal pulses defined over the interval 7/2 < t < T, that is
Aa Ms (n — 1)T nT
sn(t) = 2VNsin ik on =! = on
3. See [10], and J. J. Stiffler, A. J. Viterbi, “Performance of N-Orthogonal Codes,” JEEE Transactions on
Information Theory, vol. IT-13, no. 3, July 1967, pp. 521-22.
300 Coherent Communications with Waveforms Chap. 4
modeled after the Hadamard design for an orthogonal code. In particular, we define
2N
; C— yr
xn(t) = ¥ \(-1)" -w (:= ) C= far
2N (i- — 1)T
re
yn(t) Gh)
ce ) aew(ea ne
aN ) taal
SS
for n = 1,2,---,.N where w(t) is an arbitrary waveform of energy T/2N and nonzero
only over the interval 0 < t < T/2N, and bj.» is the i element in the n™ row of a2N x 2N
Hadamard matrix. Show that the resulting signal set {s;(t)} is L-orthogonal.
4.26 Show that for the special case of L = 2 and L = 4, the error probability performance of an L-
orthogonal signal set of M = NL signals defined by (4.145) is equivalent to that of a biorthog-
onal signal set of M signals.
4.27 Using the approach of Section 3.2.5, derive (4.198b) which represents the probability that the
receiver’s noisy estimate of the phase in the i" transmission interval lies in the k" sector of the
unit circle, that is, (6; + @g + (2k — 1)2/M < nj <6; +¢a + (2k + 1)7/M).
4.28 Consider the two sets of orthogonal functions
@2 Sin(@zt — nv)
6, (t) = — ———_
iE @2t — ni
‘ ; j — — m2.
iC wy, Sin(@jt — nz) w sin(w2(t — tT) + m7) q@, 1 [ou (n m2) |
MI, = ee NT ee et
6 © @t—nr nN a(t—t)+mn 1 wit —(n-m2) x
@2
Hint: Consider the product of the Fourier transform of ¢,(t) and 6,,(—t).
(b) Suppose that a WSS process x(t) has the correlation function
@2 SIN w2T
R,(t) =
wt
Does this process have a K-L expansion in terms of the orthogonal functions ¢,(t)? If
so, do the coefficients in this expansion represent samples of the random process at the
sampling rate 2 f; = 2w,/27?
4.29 The covariance function of the additive Gaussian noise process present in a communication
channel is characterized by
1
LEC NE miexp{—2|t—s|}, |t—s|<oo
N
si(t)= sijPj(t), i=O0,1,---,M-1
y=
4.30 Let x(t) be a zero-mean random process with correlation function given by
R,(t) = ape
Find the eigenvalues and eigenfunctions over the interval (0, T) when T < 1.
4.32 Consider the integral equation
T
nat) = | Rn(t, s)bi(s)ds, O<t<T
0
where
6 ; :
2mit 21s
Bad Onda ances (=) cos ( 7 )
Find the eigenvalues and eigenfunctions over the interval (0, 7).
4.33 Consider the binary radar detection problem:
Ay: rH=—s@e)--n), —oo <t <\00
where the signal s(t) is nonzero in the interval (0, 7) and the noise n(t) is Gaussian with PSD
given by
4
Saf) = ; Watts/Hz
(2x f)* + 16
4. T. Kailath, “Some Integral Equations with “Nonrational” Kernels,” [EEE Transactions on Information Theory,
vol. IT-12, no. 4, October 1966, pp. 442-47.
302 Coherent Communications with Waveforms Chap. 4
OPTION 1:
OPTION 2:
Figure P4.34
(a) Determine a suitable realizable whitening filter for the minimum probability of error re-
ceiver.
(b) The optimum receiver processes only that portion of the received signal lying within the
range: (i) [0, 7], (ii) [0, 00), (iii) (—oo, 0], and finally (iv) (—00, 00). Indicate your choice
and explain your answer.
4.34 A noise generator produces a WSS Gaussian random process x(t), whose correlation function
is
Rx(t) = exp(—3|t|)
You must construct a random number generator from x(t), which produces a new independent
random number generator every tenth of a second. You have two options for doing this, as
indicated in Fig. P4.34. Here the samples are instantaneous, ideal, and operate at ten samples
per sec. The variables indexed with the integer n occur at times 0.1n sec.
(a) Compute the correlation coefficient between X,(0) and X;(1).
(b) In option 1, what filter system function G(f) would you use to make w(t) a white Gaussian
random process with PSD S,,(f) = 1?
(c) In option 2, X;(n) is defined on the integers. Is X;(n) a Gaussian random process? Com-
pute E{X,;(n)} and E{X,(n + m)X;5(n)}.
(d) In option 2, what causal filter system function H(f) would you employ to whiten X,(n)
and produce Y,(n)?
(e) Are there any problems (either physical or theoretical) with these designs? Which design
do you prefer?
4.35 Consider the detection of two equilikely signals in additive zero-mean colored Gaussian noise
with autocorrelation R,,(t, s) where
a ai “yz 0
n=1
An m
1
Problems 303
for 0 <t<T where f, is the center frequency of the channel and T represents the symbol
duration.
(a) Determine the PSD if d; and dg are independent, identically, distributed random variables
taking on values +1 and —1 with equal probability.
(b) Repeat (a) if
NE
O=peO=1e
oa AEE
SST
0, elsewhere
1, OSn— 2
pit) = po =|1 LY DISS Oh
0, elsewhere
1-340, |r<t
rs(t) = ie Rae,
(b) By direct substitution of the result of part (a) into (4.374), verify that the conditional error
probability for Manchester coding is given by (4.378).
4.39 Consider a data sequence formed by transmitting with equal probability one of two equal
energy pulse waveforms, that is, pj(t) and p(t) independently in each symbol interval. It is
easily shown that the decision rule [analogous to (4.367)] for this type of sequence is given by
where
Ts
(a) Show that the conditional error probability for this type of sequence is given by
P(E/A) = serte{v (E/No)' [ru AT) + riui(T — AT) —raQT) — nit > aT)
for
A>0
and
forA <0
where
where
nT +5
Bij ~r(@)pi[t—(—1)T —8)at
(n—1)T+6
nT +5.
a= | _1(t)p2[t — (a — )T — 8]dt
(n—1)T+6
and
Problems 305
VP O<{t= 1/2
P20)=) —./P 'T/2<1 aT
0) elsewhere
Derive the expression for conditional error probability as given by (4.380).
4.41 Suppose the error probability is a function of random variable x, say P(E/x). Show that the
average P(E) = E{P(E/x)} can be well approximated by
2
P(E) = P(E/%) +“ P"(E/%)
where x is the mean of x.
4.42 Some BPSK modulators exhibit both imperfect phase switching and amplitude imbalance si-
multaneously. In this case, assume that the signal model is given by
2E : A
s(t) =,/ al + ed(t)] sin[w,.t + (1 — a), 4)
where d(t) = +1 and € and @ denote amplitude and phase imbalance, respectively.
(a) Assuming an AWGN channel, equally likely signals and ideal coherent demodulation/de-
tection, find the probability of bit error.
(b) Plot (on one graph) three curves showing the loss in effective E/ No as a function of a for
three values of €, 0.1, 0.2, and 0.3, for a ranging from 0 to 1.
4.43 A QPSK system operates in AWGN with an E,/No = 9.4 GB (resulting in bit error probability
of 107°).
(a) Find the bit error probability if there is a phase mismatch between the transmitted carrier
phase and that stored in the receiver of ¢ degrees. Assume the interval (0, 7] is known at
the receiver.
(b) Assume ¢ is 3°, 6°, and 10° and evaluate the necessary increase in E;,/
No such that the bit
error probability remains at 1075.
4.44 (a) Repeat Problem 4.43 if there is a misalignment of the transmit and receive clocks by € sec.
(b) Assume €/T = 0.05 and 0.1 and evaluate the required increase in E;/ No needed to con-
tinue to operate at a bit rate of 10~>.
4.45 As an approximation to the integrate-and-dump detector of Problem 4.7, consider replacing it
with a lowpass single-pole Butterworth (RC) filter with transfer function
H(f)=
: 2n
aed (3s |
n(t)
m(t) —e(+) an
sample
at t=nT
Figure P4.46
No and the data is to be detected by the receiver configuration shown in Fig. P4.46. Suppose
that P(f) is an ideal rectangular filter with bandwidth equal to half the data rate, that is,
VT, 0<|fl<
GU -a)
P(A)=4 VT cos{Z [2 fIT -1+e]}, +d-a@)<Ifl<
+ +a)
0, otherwise
Repeat (a), (b), and (c) of Problem 4.46 for this filter.
4.48 A binary communication system, with signals so(t) and s(t) with average power of P watts,
is observed in AWGN (No watts/Hz single-sided). The data rate is 1/T bits/sec and the binary
messages are independent and equally probable.
(a) Draw the optimum coherent waveform receiver.
(b) What is the bit error probability?
(c) Assuming an additive interference characterized by [ (t) = /2A cos(w-t + ) (where ¢ is
a uniformly distributed random variable and A is known), find the optimum receiver for
(k —1)T <t <kT in AWGN.
(d) What is the bit error probability of the receiver used in (a) subjected to AWGN and the
interference characterized in (c) if so(t) = —s,(t)?
(e) What is the bit error probability of the receiver found in (c) subjected to AWGN and the
interference characterized in (c) if so(t) = —s,(t)?
4.49 Consider the bandpass filter illustrated in Fig. P4.49(a), which exhibits an approximate symme-
try about the center frequency f,; the approximation improving as the Q of the resonant circuit
is increased.
(a) Determine the transfer function H(f).
(b) Determine the symmetric part of H(f) about f. = V1 — (1/4Q72)
f,.
(c) Determine the antisymmetric part of H(f) about f..
(d) Given the I-Q modulation-demodulation_ system illustrated in Fig. P4.49(b), develop ex-
pressions for /o(t) and Qo(t) when 6 = 6 and H(f) = 1.
(e) Repeat (d) when H(f) = 1 but 6 #0. Comment on the system engineering meaning of
your result. ‘m
(f) Repeat (d) when 6 = 6 and the bandpass characteristic H (f) is symmetrical about f, (that
is, >> 1):
Problems
307
(a)
MODULATOR DEMODULATOR
I (t)
BANDPASS
CHANNEL
Q (t)
(b)
Figure P4.49
Figure P4.50
4.50 Consider the 16-QAM signal set shown in Fig. P4.50. Assume all signals are equilikely and
that the noise is AWGN with PSD No/2.
(a) Write an I-Q waveform model using rectangular coordinates for square I-Q pulses.
(b) Write an J-Q waveform model using polar coordinates for square I-Q pulses.
(c) Find the probability of error in terms of d and No.
' (d) Find the probability of error in terms of the “peak energy” E, and No.
308 Coherent Communications with Waveforms Chap. 4
REFERENCES
1. Wozencraft, J. M. and I. M. Jacobs, Principles of Communication Engineering, New York: John
Wiley and Sons, 1965.
2. Papoulis, A., Signal Analysis, New York: McGraw-Hill, 1977.
3. Strang, G., Linear Algebra and Its Applications, 3rd ed., San Diego: Harcourt Brace Jovanovich,
1988.
4. Bracewell, R., The Fourier Transform and Its Applications, 2nd ed., New York: McGraw-Hill,
1978.
5. Gagliardi, R., Introduction to Communications Engineering, New York: John Wiley and Sons,
1978.
6. Weber, C. L., Elements of Detection and Signal Design, New York: Springer-Verlag, 1987.
7. Lee, P. J., “Computation of the Bit Error Rate of Coherent M-ary PSK with Gray Code Bit
Mapping,” JEEE Transactions on Communications, vol. COM-34, no. 5, May 1986, pp. 488-91.
8. Irshid, M. I. and I. S. Salous, “Bit Error Probability for Coherent M-ary PSK Systems,” JEEE
Transactions on Communications, vol. COM-39, no. 3, March 1991, pp. 349-52.
9. Korn, I and L. M. Hii, “Relation Between Bit and Symbol Error Probabilities for DPSK and FSK
with Differential Phase detection,” submitted to EEE Transactions on Communications.
10. Reed, I. S. and R. A. Scholtz, “N-Orthogonal Phase-Modulated Codes,” JEEE Transactions on
Information Theory, vol. IT-12, no. 3, July 1966, pp. 388-95.
11. MacWilliams, F. J. and N. A. Sloane, The Theory of Error-Correcting Codes, New York: North-
Holland, 1977.
12. Lindsey, W. C. and M. K. Simon, Telecommunication Systems Engineering, Englewood Cliffs,
N.J.: Prentice-Hall, 1973. Reprinted by Dover Press, New York, NY, 1991.
13. Van Trees, H. L., Detection, Estimation and Modulation Theory—Part I, New York: John Wiley
and Sons, 1968.
14. Papoulis, A. , Probability, Random Variables and Stochastic Processes, New York: McGraw-Hill,
1965.
15. Cooper, G. R. and C. D. McGillen, Methods of Signals and System Analysis, New York: Holt,
Rinehart and Winston, 1967.
16. Helstrom, C. W., Statistical Theory of Signal Detection, 2nd ed., Oxford: Pergamon Press, 1968.
17. Wainstein, L. A. and V. D. Zubavov, Extraction of Signals from Noise, translated from the Russian
by R. A. Silverman, Englewood Cliffs, N.J.: Prentice-Hall, 1962.
18. Brown, J. L., “Detection of a Known Signal in Colored Gaussian Noise—A New Coordinate
Approach,” JEEE Transactions on Information Theory, vol. IT-33, no. 2, March 1987, pp. 296-
98.
19. Nguyen, T., “The Impact of NRZ Data Asymmetry on the Performance of a Space Telemetry
System,” JEEE Transactions on Electromagnetic Compatibility, vol. 33, no. 4, November 1991,
pp. 343-50.
20. Simon, M. K., K. Tu and B. H. Batson, “Effects of Data Asymmetry on Shuttle Ku-Band
Communications Link Performance,” JEEE Transactions on Communications (Special Issue on
Space Shuttle Communications and Tracking), vol. COM-26, no. 11, November 1978, pp. 1639-
Bl:
Chapter 5
Noncoherent Communication
with Waveforms
Miyjan
— Hypa Ft) = si 8) +n@), Ox<t<sT
where the vector W denotes the vector of unknown parameters [1] that are not or cannot
be estimated in the receiver. As before, the noise is white Gaussian with one-sided PSD No
Watts/Hz and T denotes the duration of the transmitted message. We assume for now that
perfect symbol synchronization has been achieved. Later, we will consider and assess the
error probability in the presence of symbol sync error. Note that due to the randomness of
the vector W, each message is transmitted using a sample waveform of a random process
rather than a deterministic and known signal. As far as the receiver is concerned, it does not
310 Noncoherent Communication with Waveforms Chap. 5
distinguish between an unknown random phase introduced in the channel or a random phase
inserted at the transmitter. A message m; can be conveyed using a set or a family of known
waveforms, rather than a fixed and unique waveform. Note that for a fixed vector W, the
received waveforms are known precisely to the receiver. Hypothesis testing as described by
(5.1) is typically referred to as “nonsimple” detection as opposed to the “simple” detection
in coherent communications.
eS) > i a
/ n(t)dt=— {n2(4) Lsne(tn)}
2
(5.6)
represents the energy in the sample function n(t). The joint pdf of the vector n = (n¢(t}),
Nc(tz), ...,Nc(T), ns (ty), Ns (t2),...,Ns(T))! created from sampling a T-second segment
of n(t) is given by
PNoi 6 VM) 2
=) (Ne (tk) + 5 (tk))
Fne(ty)--sns(T)(M) = C exp | — NB : (5.7)
1. As in Chapter 4, we denote the received random process by r(t) and a sample function of that random process
by p(t). Similarly, the received random vector is denoted by r and a sample vector of that random vector by p.
2. In Chapter 5, ¢;(t) denotes the phase information when message m, is transmitted, but it also denotes the ith
basis function of the K-L expansion (consistent with the notation used in Chapter 4). The reader is therefore alerted
to the dual use of ¢;(t) and special care is taken to redefine ¢;(t) whenever it is used.
Sec. 5.1 Noncoherent Receivers in Random Phase Channels 311
since the 2 x 2BT samples are uncorrelated, Gaussian each with variance NoB. The con-
stant C is the normalization factor such that the multidimentional integral of the density is
unity. Using (5.6) to convert from sums of samples to a time integral, we have
1 T
Friclty),....ns(T)) = C exp {-5 i moar (5.8)
No Jo
which is identical to (4.7). Assuming that message m; is transmitted and conditioning on the
random phase 6, then
Sr(p/m;i, 8) = fn(p — si/mi, 0) = fn(e — Si)
Dy
S(OUS.qae p
fy (0) = V2ai(t) cos(wet + b(t) +)) at] 5.9)
No
and
8
Since the receiver does not attempt to estimate the incoming phase 6, it can be assumed
random with a uniform pdf, that is
1
fu(0) = |3 Fh ae (5.12)
0 otherwise
Expanding (5.9), fr(p/mi, @) can be expressed as
ZL E;
fr(e/mi, 0) = C’ exp | a cos@ — Zs; sin } exp (-=) (5.13)
No No
where E; = he a?(t)dt is the energy of a;(t) and the I-Q projections generated by means of
a local oscillator are
ip
oat = p(t)a;(t) V2 cos(wet + oj (t))dt (5.14a)
0
f.
isi = / p(t)ai(t)V2 sin(wet + pi (t))dt (5.14b)
0
As a special case when the transmitted information is contained in the amplitude only,
then r(f) = J/25;(t) cos(w-t +0) +n(t), that is, dj(t) = 0 Vi and s;(t) = a;(t) and (5.14)
can be expressed as the I-Q projections
r
Zci =i! Pc(t)sj(t)dt (5.15a)
0
‘}
Zsi ah Ps(t)sj(t)dt (5.15b)
0
312 Noncoherent Communication with Waveforms Chap. 5
V2 cos
opt
ro
rt)
Figure 5.1 Demodulation with random
V2 sin wt sie
where
then
Zci COSA — Zs; sind = (& cos g;) cos 8 — (&; sin g;) sin@ (5.18)
= § cos(0 + gi) |
and (5.13) becomes
2&; E;
fxs 8) ~ exp |= os + 4) |exp a (5.19)
No No
| : 2&; E;
i i a Isexp |
Sr(p/mi) mel cos( 0 + 00|
: poy =)
adexp ( sta (5.20)
Such an integral has been encountered in several engineering applications and is referred to
as the “zero-order modified Bessel function of the first kind” [3, 4]. Specifically, it is defined
through
Lutte
f(x) 8 = | e* SsBap (5.21)
=U;
Ig
Hence, using (5.22) in (5.20), the optimum decision rule sets m(p(t)) = mx when
ee
0 Noe (-=)
No (23)
is maximum for k =i. Recall from (5.17) that the envelope &; is given by
ag aie: (5.24)
314 Noncoherent Communication with Waveforms Chap. 5
where z,; and Zs; are the projections of p,(t) and ps(t) onto the i signal as given by (5.15a)
and (5.15b), respectively. For equal energy signals, the factor exp(—E; /No) can be ignored
and since J9(2&;/No) is a monotone increasing function, we need only to maximize &; or gE?
where
followed by a sampler at multiples of T seconds as shown in Fig. 5.4. When the signals can
be expressed as
N
T(E —)) se), 2 0, Vane (5.27)
j=l
as given by (4.40), then it is possible to correlate r,(t) and r;(t) with the basis functions
gj(t),i =1,2,---,N and apply the appropriate signal weights to form the various Ee.
i=0,1,---,M — 1, as shown in Fig. 5.5.
A significantly different approach for optimum detection of equiprobable equal en-
ergy signals is possible with envelope detectors. Consider Fig. 5.6, which depicts a filter
hpp,x(t) followed by an envelope detector. Since the input to the filter r(t) is centered at fo,
let the filter hp» ,(t) be matched to the signal s,(t), translated to frequency f, (“bp” stands
for bandpass), that is
or
3. Note that in Fig. 5.3, lowpass filters have been inserted at the output of the mixers. As long as the bandwidth
of these filters is much larger than the data rate, they do not impact the performance of the optimum receiver. It is
however good practice to include them in an analog implementation as the mixers are highly nonlinear devices that
generate more than the desired product at their output.
Sec. 5.1 Noncoherent Receivers in Random Phase Channels 315
V2 cos w,t
DOAYFDYrVEOO
Lg —sfifoa
foe |fot
M-1
wy,
Sy (t)
Note: From here on, a "comparator" selects the message
corresponding to maximum input value or voltage.
where we define
lo 2)
CO lo @)
Usk (t) 4 | r(t)sy(T —t+ t)V2 sin Wetdt =f rs(t)sx(T —t+1)dt (5.31b)
—o0o —0o
Since r-(t) and r;(t) are low-frequency signals and s,(t) is also a slowly varying signal,
both u,,(t) and us,(t) remain relatively constant over several cycles of cos wet. As a result,
ux(t) is a bandpass signal and can thus be expressed as
Cc
O
M
P
A
V2 cosw,t
R
A
7
O
R
V2 sin wt
Figure 5.4 Alternate implementation of M-ary noncoherent receivers for equal energy signals with
matched filters
, ()
SIGNAL
; WEIGHTING =
: MATRIX | -
ot M
V2 cos w,t h
% OX J(at
V2 sin @,t > (t)
) SIGNAL
: WEIGHTING
; MATRIX
(1
.
©
MATCHED
FILTER
Nop x!) t=T
Figure 5.6 Envelope detection
of great practical interest as they occupy the minimum required bandwidth for orthogonal
noncoherent communications. The transmitted signals are characterized by
where «, denotes the frequency of the k'” tone and @ the unknown random phase. The latter
is independent from message to message because each signal is typically generated by a
unique oscillator whose initial phase is random when the signal is selected. The pairwise
signal correlation coefficient, as defined by (4.45), reduces to
1 Lf 1 Ls ;
= ot si(t)sj(t)dt = al cos(n
fi,jt+0 — 0 )dt 6:37)
where E = PT, fi,; = fi — fj is the frequency separation between the tones, and 6, @’
318 Noncoherent Communication with Waveforms Chap. 5
BANDPASS
MATCHED
FILTERS :
ENVELOPE So
DETECTOR
ENVELOPE
r (t) ‘ DETECTOR
DOHAYFDFYFVEOON
ENVELOPE
DETECTOR i
task
Figure 5.7 Envelope detector noncoherent receiver for equal energy signals
are the random carrier phases when signals s;(t) and s;(¢) are transmitted. Because of
the random phase @ — 6’, the only way to guarantee zero correlation is for the integral to
be carried out over at least a full period of the difference sinusoid. Hence, the minimum
frequency separation to guarantee orthogonality is
1
This separation is equal to twice the minimum separation required for coherent M-FSK
transmission, 1/(27), derived earlier in Chapter 4. Given M tones, the minimum bandwidth
required for orthogonal noncoherent transmission is M/T Hz achieved by uniformly spac-
ing the tones with 1/7 Hz separation. The received signal, assuming that message myx is
transmitted, is given by
Typically, the tones are symmetrically spaced around the carrier f-, which can easily be
modeled by (5.39) by considering negative values of k. The optimum receiver is still as
shown in Fig. 5.3 where the various signals are the corresponding sinusoids. The receiver
still computes &2 for each signal using gE?= a + Zz where Z,-; and Zs; are given by (5.14)
with aj(t) = VP Vi and ¢;(t) = a;t Vi, that is
a
en ae | r(t) COS(@e + @;)tdt (5.40)
0
and
ip
Boh = V3 | r(t) sin(@e + @;)tdt (5.41)
0
Hence
il 2 T 2
E? =2P (| r(t) COS(@e + wear] +(/ resin tonya) | (5.42)
0 0
Sec. 5.1 Noncoherent Receivers in Random Phase Channels 319
SPECTRAL
r() CORRELATOR FOURIER OBSERVATIONS
TRANSFORM
i
®, (6+)
D>
>
0
DOA
VUs0
Exact evaluation of the spectral observations is difficult when one elects to implement
the spectrum analyzer receivers of Figures 5.8 and 5.9. However, it is possible to ap-
proximate the spectral estimate ©,(f) which then results in a receiver performance that
is suboptimum. There are various approaches to approximating the spectral observations
(OAK ak Onl e+, .A4 —1}.
First, one can evaluate the autocorrelation function @,(t) directly and then evaluate
the Fourier transform to obtain ®,(f). This is referred to as the autocorrelation function
4. In Figures 5.8 and 5.9, the Fourier transform is operating on the resulting even function, as discussed in
Appendix 5B.
320 Noncoherent Communication with Waveforms Chap. 5
A SPECTRAL
®v(f) |OBSERVATIONS
CORRELATOR abet ve ;
oy(+)
as)
3
0)
|=
(@)—5
28)
(ONS)
3
method. A practical realization of this method that is convenient for digital processing is
discussed in [6, 7] and summarized here. Since the Fourier transform of the random process
u(t) is essentially zero outside the frequency interval | f| < M/T, then in accordance with
the sampling theorem we have, to a good approximation
2BT
v(t) ~ vat) Sov (=) sinc [28 («= (5.47)
n \
where B A M/T and vy, = v(n/2B) is a sample of v(t) at t = n/2B. In terms of the above,
the autocorrelation function of vg(t) is given by
Tt 2BT 2BT a
Ont) i Ug(t)ug(t + t)dt = ey = Un UmSiNC E («+ =) (5.48)
tH=lm=l)
which approximates that of ¢,(t). We observe that for r = k/2B, k an integer, the above
expression simplifies to
Since the spectrum of ¢,,(t) is limited to the frequency range |f| < B, one obtains by
means of the sampling theorem
2BT-1 k k
1 pera k kf
v(f) vg (Sf) Fp oral Mies 2B Pv, (55) -5( B ). BS fea B (5.51)
These expressions for the correlation coefficients and for the approximate spectral observa-
tions ®, (k/T) form the basis for the computational procedure.
Second, one can calculate the spectrum by the discrete Fourier transform approach.
The latter takes samples v(k/2B), k = {1, 2,---,2BT} and evaluates
2BT E 2BT
X(N=) (55)00s(“L) aw rin= >» (<5) si (=) (5.52)
Sec. 5.2 Performance of Noncoherent Receivers in Random Phase Channels 321
ENVELOPE
DETECTOR
ZERO
r(t) THRESHOLD 3>
COMPARE
ENVELOPE
DETECTOR
the effect of frequency and timing synchronization errors. A bound on the performance
of orthogonal M-FSK signals is then presented in the presence of frequency error. Finally,
the performances of two suboptimum M-FSK receivers are evaluated and the effect of time
domain truncation in evaluating the spectral observations is assessed.
The M-ary receiver for equal energy signals is depicted in Fig. 5.3 and the receiver
picks the message mx corresponding to maximum &? for i=0,1,---,M —1. In or-
der to compute the probability of symbol error, we need to characterize the joint pdf
2 2 : 5 A =
Fe2 2,62, (602 815° Sup): Assuming that message m, is transmitted, then r(t) =
J 25,(t) cos(@-t +0) +n(t) with
i if
z= / re(t)s)(t)dt = if(sg (t) cos + n¢(t))s;(t)dt (5.54)
0 0
and
T if
i / rs(t)sj(t)dt = it(sg (t) sinO + ns(t))s;(t)dt (5.55)
0 0
Similarly
T : P
2 Bei). qi
{Zsi/0} = sin@ [ Sk (t) sj (t)dt 8 ay (5.57)
E : = F —_ Ss ’
and
aR 2
N
of 16 =E 10 na(tsi(at = > Es (5.58b)
0
which are also the unconditional variances. Recall that n,(t) and ns(t) are bandlimited
processes; however, assuming that BT >> 1, they can be viewed as nonbandlimited using
the matched filter argument previously outlined in Chapter 4. Since n,(t) and n,(t) are
independent, z-; and zs; are conditionally uncorrelated and hence independent since they
are conditionally Gaussian, that is
iewee : y/@0) ==
1 {-@ — E{zci/6})? + (y — ai"
Rt E;No
| iZ+k (5.59)
Consider now the pairwise correlation among the various z,; and zs;, i =0,1,---,
M — 1. Assuming that m, is transmitted, then
ifs T
E{zciZej} = E {|
/ (sg (t) cos @ + nels | (sg(tT) sin 6 + ne(e)sj(e)at]|
0 0
Ae T
E {(/ ne(si(a (| ne(a)sj(e)dr) | (5.60)
0 0
Fie 4:
Using similar arguments, it can be easily shown that
E{zsiZsj} = ElzsiJE{zsj}=0, iF j
(5.62)
E{zcizZsj} = E{zci}E{zsj}=0, t=j
which holds true even when i = k or j = k but not both. Therefore, the variables z,; and Z;;,
i =0,1,---,M —1 are uncorrelated and hence independent since they are conditionally
Gaussian. Defining 0; = tan! (Zsi/Zci), then (5.59) becomes
I tye
wEsNo ©*P |-54 , ixk
Sccis2si Cap y/0) = 1 z-+y*+E?—2E, [224-2 cos(0—6;) Pee k (5.63)
25 ex p |-=, ; ixzk
iZ
ACh
J&j a tte
2& eee)
Bese. I 25
(5.65)
ix=k
EyNo OP E;No O\ No)? ‘7
The Rayleigh and Rician densities of (5.65) can also be obtained using the results of Ap-
pendix 5A, namely, (5A.21) and (5A.27) with 0? = NoE;/2 and sex Ee. Let us proceed
by normalizing é? using
BA
Aj = — (5.66)
E;No
fa;(@) — V EsNoa)
Se (VEsNoa) (5.67)
2S (a/EsNo)
324 Noncoherent Communication with Waveforms Chap. 5
or
where u(a) is the unit step function. The variables A;, i = 0, 1,---, M — 1 are also inde-
pendent since they are nonlinear functions of independent random variables [11]. Hence
Lt ep Wan LI8 aj,°°° ,amM-—1) = fo (G0) fA; (a1) acre fAy_\(am-1) (5.69)
From this joint pdf, we can now evaluate the probability of symbol error P;(E). Assuming
that m, is transmitted and that the output of the envelope detector matched to s,(t) is Ax,
then the probability of correct symbol detection conditioned on message m, and Ax is given
by
P;(C/mx, Ax) = Prob {Ao < Ax, Ai < Ak, ---, Ag—1 < Ax, Ati < Ak, ++, Am—1 < Ax/my}
M—1 pA,
(5.70)
=[T] | fa@da=— ey"
i=0 —oo
itk
Hence
zs as - (©27))
= exp {=} f e “*Io (sia (1 - ese dag
£ at petal m Ee
aa Seite iGmenp | (=)| (5.77)
The symbol error probability is depicted in Fig. 5.11 versus bit SNR [Ep/No = (Es/No)/n
where n = log, M] for various values of M. Since the signals are orthogonal, the bit error
probability P,(£) is related to P;(E) through (4.96) and is plotted in Fig. 5.12 versus
Ep/ No.
Consider now the case of coded versus uncoded transmission. The symbol error prob-
ability performance of a noncoherent system that transmits an n = log, M bit symbol by
sending one bit at a time (so-called “uncoded transmission’’) is given by
P,(E)|
b M=2 pe
>= 5 exp -=)
2No 5.79
(5.79)
obtained from (5.77) by setting M = 2 and E; = Ey. The uncoded symbol error proba-
bility, as given by (5.78), is illustrated in Fig. 5.13 for various values of n as a function
of bit SNR. Coded and uncoded symbol error probabilities are compared in Figures 5.14
and 5.15 for n = 5 (M = 32) and 10 (M = 1024), respectively. Also shown in these two
figures are results given in Chapter 4 for coherent communication employing orthogonal
and biorthogonal signals. For orthogonal codes, we observe that the type of detection per-
formed, whether a coherent or noncoherent detector is mechanized, is not very critical for
error probabilities less than 10~*. As a matter of fact, the use of 5-bit biorthogonal codes
and coherent reception is not appreciably more effective (given a certain energy-to-noise
ratio) than the use of orthogonal codes and noncoherent reception. On the other hand, a
comparison of coded versus uncoded transmission indicates that coding the transmitted sig-
nal into n-bit orthogonal symbols (rather than transmitting an n-bit symbol one bit at a time)
is more effective in noncoherent communications. For example, if five bits of information
are sent with a symbol error probability of 10~? (typical for telemetry systems), the use of
orthogonal codes and coherent detection will reduce the signal-to-noise ratio per bit, that
is, Ep/No, by approximately 2.5 dB under that required for equivalent performance using
bit-by-bit detection. Under the same symbol error probability conditions, the use of orthog-
onal codes and noncoherent detection will reduce the required E,/No by approximately 6
dB under that required for equivalent performance using bit-by-bit detection. For ten bits
Noncoherent Communication with Waveforms Chap.5
326
15
10
9
8
if
6
5
4
3
2
1
(M = 1024) and P,(E) = 1073, orthogonal and biorthogonal coding and coherent recep-
tion reduce the required E),/No by approximately 5 dB under that required without coding.
Finally, using noncoherent reception, 10-bit orthogonal codes and P;(E) = 10-3, one real-
izes an improvement of approximately 8.5 dB over the uncoded noncoherent procedure of
bit-by-bit detection.
Finally, by a limiting procedure similar to that employed in Chapter 4, it may be
shown (see Problem 5.3) that the asymptotic (M — oo) error probability performance of the
optimum noncoherent receiver is identical to that for coherent communications as described
by (4.182).
Sec. 5.2 Performance of Noncoherent Receivers in Random Phase Channels 327
10°
102!
10m
a a
a
10°?
{o>%
107"
=41(0' 20
Figure 5.12 Bit error probability of ideal noncoherent receiver (M = 2”) as a function
of bit SNR (Reprinted from [28])
Nonorthogonal signals can occur in either one of two scenarios; the signals can be chosen
nonorthogonal at the transmitter or the latter sends orthogonal signals but due to imperfec-
tions in the receiver, the signals become correlated and thus nonorthogonal. The second of
the two cases occurs as the result of frequency or timing error in the receiver and will be
treated in more detail in a later section. The key difference between the two cases in terms
of analysis complexity is that when the signals are nonorthogonal at the transmitter, the re-
ceiver correlates the observed waveform with the transmitted signals and the noises at the
output of the integrate-and-dump filters are correlated and dependent. On the other hand,
Noncoherent Communication with Waveforms Chap. 5
328
—
aAawonarn
(E)
P,
20
when the signals are nonorthogonal due to frequency or timing errors, the receiver correlates
the observed waveform with the original transmitted orthogonal signals and consequently,
the noises at the output of the integrate-and-dump filters remain uncorrelated and hence in-
dependent in an AWGN channel. Assuming nonorthogonal signals, the received waveform,
when message m; is transmitted, is given by
UNCODED
107!
ORTHOGONAL
CODING
BI-ORTHOGONAL
105" CODING
ny
fatase
ORTHOGONAL
CODING
10+.
— — NONCOHERENT
RECEPTION
-4
oe COHERENT
RECEPTION
19%
Figure 5.14 A comparison of coded and uncoded, coherent and noncoherent, symbol
error probabilities (M = 32) (Reprinted from [28])
s;(t) and n(t) is AWGN* with one-sided PSD No Watts/Hz. It is convenient to use complex
notation to denote the transmitted signals. Let
6. It is easier to derive the results when the noise is white. For bandlimited noise, the results are identical, but the
derivation is tedious.
Noncoherent Communication with Waveforms Chap. 5
330
UNCODED
ORTHOGONAL
CODING
ORTHOGONAL OR
BI-ORTHOGONAL
CODING
w
a”
== —— NONCOHERENT
RECEPTION
COHERENT
RECEPTION
Figure 5.15 A comparison of coded and uncoded, coherent and noncoherent, symbol
error probabilities (M = 1024) (Reprinted from [28])
oo()= /2 {aj (t) cos $j (t) cos(@ct + 6) — a;(t) sin p(t) sin(@ct + 6)}
(5.83)
= J2 {a¢i(t) cos(wet + 8) — asi(t) sin(wet + 6)}
where
Sec. 5.2 Performance of Noncoherent Receivers in Random Phase Channels 331
aci(t)
4 aj(t) cos ;(t)
(5.84)
asi(t) A a;(t) sin g(t)
We will assume that the signals have equal energy E,, that is
r T
E,& / s?(t)dt = / a?(t)dt
18 Ik
(5.85)
:
0 0
Define the complex correlation coefficient /;j [13] as the correlation between any two
complex baseband signals 5;(t) and 5j(t) as (let * denote complex conjugate)
3 Nar Ree
py Zi Si(t)5; (t)dt (5.86)
Vee
Bij & =i aj(t)aj(t) sin(g(t) — pj (0))dt
oo (5.89)
1
= =| (asi (t)acj(t) — aci(t)asj(t)) dt
Es 0
Note that a;; = 1 and 6;; = 0. Furthermore, a;; = aj; and Bj; = —f;;. In the special case
where the signals are orthogonal, then a;; = B;; =0 for i # j. Recall that the optimum
receiver still computes the decision variables &? for k = 0, 1,---, M — 1 where a 2+
ae and
ii
Zeck af r(t)V 2az (t) COS(Wct + Pk (t))dt (5.90a)
0
T
Ze & / r(t)V 2ag(t) sin(wet + Px (t))dt (5.90b)
0
It can be easily demonstrated that E? is independent of the carrier phase angle @ of the in-
coming waveform (see Problem 5.4). Therefore, we will set @ = 0 from here on to simplify
the analysis. Substituting (5.80) into (5.90a) and (5.90b), we have
where
T
Neck & ifn(t)V 2a (t) COS(Wct + by (t))dt (5.92a)
0
i
ys / n(t)V 2ag(t) sin(wet + ox (t))dt (5.92b)
0
It is easily shown that n-, and ns, are zero-mean Gaussian random variables with variance
(No/2) Es and cross-correlations
N
E {ncknej} = = Es Qk; (5.93a)
Ni
E(nsknsj} = > Eset (5.93b)
No
E{ncknsj} = a EsPri (5.93c)
OE.
Zck = ,| —— Aki + Xck (5.94)
No
Dade
Zsk = ,| =— Bei + Xsk (5.94b)
No
where x-, and x;, are normalized unit variance random Gaussian variables with
E{xckxXsj} = Bj (5.95c)
and @; and B; as
QO; Boi
QI Bii
ial and p;=| (5.99)
QM—li Bu -ii
Then. the pdf of the augmented vector (z7 z!), assuming message m; is transmitted, is 2M-
multidimensional Gaussian given by (3.61) as
DER.
fantevimy=0((¥) tei) (5.100)
Are
es (31 ) (5.101)
and is the covariance matrix of the augmented vector (z, zs)’. The probability of correct
decision, assuming that message m; is transmitted, becomes
pccimy= ff JJ J] J] i of]
°((5)-Va(3)4) -
dzm—1dym-—1---dzj4+1dyj41dz;-1dyj_-1 ---dzidy,dzodyodzidy;
where R; is the region within the circle of radius Aa + hie centered at the origin. The
unconditional probability of correct symbol detection P;(C) becomes
M-1
PC) =), P.(C)m;) PGni) (5.103)
i=0
where P(m;) is the a priori probability of sending message m;. The 2M-fold integral in
(5.102) cannot be simplified any further in general. In the following sections, we will con-
sider scenarios under which P,(C) can be computed in closed form or in terms of well-
known functions.
For M-ary signals, Nuttall [14] derived the probability of error for the case when the signals
are equicorrelated and furthermore, the correlation is real and nonnegative, that is
% ee teae oF | 104
Pl; a, ij anda real and nonnegative Chel
Hence, the imaginary component f of the complex correlation is assumed zero. Nuttall
[15] provides a heuristic argument (not a proof) that for a given magnitude of the complex
correlation, the assumption of zero angle (or zero imaginary component of () realizes the
334 Noncoherent Communication with Waveforms Chap. 5
minimum error probability. It is shown [14] that the probability of correct symbol detection
is given by
le,@) [o.@) ]
exp (-3 (x? +»)
P;(C) = (1 — @) exp (-=) | if xy
(5.105)
x o( = - es)Io (Vaxy) [1 — Q (Jay, x)]""' dxay
If a = 0 in (5.105), the integral on y is unity resulting in
lee) 2 aa 2 M-\
pc) =exn (-2) [ reso (-) 0 ([Fe] [1-ex(-5)] dx (5.106)
Using the expansion in (5.72) and the integral of (5.74), (5.106) reduces to (5.76) as ex-
pected. If, on the other hand, a 4 0 but M = 2, use the integral [15]
oo Is > a ac b
[ s exp (-3 (s 4+¢ ))Io(cs) Q(as, b)ds = Q (Ss. =) (5.107)
in (5.105) to obtain
1a | ee
P,(C)
= (1 — @) exp (-=) r exp (-3" (1 —a@)
0
; aba o( ar Fr Je
x r —a - '
"VV No ee wher
(5.108)
af | — a”) exp (-=) [xe (-3* (1-«))
x IIo |x meal
No yy
( — eT (1 — QO (ax,
a?) x)) d xdx
Ee [Bs
epee BW ieee: i ee
fC) a o(\2No (1 a”), 2No (1+ ; =))
(5.109)
ae (-Be Veale
ane a («m5
which is exactly 1 — P;(E) where P;(E) is the probability of error for two nonorthogonal
signals and will be given by (5.133).
B = Bio = —Bo1)
2E; DE
U0 = No Lc) N + UcO
(5.110)
2E,
UVs0 = iNeaee = Us0
and
2E 2E
vei A a lS yao t eel
(5.111)
2E 2E
V5, 4 nar yb + us
assuming mo was transmitted. The Gaussian noise variables, u.0, Uso, Uci, and us},
are zero-mean, unit variance with E{u-ouci} = E{usous1} = (2Es/No)a, E{ucousi} =
—E{u.iuso} = (2E;/No)B. Letting &; A ,/ v2, - v5 we know from Appendix 5A that the
marginal densities of €,9 and €,; are Rician, but due to the nonzero correlation of the Gauss-
ian noises, €)9 and €,; are not independent. The probability of error, assuming message mo
is transmitted, is equal to the probability that €,; exceeds £,9 or
P;(E/mo)
= Prob(&y1 > &v0/mo)= [ i feof(X, y)dydx 6.112)
which requires the computation of the joint pdf f¢,,¢,,(x, y). This approach is outlined by
Helstrom [16] who determined the joint pdf and computed the double integral in (5.112).
Another approach to the computation of the probability of error, taken by Proakis [17],
is based on the observation that a _ ef is a special case of a general quadratic form in
complex-valued random variables, defined by
1b,
A= > (Ailxel? + Arlyel? + Asxeyy + A3xzy«) (5.113)
|g |
where A), A2, A3 are constants and {x,, yx} are pairs of correlated complex-valued Gauss-
ian random variables.
The approach followed in this book (also see Viterbi [18]) is based on whiten-
ing the observations v¢9, Us0, Ucl, Us} SO that the resulting variables Vy.cO, Uw,s0, Uw,cl;
and vy,s1 are uncorrelated. Consequently, the variables &, 9 4 a DE + Cae and &y.; A
[v2 Alaa v2 «| are Rician and independent and the probability of error reduces to the prob-
ability that one Rician variable exceeds another independent Rician variable. Defining the
vector X as
(2E;/No) I
= 0 Nae) Ze
(feat 0 se
mx | (2E,/No)a i or a
(2Es/No)B B
and
i 0 a £B
= 2E;s 0 1 —B (64g
= (%) eee 0 (5.116)
B a Oa
Due to the symmetry of the problem, it is easily shown that P;(E/mo) = Ps(E/m1) =
P;(E). Hence, the error probability is given by
P,(E) = Prob(é2, < 2.) = Prob(v2y + v2) — v2, — v2, <0) (5.117)
or
Xo x (5.120)
S121)
where
where Vy,c0, Vw,s0; Vw.cl, aNd Vy.s; are unit variance uncorrelated Gaussian (and hence
independent) random variables with means
2(Es/No)k(k + 1)
Bieter pixy 2 ols No/Es 0
Be Ba Decpt \i, (hs) Node oa
2(Es/No)kB
The optimum receiver compares ee 9 to ae ,; Where g2 ; 1s the sum of the squares of
Vw,ci and Vy,s;. The pdf of &,; is Rician with parameter s;, that is
ee
Fé, (E /mo) = & exp (-! i 1g (sig) (5.127)
where
E
56 = E*{vw,co} + E7{vw,so} = (=) (+k)
% (5.128)
st SE arc tes (east = (=) (bk)
No
Hence, the probability of error becomes
(5.129)
| fey, dydx
= Fewo(®)
which is similar to (5.112) except that the variables &, 9 and &,,; are now independent. It
can be shown that [19] (see Problem 5.6), given two independent Rician random variables
Ro and Rj, each with parameter s; and variance oe (i = 0, 1), then the probability that Ry
exceeds Ro is given by
where
a=
ae
as eer and b=
eae
5) 5) (5.131)
Og + OF 09 +O;
(57132)
2No: 1+kh=
E, Es
1 ay 1 — ir)
~
jij rik) = ( |p|?
338 Noncoherent Communication with Waveforms Chap. 5
with p =a + jf. Hence, the probability of error for two nonorthogonal signals is given by
P,(E) = o( + /1- a)
(5.133)
1
which is identical to 1 — P,(C) where P,(C) is given by (5.109). The probability of error
is illustrated in Fig. 5.16 for several values of ||. It is clear that orthogonal signals (that is,
p =0) minimize P,(£). In this case, P;(£) reduces to
jike 1 Ee
| ‘|— 5 exp (-)
P,(E)=@Q (0. (5.134)
= (5.135)
P;(E) zt i ie eee
= = -—exp|{
= exp |-—— } — =exp[— — :
: PONG)! Oo N ONy ONE
which is identical to (5.77), for M = 2, or equivalently, (5.79) with E; = Ep.
Consider the cross-correlation between two FSK signals at frequencies fo and f/f).
Then, the cross-correlation between the corresponding complex baseband signals is
Nps
p=— | V2E,/Tel?"' /2E,/Te lat
2Es 0
- (5.136)
¥ a pidnfo- ft, — IA ASOIT
|jxforT
PI6 m fo,1T
and
5 sin 2 fo,1T
(A= |e 5:137
m fo,1T ( )
The performance for any frequency separation fo,; can then be computed by first evaluating
the corresponding ( and then picking the corresponding curve in Fig. 5.16. In particular, if
fo,1 = 1/T, then |p| = 0 and P,;(E) reduces to (5.135).
When the received frequency is not perfectly known, the observed signal, assuming message
m; is transmitted, is given by
10s
110i
40>°
(E)
P,
100
Figure 5.16 Probability of error for noncoherent detection of two nonorthogonal signals (Reprinted from [17])
T
zo f r(t)v2P cos (2n(fce
+ fx)t) dt
0
7 (5.139)
— re) cos (21 (fe + fi + Af)t) cos (2 (fo + fx)t)dt + nex
0
(a)
RECEIVED
FREQUENCIES
(b)
LOCAL
FREQUENCIES
where n-x is a zero-mean Gaussian random variable with variance o2 = (No/2) Es. Simpli-
fying, (5.139) reduces to
E Ie
Zck sin (27 (fix + AFT) + nex (5.140)
~ On(fix + Af)
where again, fj, denotes the frequency difference between messages i and k, that is
P 2
oe aan) {sin? (27 (fin + Af)T) + (cos (27 (fi,x + Af)T) — 1)°|
= (seaarap) bs ra
+ann)
a £ : ae
Normalizing z,x and zs5x% by 1/o = ./2/(NoEs), the parameter se is then normalized by
2/(NoEs) and since f; =i/T for orthogonal signals, then
PN)
f= (2) eG =Ee et
aie? CEO)
eye (5.144)
5.144
pSAFT (5.145)
From (5A.27), the density of & is given by (note that 0? = | after the normalization)
where we let
1
fpGbHA sin? (1 (i — k + p)) (5.147)
m?(i —k + p)?
First, note that the detector matched to the incoming signal suffers from signal attenuation
equal to
5, i
ue a sin’ (sp) (5.148)
On ford 54k
fg. = ae sin? (x (i — k)) = ly Morr=k oy)
and signal orthogonality is restored. In this case, the Rician densities reduce to the Rayleigh
densities derived earlier. Despite loss of orthogonality, the variables &, &|,---, €y—1 remain
independent since the Gaussian random variables resulting from the noise integration are
still independent as the local signals remain orthogonal. In this case, the probability of
correct symbol detection, assuming message m; is transmitted, is given by
342 Noncoherent Communication with Waveforms Chap. 5
= 1) Se (Xi) ra: fe (X0) fe, (01) °> + Fe; Oi-1) fei i+) - >
0 0 0
M-1 integrals } (5.150)
x fey _\(Em—1)dxodx1 +++ dxj—dxi41 ++ dxm—14dX;
le.@) M-1 Xj
in
P.(E/m;) =1 -| fe, (Xi) I] (:215 (a f3¢.m).x)) dx;
2Es 42. (5.152a)
0 m=0
mAL
If (k, i) denote the difference between the codewords (bit mapping) assigned to messages
(symbols) m; and m,, that is, the number of bits in which the two differ, then the average bit
Sec. 5.2 Performance of Noncoherent Receivers in Random Phase Channels 343
When the received frequency is known precisely but the symbol epoch is not, the receiver
integrates using its own estimate of the symbol epoch which is offset from the true epoch
by Ar sec. This phenomenon is depicted in Fig. 5.19 where the integration overlaps two
successive symbol intervals. This results in signal attenuation in the detector matched to
the incoming frequency and moreover, loss of orthogonality due to signal spillover into the
remaining detectors [8]. The received signal can be modeled as
where we assumed that message m; is transmitted followed by message m ; and that phase
continuity is maintained. Since the local epoch estimate is not perfect, the receiver integra-
tors operate from At to At + T to obtain at the k"" detector (setting 6 = 0 without loss of
generality)
At+T
Zeck = if r(t)V2P cos(2m(fo + fx)t)dt
At
A
=2P {| cos(2 (fe + fi)t) cos(2m (fe + fx)t)dt (5.156)
At
At+T
+ / cos(2m(foe+ fy)t) cos(2a(fo + fx)t)dt —+ nck
T
and
At+T
Zsk =f r(t)V2P sin(2a(fo + fx)t)dt
At
;
=2P {| cos(27 (fe + fit) sin(21 (fe + fx)t)dt (5.157)
At
At+T
+ / cos(2m(fe+ fy)t) sin(2m (fe + fonat| + nsx
de
Normalizing z-x and zs, by 1/o, the pdf of &; can be expressed as
E> fx(i,i.)
(2Es/No) im Ame
ALA MU
T
MTabANE I
TT
344 Noncoherent Communication with Waveforms Chap. 5
i
LE a SO HA
1072 EN
ET
PEAS
tT
10
10
10
10
10 11 12 13
! N
-o
awe
105
107
107
10 11 12 13 a 15 16
o ! N
—
°
bl
oO
100°
10
1077
where
LC
aay es 1
rears ‘
sin? [7(i 1
— (1 — A) + Gab? sin” [1(j — KA]
< 1 ol ee 4 ae ;
POG Sh {x [@ — fj) -— (Gi — kal}
, (5.159a)
rer h)sin” {7 [@ —k)A— (j —k)]}
1
+ G2 HG Tm SY LE -D- GO, ixk, j#k
i ® ey 1 oy A
OS =e m(i —k) ;
5.159
elec
TOG OK
eG =) Ae sin Ao
‘ (i al : Ces) (5.159c)
+a — jf) —A)snQzrG-j)A+1))} ifs
ARG.
Die
B= pH
ae
[sir(i
Sao) b
— HL —)
5.159d
sgt Ac ak)
— mA(i —k)sin(2wrA(Gi —k))}, ifXk
Gi, . Vi (5.159e)
and i denotes the normalized timing error given by
At
Pa
Ships (5.160)
As acheck, we have
1
heats ened {2+4 2) 1 — 2)?
K Oe eee : 2 — 2cos(2mxa
CALS asp woa) ater ioe (5.16 1a)
4 Anx(1 — A) sin(2rx(A + 1)) — 4ax(1 —A) sin(2xx)|
which is 0/0 ! Using L’ Hopital’s rule once, we have
I
lim f2G, j,i) = lim —— {[8x?x(1 — 2)? + 2(2mA)sin(20xA)
imj x30 872x
which is still 0/0 ! Applying L’Hopital’s rule one more time, we obtain
348 Noncoherent Communication with Waveforms Chap. 5
1
lim f2@, j,i) = lim — {8771 —Aa)*?+4nr(1-—A) +4204 1)
i>j x0 87
(5.161c)
— 4n(1 —2)4m + 8272"
=i
as expected. Using a similar procedure, it can be shown that
From (5.152), the probability of symbol error, assuming message m; is sent followed by
message m j;, is then given by
ee) 2 2E, N, 2 Pr Oe
(2h ew
«to (s efttasd) (5.163)
ia 2B ya
m=0 No
mA
As was the case for frequency error, the presence of timing error produces a lack of
orthogonality which results in the symbols not being equally likely. Hence to compute the
average bit error probability, one must once again compute the probability of a particular
symbol error for a given transmitted message. Analogous to (5.153), the probability of
choosing m, when message m; was sent followed by message mj, is given by
Pal (mi)
sk /m;,mj; = [exp
= 4 Xj ©Xp |— (2+= Ba. j.) |to (mn = 120.
5) Noes ) 0 | Xi No es )
oh e (5.165)
x I] 1i 0 (FAG. jem.) a
m=0
m#k
with f,(i, j,m) as in (5.159). Finally, the average bit error probability is analogous to
(5.154)
M-1M-1M-1
P(E) = W108, M ; (k, i) Psk(E/mj, m;) (5.166)
Sec. 5.2 Performance of Noncoherent Receivers in Random Phase Channels 349
(a)
RECEIVED
FREQUENCY
AND TIME
At At -— At
(b)
LOCAL
FREQUENCY
AND TIME
Here again, the evaluation of (5.166) will depend on the mapping of the symbols to bits.
For the conventional Gray code mapping of Table 4.1, Fig. 5.20 plots the average bit error
probability versus E,/No in dB for binary, 4-ary and 8-ary FSK with 4 as a parameter.
Digital computer simulations were used to confirm some of the results.
When both the incoming frequency and the symbol epoch are not known [21], the received
signal is still given by (5.155) with f. replaced by f. + Af, as depicted in Fig. 5.21. The
inphase and quadrature integrator outputs become
T
ee
in a
é:
I
ULINN
l
(oS EN
Gane ee
ns
BS
ae ae
e
a Ee SAN ae eB
r=0 XN A= So
5
10
aan
2 a el ST
SSS
10
=
10
-7 ieSA NN
10 11 12 13 14 15 16
C
Sec. 5.2 Performance of Noncoherent Receivers in Random Phase Channels 351
EM ALLL
10-1
SENT
4=0.3
100-2
2
j
10%
eI IIE
ZiT r=0 2
rail
ae 10c
105"
OIE =0.1
pee
10-8
10-
10
iN
Wee
zit
(AO
NY, 11 a 13
i SIMULATION POINTS
14 15 16
07
UMC PTL
aa
iiaHE
Olme
07°
10-
11 12 13
fj + Af fj + Af
(a)
RECEIVED
FREQUENCY
AND TIME
FREQUENCY
AND TIME
e
Se in fe (jf — Kk + 2)A]
i Gee pp eo i
1 (5.170)
{cos(2z(i Fu phyB)
+ 292 —k
+p)G —k +p)
— cos(2m(i — j)) — cos(2a((i — j)A —j —k — p))
dB for binary, 4-ary and 8-ary FSK with both p and A as parameters. Digital computer
simulations were used to confirm some of the results. Note that when timing and frequency
errors occur simultaneously, the losses are not additive.
As required by the results in Appendix 5C, we need to evaluate the characteristic functions
of the pdfs in (5.171) and (5.172). In particular, letting 8; = ae then
ees i ef fe,(E)dE
@ Gees, (5.173)
= exp ———
1 — 2x No 1—2x
and
The bound on P;(E/my,) is given by (5C.8) and repeated here for convenience
Sec. 5.2 Performance of Noncoherent Receivers in Random Phase Channels 355
10°-
105°
P(E)5
i} >
as
PN
rie ss
Zz
as
10
7a ae eee
Sa
Saar
Be
aes
law wee
EeRar a
(ae Pee
10
Hie ae
ae
ee
PS
[ea
Dee
a
a fr ey
eee
ae
eee
ae
eee
l
ETT
MOAI
UTI
TT
VTA
ii se Se
noth) ae
Ep
Ns , dB
Figure 5.22a Bit error probability of M-FSK with both frequency and timing errors, M = 2 and p = 0.1
356 Noncoherent Communication with Waveforms Chap. 5
UII ti
TL VAT PME
IN
= oO
IAAT
°
a
'o
I
Lan N.
- Oo
wo
— Oo
©
- oOI!
~
10 12 13 14 15 16 17
Figure 5.22b Bit error probability of M-FSK with both frequency and timing errors, M = 2 and p == ()22
Sec. 5.2
at
Performance of Noncoherent Receivers in Random
5
Mal
ry
Ik
Biiiee TL
A
euil
a-
=a
oo
107 —————
aaa
Rae
Seay
er
i
107
if
—=4
Hi a
Aas!
iT
nou
Hi SIMULATION POINTS
4
[our
Son
“og is
oa
Fa
10%
10 14 15 16
Figure 5.22¢ Bit error probability of M-FSK with both frequency and timin g errors, M 4andp=0.1
358 Noncoherent Communication with Waveforms Chap. 5
Mi SIMULATION POINTS
Figure 5.22d Bit error probability of M-FSK with both frequency and timing errors, M 4 and p = (2
Sec. 5.2 Performance of Noncoherent Receivers in Random Phase Channels 359
107!
V,
'
Pe
A foo
(E) 3
P,
Figure 5.22e Bit error probability of M-FSK with both frequency and timing errors, M = 8 and p = 0.1
LA
aaML
IA
MN al iM Ul
i
360 Noncoherent Communication with Waveforms Chap. 5
i
Om
:
107
Om
(Om
Om
i
Figure 5.22f Bit error probability of M-FSK with both frequency and timing errors, M = 8 and p —=i 02
Sec. 5.2 Performance of Noncoherent Receivers in Random Phase Channels 361
M-1
Bee
P,
mat ok)
k+1
Letting x9 = 2x and using (5.173) and (5.174) in (5.175) gives after some manipulations
SE Beale) any
awe 1+kxo
(5.176)
i1—=0 i =0
i <ig<---<i, and
iy tgs tpn
1
M-1
P(E) S = eB P,(E/mn) (5.177)
Figure 5.23 illustrates the upper bound on P,(E) as given by (5.177) versus bit
energy-to-noise ratio E,/No for M = 4 and various values of p (the normalized frequency
error), assuming minimum spacing for orthogonality. It is to be emphasized that the results
in Fig. 5.23 are upper bounds and thus should not be used to predict the true error prob-
ability performance. Rather, their value is for making system comparisons and trade-offs
since the relative tightness of the bound to the exact result should be about the same in all
cases considered. As is true for most Chernoff-type bounds, they are asymptotically loose
by about | to 1.5 dB. For other bounds, see [29].
HINE
362 Noncoherent Communication with Waveforms Chap. 5
EXACT
===: BOUND
107 *
‘IN
107 ec
ail
eas ha
if
|
eC
107
a
iiihags! He
wo \
= a So-
q
S
af
}
nee
|
o
|
-
|
DS
~-
oO
oO
o 10 11 iE 15) 16
Figure 5.23. A bound on the symbol error probability of 4-FSK in the presence of frequency error
(S179)
The quantity 1 — [(J + 1)/2BT] is a measure of the fractional truncation. With truncation,
the frequency separation fj, ; (between tones i and j) must be adjusted so that they remain
envelope orthogonal at the receiver. Thus
aa ORT ain
PASSTON eI
yields orthogonal signals when truncation is employed. Hence, the minimum total transmis-
sion bandwidth B,, required for M tones must now satisfy the inequality
MCE) OM ere
aes se aaa 6 Picea
Using a result from the theory of sequences of j-dependent random variables [23], Springett
and Charles [7] are able to approximate the pdf of the spectral estimates ©, (fi; J),
1 =0,1,---,M —1 and hence the error probability performance under the assumption
P/NoB «<1 and J + 1<2BT. Under these conditions and assuming that the signal cor-
responding to f, was in fact transmitted, the pdf of ®, (f; J) at f = fi is Gaussian with
mean
{6
oy A (Ai J)} = =a
2B IN0B + Rs, (9)]
J FOBT j : (5.183)
+ oe ae Rs,(j) cos (27 (5)
where
J
Ria (jf).2P cos Baz (5) (5.184)
is quite tedious to evaluate and hence we outline only the major steps. The first term in
(5.185) can be written as a double sum, namely
sete} «Sgn HS
cgi
al! Z
(5.186)
; J ait
x E {bv (A) brv.(h)} cos 27 fi (5) cos 27 f] (=)
where €9 = 1 and €, = 2 (n £0). The elements E{dy,(j)¢,,(A)} are now considered for the
following four cases: (1) j =h =0; (2) j=0, h £0; (3) j =h £0; and (4) j Fh £0,
364 Noncoherent Communication with Waveforms Chap. 5
Eldry,D) Pr (h)}
(35) (NoB)? [2a (1+
+ 55 i 2(1 +288)|, bag (ere
2BT—H (NoB) |
[287 (1+ #0 )olRy (h), , Re Cae
» ae (NoB)? (! [oar - hn)(ay + 4h |oa (5.187)
~ (4 fn (Sa 0) j=na0
ae ee) {ear — i)) (8) + 25¢5 (2+ 8) |ea
(2-+fa)Fat?)
— (fe) tbAU Te
where
5A = (5.188)
The variance of the spectral estimates can now be obtained by weighing each j, h element of
the covariance matrix by €j€n(1/2B)?[cos 2 fi(j/2B) cos 27 fi(h/2B)], summing the re-
sulting (J + 1)? elements and subtracting the square of the mean. Although closed form
expressions can be obtained for both (5.183) and (5.185), it is simpler to evaluate them
numerically on a digital computer. Assuming that the M Gaussian distributed spectral es-
timates ®,, (fo; J), By, (fis J), >>, Pv, (fm-—1; J) are uncorrelated, an expression for the
probability of correct detection is given by
a ia 2\ {1 Of be
— w]e
PG) ce ei (= ){Serte| a aoa || dx (5.189)
where the jzz’s and o,’s are defined in (5.183) and (5.185), respectively. Also in (5.189), any
value of / #k is suitable since jz; and o; are independent of / for 1 4k (see Problem 5.8).
Numerical integration of P,;(E) = 1 — P;(C) was carried out in [7] for various values of
the normalized window width (J + 1)/2BT in percent. The results are illustrated in Fig-
ures 5.24 and 5.25 where P;(E) is plotted versus E,/No, for M = 2”, n= 1, and 5 bits.
Although the results were plotted for small percentage truncations (including no truncation),
one should be suspect of applying these results to a practical system, since the assumptions
regarding the independence and Gaussian nature of the spectral estimates break down when
J — 2BT. Comparing the no-truncation curve of Figs. 5.24 and 5.25 with the ideal nonco-
herent results of Fig. 5.11, one observes that the approximation breaks down in such a way
as to give a pessimistic estimate of the correct answer.
The bandpass filter receiver. Let’s consider the performance of the subopti-
mum receiver (binary case) when the integrate-and-dump filters are replaced by bandpass
filters as illustrated in Fig. 5.10. The two possible transmitted signals are given by
102
10 | 5 = 1/256
(E)
P;
NO TRUNCATION
4075
6.7 8.7 10.7 12.7 14.7 16.7 18.7 20.7
Figure 5.24 Probability of error performance for the spectrum analyzer receiver with
autocorrelation function truncation; n = | (Reprinted from [7])
The bandpass filters in Fig. 5.10 each have a noise bandwidth B « f, and are centered,
respectively, at the frequencies corresponding to the two possible transmitted signals. As-
suming further that Aw is large enough (B is small enough) that the BPFs don’t spectrally
overlap, then the noises at the filter outputs are uncorrelated (see Problem 5.9). Finally,
assume that B is also wide enough to assume negligible intersymbol interference (ISI).
We express the noises at the filter outputs as narrowband expansions around the ap-
propriate bandpass frequency. Thus
8 = 1/256
P(E)
Figure 5.25 Probability of error performance for the spectrum analyzer receiver with
autocorrelation function truncation; n = 5 (Reprinted from [7])
where
Aw : Aw
n(t) =Nne\(t) cos |
(x~ >) t+ | —ns\(t) sin IG ~ =) t+ | (5.192a)
Z
A
no(t) =Nn-o(t) cos |
(vx
+. >) t+ | — nso(t) sin |
(0+ >) (Pa | (5.192b)
2
Rewriting (5.191) in polar form gives
Sec. 5.2 Performance of Noncoherent Receivers in Random Phase Channels 367
A@
r(t) = &(t) cos (2.ot >) t+. + wo e Od (5.193)
where
where o? = NoB is the noise power at the output of either BPF. Since the noises nj(t)
and no(t) are uncorrelated (and thus independent since they are Gaussian), €; and & are
independent random variables, that is, at any sample time, the envelopes of the processes
r,(t) and ro(t) are independent.
To compute the probability of error, we first compute the conditional probability of
error assuming s;(t) was transmitted. Under this assumption, BPFo has no signal in it. Thus,
f(x) should be computed with P = 0, yielding the Rayleigh pdf
x2
The probability of choosing so(t) when indeed s;(t) was transmitted is the probability that
&, — & is less than zero (the threshold) at the sample time, that is
Similarly if so(t) were transmitted, then f¢,(y) should be computed with P = 0 and the
conditional error probability would be identical to (5.198) with the subscripts “1” and “0”
reversed. Thus, because of the symmetry of the problem, the two conditional probabilities
are equal and the average error probability P;(£) is equal to either one of them. Substituting
(5.196) and (5.197) into (5.198) gives
368 Noncoherent Communication with Waveforms Chap. 5
lee)
y y~2 +2P V2P y /(ee) x ( 52 )
PUR)
3° (E)= [ peesa exp (ey 452 )o( 2 hae — expPile |——
seo ] dx dy
———
ae
exp(—y2/207)
FT
i
(1/2) exp(P/207)
J2P
ae
i ie 1 ]
= 5 exp (-3) a 5 exp (-35NRer)
where SNRzp is the ratio of the signal power P to the bandpass (input) noise power NoB.
Comparing (5.199) with (5.79), we observe that the suboptimum implementation gives
“essentially” the same performance if the bandpass bandwidth B of the filters in Fig. 5.10 is
chosen equal to 1/7, that is, the bandpass noise bandwidth of an integrate-and-dump filter.
By “essentially,” we imply that we are ignoring the portion of the data spectrum that falls
outside the bandpass bandwidth B.
where the g;’s are independent Gaussian random variables each with mean jz; and variance
of and n(t) is AWGN with one-sided PSD No Watts/Hz. We further assume that the var-
ious signals {8}, j(t), 52,;(¢), +++, 8s,;(t)} are orthogonal with equal energy Ej, which is a
function of the message m ;, that is
1h
/ Si, j(t) 8), jdt = E 6.1 (5.201)
where 6; is the Kronecker delta function. The MAP decision rule selects message mj,
which maximizes the general likelihood function A(r(t)), given by
ie.) foe) 9) ap J
ar) = | of Foe festenexe |ef r(t) >> gisi, (at
=—6o) —0o 0 0 i=l
J) Sie (5.202)
J
The exponential term is the likelihood function assuming a perfectly known signal in
AWGN. The general likelihood function is its average over the unknown parameters, the
Gaussian amplitudes in this specific instance. Defining
if
uae LOse@ar,t = 1,2,..7, Jandy =0,1,2.7,M@—1 (5.203)
0
then
J ee)
ps 8;2
A(r(!) = I]ih
ae fc;(gi)exp
p }—giWi,;
No ND — E;ae } dg;
§ (5.204)
sow=[]fxHG; JiL
—
M8 dg, (5.205)
(5.207)
where
(5.209)
l o?
7S — (2u,+4—-wW;,; (5.210)
at ne Deda init NG wt
Hence, the MAP decision rule selects message m;, which maximizes
(5.211)
ileJe (2)
i=1
Assuming that the signals have equal energy, then E; = E and asa result, p;,; = pi. In this
case, we need to only maximize
(5.222)
370 Noncoherent Communication with Waveforms Chap. 5
Z a2 4
SeMs ge (5.213)
J
op
(#5‘ Mis) (5.214)
As a special case, let the means be zero (that is, 4; = 0 for all i.) Then the MAP rule reduces
to maximizing
df 2
ue tJ (x73)
No
(5.215)
2.
On the other hand, when the means are nonzero but the variances are equal (that is, 07 =
o), then we need to maximize
Sy (4 + aM) (5.216)
J Li 1 2
Both rules given by (5.215) and (5.216) will be very useful in deriving the optimum detec-
tion structures in Rayleigh and Rician channels, respectively.
=a) (5.219)
all /
Sec. 5.3 Noncoherent Receivers in Random Amplitude and Phase Channels 371
a TROPOSPHERE
b C, ey
cS G e a
a MI ~ ie ee
RECEIVER —
TRANSMITTER
Assuming that the delays {t;} are all small relative to the time constant (which is approxi-
mately the inverse of the bandwidth) of the complex signals {a,(t) exp(j@x(t))}, then r(t)
can be approximated by
where
i we CL COS dj (5.223a)
all /
rs =) cr sin gy (5.223b)
all /
Using the independence of the {7}, it is easily shown that
1 (5.225)
On = 9%, = 5 > Elci} = o*
all /
and
E{rers} =0 (5.226)
Assuming that the number of scatterers is “large,” it can be argued via the central limit
theorem that r, and r, are Gaussian, independent with zero-mean and equal variance o*.
Rewriting r(t) as
r(t) = rv
2ag(t) cos(Wct + x(t) + 9) (5.227)
where
then from Appendix 5A, r is Rayleigh distributed and independent of 0, which is uniformly
distributed over (0, 277). Thus, the Rayleigh model is the result of a channel that has many
independently located scatterers in the propagation path.
In order to derive the optimum decision rule in a Rayleigh channel, consider the signal
model of (5.222) in the presence of noise, namely
r(t) = rev 2az(t) COS(@ct + by (t)) + rsV2ag(t) sin(Wet + de (t)) + n(t) (5.229)
where n(t) is AWGN with one-sided PSD No Watts/Hz. Comparing (5.229) with (5.200),
the optimum decision rule for equal energy signals can be written by inspection of (5.216),
with xj =0, J =2 and s1;(t) = V2a;(t) cos(wet + o;(t)), 82,;(t) = V2a;(t) sin(wet +
gi (t)); namely, the receiver sets m(r(t)) = mx if and only if
2 2 2
& = Wr, + Wo;
if 2
= (/ r(t)v 2a; (t) COS(@ct + ; (dr)
0 (5.230)
T 2
=f (/ r(t)V2a;(t) sin(wet + dr)
0
is maximum for i =k. Note that W,,; = ze; and W2,; = zs; where Z¢i, Zsi are defined by
(5.14). The decision rule of (5.230) is identical to the optimum detector in noncoherent
channels with a fixed amplitude. The optimum receiver can thus be alternatively derived by
noting that the noncoherent detector with a known amplitude is the optimum noncoherent
detector for random amplitude channels, if we condition on a known amplitude. Once we
condition on the amplitude, it can be ignored as it does not contribute to the maximization
process and as a result, averaging over its distribution does not alter the structure of the
optimum detector. Note however that even though the structure of the detector is not altered,
its performance will be a function of the amplitude pdf, as will be shown shortly.
SATELLITE
addition to the Rayleigh component, as depicted in Fig. 5.27. Such channels are referred to
as Rician and the received signal, assuming message m, is transmitted, is given by
Coherent detection. When the phase 6 is known (that is, coherent detector), the
received signal can be expressed as
and
Var(re} = Var{rs} = 07 (5.234)
Comparing (5.232) with (5.200), the optimum receiver can be written down immediately
using (5.216) with J = 2, 4; =acosd, 42 = a@ sin 4, that is, set m(r(t)) = mg if and only if
acosé
——
|
+ —W,;]
: a sind \*
( ro2 46 No is) +(—W2;+
( Det ) (3.239 )
is maximum for k=i, where Wj = ie r(t)/2aj(t) cos(wet + oj(t))dt and W2,; = ie r(t)
2a; (t) sin(@cet + ;(t))dt . Equivalently, the decision rule can be modified to maximize
(5.236)
Noncoherent Communication with Waveforms Chap. 5
374
DOHPYPDPVEO
a qt
A(r(t)/5, mj) =] | (—-) exp (#4) (5.237)
eh a 4pi,j
Averaging over a uniform 4, (5.237) reduces to
P4 2
A(r(t)/m;) = Koki ¥2s) 7 (Ka,/W?, BSWw?) (5.238)
Sec. 5.4 Performance of Noncoherent Receivers 375
where K,, K2, and K3 are positive constants and are independent of j for the equal energy,
equilikely case. Since the likelihood function is a monotonic increasing function of the en-
velope (or equivalently, envelope squared), the optimum receiver makes its decision on the
basis of samples of the envelope alone, that is, Ww? ae Ww? j (see Problem 5.10). The receiver
need not know o7/aNo, which is a great practical advantage over a coherent receiver.
PAE se
5 ( eG” y+ M-1 3cm(——5)e7| awe
EET (e=)}
r No (5.239)
P(E)y=
3 ( 4)[|
P,(E/r)—
5 ( jn) Z exo(-7)a
Pons
5.240
M-\ pet ( )
= _ |Cy
dm m+
1+ m(Eav/ No)
where Egy/No & 207(E;/No). Note that P,(£) in Rayleigh channels is no longer an ex-
ponentially decreasing function of Eg,/No as in AWGN channels, but is a function of the
inverse of Eg,/No. This means that for a fading channel, the transmitter must provide a large
amount of power in order to provide a reasonably low probability of error at the receiver. In
many applications, the required transmitted power is not permissible for various reasons.
An alternative solution is obtained by means of diversity techniques [24, 25], which insert
redundancy in the channel. Depending on the nature of the fade, time, frequency, spatial
diversity, or combinations thereof may be employed (see Problems 5.11 to 5.17).
For the binary case (that is, M = 2), P;(E) as given by (5.240) reduces to
It would be interesting to assess the effect of signal fading, in the presence of perfect carrier
coherency. As a result, assume that the receiver performs an exact estimate of the carrier
376 Noncoherent Communication with Waveforms Chap. 5
phase and uses that information in the detection process. Then, the conditional probability
of error, assuming orthogonal signals, is given by (4.59), that is
piein=o(r E
| (5.242)
No,
and
ee if r2 jks 1 My 210)
PCE = — exp |-—; —_—e™’ '*dydr (5.243)
ae i o? o( 20°] SrENV20
Performing the integration, one obtains
and for comparison, we list the other two previously derived results for orthogonal signals,
namely
Es
JE3)) =O) ae M =2, Coherent Channel (5.245)
0
and
1
P,(E) = = exp'| — *), M=2, Noncoherent Channel (5.246)
2 2No
The performance of these four channels is depicted in Fig. 5.29. Note that at high E,/No,
coherent detection offers about 3 dB advantage over noncoherent detection. Simultaneously,
detection in random amplitude channels (Rayleigh channel with perfect phase measure-
ment) offers about a 3 dB gain over detection in random amplitude and phase channel.
Binary correlated signals. Other results have been derived in the literature for
correlated signals in Rayleigh channels. Turin [26] has derived the performance of corre-
lated signals for M = 2 and is given by
te
P(E) = 5 (a oe) (5.247)
“Ve pe
with = pio denoting the correlation between the complex baseband signals as given by
(5.86) and
A (Eav/No)
Se 5.248
(Eav/No) See ( )
In Rician channels, two different receivers can be employed depending on whether the phase
of the specular component is known or not.
Noncoherent detection. _Let’s consider the noncoherent receiver that selects the
signal corresponding to maximum &? = ja + roe
Sec. 5.4 Performance of Noncoherent Receivers 377
1.0
(E)
P, NONCOHERENT
CHANNEL
COHERENT
CHANNEL
Figure 5.29 Probability of error for the coherent, noncoherent, random amplitude and Rayleigh channels
M-ary Orthogonal Signals. Here again, if we condition on the fade r, then the
performance is identical to (5.77). Averaging over the Rician variable r, one obtains
M-1 1 CO >
P(E) = 0 (-1)"4
1m-1Cm m+ 1 0 o2
m=1 (5.249)
eae 1 2 2
= m
5)No we 0(5)
2 je
ey pela
ar
fi
x exp | pa +a) exp] ar
M-\
P(E) =) (-1)" |
thy-1Cm sien xe | 5.250
m=1
| Hrd Ste aby |e
where Bf is the SNR in the scatterlike signal, that is
E, '
BSA 207No (5.251)
and y? is twice the ratio of the power in the specular component to the power in the Rayleigh
component, that is
yas (5.252)
oO
As acheck, if y = 0, the Rician channel reduces to a Rayleigh channel and (5.250) reduces
to (5.240), as expected. Note that the average symbol SNR in a Rician channel is given by
for =(a?+20)Ft=(1+2)e
Z
Eav Es We
(5.253)
which for y = 0 reduces to f, the average symbol SNR in a Rayleigh channel. In the binary
case, (5.250) reduces to
IRUa v= ——.ex
Casde
Depo. {-acne
2(B +2) M =2 Rician Channel (Noncoherent) (5.254)
Here again, when y = 0, (5.254) reduces to (5.241) as expected. The performance in Rician
channels for M = 2 is depicted in Fig. 5.30 as a function of 8 for various values of y. For
y =0, the performance reduces to that in a Rayleigh channel. On the other hand, when
y = 00, it reduces to the performance of binary orthogonal FSK in an AWGN channel.
Binary Correlated Signals. The performance of binary noncoherent reception in
the presence of correlated signals has also been derived by Turin [26] and the probability
of error is given by [with 6 defined by (5.86)]
where
Te
— V0 = lad = wa?) 5.256
1 —plp\? eee
nd ie IA) — w2 1A)
w= 5.257
tial ( )
0.5
107"
102,
a
a”
107°
10-2
107” \
0 10 20 30 40 50
B, dB
Figure 5.30 Probability of error for binary orthogonal signals, Rician channels (Reprinted from [27])
and uw A B/(B +2). Note that when |f|=0, u=0, w = V2 and v=y/p/2. Since
COO X= exp(—x?/2), (5.255) reduces to
2 2
yu I yu
P(E) =exp | = 501+ wexp |-7 |
, (3.259)
Se hate {-;255 |
pd 2(B + 2)
which is identical to (5.254). On the other hand, when y = 0, v = 0 and (5.255) reduces to
Noncoherent Communication with Waveforms Chap. 5
380
(E)
P,
2
(a +20 )E, dB
No
(a) (b)
Figure 5.31 Probability of error, noncoherent receiver: (a) y = 0 and y = 00, (b) y =2
(Reprinted from [26] © IEEE)
1 rf L = bole
P(E) = 0(0,0) ~ 5 fry
(5.260)
=53(1- H er |
2 v1 — p?|p/?
which is identical to the performance of the Rayleigh channel as given by (5.247). The
performance for noncoherent detection is depicted in Fig. 5.31 for y =0, oo and y = 2 for
various values of |p|, as a function of Egy/ No. Note that orthogonal signals (that is, || = 0)
provide the optimum performance for all y and Egy/No.
Coherent detection. In the coherent Rician receiver, the decision rule is modi-
fied by the addition of the appropriate biases as given by (5.235). Since the phase 6 is known
perfectly, it can be set to zero without any loss in generality.
Binary Orthogonal Signals. _ For the binary case, the rule reduces to
a pO)
Wye 2 en 28
Woo\? 2, 0 (a Wis 2
Was 2
etree le | aed 5.261
(5 72) (=22) a Ea No ~ No ( )
Sec. 5.4 Performance of Noncoherent Receivers 381
where
Wen =tol
ee i,” r(t)V2a0(t) (cos
ee )(Wet + do(t))dt (5.262)
and
and E{r.} =a, E{rs} =0 with Var{r.} = Var{rs} = 0. Assuming that the signals are or-
thogonal, then
Wio=rcEs +m (5.265a)
W20=PrsEs + n2 (5.265b)
where n;, i = 1, 2,3, 4 are zero-mean Gaussian independent random variables with vari-
ances NoE,/2. From Appendix 5A, we know that R; is Rician where
(5.266)
with parameters
(5.267a)
> a \2
(5.267b)
nf bla a
and
| D0) NoE Ss Esk
2
= — | E’o* + = 1+ (5.268a)
°0 No ( : D) ) IN| P)
(5.268b)
The conditional probability of error, P;(E/mo), is then the probability that R; > Ro, which
is given by (5.130) for two general Rician random variables. In our case, the parameters a
and b from (5.131) become
Tee ane
(5.269)
“opto? BIB +2)
rad ss yp +1)2 (5.270)
~optor B(B+2)
382 Noncoherent Communication with Waveforms Chap. 5
and
a ee (S271)
og +o; +2
Substituting all these results in (5.130) and noting that P;(E) is identical to P;(E/mo)
because of symmetry, then
P(E) = (ats
y 4
y(B +1)
: JB(B +2) J/B(B +2)
(5.272)
(S) y? (6? +26 +2) ; (v B+ )
- (55) ex ~9BUS
ED) tle aoa)
which was also derived in [27].
Binary Correlated Signals. The performance of binary coherent reception in the
presence of correlated signals has been derived by Turin [26] and is given by (5.255) with
[p defined by (5.86)]
pV 1 — pl?
u= | —- ———_—_ (5.273)
v1— w*lal?
Peed pea/ 1 — |p|?
pk eo (5.274)
v¥ l= p7lpr
and
» —_ |BY2 2 [8G =
SW
1A?) + 20 az.—Re())]
i
eo
467(1 — |p|?)
Here again when p = 0, thenu = 1 — w=2/(8 +2), w=1+ m= (28 + 2)/(B6 + 2), and
v = (y/2)./(B
+ 2)/B and P;(E) reduces to (5.272) since uv = fa, wv = Vb, and so on.
The performance for coherent detection is depicted in Fig. 5.32 for y = 0, co and y =2.
For comparison purposes, various curves from Fig. 5.31 and Fig. 5.32 are replotted in Fig.
5.33. The results indicate that when fading does exist, there is vary little gain obtained by
the use of a coherent receiver. Of course, for a specular signal alone (that is, y = 00), the
use of a coherent receiver can result in up to 4 dB improvement over a noncoherent receiver.
But the performance of the coherent receiver for y = oo is already excellent so that the extra
advantage obtained from a coherent detection might not be needed.
In conclusion, a system that is to operate through fixed and fading paths will be for
all practical purposes optimal if it is designed to use uncorrelated signals and noncoherent
reception. The latter has the practical advantages that it does not require estimates of 5 or
o*/aNo at the receiver.
Sec. 5.4 Performance of Noncoherent Receivers 383
Re@)= +0.95
+0.9
+0.8
+0.6
+0.4
Re(p) >0
Re@)= 0.95
0.9
50
Figure 5.32 Probability of error, coherent receiver: (a) y = 0 and y = 00, (b) y =2
(Reprinted from [26] © IEEE)
384 Noncoherent Communication with Waveforms Chap. 5
0.0005
0.0002
10 2.y=2, Re(p) =9 NC
3.¥=2, Re(f) =9, ¢
5. Y = 0, Re(f) =0, Cc
0.0005
0.0002
10°
0.0005
0.0002
Figure 5.33 Comparison of probability of error curves for several important cases (Reprinted from [26] © TEEE)
Appendix 5A
Useful Probability Density Functions
In various optimum receivers, we encounter nonlinear operations on Gaussian random vari-
ables. Specifically, in noncoherent and later on in Chapter 6 partially coherent communi-
cations, the optimum receiver requires squaring operations on Gaussian or conditionally
Gaussian random variables. In this appendix, we summarize key results in this area and de-
fine useful densities, namely, the Chi-square, the noncentral Chi-square, the Rayleigh and
finally, the Rician density. A more detailed treatment of this topic can be found in general
probability theory books [11]. Here, we merely present the results as taken from Proakis
[17].
Vie Xe (5A.1)
where the X;, i = 1, 2,---,n are statistically independent and identically distributed (1.1.d.)
Gaussian random variables with zero-mean and variance o°. Then, the characteristic func-
tion of Y is given by
1
Wy(jo) 2 E {el = $e ee RS (SA.2)
(1 = j2wo?)"”
which is the consequence of the statistical independence of the X;’s. The pdf of Y is ob-
tained by the use of the inverse Fourier transform [5], that is
i n/2—1,—y/20
fyry)= onan (nJ2)> ae gy) (SA.3)
1 3 aft
iGan r="
This pdf is called the Gamma or Chi-square pdf with n degrees of freedom and is illustrated
in Fig. 5A.1 for several values of n. The first and second moments of Y can be computed to
give
E{Y} =no*
E{Y?} =2no4 +n?o* (5A.5)
pan= 2no
oy
Noncoherent Communication with Waveforms Chap. 5
386
ty
(y)
Figure 5A.1 Pdf of a chi-square-distributed random variable for several degrees of freedom
pa ee
Fy(y) =1—e 2/9 2 i (Gen) (5A.7)
When the Gaussian random variables of (5A.1) have nonzero means m;, i = 1, 2,---,n but
still identical variances o7, the characteristic function of Y is no longer given by (5A.2) but
becomes
n 2
Wy (jw) = Mae
Di | (5A.8)
(1 — j2woa2)n/2 o| 1 — j2wo?
As expected, (5A.8) reduces to (SA.2) when all the means are zero. The inverse Fourier
transform of this characteristic function yields
1 py \(-2/4 st+y s
fr) = = (5) exp |- as |5-1 (v=). y>0 (A9)
Sec. 5A.3 Rayleigh Density 387
By ny (54.10)
and I,(x) is the k'*-order modified Bessel function of the first kind [3] which can be ex-
pressed by the infinite series [4]
eA =) Fire
K(x) dLene
ee ac+740’ x>0 (5A.11)
The pdf, as given by (5A.9), is typically referred to as the noncentral Chi-square density
with n degrees of freedom. The parameter s? is called the noncentrality parameter of the
pdf. The cdf is obtained by integration as
Y 1 suy(n-2/4 stu 5
Fro)= f pee) exp |- =p |1-1 (va) du (5A.12)
Here again, when n is even (k = n/2), the cdf can be expressed in terms of the generalized
Q-function Q;(x, y) as
where
(5A.14)
2 b2 gall b J
Ey, = no> + s*
which reduce to (5A.5) when the means are zero (that is, s = 0).
(5A.17)
where the X;, i = 1, 2,---,n are as defined earlier and the means are zero. Then, the pdf of
R can be easily obtained from the Chi-square pdf with a simple transformation to give
pal r2 }
fa) = saan {ae} r>0 (5A.18)
Here again, when n is even (let k = n/2), the cdf of R can be obtained in closed form as
pa ee ee
Fair) =1~ 00 |-3-3 07 (53) ts!) (SA.19)
r i
RON SE Vere (Re ee (5A.21)
rn/2 r2 + 52 rs
Fal) = yaaa | = ce (=), r>0 (5A.23)
with corresponding cdf for n even (k = n/2)
Ser
Fr(r) =1— Ox (=,
0 0
=) (5A.24)
where s2 is defined in (5A.10). The m= moment of R is
Som:
Far) =1-0(=,-)
ST 1G
(5A.28)
When s = 0, the Rician pdf and cdf of (5A.27) and (5A.28) reduce to the Rayleigh pdf and
cdf of (5A.21) and (5A.22), respectively.
where X, and X> are i.i.d. Gaussian random variables with variance o2 and m; =m but
mz = 0. We already know that R is Rician distributed if m 4 0 but Rayleigh distributed if
m = 0. It would be interesting to derive the joint pdf fr@(r, 8) for both cases. Since R and
© are two functions of two random variables, it is easily shown [11] that
or equivalently
1
Jr.00,9)= a exp
&x |-a2 (r?— 2mr cos @ + m*I. r>0,-—7 <6 <7(5A.31)
210
if r2 +m? I mr St)
rhc 2 se ola). Views
which is the Rician pdf as expected with s = |m|. On the other hand, the marginal pdf of ©
becomes
Completing the square in the integral and then substituting x = r — mcos 6, (5A.33) reduces
to
1 m? sin? 0 oe) 2
fo@) = agi OP | |f (x +meos@)exp |- 25
2
|ds (5A.34)
Miss —m cos6 oO
or
m2
Watt
(9)
0 0.17 0.2n 0.31 0.41 0.51 0.6n O0.7n 0.81 0.97 1.07
A)
2 Tit
a = / / r(t)r(s) {cos w;t cos w}s + sin w;t sin ws} dtds
0 0
rater (5B.1)
=|! ifr(t)r(s)
cos w;(s — t)dtds
0 JO
Let
e- In yoih
ap =| [ f(t, s)dtds (5B.3)
Divide the integration on'‘T? square region in the , s plane into two integrations on triangu-
lar regions as shown in Fig. 5B.1. As such, (5B.3) can be written as the sum of two integrals,
namely
&? lst T ps
c= | / fit.sydsdt + | / f(t, s)dtds
DP o Jo 0 JO
(5B.4)
Tieet Pom
=i) i S(t, s)dsdt +f ifJ (s, t)dsdt
0 0 0 0
where the second integral is obtained by switching ¢ and s. But from (5B.2), f(t,s) =
f(s, t), thus
dt’
E? et
52 f | f.srasat
eye OBS)
= 2 | / r(t)r(s) cos w,(s — t)dsdt
0 0
Letting s =f + T, then
Es T 0
AD a | / r(t)r(t + tT) cosw,tdtdt
Bago (5B.6)
T ft
= 2f / r(t)r(t — t') cos w;t’dt'dt
0 JO
This is an integration in the triangular region of the r, t’ plane using a vertical strip of width
dt extending from t = 0 to t =T. If, instead, we choose for our integration in the same
triangular region a horizontal strip of height dt extending from t = t’ to t = T as shown in
Fig. 5B.2, then we get
£2 She Hl
or = / ifr(t)r(t — t') cosw,t'dtdt’
0 v!
z (5B.7)
ii
= 2 | COs ojr'at’ | r(t)r(t — t’)dt
0 ue
Note that ¢,(t) is not an even function due to the finite duration. Define the modified
function @, ,(T) as
Hence, the Fourier transform of ¢.,,(t) (which is identical to its cosine transform) is iden-
tical to the cosine transform of ¢,(t). This technique provides an efficient computational
method for evaluating the cosine transform of ¢,(t) using fast Fourier transform (FFT) rou-
tines on @¢ -(T).
Appendix 5C
Generalized M-ary Symbol Error Probability Bound
Consider an M-ary communication system whose decisions are made based on a relation
among M outputs xo, x1, ---, xw—1. Let these outputs be represented by independent
random variables with pdfs as follows:
Rice hi Xan)
: (5C.1)
he ar Cin) for B=; l,---,n—I1,n+1,---,;M—1
That is to say, for some particular n, the random variable x; has a fixed pdf whereas the
remaining M — | r.v.’s x;, i #n, all have the identical form pdf (perhaps different than that
for x,)! which, however, depend on a parameter ¢;, that varies with both i and n. Assuming
that message m,, is transmitted, then a correct decision is made at the receiver when x, > x;
for all i #£n. Then, the conditional probability of a correct decision is
coo M-1
jeu all[1 — Prob{x; > o/mn}] fi(oda
ae
If Prob{x; > a/my} is hard to evaluate, then use the Chernoff bound
| A(xj—a@)
Prob{x; = a@/mn} < E {e /m, | (G3)
=e *E {ei /m,} for any 4 > 0
Define
gi(a)
SE {e*"/m,} = iL e*? f\(B)dB
(5C.4)
1. If the two pdfs have identical form, then we shall ignore the “0” and “1” subscripts on them and simply
write f (x) or f(x; ¢), as appropriate. An example of where the two pdfs are, in principle, different in form would
correspond to the case of ideal (zero frequency error) noncoherent detection of M-FSK, in which case f)(x) would
be Rician and fo(x) would be Rayleigh. Also in this ideal situation, fo(x) would not be dependent on a parameter
¢ which varies with the random variable being characterized.
396 Noncoherent Communication with Waveforms Chap. 5
Then
(oe) M-\
ef no, 2)
ft [] [esos tw] ffil@de
= f: [+ Septem (SC)
oe) M-\
rz k=1
k
¥ 7 lil20(A; 7 fi(a)da
ij=0 i2=0 =O
ip .igsipAn and
ip <ig<ir<iz
or using (5C.4), simplifying and minimizing over the Chernoff parameter, we get
M-\
Ps(E/mn) < min oS(181
gi(—Ank)no
M—\M-\1 ell is
(5C.8)
es ‘
i;=0 i2=0 ix=0 j
a
80(An; Sijn)
Assuming equiprobable signals, then the average probability of symbol error is given by
1 M-1
P(E) = = DT Ps(E/mn) (5C.9)
n=0
Note that if the parameter ¢;, is independent of i, that is, all x;, i #4, have identical pdfs
fo(x), then
—1M-1 M-\1 ek
fy tasetgAn and
i] <ig<-+<ig
where
oe | oP fy(B)aB (5C.11)
Problems 397
PROBLEMS
5.1 Let z be a complex Gaussian random vector of dimension N, with
DIX,
where x and y are the real and imaginary parts of z. Let m, = E{z} denote the mean and
Kyz* = E{zz*"} denote the covariance matrix. Assume further that E {zz" }= [0] (the all-zero
matrix). Show that the probability density function of the random vectors x and y is given by
where w = wi + jW2.
5.2 Consider the following binary detection problem:
Ay: r=sin
Jaye jeer
where r is the observed vector, n is a zero-mean, complex Gaussian noise vector with
E {nn*”} =2NoL
5.3 Show that the asymptotic (M —> 00) error probability performance of the noncoherent receiver
as given by (5.77) is identical to that for coherent communications as described by (4.182).
5.4 With r(t) given by (5.80), verify that the decision variables &, k =0,1---, M — 1 are inde-
pendent of the incoming carrier phase 0, as claimed below (5.90b).
5.5 Perform the matrix multiplication TK,x7! and verify that the whitening transformation T as
given by (5.121) does indeed produce Kx, = Z.
5.6 Consider two independent Rician random variables Ro and Rj, each with parameters s; and
variance Ge (i = 0, 1), that is
2 2
ij Sea stele Tj{Sj ;
intr) = Fes] - eos |(22), r, >0
1
We wish to evaluate
Oo) Ss ro
P =| Q (2. *) fRo(ro)dro
x 0 oO) OF}
(b) Show that the Laplace transform (L.T.) of /o(2./Bt) is given by (1/s) exp(B/s), that is
[o@)
1
i In(2./ Bt)e ‘dt = —e8/5, Re(s) > 0
0 AY
or, equivalently
1 ot+joo b/s
ts ©
Io(2/Bt) = al Gis GPS)
2} Ja—joo Ss
(c) By inserting the expression for /o in part (b) into the definition of the Q-function, show that
o+joo past&
Q (v2a, v2B) = etary | () ds, o>-l
2nj -—jo §— 1
(d) By substituting the above expression in (c) for the Q-function into the expression for P in
(a), interchanging the order of integration and simplifying, show that
(e) Define
2 2
aA a] A SQ
== ee 2 2 2}
09 + OF 09 +9;
2
ey0 Oo i
D
A A
o= 2 5) cc) = 5) 5)
Problems 399
( 1 )( s 1 co
c, |—— = =
OS Al S—Cco S— Ll sto
to separate the expression for P in (d) into the sum of two integrals. Make the change of
variable eG — 1)=(p — 1) in both integrals. Using the results of (c), simplify the first
integral to show that it reduces to Q(./a, Vb). Using results from (d), show that the second
integral is given by
Therefore
show that the mean of the spectral estimate ®,, (fi; J) from (5.183) is given by
Mi 2 EAPy,
(ft; DY= ) yor
ME(t+ abe [1- ee), n=}
P_ J+ pee
oo Listen [air — 1 as
where k/T is the signaling frequency. A similar independence on / for /#4 k can be shown for
the variance of ®, (fi; /), but the mathematics required is cumbersome.
5.9 Consider a WSS random process x(t) which is input to two different linear filters G(f) and
H(f) resulting, respectively, in outputs y(t) and z(t). If G(f) and H(f) do not overlap in
frequency, that is, G(f)H(f) = 0, show that y(t) and z(t) are uncorrelated.
5.10 Derive the optimum noncoherent receiver for the Rician channel. Follow the approach as out-
lined by (5.237) and (5.238). Compute the constants K}, K2, and K3 for the unequal energy
case. Simplify your results for the equal energy case.
5.11 One form of combating fading is to employ diversity. In this case, multiple channels are used to
transmit the signal using different frequencies (frequency diversity), multiple antennas (spatial
diversity), different time slots (time diversity), or different polarizations (polarization diversity).
400 Noncoherent Communication with Waveforms Chap. 5
ny (t)
Rayleigh channel 1
(rq, 84)
Rayleigh channel 2
(rp, 82)
DECISION
i COHERENT
HO), Uneees WS RECEIVER
Rayleigh channel L
(TL; 81)
Figure P5.11
where
i (Eav/ No)
1 + (Egy/ No)
and (Eqy/No) is the average SNR per channel and is given by
Bo E
No No
(b) When (Eqy/No) > 1, use the approximations (1 + w)/2 ~ 1, (1 — )/2 ~ 1/4(Eqy/ No)
and
Ll
> eee
= k L
to show that
8 1 Eo
mi2= (ea) L )
Problems 401
Rayleigh channel 1
(ry, 84)
Rayleigh channel 2
(Tp, 89)
i=0,.1,..., M-1 NONCOHERENT DECISION
RECEIVER
Rayleigh channel L
(rL, 8)
Figure P5.15
5.13 Repeat Problem 5.12 using binary, orthogonal FSK and show that the same results hold, with
(Eav/No) replaced by (Eav/No)/2.
5.14 The transmitted electric field of a communication signal is circular polarized and used to pro-
— dual diversity in a Rayleigh fading channel, using binary orthogonal FSK signal. Assume
Ora 2:
(a) Assuming a coherent receiver, find the error probability if E,/No = 20 dB.
(b) Compare the result in (a) to the case where no diversity is used.
(c) Repeat (a) if the channel adds only white noise, that is, does not fade.
5.15 Repeat Problem 5.11, assuming that the phase information about the various channels is not
available at the receiver. In this case, the receiver combines the signals noncoherently as shown
in Fig. P5.15. Draw the optimum noncoherent receiver.
5.16 Show that the performance of the noncoherent receiver (derived in Problem 5.15) for binary,
orthogonal FSK is given by
n4(t)
Rician channel 1
(r4, 04; O44, 54)
Rician channel 2
(ro, 89; a9, 52)
cd COHERENT DECISION
ne RECEIVER
Rician channel L
(r, OL; OL, 5)
Figure P5.18
Rician channel 1
(r4, 64; Oy, 34)
Rician channel 2
(rp, 89; a, 59)
NONCOHERENT DECISION
i=0, 1,. ., M-1 RECEIVER
se ign
Rician channel L
(r_, 8; &, d,)
Figure P5.19
5.19 Repeat Problem 5.18 assuming that the phase of the specular component in channel i is not
known for all i and that the 4;’s are independent from channel-to-channel. Draw the optimum
noncoherent receiver as shown in Fig. P5.19.
5.20 In noncoherent combining,” one decision rule in M-ary communications is to decide on mes-
sage m; (j =0, 1,---,M — 1) if and only if
yi {iG 1h 2
>a:Sp = > |
(| rg(t)V2a;(t) Ccos(wet + bi(opat)
0
T 2
2. W. C. Lindsey, “Error Probabilities for Rician Multichannel Reception of Binary and N-ary Signal,’ JEEE
Transactions on Information Theory, October 1964, pp. 339-50.
Problems 403
is maximum for i = j, where r;(t) is the received signal in channel k. The combining rule is
optimum for Rayleigh channels but not for Rician channels. Assuming that the channels are
Rician with respective parameters (rx, 0,3 ax, 6x) fork = 1,2,---,L
(a) Show that the probability of error for binary (M = 2) orthogonal FSK is given by
I Ss =
BO- Gea) ?\-s5]2g\ +) k=0
1 ib,dbp ee 1+ k
P(EY = —— eats
oe (eA) Dal k Nees
where 6 = Ey/No as all the signal power is in the Rayleigh component [see (5.253)].
Hint: see [25].
BPA Given the M-FSK signal set
with $;(t) = /2/T cos[(w, + w;)t + 4] fori =0, 1,---, M — 1 and 6 a uniformly distributed
random variable. Find the minimum frequency spacing required to obtain orthogonal signals if
(a) Coherent detection is used,
(b) Noncoherent detection is used,
(c) When T = 10~’ sec and M = 1024, what is the required minimum bandwidth for coherent
detection?
(d) Repeat (c) if noncoherent detection is used.
5.22 A BFSK transmitter sends one of two orthogonal signals
for 0 <t <T. Here, @ is a uniformly distributed random variable. The receiver knows @;, E,
T, and No exactly.
(a) Sketch the optimum waveform receiver.
(b) What is the message error probability and the energy per bit required if E,/No = 2 dB and
T = 290° Kelvin?
5.23 For the binary signals defined in Problem 5.22, determine and sketch the optimum receiver if
(a) £@) = [8@ + %) + 8@ — §)].
(b) Repeat Problem 5.22(b).
5.24 Assume the same signal model as in Problem 5.22. The channel model is as shown in Fig.
P5.24. Assuming the multiplier “r” is a Rayleigh distributed random variable,
(a) Determine the optimum decision rule for noncoherent detection.
(b) Sketch the optimum waveform receiver.
(c) Derive an expression for the average bit error probability.
(d) Sketch the average bit error probability if o = 1 (o is the parameter which characterizes
the Rayleigh random variable).
404 Noncoherent Communication with Waveforms Chap. 5.
5,(t) r(t)
REFERENCES
1. Helstrom, C. W., Statistical Theory of Signal Detection, 2nd ed., Oxford: Pergamon Press, 1968.
2. Papoulis, A., Signal Analysis, New York: McGraw-Hill, 1977.
3. Abramowitz, M. and I. A. Stegun, Handbook of Mathematical Functions, New York: Dover
Publications, 1970.
4. Watson, G., Theory of Bessel Functions, London and New York: Cambridge University Press,
1944.
5. Bracewell, R., The Fourier Transform and Its Applications, New York: McGraw-Hill, 1978.
6. Charles, F. and N. Shein, “A Preliminary Study of the Application of Noncoherent Techniques
to Low Power Telemetry,” Pasadena, CA: Jet Propulsion Laboratory, Technical Memorandum
3341-65-14 (Reorder No. 65-815), November 1965.
7. Charles, F., N. Shein and J. Springett, “The Statistical Properties of the Spectral Estimates Used
in the Decision Process by a System Analyzer Receiver,’ Proceedings of the National Telemeter-
ing Conference, San Francisco, CA, May 1967.
8. Chadwick, H., “Time Synchronization in an MFSK Receiver,’ Pasadena, CA: Jet Propulsion
Laboratory, SPS 37-48, vol. III, 1967, pp. 252-64. Also presented at the Canadian Symposium
on Communications, Montreal, Quebec, November 1968.
9. Cooley, J. and J. Tukey, “An Algorithm for the Machine Calculation of Complex Fourier Series,”
Math. of Computation, vol. 19, pp. 297-301, 1965.
10. Simon, M. K. and J. C. Springett, “The Performance of a Noncoherent F SK Receiver Preceded
by a Bandpass Limiter,’ JEEE Transactions on Communications, vol. COM-20, no. 6, December
1972, pp. 1128-36.
11. Papoulis, A., Probability, Random Variables and Stochastic Processes, New York: McGraw-Hill,
1965.
12. Gradshteyn, I. S. and I. M. Ryshik, Table of Integrals, Series and Products, corrected and en-
larged edition, Florida: Academic Press, 1980.
13. Weber, C. L., Elements of Detection and Signal Design, New York: Springer-Verlag, 1987.
14. Nuttall, A. H., “Error Probabilities for Equicorrelated M-ary Signals under Phase-Coherent and
Phase-Incoherent Reception,” JRE Transactions on Information Theory, July 1962, pp. 305-14.
15. Nuttall, A. N., “Error Probabilities for Nonorthogonal M-ary Signals under Phase-Coherent and
Phase-Incoherent Reception,” Waltham, MA: Litton Systems Inc., Tech Report TR-61-1-BF, June
1961.
16. Helstrom, C. W., “The Resolution of Signals in White Gaussian Noise,” JRE Proceedings, vol.
43, September 1955, pp. 1111-18.
17. Proakis, J. G., Digital Communications, New York: McGraw-Hill, 1983.
18. Viterbi, A. J., Principles of Coherent Communications, New York: McGraw-Hill, 1966.
19. Stein, S., “Unified Analysis of Certain Coherent and Noncoherent Binary Communication Sys-
tems,” [EEE Transactions on Information Theory, vol. IT-10, January 1964, pp. 43-51.
References 405
20. Nakamoto, F., R. Middlestead and C. Wolfson, “Impact of Time and Frequency Errors on Pro-
cessing Satellites with MFSK Modulations,” JCC 8] Conference Record, Denver, CO, June 1981,
pp. 37.3.1-37.3.5.
Die Chadwick, H., “The Error Probability of a Spectrum Analysis Receiver with Incorrect Time and
Frequency Synchronization,” International Symposium of Information Theory, Ellenville, NY,
January 1969.
22 Wittke, P. and P. McLane, “Study of the Reception of Frequency Dehopped M-ary FSK,” Re-
search Report 83-1, Queen’s University, Kingston, Ontario, March 1983.
23% Frasen, D., Nonparametric Methods in Statistics, New York: John Wiley and Sons, 1957.
24. Lindsey, W. C., “Error Probabilities for Rician Fading Multichannel Reception of Binary and N-
ary Signals,” IEEE Transactions on Information Theory, vol. IT-10, October 1964, pp. 339-50.
2d: Lindsey, W. C., “Error Probabilities for Incoherent Diversity Reception,’ JEEE Transactions on
Information Theory, vol. IT-11, October 1965, pp. 491-99.
26. Turin, G. L., “Error Probabilities for Binary Symmetric Ideal Reception through Nonselective
Slow Fading and Noise,” Proceedings of the IRE, vol. 46, September 1958, pp. 1603-19.
Dike Van Trees, H., Detection, Estimation and Modulation—Part I: Detection, Estimation and Linear
Modulation Theory, New York: John Wiley and Sons, 1968.
28. Lindsey, W. C. and M. K. Simon, Telecommunication Systems Engineering, Englewood Cliffs,
N.J.: Prentice Hall, 1973. Reprinted by Dover Press, New York, NY, 1991.
Hug). Hinedi, S., M. K. Simon and D. Raphaeli, “The Performance of Noncoherent Orthogonal M-
FSK in the Presence of Timing and Frequency Errors,” to be published in JEEE Transactions on
Communications.
Chapter 6
Partially Coherent
Communication
with Waveforms
When the incoming carrier phase is estimated in the receiver, the latter can employ coherent
detection schemes as discussed in Chapter 4, provided the estimation is performed perfectly.
In many practical communication systems, the estimation results in a residual carrier phase
error, the amount of which depends on various parameters such as signal-to-noise ratio, fre-
quency stability, the estimation algorithm itself, and so on. Given the statistical description
of the phase error, the communication engineer is faced with two choices: design the de-
tector assuming a perfect carrier reference (that is, coherent detection) and then account for
any degradation due to the nonideal nature of the estimate, or as a second alternative, imple-
ment the optimal detector which exploits the available statistical information on the phase
error to obtain an improved detector. Such a scheme, which utilizes the carrier phase error
information, is referred to as partially coherent detection as it is neither coherent nor non-
coherent in the absolute sense. If the estimation is improved to the point where it is perfect,
then we expect the partially coherent receiver to approach the coherent receiver of Chap-
ter 4. On the other hand, if the estimation is degraded to the extent of providing no credible
information about the phase error, then we expect the partially coherent receiver to reduce
to the noncoherent detector, as discussed in Chapter 5.
In partially coherent communications, the structure of the optimal detector depends
heavily on the statistics of the phase error. This is to be expected since the optimal detector,
by definition, fully exploits the available information. The phase error itself is typically
a random process that we assume is slowly varying over a symbol (or multiple symbols)
and can thus be modeled as a random variable. Assuming that the communication system
transmits an unmodulated carrier for the sole purpose of aiding synchronization in the
receiver, a phase-locked loop (PLL) can then be used to track the incoming carrier phase
0. For a first-order PLL, the stationary pdf of the carrier phase error ¢, defined as 9 — A)
where @ is the locally generated carrier phase estimate, is given by the Tikhonov density [1 )
ea
AO) ’ |p| < os
fo) = Io (pg)
27
; (6.1)
0F otherwise
where pg denotes the carrier loop signal-to-noise ratio (SNR) and is given by
Partially Coherent Communication with Waveforms 407
4.0
3.5
3.0
{,) 2.0
1.0
0.5
(6.2)
In (6.2), P- denotes the power of the unmodulated carrier (which is unrelated to the power
of the information bearing signal), No is the one-sided PSD of the white Gaussian noise,
and B, is the one-sided bandwidth (in Hz) of the PLL. The synchronization aspect of the
receiver will be the topic of Volume II of this book and it suffices to state here the tracking
performance as characterized by the pdf of (6.1). Note that (6.1) is the exact pdf for a first-
order analog PLL but can also serve as an approximation for a second-order PLL, provided
that py is sufficiently large. Moreover, since the phase reference is arbitrary, we may set
@ =0 without loss in generality so that the received phase 6 is identical to @ (that is, 6 = @),
as depicted in Fig. 6.2. Partially coherent reception is a more general mode of operation than
previously considered since
408 Partially Coherent Communication with Waveforms Chap. 6
V2 cos(w,t + 6)
V2 cos wet
exp(pgcosd) _ 1
i = — (6.3)
pe>0 27 I10(pg) 20
pg>
Sane DeOe 5
=27 Io(pg)
(6.4)
corresponds to coherent communication as ¢ = 0 with probability one. Since, in principle,
p@ can take on any value between 0 and oo, all intermediate cases are possible.
where
26; Su Qe 4
ie A (= cos @; + vs) + (= sin 6) (6.8)
Sec. 6.2 Performance of Partially Coherent Receivers/One Symbol 409
and
(2&;/No) sin ¢;
tan yj; & (6.9)
(2&;/No) cos ; + pg
Using the definition of Jo(x) as given by (5.22), (6.7) can then be expressed as
1o(x;)
Sr(p/mi) ~ To(og) (6.10)
Recalling once again that Jo(x) is a monotonically increasing function of its argument, the
MAP receiver sets m(p) = m; if and only if x Ais maximum when k = i. Using z,; and Zs;
as given by (5.17), the rule reduces to maximizing
oD) 2 2 2
Xx; = No ci + Po No si .
For the case where ¢;(t) = 0 Wi and a;(t) = s;(t), then the I-Q projections z,; and Zs;
are given by (5.15) and are repeated here for convenience:
T
la = / Pe(t)si(t)dt (6.12a)
0
me
Zsi = / Ps(t)si(t)dt (6.12b)
0
1. Note [3] that if we define X; = (2/No)zci and Y; = (2/No)zs;, then the rule of (6.11) becomes (X; + po)” +
Yy or after expansion, (Xx? + Vf} + 2p4X;i, which indicates a weighting of 2g and 1 for the coherent and
noncoherent contributions.
Partially Coherent Communication with Waveforms Chap. 6
410
r%
(ing)80"
C
fe)
M
=)
A
R
A
T
r fe)
V2 cos(w.t + 6) R
r.(t)
S?)
>
LPF :
r (t) ®
KA Pr
V2 sin(wet + 6)
Figure 6.3a Two alternate implementations of the optimal partially coherent receiver
When the signals are antipodal, so(t) = —s,(t) and as a result, the noncoherent contribution
(that is, i + Za) to the partially coherent detector is independent of which signal was
transmitted since the squaring operation destroys the sign of the signal. Thus, the optimal
receiver for antipodal signals is identical to the coherent receiver whose performance in the
presence of carrier phase jitter was computed in Chapter 4 and given by (4.363) and (4.364),
Sec. 6.2 Performance of Partially Coherent Receivers/One Symbol 411
DOHYFVDYVEOO
V2 cos(w,t + 8)
(b)
Figure 6.3b Two alternate implementations of the optimal partially coherent receiver
412 Partially Coherent Communication with Waveforms Chap. 6
that is
The performance is actually plotted in Fig. 4.46 for various values of loop SNR, pg. When
the latter is less than 14 dB, significant degradation starts to occur and gets worse with
decreasing py. For the limiting cases of noncoherent (pg = 0) and coherent (pg = oo) de-
tection, (6.14) reduces to
1
lim P,(E) = = (6.15a)
po 2
IE
lim P,(E)=@Q ( ) (6.15b)
po oo No
59 = p3 (6.17)
Since both z,; and zs; have identical variance given by (5.58), then oO; and oe [equivalent to
o” in (5A.27)] are equal and given by
2, 2 2 y NoE;s
Ofs E Ss
0; = =9% =i (=)
— = 2s (=)
aatek (6.18)
Because of the symmetry of the problem, P;(E/m}) is identical to P,(E/mo). The condi-
tional probability of error, P;(E/@), is then the probability that one Rician variable exceeds
another independent Rician variable, which is given by (5.130). Hence
where
2 2
P04 waa
ih Daas ea (32) oe
Sec. 6.2 Performance of Partially Coherent Receivers/One Symbol 413
and
2
GE)
2Es
a1; P5 +4 (#)
s
po cos
i
The probability of error is then obtained by averaging over the pdf of ¢, that is
P(E)= |
base
PO uo | _ (PsE/G) dip,
27 Io(p¢)
62D)
If we take the limit of (6.22) as pg approaches zero, we obtain the result for noncoherent
reception of binary orthogonal signals, namely
which is identical to (5.79). To obtain the limit as pg — oo, it is more convenient to return
to the optimal detection metric as given by (6.13) and note that we are only interested in zo
and z,1. In this case, it can be shown that (see Problem 6.9)
which is identical to (4.59). Note that for large E;/No, the limiting cases are not very
far apart, since asymptotically Q(./E;/No) approaches exp(—Es;/2No)//27 Es/No [see
Appendix 3B, (3B.12)], so that for transmission with orthogonal signals, the amount of
coherence is not a crucial matter for performance, as it is with antipodal signals.
To compute the performance for arbitrary binary signals, we need to evaluate Z¢0, Zs0, Zc1,
and z,1 in the presence of carrier phase error ¢. Assuming that mo is transmitted, then from
(5.86) through (5.90), we have (let 6B4 Boi, a A a1)
XL =Zit+Y; (6.26)
where
é)
Z A —Zck + bp (6.27a)
No
5
ae Now (6.27b)
414 Partially Coherent Communication with Waveforms Chap. 6
Zw,0
and since the variables have been normalized, n=ao Cn= 1. The conditional probability of
error is then given by (5.130), that is
a;Ree
= 5 {a? + 6?Ben + 2a orfi cos(o — ¢)}, oul
we (6.39)
and
jp
og =a,
= /_—_8 (6.40a)
kV (2Es/No)
Bo = (1 + k)(Es/No) (6.40b)
By = —V/(1 —k)(Es/No) (6.40c)
= B
=t —— 6.4
SS at G4) (One)
-| B
¢) = tan (4) (6.40e)
Pe(E)= [Py(E/6)
My
nx
POPP
exp(0g COS f)
27 Ip(p¢)
ag 6.41)
which is evaluated numerically in Figures 6.4 and 6.5 for 8 =0 and @ between 0 and
—1. The error probability is plotted versus pg where E;/No is selected so that the error
probabilities for coherent reception of antipodal signals are 10-7 and 10~°, respectively.
These correspond to E;/No = 6.8 dB and 9.58 dB, obtained from Fig. 4.46 with pg = oo
The curve a = —1 corresponds to antipodal signals while the case w = 0 corresponds to
orthogonal signals. The effect of nonzero imaginary signal correlation (that is, nonzero £) is
illustrated in Fig. 6.6 for E;/No = 9.58 dB. It is clear that the best performance is obtained
when B = 0.
With M-ary orthogonal signals, the receiver selects the message corresponding to the maxi-
mum x,. Assuming that message m, is sent, then x,, 1 = 0,1,..., M — 1 are all condition-
ally Rician distributed with parameters as given by (6.16) and (6.17), that is
2Es 2 1 yee
9? = (78)+ 6 + ACR) PG COS, ae (6.42)
Po l - k
Furthermore, z,; and zs; have identical variances for i = 0, 1,..., M — 1, given by o2=
2E;/No. The conditional probability of correct detection is then given by
ul
=oooooo
bh
©
ODN
ReRRRRRR
0
-0.
-0.
owou -0. wr
nou
RRee
"%
Figure 6.4 Performance of the partially coherent receiver for M = 2 and real nonpositive
a (B =0, E;/No = 6.8 dB) (Reprinted from [3] © IEEE)
Since the signals are all orthogonal, the variables x, are all independent. Hence
CO
The variables x,, i =0,1,...,M —1 and i #& are all identically distributed. Therefore,
(6.44) becomes
fone) Xk M-\
P;s(C/mx, &) = [ fx,(XK) | f,,(s04x0| dx, (6.45)
0
Using the definition of Q(a, b) as given by (5A.15), the inner integral reduces to
SQ Xk
i fy, (0) dxpx0 =1—- Q (= a )
—,— (6.46)
Substituting (6.46) into (6.45) and simplifying, we have
Sec. 6.2 Performance of Partially Coherent Receivers/One Symbol 417
Figure 6.5 Performance of the partially coherent receiver for M = 2 and real nonpositive
a (B =0, E;/No = 9.58 dB) (Reprinted from [3] © IEEE)
Ps(C/mx,
&) = Ps(C/)
00 pie 6.47
=i erp (-£ = )vent Ove, 14a oy
0
where
2)
%
Figure 6.6 Effect of nonzero B (M =2, E;/No = 9.58 dB) (Reprinted from [3] © IEEE)
o4 (o°e) 24 2
P;(E)= 1 - | [ y exp & = )wen
(6.50)
_1 €XP(o¢g Cos P)
x [1 — Q(c2, y)|” dydo
27 I9(P¢)
The probability of symbol and bit error [obtained using (5.154)] can then be computed
numerically for various values of M and loop SNR pg.
Sec. 6.3 Optimum Partially Coherent Receivers/Multiple Symbol 419
where 7i(t) is a complex additive white Gaussian process with single-sided PSD 2No. The
receiver demodulates 7(t) with the complex reference signal exp { 4] (cet = 6) to pro-
duce the complex baseband signal S,e/? + n(t)e Je +4) which is then passed through an
I&D, resulting in the complex detected sample?
2. As in Section 4.1.6, it is necessary to introduce a notation for characterizing discrete time when dealing with
observations that span more than one symbol interval. As such, we shall once again use a subscript, for example,
k, on a variable to denote the value of that variable in the k"* transmission interval. To avoid confusion with the
previous notation of 6, to denote the k' message phase 27k /M (k'® hypothesis), we introduce a new Greek
symbol, that is, Bx, to denote the k'h message phase. Whenever it is necessary to represent a variable both in discrete
time and/or by an associated hypothesis, we shall use a parenthetical superscript for the latter.
3. It is convenient in what follows to introduce a normalization of 1/7, into the definition of the complex I&D
output sample. This normalization is denoted by the use of a script typeface.
420 Partially Coherent Communication with Waveforms Chap. 6
f . 1 P&tDTs | ‘
Vy, = Spel? + 7 / n(tye /°'dt |e-/
s k JkT, (6.53)
= spel? + N.e/?
where NM, is a zero-mean complex Gaussian random variable with variance 2No/T;, or
equivalently, a variance a, = No/T; for the real and imaginary components. For an obser-
vation of the received signal over the interval (k — N; + 1) T; < t < (kK + 1)T;, the sequence
v= (Viewer Ve-N, +21 tes View Vx)conditioned on ¢ is a sufficient statistic for making
an ML decision on the transmitted sequence § = (SKN, +15 SEs Niland Pika 5k).
For the assumed AWGN model, the a posteriori pdf of v given S and ¢ is
Fe 22
fy gVis.
9)= (2702) exp |— |b — 5e’9|
Vie Do
where
Nl
Iim-fu oe Ve-iSy_;
a A tan aes (6.55)
Re|ss nat
i=0
Fy P= | Fy 08.0 foods
~ uw ~
1 l a Raigad
= an Seer i Y =H 5 =i] :
Io(Po) (27202)™ of 20; 2, | ith che Sa ] (6.56)
a it iste Nox! :
Xda b=
o2
yy Vai},
TI kj
|cosa+ P¢ pgos
n | + Vi_.5*
k-iSp_; :
sing
Sau i=0
Since
N-1 N-1
ye Ve-i5f_;| COS = Re 3 Pesa fy
i= i=0
N-1 ees (6.57)
» Ve-iS¢_;| Sina = Im se
= i=0
Sec. 6.3 Optimum Partially Coherent Receivers/Multiple Symbol 421
Ce
Wh l Se
l if +vei?
ty eee(eS |pee (6.58)
\ 1o(P$) (2702) P| 22 a+ |Sk—i| ki]
1 Ned : heels :
x Ip — Re ey Vea |
+ wet) ae (is ate mst.)
; eae i=0
Note that for M-PSK, Kak is constant for all possible transmitted phases B,.. Thus, maxi-
mizing Sys) over S$is equivalent to finding
Ne=1_ 2 N,—I 4
max (Re|1 ni ae ooo) + (is & ist) (6.59a)
1=0
or equivalently
Ne=1 2
max Se Ve-iS¢_, + 060°
: i=0
(6.59b)
Ne—1_ 2 Ne
= max ys Vets, Ae 20907 Re |SD aid
e i=0 i=0
ease, 1 p&-itDTs
Saas = — F(t)Sf_; exp (—j@ect)dt (6.60)
on No J(k=i)T,
where we have arbitrarily set 6 =0 and thus 6 = , we see immediately that this met-
ric is the generalization of (6.11) for the case of multiple symbol observation (N, > 1).
As for the single symbol observation case, the first term in (6.59a) represents the compo-
nent of the decision metric associated with noncoherent (differential) detection, whereas the
second term represents the component of the decision metric for ideal coherent detection.
Thus, the partially coherent decision metric represents a hybrid of the two extremes with
a weighting of the two terms proportional to py. Note that for any nonzero value of pg, a
decision rule based on this metric is unique. For pg = 0 (differentially coherent detection),
there is a phase ambiguity associated with it since the addition of an arbitrary fixed phase,
say 0q, to all Ns estimated phases Oy N41, aie 6r_1, 6, results in the same decision for
Ok—N,+15 +++» Ok—1, %. In Chapter 7 we shall see that by letting 6, = 6,—n,+1 and differen-
tially encoding the input phases at the transmitter, that is,
where now A; denotes the input data phase corresponding to the k'® transmission interval
and 6; the differentially encoded version of it, then we can turn the decision rule into one in
which the phase ambiguity is resolved. For the moment, we assume pg # 0 and thus there is
no formal requirement for differentially encoding the data phase symbols.
422 Partially Coherent Communication with Waveforms Chap. 6
I a
WO esl rey Dd) w(u, ti) Pr {i}F > n16, 9} (6.62)
where w(u, U) denotes the Hamming distance between u and G, @ is any input sequence,
(for example, the null sequence (0, 0,...,0) = 0), and Pr {7 > ld, >} denotes the condi-
tional pairwise probability that 6 is incorrectly chosen when indeed @ was sent.
Finally, the decision statistic 7 is defined from (6.51) and (6.59b) by
Ny] 2 |?
= y ¥
Y-je
—jORj Poon
JPk-i + —— .
0 V2Es/Ts (662)
and the corresponding error statistic 7 is identical to (6.63) with each 6, replaced by By.
4. By UEP we mean that the probability of error conditioned on having sent a particular phase is independent of
which phase is indeed sent.
5. The result in (6.62) is a special case of the more general result in (4.88) wherein M is replaced by MNs and
because of the UEP condition, the double sum collapses to M times a single sum. Furthermore, a dependence on ¢
is now included in the notation for the bit and pairwise error probabilities.
r
poyoieyy % Aryaq my, Aejaq Avlaq ttn, ABloq
| 'n-4
aN L L = J t
SSouppY
= af 6 hee “yWN
t9/-? aS
ZS gS See oO
e e e
AMvy-orenbs
adoyaaua 10}99)}9p
¥490TO oredwoa
pur
9S 199]24}
1SOBIE]
“WN sajaAo
dod "LN
gqeuq ae399 aun
[ea
: : E ;
Aejap
) 0 Og)
*N-¥61+ N-F,
TH 062 60 A 61-8Vp
aanS1q
"9 OONG UONRIUOW[dWI
JOB JOATOo0I
Jo} a[dry[nus joquiés uonoajep
yo Aqened ls19Yo9 MASd-W
423
424 Partially Coherent Communication with Waveforms Chap. 6
To compute Pr {7 > ni|6, >}, we use the approach taken in Section 6.2.2 for evaluating the
probability of one Rician random variable exceeding another independent Rician random
variable. In particular, letting 7 = |z) ? and 7 = \z2|? [see (6.63) and the statement below it
for the definitions of z; and zz], then®
8s
bol a=
| es
=
E
are) |N4 ie
Gore je] (nee
Es
Ty \' MEL[Ngee ot
Po Po
EOS
?
S 2 |e
|
A ={z\*=
E
— 16 a+—*—e e J?|
pee aes =
5 [a2 =|
ee TN oree a No
N-
A =|z.
2 - HP = =l22
D - Zl? = Ns —; (6.66)
eC
SON oe
NEO v faeLarg p,
ee= args
6) 2 areaoa =arg |
Neel?+ Po
PI A arg =z2 =ar
@, {se!#+ Po
; 2Es;/No its S 2E;/No
and
Nsw) :
5 A ae of ($x-i- 4-1) (6.67)
b Es I Po
= Ns|1+— ;
te 2No |‘| TN tana po
| p W) d
- (N; — |6| cos v)
2Ns (N2 — |5|*) (aaa) é (6.68)
Es
+ ON V Ne lol
2
ae
I
(22Po
) iwsc050 — bl eos + vp
yf NSO loa
We now consider some special cases of practical interest.
be — k br—1 — Pk-1 )
0 TU 0
4 0 0
a4 4 —2
E i
b= — 442 Bt eos + 7 Bo
2No E,/No 4 \ Ey/No
(6.69)
a LY
poe
2No | 4 \ Ep/No
where E,/No = (Es/No) / logM is, as always, the bit energy-to-noise spectral density
ratio. For the third error sequence, both a and b approach infinity (the ratio a/b, however, °
approaches unity) as 6 approaches —2. Thus, we must evaluate (6.64) separately for this
case. It is straightforward to show that
a/b—>1
I | Oe
, lim, f(a, b) = 5 erfe (Ne cos ‘) (6.71)
/2E
glim, f(a, b) = 5 erfe (ae cos ) (6.72)
Finally, noting that the Hamming distance w(u, @) is equal to | for the first two error
sequences and is equal to 2 for the third sequence, then substituting (6.72) and (6.64)
combined with (6.68) and (6.69) into (6.62) gives
Finally, the upper bound on average bit error probability P, (E) is obtained by averaging the
upper bound in (6.73) over the pdf in (6.1). Figures 6.8 and 6.9 are plots of this upper bound
426 Partially Coherent Communication with Waveforms Chap. 6
Coherent
Probability
Error
Bit
Average
Bound
Upper
on
on average probability versus E»/No in dB for values of pg = 7 dB and 10 dB. For purposes
of comparison, we also show the exact results for the conventional ideal coherent metric (or
optimum partially coherent metric with NV; = 1) operating in a noisy carrier synchronization
environment as given by (6.14). We see that even with only one additional observation
bit interval, considerable savings in E,/No can be achieved at a fixed error probability,
particularly in the region of the knee of the curve where the system begins approaching its
irreducible error probability’ asymptote.
that is, there are >) sequences that yield a value 5 = —(N, — 27). Furthermore, the
Hamming weight associated with each of the (ae) sequences that yield a value 6 =
7. We have seen in Chapter 4 (for example, Fig. 4.46) and also earlier in this chapter that conventional BPSK
systems exhibit an irreducible error probability (i.e., a finite error probability in the limit as E,/No approaches
infinity) when a noisy carrier synchronization reference with fixed power is used as a demodulation signal. This is
observed by examining a curve of P,(E) vs. E,/No with loop SNR, pg, held fixed. The value of this irreducible
error probability is given by P,(E)|;,, = Ser fo(@)d@. Unfortunately, increasing the observation interval and
using sequence estimation does not reduce this error. Thus, any gain due to performing block detection must take
place prior to approaching this asymptotic limit.
Sec. 6.4 Performance of Partially Coherent Receivers/Multiple Symbol 427
10°.
\
\
Coherent
\ Ng
Bound
Upper
Probability
Error
Bit
Average
on
—(N, — 2i) is w(u, @) = N, —i. Finally then, using the above in (6.68) and (6.71) and
substituting the results in (6.62), the conditional bit error probability is upper bounded by
PE
» (Cay
eyeSete
on PINee Leaale gh s —t)
fee f (a, b Bi)
ny (6.74)
= 1see tee ( 1 )Fab
S'(N,-
witha) 4 aE
where
Pre. Ee po ih( Po if
ln|ao ik: Gand 08 + ai \2B /No (6.75)
: 5 Ns —1 Po
se E i(Ns SO aves, (-2%,) |
The upper bound on unconditional average bit error probability is now obtained by averag-
ing (6.74) over the pdf in (6.1). The numerical results are illustrated in Figures 6.8 and 6.9
for values of VN, = 4, 6, and 8. We see that as N, gets large, the curves appear to approach
an asymptote. This asymptotic behavior is analytically evaluated as follows.
For N, large, the first term in (6.74) when integrated over the pdf in (6.1) approaches
428 Partially Coherent Communication with Waveforms Chap. 6
the irreducible error probability Py (£)|;,-- = Sere fo(~)d¢. Also, the dominant term in the
summation term of (6.74) corresponds to i = N, — 1, that is, 6 = Ns — 2. Thus, for large Ns,
the second term of (6.74) approaches f (ay—1, by—1) where
paaee
an,-1
ee
2No
[w, #2V(,—D} (6.76)
Since, from (6.76), ,/by,—1 >> //bn,—1 — /@n,—1, then using the asymptotic form of
(6.64) for a and b large (see [5], Appendix A), namely
1 =
f(a, b) = ~ erfe vb- Ja (6.77)
D 2
1 E
f (an,-1, On,-1) = 5 erfc || (6.78)
independent of @. Finally then, for large N,, the asymptotic behavior of the average bit error
probability is approximately upper bounded by
namely, the sum of the performance for ideal coherent detection and the error floor.
Equation (6.79) is in very close agreement with the curves for N, = 8 in Figures 6.8
and 6.9.
The corresponding values of a and b for each of the first five cases are given as follows:
Sec. 6.4 Performance of Partially Coherent Receivers/Multiple Symbol 429
Po
+[2+ (sarin, )e# ||
b\ _£» Peo 1s tek
E |~ No 2 . Gane eer Gad) |
+|v4 —( ef -)cos + sind) ||
(6.80)
1 (INsEp
lim f(a, bp) =—Pine (cos@ + sind) (6.81)
b> +) Ns 2No
1 2Ep .
ae 5 J4 b)=-=Ree (cm (cos @ + sin ») (6.82)
\ 0
1 4Ep
slim ie b= one ( : oO (6.83)
\ No
Evaluating the Hamming distances for the 15 error sequences and substituting the above
results into (6.62) gives
Partially Coherent Communication with Waveforms Chap. 6
430
1071
' ine)
=<oO
Ideal Coherent
a =, w Block-by-Block
Detection
Probability
Error
Bit
Average
Bound
Upper
on
|
PL(E |?) < y {6f (a1, b1) + 2f (az, br) + 2f (a3,b3) + 4f (a4, b4) + 4f (as, bs)}
1 of 2Ep , f 2Ep ;
+7 yertc No (cos @+ sing) } + erfc Ne (COS SSID) (684)
_( /4Ep
rete([oo
Figures 6.10 and 6.11 are the comparable plots to Figures 6.8 and 6.9 for the M = 4
(QPSK) case. The analytical exact result corresponding to the ideal coherent metric operat-
ing in a noisy carrier synchronization environment is now
| LE
PRES / aoe (ie(cos @ + sin ®)fo(g)dg
(6.85)
+fee jen
erfc eb (cos@ — sing) } fo(d)dd
No
Here we investigate the sensitivity of the average bit error probability (in terms of its upper
bound) of Fig. 6.7 to a mismatch between the true loop SNR, pg, and the estimate of it,
Problems 431
107!
‘S ro
—sS; x)
Ideal Coherent
axae aN
Block-by-Block
Detection
' oO
Bound
Upper
Probability
Error
Bit
Average
on oO
—_
/¢, supplied to this implementation. In particular, we shall evaluate for the special cases of
Section 6.4.2 the upper bound
8
where P,, (E |, Ay) is given by the upper bound in (6.73) or (6.74) with pg replaced by
be = Po [1+ (bo — p9)/P¢| & eg [1 + €9]. Figures 6.12 and 6.13 are illustrations of (6.86)
for M = 2, pg = 10 dB, and N; = 2 and 8, respectively, with fractional mismatch eg as a
parameter. We observe that even with mismatches as much as 50% (€g = £0.5), there is
negligible effect on the error probability performance. Thus, we conclude that the block-by-
block detection receiver is quite insensitive to mismatch in the parameter /g.
PROBLEMS
6.1 The random phase inserted into the transmitted signal s;(t) = aj(t) cos(wet + 6j(t) + @) is
governed by the probability density function
]
f@)= 4 |sin gl, lp| <7
for all i=0,1...,M — 1. Determine the optimum receiver when white Gaussian noise is
present.
432 Partially Coherent Communication with Waveforms Chap. 6
By = 4 0B, €g = -0.75
Ay = 10 dB, eg = 0
Coherent
By = 16 OB, eg = 0.75
Probability
Error
Bit
Average
Bound
Upper
on
where @¢ is a random variable with the Tikhonov density given by (6.1) and n(t) is AWGN with
PSD No/2.
(a) Derive the decision rule that minimizes the probability of error.
(b) Express the probability of false alarm (choosing H; when Ap is true) in terms of the
Marcum Q-function.
(c) Express the probability of detection as an integral of the Marcum Q-function.
6.3 Consider the M-PSK communication system and assume that
fori =0,1,..., M — 1 where @ is a random variable with the Tikhonov density given by (6.1)
and n(t) is AWGN with PSD No/2.
(a) Show that the optimum receiver is independent of pg.
(b) What is the minimum number of correlators required in implementing the optimum re-
ceiver?
(c) Express the probability of error as a multidimensional integral of the multidimensional
Gaussian density.
6.4 (a) Derive the optimum receiver in an AWGN channel (No one-sided PSD) for M-DPSK
Problems 433
Ay = 10 dB, eg =0
By = 16 cB, Ey = 0.75
By = 4 dB, €g = -0.75
Coherent
Bound
Upper
Average
Probability
Error
Bit
on
signaling if the phase uncertainty is given by the Tikhonov density. Assume signal models
of the form
SH=H=VZEcosaoct. OUSts Tf
and the channel randomly phase shifts so(t) or s;(t) through @ radians with probability density
function
and then adds white Gaussian noise (No one-sided PSD) to the resulting signal to produce the
received signal
r(t)=s(t,@)+n(t), O<t<T
fa(8) = “PCSES”
7 19(p5)
alc
and that the 6;’s are independent from each other and all other random variables in the model.
(a) Derive the optimum receiver when L = 2.
(b) Generalize your result to an arbitrary L.
(c) Simplify the optimum receiver if ps = 0.
(d) Simplify the optimum receiver if ps = oo.
6.9 Verify for binary orthogonal signals that the limit of (6.22) as pg — oo is indeed given by
(6.24).
6.10 Verify that the whitening transformation as given by (6.31) produces Kx, = TZ as claimed in
(6.33).
6.11 Consider the received signal of (6.52) corresponding to the transmission of M-PSK over an
AWGN channel with unknown phase @. Assuming binary modulation (M = 2), then passing
r(t) through an I&D filter produces at time t = (k + 1)T), the received sample
ee
pe 2h e
sine A
Th
where 6, denotes the transmitted information phase which takes on values of 0 or z and nx is
a sample of a zero-mean complex Gaussian noise process with variance 0,2 = No/ Tp. Consider
a received sequence r = (Feehy 7x-2,-.-,F%k—n,) Of length N, and assume that the unknown
channel phase @ is constant (in time) over the length of this sequence.
(a) Write an expression, analogous to (6.54), for the conditional pdf of F given the information
phase sequence @ = (0x—1, 9-2, ..., ®%—N,) and channel phase @. Show that this pdf can
be expressed in the form
Ny
fro © 10,0) = Fexp fRe lieuer ee)
i=]
ee VJ 2E,/Tp 2) V2E,Tp
Oo; N0
(b) Show that the average of the conditional pdf feo F|8,0) over the information phase
Problems 435
sequence is given by
Ns
Fro © 19) = Eg {fro F 10, Oy) =F | [cosh (a Re {7-1e 7°})
t=!
For a given observation (received sequence), we would like to determine the maximum-
likelihood estimate, Our. of the channel phase 6 and then use this as a demodulation
reference phase resulting in a pseudocoherent receiver (a form of partial coherence).
(c) Show that 6u L defined by
is the solution to
Ns
)- tanh (a Re {7-1e7/°} Re {jr ie /?}) =0
i=1
(d) For small SNR, we can approximate tanh x by x. Making this approximation in the result
of part (c), show that the maximum-likelihood phase estimate is given by
1/2
(e) Sketch in block diagram form the complex representation of a pseudo-coherent receiver
that implements the above phase estimation and performs detection of 6; using Out (actu-
ally e/ but) as a demodulation reference.
6.12 Redo parts (b)-(e) of Problem 6.11 for the case of arbitrary M = 2”. In particular:
(a) Show that the average of the conditional pdf ffi9 (f|@,0) over the information phase
sequence is now given by
N, 3-1
fr 10) = Eg {fe @18,6)} = FT | Y>cosh (aRe fixie 104271} )
i=! /=0
where F’ is a constant independent of the data and a is as previously defined.
(b) Show that bur is the solution to
Me
M
O+2"/M) |
_je
Ds sinh (a Re {7_je-/0+27"/)}) Re {j7,J
Ns
4
DY cosh (a Re {7—j;e7/@+27!/M)})
i=1 2
=O
(c) For small SNR, we can make a first-order approximation of sinh x by x and cosh x by 1.
Making these approximations in the result of part (b), show that it reduces to
IS ! Ny i
% Re tfB pede Re {ifBs et (Oui+2n1/M) =
1=0 i=1
which does not yield a unique solution for 6u L since the left-hand side of this equation is
identically equal to zero for any @y, and M > 2 (show this).
436 Partially Coherent Communication with Waveforms Chap. 6
(d) To get an approximate solution for 6u_, we must take the next higher order terms in
the Taylor series expansions of sinh x and cosh x, namely sinh x = x + x>/6 and cosh
x = 1+x?/2. Using these approximations in the result of part (b) and recognizing that
(1 + x2/2)~! = (1 — x?/2) for small x, show that, for QPSK (M = 4), the maximum-
likelihood phase estimate is given by
1/4
Ny /
a mm Vi
clout — i=l
Ns
<4
Thi
ji=1
1/M
Neat /
. ye,Peei
elOmL — i=l
N65,
om Tey
i=]
6.13 Reconsider Problem 6.11 where now the unknown channel phase 9(t) varies with time in
accordance with a frequency offset, Af, due to Doppler. In particular, 6(t) at time t = kT),
is related to 6(t) at time t = (k —i)Tp by 0(kT,) = 6 ((k —i)T,) + 207i Af Tp. As before, it is
desired to find an ML estimate of 0(t) now, however, in the presence of frequency offset.
(a) Show that the conditional pdf of the received vector r = (Fe-1, tee 7k—N,) given
the information phase sequence 6 = (6-1, Cpa e wr Glee N,)s channel phase 6(f) at time
t = kT), and frequency offset Af can be expressed in the form
Ns
fioar (£10, 6 (kT) ,Af) = F exp feRe ogame mle
=
| [oosh(aRe|
=F | | cosh(aRe |Ry_je
1° 4™
!)
where
Reyeipge
(c) Assume that the frequency offset is completely unknown and as such has a uniform pdf
between — fax and fmax Where fmax corresponds to the maximum expected Doppler fre-
quency shift. Compute the conditional pdf fi9 (if|6 (k7;)) by averaging the result of part
(b) over this uniform distribution.
(d) Show that the ML estimate, 6uL, of the channel phase 6(t) at time t = kT} as defined by
9A (kTs)a1
Seneaat
= max -le
~'feo (F(z 19 (KT)) == max -1 "In fay (F10
=
(KTS)
References 437
is the solution to
a ; io ‘
Smax Ns
/ (aRe ey
B sinh ( TAO (Re Wee ol)
Simax i=
sy
(e) For small SNR, we can approximate sinh x by x and cosh x by unity. Making these approx-
imations in the result of part (d), show that the ML phase estimate is given by
1/2
, Yi
ie_;, Sine (4 fmaxi Tp)
el? &To) nx — | t=!
REFERENCES
1. Lindsey, W. C., Synchronization Systems in Communication and Control, Englewood Cliffs, N.J.:
Prentice-Hall, 1972.
2. Lindsey, W. C. and M. K. Simon, Telecommunication Systems Engineering, Englewood Cliffs,
N.J.: Prentice-Hall, 1973. Reprinted by Dover Press, New York, NY, 1991.
3. Viterbi, A., “Optimum Detection and Signal Selection for Partially Coherent Binary Communi-
cation,’ JEEE Transaction on Information Theory, vol. IT-11, pp. 239-46, April 1965.
4. Simon, M. K. and D. Divsalar, “Multiple Symbol Partially Coherent Detection of MPSK,” sub-
mitted to the JEEE Transactions on Communications.
5. Schwartz, M., W. Bennett, and S. Stein, Communication Systems and Techniques, New York:
McGraw-Hill, 1966.
Chapter 7
Differentially Coherent
Communication
with Waveforms
in the system solely as a means of resolving the phase ambiguity associated with the demod-
ulation reference signal supplied by the carrier synchronization loop. Stated another way,
differential encoding is not an essential element of the coherent detection process provided
that another means for resolving the phase ambiguity of the carrier demodulation reference
signal is employed. On the other hand, for differential detection, the inclusion of differential
encoding at the transmitter is absolutely essential! This will become clear shortly.
Perhaps the earliest known (to the authors) citation of the above ideas, herein re-
ferred to as classical multiple differential phase-shift-keying (M-DPSK), is due to Fleck and
Trabka and appears in a 1961 document edited by Lawton [1]. The receiver structures pre-
sented in [1], which were based on observation of the received signal plus noise over an
interval of two-symbol duration, were obtained without regard to their sense of optimality.
We begin our discussion by showing that classical M-DPSK is indeed optimum (based on
maximum-likelihood (ML) considerations) for a two-symbol observation interval. Later in
the chapter, we shall show that if the channel phase characteristic can be assumed constant
over N,7T; seconds, N, > 2, then by extending the observation interval to N,7; seconds and
applying an appropriate maximum-likelihood detection rule, one can further improve the
performance of the system. In fact, by making N, larger and larger, one can approach the
performance of a coherent M-PSK system with differential encoding (see Section 4.1.6).
More about this later on.
1. It is understood that the tan! operation used to define the phase nj; in Fig. 7.1 is meant in the 4-quadrant sense
and not in the principal value sense.
440
O[Npour
uz esooyD
ISoTTeUIS
ur
(4)4 ainjosqy
oneA
[apoio
[eo
IO}eT[DSE
eansiy
TL woaundg J9AIVO0I
JO} [edtsse[d ASdd-W
Sec. 7.1 Classical M-DPSK 441
at the final decision on the information phase A6;. As such, the differential detector intrin-
sically includes the differential decoding operation and thus no separate differential encoder
is required.
The receiver of Fig. 7.1 can also be redrawn in a different form as in Fig. 7.2a, which
takes on the simpler form of Fig. 7.25 when M = 2. This receiver was suggested in [4] as
well as several other places and its equivalence with Fig. 7.1 was demonstrated in [5] (see
Problem 7.1).
The decision algorithm used in the receiver in Fig. 7.1 is quite analogous to that in Fig. 4.32b
in that the decision regions are still defined as wedges uniformly spaced around a circle (see
Fig. 4.22). Here, however, the phase associated with the signal that defines each of these
regions is the value of the phase difference 6; + 6 — (6;-; + 9) = 0; — 6;-1 = AQ; and the
decision variable that falls into these regions is the phase difference yy; = nj — ni_1 modulo
27. In particular, if 0; — 0;_; = AO; = Bx 1s the value of the phase communicated in the ith
transmission interval, then a correct decision is made when y; falls in the Ee wedge defined
by the angular interval (8, — 27/M, By +72/M). In mathematical terms, the conditional
probability of a correct symbol decision given that 6; — 6; = A0; = Bx is
ca
TACO)Wea 0 io0 eal a) Cea 2) |e
Cai Bee, M
Bx G2)
apr(ii—alsi}=] fxooax
a/M
where x = Wj — Bx. Thus, analytical evaluation of symbol error probability reduces to find-
ing the pdf of X.
In [1], the pdf fx (x) was determined by first expressing it in the form of a convolu-
tion integral based on the pdf’s of n; and nj;—) (since n; and n;— are statistically indepen-
dent) and then evaluating this integral using the method of characteristic functions with the
result
Since this pdf is an even function of x and independent of the particular phase 6, transmit-
ted in the i" interval, then P,(C|k) is independent of k and hence the average symbol error
probability is given by
442
- YLT
UIS W I-W""“TT0=7'
{74}
YLT
eye) W I-W“ZTT0=7'
[eo]
JO}e]]DSO
aaNS1yBZ" Uy aJeusoye
WO}JO emBL{
I'L
Teo]
10}eT[DSO
BANBIYqZ*Z,
= UY JUgTeAINDaJoATa001
0} OINB1JZL IO} W = Z
443
444 Differentially Coherent Communication with Waveforms Chap. 7
1 M ma/M ;
Safi fu (Ww) dy
m/M
where we have arbitrarily assumed 6, = 0 and hence x = —. Substituting (7.3) into (7.4),
we observe that P,(E£) is in the form of a double integral, which must be evaluated numeri-
cally.
More recently, the evaluation of (7.2) has been carried out as a special case of the
more generic problem of finding the probability distribution of the phase angle between two
vectors, each perturbed by additive Gaussian noise. The solution to this problem is discussed
in great detail in [4] and summarized here in Appendix 7B. For the M-DPSK receiver of Fig.
Hel @asenl applies” with yy; = 2/M and w2 = z. Thus, from (7B.3) and (7B.4)
sin
:
Z 7/2 exp |- 7g
45[1 — cos 7,
It
cost]
=0= dt} (7.5)
4x J_n/2 1 — cos 4, cost
2
sin = (eo exp -4 [1 = COS 7,cos ]|
ce on a eR a ee ca
4x J_x/2 I — cos ¥, cost
Substituting (7.6) into (7.5), the symbol error probability of (7.4) becomes
Es
sin Z #/2 eXp {—# [1 —cos %I cos]
PCE) = Sit f i es eee es (oT!)
1 —cos 47 cost
Uv
21 —7 {2
A exp (— A?)
PAB) = ee ee (7.8)
ae lee a1 EH
2. Since, as mentioned above, P; (C|k) is independent of k, then for convenience we again take k=0 (Bk =
Bo = 0), which implies that the transmitted signal is the same in both the (i — 1)*' and i" intervals.
Sec. 7.1 Classical M-DPSK 445
M Q
3 0.9186
4 0.9768
8 0.9987
16 0.9999
where
A == J2E,/N
5 0 sin
1 (——)
=e
2M (7.9)
é
This expression is simpler to evaluate but less accurate than (7.8). Bussgang and Leiter [6]
derived an upper bound on P;(E£) which for large EF /No is itself a good approximation. The
result is
P,(E)
& erfe — sin a) (7.11)
ell JI+2E;/No M
When both E,;/No and M are large, (7.11) resembles (7.10) (since for small x, sinx = x);
however, for smaller values of E/No, (7.11) is the more accurate of the two.
Although all three of the above approximations are good for large E/No, none are
asymptotic in the sense that the error in using them becomes vanishingly small as E;/No
becomes infinitely large. Pawula et al. [4] obtained a true asymptotic (for large E;/No)
formula for symbol error probability. Their result is
M E
P;(E) Ile 1+ cos7/M erfc {| _/— [1— cos =| wo M>3 (7.12)
2cosm/M No M
The Fleck and Trabka approximation, that is, (7.8), is closest in form to (7.12). To compare
the two, we form the ratio Q of (7.8) to (7.12) by first neglecting the 1/8 factor in the
denominator of (7.8) and then employing the asymptotic expansion of the erfc function in
both (7.8) and (7.12). The result is
The special case of binary (M = 2) DPSK, often simply called DPSK, is a case where
the error probability performance can be found in closed form (as opposed to an integration).
In particular, letting M = 2 in (7.7) gives the simple result
This result was independently obtained by Lawton [7] and Cahn [8]. Figure 7.3 is an il-
lustration of the exact expression for symbol error probability of M-DPSK as obtained
from (7.7) or (7.4) versus E,/No in dB for M = 2, 4, 8, 16, 32, and 64. Superimposed on
these curves are the analogous results for the symbol error probability performance of co-
herent M-PSK as obtained from (4.131).? Asymptotically for large M and large E,/No,
we see that differential detection of M-PSK requires 3 dB more E,/No than coherent de-
tection of M-PSK. This result can be obtained quite easily by comparing the effective
SNR’s for these two detection schemes as determined from the arguments of the erfc func-
tions in (4.139) and (7.13). In particular, for coherent detection the effective symbol SNR
is given by (Es/No)er¢ = (Es/No) (sin? a /M) = (Es;/No) (1 /M)* whereas for differen-
tial detection (Es/No)ege = (Es/No) (1 — cos /M) & (Es/No) [1 — (1 — (t/M)? /2)] =
(Es/2No) («/M)°.
7.1.2 Bit Error Probability Performance
Assuming a Gray code bit to symbol mapping as was done in Chapter 4, the bit error prob-
ability performance of M-DPSK can be related to its symbol error probability performance
evaluated in the previous section by the simple approximate relation given in (4.140). It is
also possible to obtain an exact expression for the bit error probability performance using
the method of Lee [11] combined with the results of Appendix 7B. For example, for 8-
DPSK, Lee obtained the result
2 137 TU
Po(E) = 5 lF(=) bs 3 (=)| (7.15)
where F(y) is given by (7B.4) with U = E,;/No. This result was later generalized by Korn
[12] to other values of M. In particular
Pes |ad(<2)
r(e)=5[F el Vly F(=)
peas Vek F(3)-F()]:
of eV ie M=16 (7.16b)
3. Using the results of Appendix 7B in the degenerate case where one of the two vectors is noise free, Pawula
et al. [4] have been able to arrive at a simpler expression for the symbol error probability performance of coherent
M-PSK, namely
1 m/2—n/M E 1
P(E) = — | ex |— sin? &sec? ao
: X J_x/2 : No M
1071
10-2
me
WW
qa»
10°3
10-4
—— COHERENT
DETECTION
-—- M-DPSK
10-5 Figure 7.3 Symbol error probability
performance of M-PSK and M-DPSK
b’ 0’ (Reprinted from [10])
which represents the Gray code approximation of (4.135). When compared with coherent
detection, some typical values of E,/No degradation at P, = 10-5 are 0.75 dB for M =2,
2.2 dB for M = 4, and 2.5 dB for M = 8.
Suppose now that because of channel bandlimitation, the amplitude of the received sinu-
soid is time-varying and the power spectral density of the received noise is not flat for all
frequencies. In particular, the received signal of (7.1) now takes the form
where n’(t) is additive Gaussian noise with power spectral density [S,(f — fo)+
Sa(f + fo)]/2 and S,(f) = F {Rn(t)} is a symmetric function. After demodulation with
inphase and quadrature reference signals as in Fig. 7.1 and passing through the integrate-
and-dump (I&D) circuits, we get the lowpass signals
448 Differentially Coherent Communication with Waveforms Chap. 7
where
] (i+1)Ts
Aj = — A(t)dt
(7.20)
(407, G+T,
Ni j= / n'(t) (v2cos wt) dt; Nij= / n'(t) (v2sin wt) dt
: ie F Hie
If, as in (7B.1), we relate the complex form of these I&D outputs to the geometry in
Fig. 7B.1, then the association is now
iT i+1)Ts
G_DT. ho
iT;
Ts R(u — v)dudv
; =0
o2 (
For binary DPSK, Pawula and Roberts [13] have been able to evaluate the bit error
probability performance in closed form (in terms of tabulated functions). For the binary
case, the error probability can be written in the form
—n/2
P, (BE) = Pri Ag=0} ([ fu wav)
32/2
m/2
+ Pr{Ad=n) ( fu Wav) (7.22)
—n/2
—n/2 m/2
a (/—3n /2 e wav) eR (=a) fe way)
where Po and Pa are the a priori probabilitiest of the two possible information phases trans-
mitted in thei" interval. Evaluation of the integrals in (7.22) follows the results correspond-
ing to Case 2 in Appendix 7B. For the case where the limits of (7B.3) are separated by z
radians, as is true in (7.22), Pawula and Roberts [13] have been able to evaluate the integrals
in closed form. In particular
4. In their analysis, Pawula and Roberts [13] allow different a priori probabilities for the information phases 0
and zr.
Secaval Classical M-DPSK 449
Wo
Bl) ee wl =i} fu (Wd
Wo-m
rsin Yo e€“Io(B) .
+ ’
24/1—r2 cos? Wo
elsewhere
where Je(k, x) is the Rice Je-function [14] defined by
] Ww V r _u
PoE) = 5 {1
Sate (Gu) ]+5 0% - Ave Io(V) (7.26)
where U, V, and W are defined in (7B.2) using (7.21). From (7.26), we observe that the error
probability does not depend on the noise correlation r so long as the a priori probabilities
Po and P, are equal, whether or not there is a power imbalance. When there is no power
imbalance, V = 0, and (7.26) reduces to
|
P(E) = ; (1+ (Pr — Po) rjerZ (7.27)
which was previously obtained by Lee and Miller [16]. When in addition Po and P, are
equal, then (7.27) reduces to (7.13) as it should.
The effect of bit timing error on the bit error probability of an otherwise ideal DPSK receiver
was first considered in [17] and later in [18]. We present only the results from [18] since
they are based on the approach taken in [4] and are thus in simpler form. In particular, if we
denote the bit sync error, as in Section 4.3.3, by ¢ = AT» where A represents the normalized
(with respect to the bit time) error, then the block diagram of Fig. 7.1 applies if we replace
450 Differentially Coherent Communication with Waveforms Chap. 7
the lower and upper limits of the I&D’s by (i — 4) Tp and (i + 1 — A)T), respectively. Having
done this, the outputs of these circuits become
Vei = PT;
V [(1 — A) cos (6; + 8) + Acos (6-1 + 8)] + Ne,i
(7.28)
V,,i = WPT, [(1 — A) sin (6; + 8) + Asin 6;-1 + )] + Ns,i
where
(i+1-A)Tp
Noi & } n(t) [v2cos wet dt,
(i—A)Tp
(7.29)
(i+1-A)Tp
Nowe / n(t) |-v2 sin wet dt
(i—A)Tp
are independent zero-mean Gaussian random variables with variance o*= NoTp/2. The
equivalent phase nj; for the i" transmission interval is found from the arctangent of the ratio
of V, ;to V..; and using (7.28) can be expressed in the form
where
A similar expression can be written for the equivalent phase, n;_1, in the (i — 1)‘ transmis-
sion interval. Once again the geometry of Fig. 7B.1 applies, where now
r=p=0
9! ?
ee aisAil) Fh|se)202 No 2
(7:33)
2\202
For DPSK, the information phase, A6;, to be communicated in the ith transmission
interval takes on a value of 0 or z. For each of these values we must compute the conditional
bit error probability when A6;_; takes on these very same values (with equal probability).
To do this, we apply the results of Case 2 in Appendix 7B, in particular, the special case
of these results given in (7.23)—(7.25). To begin, we observe from (7.34) that for A6;_;
equal to either 0 or 7, AD = A@;. Furthermore, for A® = 0, we have wo = —2/2 while
for A® = 2, we have wy = 7/2. Thus, the above conditional bit error probabilities are
evaluated from (7.23) as
P, (E|0,4)=Pr{-—32 <y<-~}=-|1-
2
1
1 ‘
Bre
ao
(2.
ao
ao)
(7.35)
Pp (E| 2,/Ay=Pr{->
A) r a<
1 1
<tl=5|i=
eae
I |, Be, ol(
pee)
Pite
(Bx 2)
ee
wie A6;-1 =0
Ng lb 2a Aye, Abeay ser
i Ep 0; AG;-7 = 0
Pig Nea Sin A Gi) arr
(7.36)
Ep. {V5 20G= 2); -AG-1=0
EBaciingg: os [iden) AGj-1=7
En [200 = a)} AG) =0
Pe = 0; AGj;-1 =
The average error probability conditioned on a given timing offset A is then obtained by
452 Differentially Coherent Communication with Waveforms Chap. 7
=;[1- (0 x) 737)
+5 ce ee)
+5 i fy - POO te ( 1-200 -2 F)
a ; Ae (0.(ha |
Ep
+ exp |
x [4A(1 — ai}
which clearly places in evidence the degradation in performance due to bit timing error
relative to the ideal (perfectly timed) performance of (7.14). Note that when A = 0, D
reduces to unity as expected. Assuming a Tikhonov pdf for the normalized timing error as
was done in Chapter 4 [see Eq. (4.324)], the average bit error probability performance of
DPSK in the presence of timing error is obtained by averaging (7.39) over this pdf and is
illustrated in Fig. 7.4 for various values of rms normalized timing error o}.
In designing receivers that employ error detecting and error correction codes, the occur-
rence of multiple consecutive errors, that is, burst errors, is an important consideration. For
Sec. 7.1 Classical M-DPSK 453
uncoded coherent receivers that make bit by bit decisions, such as those discussed in Chap-
ter 4, the probability of a burst error is simply the product of the bit error probabilities over
the length of the burst, or equivalently, the average bit error probability raised to the power
of the burst length in bits. The reason for this is that in a coherent system, the bit decisions
are independent of one another. For uncoded DPSK, since in effect the phase of one signal
acts as the phase reference for demodulation of the next signal, the decision statistics are not
independent from bit to bit and hence the probability of two or more consecutive errors is
not the product of the individual bit error probabilities over the length of the error sequence.
The analysis of double error rates in DPSK systems was first considered by Salz
and Saltzberg [19]. Although their results were derived for a suboptimum receiver (to be’
discussed later in this chapter), the application of their approach to the optimum DPSK
receiver of Fig. 7.2b follows directly. In particular, let
Veg = PT,
V 0086; + Nei, Vo,i = PTp
V sin; + Ny.i
Here again, N..; and N, ;are independent zero-mean Gaussian random variables with vari-
ance 0” = NoTp/2. The decision on the information phase, A6;, communicated in the me
transmission interval (or, equivalently, the decision on the binary bit a; = cos A@;) is then
The approach taken in [19] is to first determine the joint pdf of Z;-; and Z; condi-
tioned on V,,;—; and V, ;-; being fixed. This conditional joint pdf is Gaussian and, since for
454 Differentially Coherent Communication with Waveforms Chap. 7
1
f (Zi24: Zi\ Ve.i—1, V5) =
2x0?| (Ve,i-1) + (Vsi-1) |
2 y (741)
1 (Zi-1 = VP Tp 008 6;Ve,i-1) + (Zi — /PTp 0s6;-2Ve,i-1)
x exp 4 ———s>
20? (Vei-1)° fe (Vea)?
Averaging (7.41) over the joint pdf of V.,;-; and V;;-; which is also Gaussian,
namely
2
TAY iain) ee exp |x (V1 -— VPT»c0s6}-1) + (Ves) }(7.42)
we obtain the unconditional joint pdf f (Z;~;, Z;). This pdf is a function of the three
transmitted phases 6 = (6;—2, 6;—-1, 9;) which for binary modulation takes on eight possi-
ble values. Associated with each of these transmitted phase sequences is a double error
probability, Pp, (Fi-1, Fi); k= 1,2,...,8. For example, if @ = (0, 0, 0), or equivalently,
A6;-; =0, AG; = 0, then a double error (two consecutive bit errors) occurs when Z;_; < 0
and Z; < 0. This double error probability is given by
0 0
Pp (Ei-1, Ei)lo=c,0,0) & Poi (Ei-1, Ei) = f (Zi-1, Zi) dZj-1dZ; (7.43)
—oo J—00
which is an integral over the lower left quadrant of the Z;_,, Z; plane. Salz and Saltzberg
[19] show that, with appropriate changes of variables, (7.43) together with the other seven
integrals of this type can all be expressed in the form of two integrals each of which in-
tegrates over the upper right quadrant of the Z;_;, Z; plane. Since each phase sequence
6 occurs with probability one-eighth, then averaging P,, (E;-1, E;); k=1,2,...,8 over
this equiprobable distribution gives the desired result for average double error probability in
the form of a quadruple integral, namely
Pt Ee >» Pil
oyEe Bi) =e Fo/Ma(& aide: fae
, (7.44)
Ey.) Cia): a Ga) 2, .2
x exp |_5 [fc aa + x4 + yp + 2x1 | } dxidx2dz\dz2
Finally, it is also shown [19; Appendix I] that this quadruple integral can be simplified to a
single integral, which yields the desired result
Sec. 7.1 Classical M-DPSK 455
Pp EEE =e
RNG | 1—erf-|
Ep
./—cosé
;
4 0 No
2
+ |am cos getEe/Nayeos?6 11 _ ore (|Be cos @ (7.45)
No No
E
x +a (soo) |
\ No
Oberst and Schilling [20] have also considered the double error probability perfor-
mance of DPSK and arrived at (7.45) in a somewhat simpler manner. Although they too
considered the same suboptimum configuration as Salz and Saltzberg [19], their approach
and mathematics is nevertheless applicable to the optimum DPSK receiver of Fig. 7.2b.
Oberst and Schilling [20] base their approach on the observation that DPSK can be viewed
as the signal in the (i — 1)*' transmission interval (i — 1)T, < t < iT; acting as a noisy de-
modulation reference for the signal in the ith transmission interval iT, <t < (i + 1)Th. They
referred to these two signals, respectively, as the “reference” and the “data.” The clever trick
employed in [20] is to switch the names of the “data” and “reference” signals for the in-
tervals (i — 2)T, <t < @@ — 1)7T; and (i — 1)T, <t <i7}; that is, one should think of the
signal in (i — 2)T, < t < (i — 1)Tp as the “data” and the signal in (i — 1)T, < t <iTy) as the
“reference.” This allows one to view the two demodulations ordinarily occurring at t = iT)
and t = (i + 1)7) as equivalently having been performed independently (with regard to the
additive noise) at t = (i — 1)7, and t = (i + 1)T) by acommon reference signal. In mathe-
matical terms, the probability of error for the i" bit conditioned on the phase estimate nj;_1
associated with the previous bit is given by
1 E
Pp (Ej \ni-1)= ~yyerfc (Ji
No
cos n-1] (7.46)
where
es
i Mirae eet?
I Ep nj-1e —(E,/No)
sin? PV ni—1 hl ert Ep COS nj-]
a —,_2 |xNo
—— cos Ni-1 —No Ni ; (7.47)
= Snare
Based on the above discussion, the double error probability conditioned on 7;— is then
1 E =
Pp (Ei-1, Ei lni-1) = ;
2;
erfc (2
No
cos n-t)] (7.48)
Averaging (7.48) over the pdf in (7.47) produces, after some manipulation, the identical
result to (7.45).
456 Differentially Coherent Communication with Waveforms Chap. 7
FEE;
THEORETICAL
O SIMULATION
The conditional error probability Py, (FE; |Ei—1) = Pp (Ei-1, E;) /Pp (E;—1) repre-
sents the ratio of double to single error probabilities, which, in view of (7.14), is given
by
Pp (Ej-1, Ei) _
P, (E; |Ei-1) = 2e%/Nop, (E;_, Ei) (7.49)
Py (Ei-1)
Figure 7.5 illustrates this ratio as a function of E,/No obtained theoretically from (7.49)
together with (7.45) as well as by simulation results [20]. The monotonically decreasing na-
ture of P, (E; |E;—1) with Ep/No indicates that the ratio of double errors to single errors
does not increase as SNR increases. As a final note, we also point out that using Oberst and
Schilling’s viewpoint, it is straightforward to show that nonadjacent decisions are mutually
exclusive and thus the conditional probability P, (£; |E;—,%) , k > 1 is equal to the uncondi-
tional bit error probability, namely, (7.14).
ter varies as an exponential function of E,/No. Aside from the 3 dB differences in SNR,
the performance similarity between optimum coherent detection of PSK and FSK and the
performance similarity between optimum differential detection of PSK and noncoherent de-
tection of FSK is more than accidental.
We saw in Chapter 4 that both optimum coherent detection of PSK and optimum co-
herent detection of FSK resulted in a correlation receiver with the same form of decision
variable, the difference between the two being the correlation coefficient, y, of the signals
[see (4.52)] which affects the mean of this variable. In the case of FSK, y = 0, whereas for
PSK, y = —1. Since the effective SNR in both cases is proportional to (1 — y), then the
above difference in the value of y accounts for the 3 dB difference in E,/No performance.
The explanation of the similarity in performance between optimum differential detection
of PSK and noncoherent detection of FSK is perhaps not as obvious since indeed the two
receiver structures appear to be quite different. In what follows, we shall expose the under-
lying thread that ties these two different modulation and detection forms together. The basis
of this thread is a special case of a unified analysis of certain coherent and noncoherent bi-
nary communication systems performed many years ago by Stein [21], which still stands as
a classic contribution in the field.
Consider the optimum DPSK receiver of Fig. 7.2b and the optimum NCFSK receiver
discussed in Chapter 5, redrawn here as Fig. 7.6 for the purpose of ease of reference and
similarity of notation. We are interested in comparing the error probability performances of
these receivers for both messages mg and m,, which are defined below for the two different
modulation/detection schemes.”
The rules for deciding between these two messages (see Figs. 7.2b and 7.6) are
mo
mo
2 2 a\
Z} = [(Wer)? + (Veni) |= [(Veou)” + (Yori) ]= 0 = NCFSK (7.516)
mM
where
5. We use a parenthetical superscript notation here to characterize the particular message being referred to so as
to avoid confusion with the subscript notation used for denoting discrete time.
458 Differentially Coherent Communication with Waveforms Chap. 7
v2 cos 27f ue
—¥2 sin2afr
fo <
r(t) Ja Sti f
<
x
2n (f — fi)
an (fF — fi)
Vises & = 0,1
For DPSK, V.,; and V;; are independent Gaussian random variables with variance o*=
NoTp/2, while for NCFSK, Ve1,i, Vc2,i, Vsi,i, and Vs2,; are independent Gaussian random
variables also with variance o? = NoT,/2. The means of these random variables when
messages mo and m, are transmitted are as follows:
V PT, cos @; k=
Eat 2 J mo orm
ci] eae Chae
25,10; mo heel’:
E Waal = {yen con my” E (Vad = [Yen sine i
It is at this point that we introduce a clever algebraic trick suggested by Stein [21] that
allows (7.51a) and (7.51b) to be more readily identified as being equivalent. In particular,
we rewrite (7.51a) as
doVei-1 aPmame
Voi
oe ae a
Vs i-1 Ar Vs.i oi
2 2 (7.54)
The random variables V.4,;, Vs+,i; Ve—,i, Vs—,i are all independent Gaussian with variance
a? /2 = NoT)/4 and means as follows:
460 Differentially Coherent Communication with Waveforms Chap. 7
J PT; cos 0; 6; =0
E{Ve+,i} = |—J/PT,cos0; 6; =n J"
0; mM},
J PT, sin 6; 6; =0
ee —VPT, sind;
= | 6 =
0; mM),
(7.535)
0; mo
ELVem) =| VPnycos 6; =0 pe
—/PT;, cos 0; GP 6 Ee
0;
B{V,2c} =| VPipsine G=0|
—/PTy sin 0; Ch .
Comparing (7.53) and (7.55), we observe the similarity between the squares of V.+,; and
Vetis Vs+,i and Vs1,;, Ve—; and Vo2i, Vs—; and V;2,;. Finally then comparing Z; of (7.54)
with Z/ of (7.51b), it is clear that they have the same probability statistics except that No
should be replaced by No/2 in the former since, as shown above, each Gaussian component
in Z; has half the variance. Equivalently, DPSK and NCFSK have the same error probability
performance except for a factor of two in SNR.
In order for this approach to have meaning, one must extend the previous assumption on the
duration of time over which the carrier phase is constant to be commensurate with the ex-
tended observation interval. For observations on the order of three or four symbol intervals,
this is still a reasonable assumption in many applications.
The theoretical framework in which we shall develop this so-called multiple-symbol
differential detection technique [22, 23] is once again the ML approach to statistical detec-
tion and is analogous to that used in Chapter 6 for deriving the optimum receiver of partially
coherent M-PSK based on an observation of N, symbols. As such, the resulting receiver
configurations derived here will be the optimum block-by-block detection receivers for dif-
ferentially coherent M-PSK. In the next section, we derive the appropriate ML algorithm.
Clearly, classical M-DPSK must result as a special case of this more general model when
Ne.
Consider as before the transmission of M-PSK signals over the AWGN channel. The chan-
nel introduces the unknown phase, 6, which in the absence of any information is assumed to
be uniformly distributed in the interval (—z, 7). The received signal in the k"” transmission
interval then takes the form of (7.1) (with i replaced by k) which can be expressed in terms
of the complex baseband signal s, = V2 Pel % by [see (6.52)]
where v(t) is a complex additive white Gaussian process. The receiver demodulates 7r(t)
with the complex reference signal exp(—jw@-t) to produce the complex baseband signal
5peJ? + n(t)e~/%", which is then passed through an I&D, resulting in the complex detected
sample 6
ee eo At Se
Vp = Spel” + T n(b)ew dt = spe 4 Ne @eai)
8 JKT;
Consider now an observation of the received signal over an interval of length N;7; (Ns is
herein referred to as the block length) and assume that the unknown channel phase, 6, is con-
stant over this interval. In particular, for an observation of the received signal over the inter-
val (k — N, +1) T; <t < (kK + 1)T,, the sequence v = (Viena Ve—Ny+2s Arar Ves Vi)
conditioned on @ is a sufficient statistic for making a ML decision on the transmitted se-
quence $s= (Se—N,+1) Sk=N, 42) We yee sk =
For the assumed AWGN model, the a posteriori probability of V given s and @ is
6. We apply the same normalization of 1/7, to the complex I&D output sample as was done in Section 6.3.
Again, we denote this normalization by a script typeface.
462 Differentially Coherent Communication with Waveforms Chap. 7
Averaging (7.58) over the uniform pdf of @ and simplifying gives the conditional probability
noting as before that for M-PSK, 5,|? is constant for all transmitted phases. Then, since
eae
La is independent of the transmitted signal and since /o(x) is a monotonic function
2
Ns-1 |
of x, maximizing AVS) over S is equivalent to finding max | )> Vy~iSz_;| » which
S a
is identical to (6.59b) with p set equal to zero. Defining the transmitted phase sequence
6= (Ox—N,+15 Ou Nek One so, Ok Ts Ox), we obtain the decision rule:
A Ns=l . >
choose @ corresponding to max Se Ve_je
1 =!
i=0
where 9 = (net GeON EDs curbs 6,-1, 4x)is the decision sequence. Note that the actual
complex I&D output in the (k — i)" interval, Vg, is related to Ve-i by (see Fig. 7.2)
r, (k-i+1)T;
Ve—i = Vek—-i + I Vs,k-i= = Ves Va
ae awe r(t)exp{—ja@,t}dt (7.61)
Niel 2
choose @ corresponding to max Sy Vj, je 1 %-
70
Sec. 7:2 Optimum Waveform Receivers 463
Note that, as for the two symbol observation case in Appendix 7A, the differentially
coherent decision rule for an N,-symbol observation also has a phase ambiguity associ-
ated with it since the addition of an arbitrary fixed phase, say @,, to all estimated phases
Ox N.+1> Oy NALD gas 6, ileby results in the identical decision rule, that is,
N,-1 2
choose @ corresponding to Ws se Vy exp {—j (@-; + 9g)}
t=)
Thus, as for classical M-DPSK, we must employ differential encoding at the transmit-
ter to resolve this phase ambiguity. Letting 6, = —0,—N,+1, then the above decision rule
can be rewritten in terms of the input information phases, AO,_; = O,—; — O—j-1; i=
Dal sa Ng = 2.88:
choose the information sequence A@ = (AOx—n,+2, AOk—N, +3, +--+» AOx—1, AO)
INe—1 N,=i=2 2
(7.62)
: Ns=2 _ N;—i—2 2
A receiver that implements the above decision rule is illustrated in Fig. 7.7. Note that
an observation over N, symbol intervals is used to make a sequence (block) decision on
N, — 1| input information phases. As such, the Ns-symbol observation blocks overlap by
one symbol since the last symbol of a given block serves as the phase reference for the next
block. Thus, for this type of block-by-block detection, the rate at which decisions are made ~
is identical to that of the symbol-by-symbol detection scheme of classical M-DPSK.
Some special cases of (7.62) are of interest. For Ns = 1, that is, an observation of the .
received signal over one symbol interval, (7.62) simplifies to
$5 eA
A= i (7.63)
which is completely independent of the input data phases and thus cannot be used for
making decisions on differentially encoded M-PSK modulation. In fact, the statistic of
(7.63) corresponds to the classical case of noncoherent detection (see Chapter 5) which is
not applicable to phase modulation.
Next, let NV, = 2, in which case (7.62) becomes
i 2s es 2 San is ;
alent Vie I A% 2 =|%-1 +|%| +2Re {HiVg oe/4% (7.64)
This results in the decision rule
which has a receiver implementation (see Fig. 7.8) that is the complex form of Fig. 7.2b, as
one would expect based on the results in Appendix 7A.
Aeyod 1-3 | Ny Kejoq
| inn
L L Aejaq a kejoq
464
cS
=? < eV
1 w—1-¥
o=u
odojaaua
10}9919p
ae oreduioa
pue
199195
ay?
IsoZIe'T]
W |» sapoAo
tad ‘L(1—“N)
a]qeuo i 2,949 oun
Avyap
S.9V
UO1s
jo
§=L*L
wo
uonoa
sans1
jooiq
©[Zz
ajdyj
paiuti
jenuas
joquiA
ySdq-
(4a
uonej
0/9
YSII9}se
Sa}OUaPp
xatduroo
a}e8n{uoo
uoneoydynur
u
asooy.)
x
*L(I+3)
Ip(e)
.
=
J='ev
{“Stoy
ST SOBIE]
aang
gt, UNUINdQ JoAIg091 xa[dwOd)
(WiI0J
10j [edIssejo y¥Sdd-W payurday)
wos [ZZ]© (JAI
465
466 Differentially Coherent Communication with Waveforms Chap. 7
2 3 2 wf aes
B
oF i +2Re {VaVite /A%} (7.65)
= \%-2| “4 x1
Note that the first and second terms of the metric used in the decision rule of (7.65) are
identical to those used to make successive and independent decisions on AO and A®&_},
respectively, in classical M-DPSK. The third term in the optimum metric is a combination
of the first two and is required to make an optimum joint decision on AO and A@_1.
Clearly, a receiver implemented on the basis of (7.65) will outperform a classical M-
DPSK receiver. Before demonstrating the amount of this performance improvement as a
function of the number of phases, M, we first present the implementation of the optimum
Ns = 3 receiver in accordance with the above form of the decision rule. Figure 7.9 is a
parallel implementation and Fig. 7.10 is a series implementation that although simpler in
appearance requires envelope normalization and additional delay elements. It should also
be noted that in the parallel implementation of Fig. 7.9, the apparent M? phasors needed to
perform the phase rotations of the output V;V,"_, can be accomplished with only M phasors
since the sum angle A@ + A6,—; when evaluated modulo 27 ranges over the same set of
M values as does AO; and AO;_1.
To obtain a simple upper bound on the average bit error probability, P,(£), of the opti-
mum block-by-block M-DPSK scheme, we use a union bound analogous to that used in
Chapter 6 to derive the performance of the optimum partially coherent block detection
receiver. In particular, the upper bound on P;(£) is the sum of the pairwise error prob-
abilities associated with each (N, — 1)-bit error sequence. An exact evaluation of these
pairwise error probabilities is carried out in the next section. Mathematically speaking,
if A@ = (AOx—N, +25 AGkoN.
23; «05, AOpen AO) is the sequence of N, — 1 information
phases and A6 = (Adi—n42, AOk_N, 435 ae A6z-1, Ab) is the corresponding sequence
of detected phases, then if u denotes the sequence of b = (Ns — 1)log)M information bits
that produces A@ at the transmitter and a the sequence of b bits that result from the detec-
tion of A@, then the bit error probability has the upper bound [see (4.88) with M replaced
by M™s~1}
irae
PAE) SF aamct 2, 2 wee, ) Pr {fi > 0140} (7.66)
A0# A6
“ev
S‘g
gsoouD
"eV
= “d
aN
BEN |
SOSIE] ST
467
aINBI
GL [a[feseg ueyeyuaWe[dun
Jo afdynw foquiAs JeNUsIaysIp‘1oONIEp
SV = ¢ pajutiday)
woy [ZZ]© ‘(aaal
468
Ogs—?
st 1SoBIR]
MeT-arenbs
adojaauq
10}93}2q
aansiy
OTL SIUIS uoneyuswajd
Jo ajdninw
ut [oquiAs [enuasaygip‘10j0019p
“yy = ¢€ pautiday)
wos [ZZ]© (AqagI
Sec. 7.2 Optimum Waveform Receivers 469
where w(u, ) denotes the Hamming distance between u and &i and Pr {4 > n|A@} denotes
the pairwise probability that A6 is incorrectly chosen when indeed A@ was sent. The
decision statistic 1 is defined in (7.62) and the corresponding error statistic 7 is identical
to (7.62) with each AO replaced by Aby. For symmetric signaling sets (such as M-PSK),
(7.66) satisfies a uniform error probability (UEP) criterion, that is, the probability of error
is independent of which input phase sequence A@ is chosen as the correct sequence. Under
these conditions, (7.66) simplifies to [see (6.62) with ¢ = 0]
1
P(E) < W,Sane Sy w(u, G) Pr {7 > n|A@} (7.67)
= D
A64A6
where A@ is any input sequence (e.g., the null sequence (0,0, ..., 0) = 0).
In (7.70), it is understood that the summation in the exponent evaluates to zero if the upper
index is negative.
Note that for any given N,, M, and data sequence A@, 5 can be evaluated for each
error sequence A@. We now demonstrate the evaluation of (7.67) and (7.69) for some
special cases of interest.
Special cases.
Case 1. Classical DPSK (N; = 2, M = 2)
From (7.70), we immediately get 5 = 0 and thus from (7.69)
b 2Es
Paes CHAD
2 I 2E; 2E,
mi niao}=3|1-0( a0) +0(0 =)| CAP)
470 Differentially Coherent Communication with Waveforms Chap. 7
Since for the binary case the pairwise error probability is indeed equal to the bit error
probability, then we have from (7.72) and (7.73) that
(ae || 33
x 0 +1 & [343]
x 1 +1 &[34v2]
Since the Hamming distance w(u, W) is equal to unity for the first two error sequences
and is equal to two for the third sequence, then using the above results in (7.69) and (7.70),
the upper bound on bit error probability as given by (7.66) is evaluated as
-o((8E-9 EE)
To see how much performance is gained by extending the observation interval for
differential detection from N; = 2 (classical DPSK) to Ns = 3, we must compare (7.75) to
(7.74). Due to the complex form of (7.75), this comparison is not readily obvious without
resorting to numerical evaluation. On the other hand, by examining the asymptotic (large
Es/No) behavior of (7.75), we can get an immediate handle on this gain.
When both arguments of the Q-function are large, the following asymptotic approxi-
mations are valid [24]:
Re 2
O(@,6) =1- fF 500 |-ol,
e a>Bp>|
(7.76)
ee Wea et (B —a)?)-
O10 ON ae F ol#2 B>a>l
Sec. 7.2 Optimum Waveform Receivers 471
2
z 1 |
Pr {if}> |A0} = 5 eae eeb i; a ao
lb = (1.77)
Pr {7 > n|A@} =
— y NP 18)?
2/2 Fn, — 181) \ in,+ \/N2 — 18?
1/4
Ns Sey +4/Neal}
N2 — |6|? MEA hee (7.78)
+
SSS= KP leaan Dia plon
Ns — ,/N2 — |6|? Y
General asymptotic results. In the general case for arbitrary N,, the dominant
terms in the bit error probability occur for the sequences that result in the minimum value
of N, — |6|. One can easily show that this minimum value will certainly occur for the error
sequence Aé having N; — 1 elements equal to the correct sequence A@ and one element
with the smallest error. Thus
= Ns — |6|max
where
2 = ae A 6k) min ee ok
dain Ne Waa AY = 4sin 7; (7.81)
and (A& — A6x) min denotes the smallest possible value between the iit symbols of the data
and error phase sequences, that is, 27 /M. Also note that for |5| = |6|max, (7-70) reduces to
1 A
PRE is= aT
a 1ylogy M Se w(u, &)
Ad4AO (7.83)
1 Ns+ |6|max E;
is Ie
as ee ———
(Ng — |8lmax)
2 ime5 ( 15 |max (Ns ae lorax) 2No ‘eae
where w(u, ti) corresponds only to those error sequences that result in |6|max.
For the binary case (M = 2), we have from (7.81) that Bae = 4 and hence N; —
\5|max = 2. Similarly, it is straightforward to show (see [22, Appendix B] and Problem 7.10)
that the sum of Hamming distances required in (7.83) is given by
> w(u, a)
i
= ZN
he G1):
eae
Ns = 2
(7.84)
Ad4AO
Thus, (7.83) simplifies to
2 N, — 1 1 Es
Pi (BE) = ied ( me ;)E exp =] (7.85)
No
1
P)(E) exp {= (7.86)
[e.
7 No
which when expressed in terms of the first term of the asymptotic expansion of the comple-
mentary error function, namely
]
erfe x & exp (—x?) (7.87)
JX
becomes
|E
P, Serfc a (7.88)
From the results in Chapter 4, the exact bit error probability performance for coherent
detection of BPSK with differential encoding and decoding [see Eq. (4.206)] has an asymp-
totic upper bound identical to (7.88). Thus, as one might expect, the performance of multiple
symbol differentially detected BPSK approaches that of ideal coherent detection BPSK with
differential encoding in the limit as the observation interval (decision memory) approaches
infinity.
7. Note that (7.85) is not valid for N, = 2 since in that case |5|max = 0 [see (7.80) and (7.81)] and thus the
inequalities of (7.76) are not satisfied.
Sec. 7.2 Optimum Waveform Receivers 473
A similar limiting behavior as the above may be observed for other values of M. In
particular, it can be shown (see [22, Appendix B] and Problem 7.9) that for M > 2 anda
Gray code mapping of bits to symbols, the sum of Hamming distances corresponding to
I5|max iS given by
Ds naa) =r)
A De
Res
Ns = 2
(7.89)
A64A6
Using (7.89) in (7.83), we get (for N; > 2)
Es
x exp aN NS —="1S5 bax)
1 cos 4,
Og Se
S ‘ IU ate
(log, M) 21 x, SiN 347 ,/COS 44 (7.92)
P,(E)
b < ———— e expp j - —No
Es sin? —M=
(logy M),/2%¢sin? %
(7.93)
1 Big)
= erfc — sin —
log» M No M
which is identical to the asymptotic bit error probability for coherent detection of M-PSK
with differential encoding and decoding [see Eq. (4.134), (4.135) together with the asymp-
totic form of (4.200)].°
Figures 7.11, 7.12, and 7.13 are illustrations of the upper bounds of (7.90) and (7.92)
for M = 2, 4, and 8, respectively. In each figure, the length (in M-PSK symbols) of the
observation interval is a parameter varying from Ns = 2 (classical M-DPSK) to N; = oo
8. It should be noted that the result in (7.93) can be obtained by observing that, for large Ns, (7.82) satisfies
Jb > Vb — Ja > 0. In this case, (7.69) can be approximated as in (6.67) (also see [25], Appendix A) by
SIMULATION POINTS
BN 2
O Ns =3
A N,=5
Ng = 2
(ideal coherent detection). Also indicated on the figures are computer simulation results
corresponding to the exact performance. We observe that, for example, for DPSK, extending
the observation interval from N,; = 2 to N; = 3 recovers more than half of the E,/No loss
of differential detection versus coherent detection with differential encoding. For M = 4, the
improvement in E,/No performance of N; = 3 relative to Ns; = 2 is more than 1 dB, which
is slightly less than half of the total difference between differential detection and coherent
detection with differential encoding.
In the previous section, we showed that for sufficiently large block size, Ns, the bit error
probability performance approaches that of coherent detection of differentially encoded M-
PSK. Unfortunately, the complexity of a receiver implementation of the type illustrated
in Fig. 7.9 for Ns =3 grows exponentially with N;. To see this, we first observe that,
for general N,, the decision metric [such as that described by (7.65) for N; = 3] must be
evaluated for Ms~! hypotheses (the number of possible A@ vectors of length N, — 1).
For each of these hypotheses there are ee terms in the metric [see, for example, the
decision rule below (7.65) which contains S = 3 terms within the braces]. Thus, we
N;—1
must calculate ( 5 )phase angle sums of the form AO_, + A@—g41 +... + AG—p
Ns F Ns Di.
for 0 < p<q< WN; —2, perform 7 phase rotations and DON fos 1 real additions.
Ns \ .
Since (Ye)is on the order of N?, the total number of operations required to implement
SIMULATION POINTS
SIMULATION POINTS
Gl Ne. =\2
O Ns, =3
A Ns, =5
N, = 2
at the receiver. We present here without proof (see [25] for the details) a description of this
algorithm.
Consider the vector of complex I&D samples” V= (v1,V2, RSE Vw.) corresponding
to a particular observation (block) of Ns; symbols. Let vj = tan! (Vs riVeni) denote the
argument of each component of this vector [see (7.61)]. For each component in the vector
V, define the complex quantity
8 = Zu) (7.96)
The complex quantities {g;;i = 1,2,..., Ns} are referred to by Mackenthun as the remod-
ulations of the received complex samples {V;;i = 1, 2,..., Ns} in [0, 277/M) ordered by
the magnitude of their phase angles. Augment the set of values of g; by including a 217/M
phase shift of each as follows:
Bi =e JIM
Gg, ys i=Nst+1,Ny+2,...,2Ns (7.97)
The extended set of 2N; values of g; represent all possible remodulations of {V;; i = 1, 2,
.., Ns} in (—22/M, 21/M) indexed by their phase angle, going clockwise around the
circle starting at e/?”/”. Next form the set of N, sums
q+Ns-1 2
figA eg, | ge eee (7.98)
l=q
q =max ng (7.99)
Arranging the phases 6x(i); i= 1, 2,...Ns in order of subscript value k(i) produces the joint
decision @ = (41,Coins 6.) on the sequence of transmitted phases. Finally, the decision
A6 = (Ad, NOs ct Ady, )on the information data phase vector has components A6; =
6; — 6;-1; t= 42h Ny where 6;: i=1,2,...Ns are the components of6 found in
(7.100).
As an example, suppose that M = 4, N, = 6 and the ¢;’S are ordered such that 0 <
62 < 66 < 05 <b1 <3 < f4 < 2/2. Then
81 = Z4, 82 = Z3, 83 = Z1, 84 = Zs,Bs= Zo, 86 = Zr
appears as in Fig. 7.14 (assuming a +1 as a reference bit, i.e, b-1 = 1) where p(t) is a
unit rectangular pulse of 7;,-sec duration. We observe that S(t) of (7.104) can be viewed
as a sequence of overlapped 27)-sec pulses of the form given in (7.103). Thus, if s(t) =
J/2P S(t)e/%! denotes the complex transmitted DPSK signal, then from Fig. 7.14 we see
that S(t) can be represented as the overlapped sequence So(t), 51(t), —So(t), —So(t), —81(0),
478 Differentially Coherent Communication with Waveforms Chap. 7
50(t), 51(t), —51(t), 51(t), —So(t) where the interval of overlap has duration 7p sec. In short,
the DPSK signal has been represented as a sequence of 100% overlapped binary orthogonal
signals.
Note that except for the polarity of the p,(t)’s in the overlapped pulse sequence
of Fig. 7.14, there is a one-to-one correspondence between the input data bit a; in the
interval i7, < t < (i + 1)T> and the transmitted pulse p,(t) in the interval (( —1)7T, <t <
(i + 1)Tp. In particular, a; = +1 corresponds to transmitting po(t) (or —po(t)) anda; = —1
corresponds to transmitting p(t) (or —pi(t)). As such, an optimum decision rule for the
information data sequence {a;} can be obtained by noncoherently detecting the sequence of
overlapped signals. In particular, based on an observation of the received signal, r(t), over
the 27)-sec interval (i — 1)7, < t < (i + 1)Tp, the £1 decision on a; is made according to
the rule
Pre 235 2
(i+1)T) (i+1)T,
i; r(t)s{(t)dt Hl r(t)S5g(t)dt (7.105)
(i-1)T, a (i—1)T,
The decision rule in (7.105) represents an optimum noncoherent decision on a pair of binary
orthogonal signals [So(t) and s;(t)] whose duration is 27) sec and is illustrated in Fig. 7.15.
Thus, the error probability associated with this decision is identical to (5.79) with Ey = PT,
replaced by 2E, = P (2T;), namely
The notion of signaling with overlapped sequences of orthogonal signals can be generalized
to M-ary modulation. !° Here, however, the overlap is less than 100%. In particular, consider
a set of M complex signals {5,(t); k=0,1,..., M — 1} defined as in (7.102) in terms
of a set of M pulses {px(t); k =0,1,...,.M — 1} which in addition to being orthogonal
over the interval 0 <t < Pots (T; = Tp logy M) have the following properties. The
px(t)’s are all identical in the interval 0 < t < mals and differ only in their polarity in
the interval T; <t < Wot. A set of functions that has these properties can be generated
from a Hadamard matrix [1] by using the rows of this matrix as generating sequences for
rectangular-shaped waveforms. An example of such a set of pulse waveforms for M = 4 is
illustrated in Fig. 7.16 corresponding to the 4 x 4 Hadamard matrix
Geawiones
(7.107)
10. This generalization was suggested by Dan Raphaeli while a Ph.D. student at Caltech and appears in an
unpublished document.
479
[euosoy}I0
sas[nd
Aseuiqg paddeypsoao Jo Suto} Ul ULIOJOARM YSqq puegaseg e jo uoNRlussaiday py sansiy
——
(1)'d
tea (yd
——e}—— ('d-
—e}=— ('c-
—>}=—_ er
—e}— (y- —e}=— u- —>}=— a —> |
480
A-
7rous
21+)
of 1p(0)
(J)4
sploeq
'p = °z}u3s
— {'z
11+)
we),
dp
P'Sso9
aansiyqsz uonrjusu
JO sjdwy
YSdq JOAINA1
se juardYOouoU
uoNoa}9p
Jo Areulg Jeuosoyrio
sjeusis
Sec. 7.3 DPSK 481
Po(t)
q
p(t)
Pp, (t)
p(t)
-1
Suppose now that we were to transmit the same information data sequence as be-
fore (see Fig. 7.14). Consider grouping these bits into symbols consisting of n = log, M
bits apiece corresponding to (qj, dj+1,...,4@j+n—1) in the interval i7, <t < (i + n)T) and
assign px(t) (or —p,(t)) as appropriate to the interval (i— 1) Tp <t < i +n)T} ac-
cording to the following rule. Let (k, k2,..., k,) be the binary (0, 1) representation of the
integer kwhere | < k < M. Then, p,x(t) (or — px(t)) is assigned to the binary (+1) sequence
aj = 1 — 2k, aj4, = 1 —2kp, ..., Gj4n—1 = 1 — 2k,. For the n = 4 example, we would have
Using the above assignment, the overlapped pulse sequence for the data sequence of
Fig. 7.14 is illustrated in Fig. 7.17. Note that the overlap interval here is T;/3 = 27/3
Sec:
An optimum decision rule for choosing groups of n information bits (i.e., data sym-
bols) is again obtained by noncoherently detecting the sequence of overlapped signals. In
particular, based on an observation of the received signal, r(t), over the n (#4) Tp-sec
interval (i_ wa) Ty <t < (i +n)Tp, the joint +1 decisions on a; and a;+) are made ac-
cording to the rule. Decide a; and a;+) corresponding to p;(t) as determined by (7.108)
where
2
(i+n)Tp
aan r(t)sy(t)dt (7.109)
ab 5| )T
The above decision rule represents an optimum noncoherent decision on a set of n orthog-
onal signals, each of duration n (wir)(i (Cai)T; sec. Thus, the symbol error prob-
ability associated with this decision is identical to (5.77) with E; replaced by (4) 1B
namely
PAE
3(E) = ba(= UN
es jMol C vex | aLad
- (; mana(=)}
\ (33 Es (7.110)
The bit error probability is related to the symbol error probability of (7.110) through (4.96).
In view of the above, we see that overlapped noncoherent M-ary orthogonal modulation
can achieve an improvement in E/ No by a factor of (4). Clearly, as M increases, the
improvement factor diminishes and is maximum for M = 2 where it corresponds to 3 dB as
previously shown.
Another way of viewing the Pray improvement factor in E/No for overlapped non-
coherent orthogonal signaling is to consider the equivalent detection problem for nonover-
lapped signals. In particular, suppose that we ignore the overlap region, that is, the region
where all transmitted signals are defined to be identical, and just define the complex signals
5,(t) as in (7.102) but only in the interval (Gat:)ote (#45)
Ye T;. Then a noncoherent
decision on this new signal set {si (t)} defined over a 7;-sec time interval should produce
the same error probability as the overlapped orthogonal set {5,(t)} defined on a (4) Ts-
sec interval. The difference is that the members of the nonoverlapped set {5,(t) } are not
orthogonal. To see what their correlation properties are we proceed as follows. We can write
5,(¢) in terms of 5, (t) as
l
Wow) 2s
H(t) = ( nate (7.111)
Using (7.111), it is straightforward to show (see Problem 7.13) that the 5,(t)’s are transorth-
ogonal (see Chapter 4), that is, their off-diagonal normalized correlation is equal to -Wh.
As such, {5,(t)} represents a simplex set that produces an wo improvement factor in E /No
performance relative to an orthogonal signaling set.
ae
a
Oe
fo—
a
i.)
=
aoe
«LIL,
ejep
Jo aouenbas
poddejisao
yeuosoyyio
asnd
Areulg
ANSIYy
Ase-p
uONRIUasaIday
aouanbes
&se
(1)'d
483
484 Differentially Coherent Communication with Waveforms Chap. 7
sample once
per bit
Lowpass
Bandpass
Filter Filter
Fig. 7.2b. The details of these analyses follow along the lines given in Park [2]. A more
exact evaluation of the performance of ACD-type differential detectors is given in Chapter
8 as a prelude to the discussion of double differential detection.
As stated above, we assume that the IF bandwidth, W (B = W/2 is the equivalent lowpass
bandwidth) of the input bandpass filter is sufficiently wide that the desired signal will pass
through this filter undistorted and with negligible ISI, that is, 1/T, << B <,/2z. Thus,
from Fig. 7.18a, and (7.1) _
r(t)=~V2P cos (@,t + 6; + 0) +:n(t)
(7.112)
= J/2P cos (wet +6; +0) + V2[n¢(t) cos (wet + 6) —ng(t) sin (wet + 0)]
where the hat denotes filtering by the input bandpass filter. For ease in analysis, we shall
assume that the product of the radian carrier frequency and the bit time, a7), is an integer
multiple of 27 and B is an integer multiple of 1/27;,. Under these conditions, the output, Vj,
of the I&D is (ignoring second harmonic terms)
(i+1)F
V; = / r(t)r(t — Tp)dt = PT, cos AG;
iTp
(i+1)T,
4 vP | [n-(t)
cos6-1 + ne(t — Tp) cos 6;] dt
ee (7.113)
(i+1)Tp
~ vP | [ns (t) sin 6-1 + ns(t — Tp) sin 6;| dt
iTp
(i+1)T,
ee ‘| [nc(t)nc(t — Th) + ne(t)ne(t — Tp)] dt
iTp
Since for DPSK the transmitted phase only takes on values of 0 and zr, then (7.113) simpli- 7
fies to
(i+1)Tp
V; =, r(t)r(t — Tp)dt = PTp cos AG;
iTp
(i+1)T,
3 vP | [tne(t) tne(t — T,)]dt (7.114)
1Tp
(i+1)Tp
+/ [ne (t)nc(t — Tp) + ns(t)ns(t — Tp)| dt
iT)
where the plus sign is associated with a transmitted phase equal to 0 and the minus sign with
a transmitted phase equal to 2. Assuming an ideal (unit height rectangular frequency charac-
teristic) filter for the input bandpass filter, then n,(¢) and n;(t) are statistically independent
with autocorrelation function
sin 27 Bt
Ry(t) = E {ne(t)nc(t + t)} = E {ns (t)ns(t + T)} = NoB ( aT BT ) (7s 103)
486 Differentially Coherent Communication with Waveforms Chap. 7
In view of the assumption that B is an integer multiple of 1 /2Ty, then R, (Tp) = 0 and hence
the first two moments of V; are given by
ee Ad; =0
Ue = sli eNO} 70
Since we have assumed that B >> 1/7», then the above integrals can be approximately
evaluated, resulting in
(7.118)
Comparing (7.118) with (7.14) we can readily identify the degradation in performance of
the suboptimum configuration with respect to that of the optimum one.
We remind the reader that care must be exercised when using (7.118) that all of the
above assumptions are met. In particular, if B is made too large, then V; becomes decid-
edly non-Gaussian. Similarly, if B is made too small, then the assumption of zero signal
distortion becomes inappropriate. In addition to these conditions, it is also important not
to let Ep/No become too large relative to BT), else the Gaussian assumption again be-
comes incorrect. In particular, if for BT), fixed (and large) we were to let E,/No — oo, then
from (7.118) we reach the erroneous conclusion that the suboptimum differential detector of
Fig. 7.18a asymptotically performs 3 dB worse in SNR than the optimum detector of coher-
ent PSK, the latter being asymptotically equivalent in performance to the optimum detector
of DPSK (see Fig. 7.2b). The range of BT, and E,/No values over which the Gaussian
assumption provides approximately valid results for error probabilities of practical interest
will be illustrated in Chapter 8, where the performance of this configuration is revisited in
the context of a discussion on the behavior of differential detectors in the presence of fre-
quency error.
Sec. 7.4 Suboptimum Waveform Receivers 487
where b; = cos 6; is the binary (+1) amplitude equivalent of the transmitted phase modu-
lation and p(t) is a unit rectangular pulse in the interval (0, 7). After passing through the
bandpass filter, we get
where
A sind. Br’
b(t) =2B thesesan BY
p(t) LEY, f (7A )
Note that because p(t) extends before and beyond the 7;-sec duration of the current symbol,
ISI exists and we must sum over all past and future data bits. To obtain a detection sample
for making a decision on the information bit a; = bj-1b; = cos A@;, we multiply 7(t) by
7(t — Ty), filter out the second harmonic of the carrier, and sample the remaining lowpass
component at t = (i + 0.5) Tp, giving
(7122)
x lveKB (G — k — 0.5) Ty) + ne (i = 0.5) Th)
k=—0o
Computing the bit error probability based on comparing (7.122) (which accounts for
the total ISI) with a zero threshold is in general a difficult if not impossible analytical
task. Thus, we must make some simplifying assumptions from which we can obtain an
approximate result. One such simplification, suggested in [28], assumes that B = 1/27} and
that the bandpass filter is cleared or reset at the end of each bit time to eliminate ISI. Under
this assumption, (7.122) simplifies to
488 Differentially Coherent Communication with Waveforms Chap. 7
bib(15/2) + nc
Vj = [VP + 0.5) T»)]
(7.123)
x [VPbi-1P (Th /2) + ne (i — 0.5) Ty)
Again comparing (7.126) with (7.14), it is a simple matter to see that, under the above
assumptions, the degradation in the performance of the suboptimum receiver relative to that
of the optimum receiver is a reduction in SNR by the factor 10 logy9 (4/27) Si? (1/2) =
—1.2 dB.
A second approach to approximately computing the bit error probability performance
of Fig 7.18b is to still make the assumption of neglecting ISI to simplify (7.122) but us-
ing a bandwidth B = 0.57/7,, which in [29] is shown to minimize ISI. In this case, the
noise variables n- ((i + 0.5) 7,) and n,. ((i — 0.5) T;) are still zero-mean Gaussian, each
with variance NoB =0.57No/Tp, but they are no longer independent and likewise for
ns (i + 0.5) Tp), and ns ((i — 0.5) T,). However, the noise variables n; ((i + 0.5) T,) +
Nc ((i — 0.5) Ty) and n- ((i + 0.5) Ty) — ne ((i — 0.5) Tp) are indeed independent and like-
wise for ns ((i + 0.5) Tp) + ns (i — 0.5) Tp) and ns ((i + 0.5) T,) — ns (Ci — 0.5) T). Thus,
remembering the analogy between the noise terms in DPSK and those in noncoherent FSK
as discussed in Section 7.1.6, and noting from (7.124) that p(7,/2) = (2/7) Si (0.577), we
can again immediately write down the bit error probability as
|
,(E) ==|
P(E) exp |="=Ep | 2
———_
se: ,
$2
i2( 1
(ara || se et Ey | (7.127)
en| 0.8052
which represents a degradation of 10 logy (2/0.5727) Si* (0.57) = —1.0 dB in SNR rel-
ative to the optimum DPSK detector.
1,2,..., Ns — 2 in the metric of (7.62), are replaced by past decisions on these quantities,
that is,
N,;—2
7 = Ve_—-Nn,+1 + Ve exp |-;(asaE ye sin)
m=1
} (7.128)
N,—2 & Ns—i-—2
=- SS Ve; exp |-;(DS sii
i=] m=0
Since the only unknown in (7.128) is now the current symbol A@,, the detector is once again
of the symbol-by-symbol decision type. While suboptimum, this multiple decision feedback
type of detector should still improve bit error probability performance relative to that of
classical M-DPSK.
As a simple example of the above, consider a three-symbol observation interval (NV, =
3). The optimum decision rule for this case is given below (7.65), which for the suboptimum
detector now becomes
choose AG corresponding to
where Ab _1 represents the previous decision that is fed back. Since the middle term in the
braces doesn’t depend on Ag, it can be eliminated, resulting in the yet simpler decision rule
ae NS — j (Ab +AO
max Re |Va see 8 Vi Vee j (Ab 1+ |
AG
The bit error probability performance of the multiple decision feedback receiver was
analytically computed in [32] for DPSK and 4-DPSK under the optimistic assumption of no
errors in previous decisions, that is, perfect knowledge of the previous transmitted symbols.
The results are given by an expression analogous to (7.68), namely
In the limit as Ns — oo both performances approach that of coherent M-PSK with differ-
ential encoding/decoding. This is the same conclusion as was reached for the optimum ML
block-by-block detector.
Figures 7.19 and 7.20 are illustrations of the above theoretical results for N, = 2, 3,
and 4. Also shown are two sets of simulation results; one that assumes perfect decision
490 Differentially Coherent Communication with Waveforms Chap. 7
10-15
--=- THEORY
40g SIMULATION POINTS
10-2
ae
ne 10°F EQS. (7.111)
AND (7.112)
PSK
10-4
feedback as does the analysis, and one that takes into account the effects of errors in the past
decisions, that is, the true situation. For the case of perfect feedback, the theoretical and sim-
ulation results agree exactly as one would hope. For the case of errors in the decisions being
fed back, the simulation results, of course, yield a poorer performance than the theoretical
results which assume perfect decision feedback. However, over the range of system param-
eters illustrated in Figs. 7.19 and 7.20, no error propagation was observed in the simulation.
Furthermore, the results are quite comparable to those of the optimum ML block-by-block
detector. For example, for Ns; = 3 and P,(E) = 10-5, the difference between the two is only
about 0.3 dB.
10-1
=-—- THEORY
, 400 SIMULATION POINTS
10-2
DETECTED
SYMBOLS
= FED BACK
Ww 49-3 FED BACK \
a2 EQS. (7.111) \ ea 4-DPSK
AND (7.113)
10-4
By choosing an initial phase 6) equal to 7/4, the set of allowable values for 0; in
SDPSK is restricted to (1/4, 37/4, 52/4, 77/4) in every symbol interval and since the
magnitude of the maximum phase change is limited to 2/2, we see that either the I or
Q components of the transmitted signal, but not both, must change polarity from symbol -
to symbol. Thus, the envelope of the transmitted signal never undergoes a complete phase
reversal (change of z radians) as is typical of conventional DPSK. The upshot of this mod-
ification is that SDPSK suffers from negligible spectrum “regrowth” after being filtered
and passed through a nonlinear device. Specifically, laboratory tests [33] show that when
SDPSK and DPSK are bandpass filtered and then passed through a hard limiter, spectrum
regrowth for DPSK is nearly 100%, whereas for SDPSK it is significantly lower. On a lin-
ear channel, unfiltered DPSK and SDPSK have identical bit error probability performance
[see (7.14)] since in both cases the two possible symbols are separated by z radians. Fur-
thermore, the two modulations have identical power spectral densities. Finally, Winters [33]
shows that SDPSK produces a lower average bit error probability than DPSK in the presence
of timing error and frequency uncertainty.
Although the desirable attributes of SDPSK with regard to minimum spectral re-
growth after bandpass filtering and hard limiting were demonstrated almost a decade ago
[33], it has only recently been suggested by Alexovich, Kossin, and Schonhoff [35] to in-
clude the filtering and hard limiting as an inherent part of the modulation scheme itself.
Such a bandpass limited symmetric DPSK (BPL-SDPSK) produces a bandwidth efficient,
constant envelope modulation. By choosing the bandwidth-time product B7) equal to unity,
it was found by simulation that the power spectral density resembled that of minimum-shift-
keying (MSK) (see Chapter 2). Furthermore, the bit error probability performance of BPL-
SDPSK with B7; = 1 was shown to be comparable to and in some cases better than that of
various forms of differentially detected MSK [36], [37], [38], [39]. A detailed discussion of
MSK performance will be given in Chapter 10.
492
P@soo Sn
@vuts
(3)4
Voutsz7p—
[R00]
10}e][1IISO
Gig; O<t<T;
an =|
i. Tere in Ose
IND
is the transmitted data phase waveform with 6;,, 0;, independently taking on values from the
set ofM possible phases B,, = 27/M; m=0,1,...,M —1. Since the signal now depends
on an additional unknown random variable, that is, 9, the conditional pdf of (4.8) is written
as
GG @ ORG) dt
f-(o1m,0) =Cexo|
No
2 aa) Dy Parte Pe
= Comp) Bans
ORT loisee
ne (7A.3)
No
a
en |
2 Jo ; p(t)
si
i; (t; 8)dt
(t; 6)
No
oT,
Dia pt) sp 50)
at
fr (p ino) =Ciexp| POE (7A.4)
0
where C’ is a new normalization constant that is independent of i. Using (7A.1) and (7A.2),
the numerator of the exponent in (7A.4) can be expressed in the following convenient
form:
Differentially Coherent Communication with Waveforms Chap. 7
494
OT; ge
2exp {j (wet + 6;(t) + 8)} ar|
/ p(t)s;(t; 0)dt = VP Re {| p(t)
0 0
Ts
= /PRe [ oervdexp {i loet + 6)} at
$$
Uy (7A.5)
OT,
+| p(t)V2exp {j(@et + 6;,)} dt |exp (j9)
U;
(7A.6)
DU) oie S
f-foim,0)=C'sxp| 20 lip + Osos +o (74.7)
Averaging (7A.7) over the uniformly distributed phase, 6, gives
f(oim)
=| fr (p mi.)
epi
fo(9)d0
alee ORR Te :
= 015[e227 ldo + ifm 0 +09}a (7A.8)
ON Rae D,
-'tn( |e + a)
No
where as before /o(x) is the zero order modified Bessel function of the first kind with
argumentx.
For equiprobable messages, the optimum (maximum-likelihood) receiver chooses that
message as having been transmitted corresponding to the maximum of f; (p |m; ) over 1.
In particular, letting m (r) = m; denote the decision made by the receiver as a result of
observing r, then since Jo(x) is a monotonically increasing function of its argument, we
have
App. 7A The Optimum Receiver for Classical M-DPSK 495
*s
i=max |— f,(p|m;)
= max atsdo + H1|
Sale
l Ll
Ts
= max! [ p(t) V2exp {7 (Wet + Gig)}dt (7A.9)
10,1]
8 2
+f p(t)/2 exp {i (@ct + 6;,)} dt
Ts
where max! g(i) denotes the value of i that maximizes the function g(i). We note that
Ll
adding an arbitrary phase, say 6,, to the decision statistic in (7A.9), that is,
Ts
i = max! / p(t)/2 exp {ji(Wet + Big + Oa)} dt
io,t} 0)
ote 2
+/ p(t)/2 exp {j (wet + 6), + Oa)} dt
Ts
= max ~' exp (ja) if
10,1] 0
p(t)V2exp {j (wet + Gig)}at
(7A.10)
hip
+f pia) VZexp{J (vet +6,)} dt
Ts
Ts
= max = if p(t)/2 exp {j (wet + Gi) }dt
10,1) 0
ye 2
doesn’t change the decisions made on 6;, and 6;,. Thus, the decision rule has a phase
ambiguity in that the receiver is as likely to choose the pair (6j) + 0a, 61, + 9a) as it is to
choose the pair (Gi9, Oras To resolve this ambiguity, we employ differential encoding in
which the information (data) phase is the difference of two adjacent transmitted phases.
Letting Ad; A 6;, — 9;, and also setting the arbitrary phase 6, = —6io, (7A.10) becomes
Ts 2Ts 2
Ts 2Ts
= max dt +exp ae)
~' i! p (t) V2exp (jact) | p(t) V2exp (jact)dt|74-11)
i 0 Ts
Vo Vi
ye ait
= max "|| +h + 2Re[oVi'exp (7060 |
Ll
496 Differentially Coherent Communication with Waveforms Chap. 7
ee po Ac
Since Vo and v1 are independent of i, we get the equivalent decision rule
W=VJU2- V2 (7B.2c)
FW) = ke as W
sin (A® — p)
ue _n/2, 4% LU —Vsint — Wcos (A® — wv) cost
(7B.6)
rsinw — pcosw
1—(rcosyw+psiny) cost
where
_ U-—Vsint
— Wcos(A®— wv) cost
E= Mee SW SOA eS (7B.7)
1 — (rcos yw + psin yw) cost
Note that for all three cases, the probability distribution function is expressable as a single
integral of simple (exponential and trigonometric) functions. Further simplications of the
integrands of F(y) are considered in [4] at the expense of complicating the integration
limits.
PROBLEMS
7.1 The optimum receiver for M-DPSK is illustrated in Fig. 7.1, which has the complex form of
Fig. 7.8. With reference to Fig. 7.8, the decision rule can be written as
Problems 499
Ad; = Bn
Hoc ~~ iB:
where n = min : {argMetre
sede ~ ~ 2m1
= min {arg1% — arg \%-1| = fi|; B= a
(! I i
An equivalent receiver to the above that makes direct use of the inphase and quadrature outputs
Vox and V;x is illustrated in Fig. 7.2a. The decision rule for this receiver is given by
A6; = Bn
where n = max
I
~! {(Vs.Vine + Vik Ver-1) cos Bj + (Wx Ven—-1 — Vek Vs4-1) sin pi|:
201
ie
(a) Show that the above two decision rules are completely equivalent. Hint: Note that the value
of i that results in ae {©;} is equal to the value of i that results in max {cos @;}.
(b) Show that the Peceiier of Fig. 7.2a reduces to that of Fig. 7.2b for ke DPSK (M= 2),
that is, they implement the same decision rule.
7.2 (Wideband IF Filter, Sampling Detector) A differential detector for binary DPSK is illustrated
in Fig. 7P.2. Assume that the IF bandwidth of the input bandpass filter is sufficiently wide that
the desired signal will pass through this filter undistorted and with negligible ISI. As such, the
output of the bandpass filter can be represented as
r(t) = V2P cos (@ct + O(t) + 8) +n (t) cos (Wet + 6) — n2(t) sin (wet + 8)
where 6@ is the unknown phase introduced by the channel, 0(t) = 6; is the transmitted phase in |
the i” bit interval i7, <t < (i+ 1)Tp corresponding to the information phase A@; = 6; — 6;—1, ‘
and n\(t), n2(t) are uncorrelated zero-mean Gaussian noise processes each with variance
o* = NoB. Here B is the equivalent lowpass bandwidth of the input bandpass filter. Also
assume that the transfer function of this filter is symmetric around its center frequency, w-,
and that the lowpass equivalent of this transfer function is denoted by H(w).
(a) Assuming that the lowpass filter merely eliminates second harmonic terms of the carrier, :
show that a sample of the output of this filter taken in the ith bit interval can be written in’
the form
z= >Re {ziz2}
l *
where
r(t) |Wideband
Lowpass
Bandpass
Filter
Filter
Figure P7.2
500 Differentially Coherent Communication with Waveforms Chap. 7
with 13(t) 4 ni(t — Tp), n4(t) A no(t — Tp), and nj; 2, n3, n4 respectively denoting sam-
ples of the noise processes nj(t), 2(t), n3(t), n4(t). Also, we have further assumed that
Onlp== 2.
(b) Show that the correlation matrix of the noise samples 11, n2, n3, 4 is given by
PO yo @
— Ta Oe ln 0
[ninj] mee Oren te ea)
OR Onn Oana
7.3 For threshold decisions made on random variables, z, of the form given in part (a) of Prob-
lem 7.1, Stein [21] has shown that the average bit error probability P,(£) is given by
Ni + M2 — JM — Np)? + 4NiN2p2,
N\N2 (1 es p32)
with
Se
nei?
5 lex > NeSog;
2 af
Ox Sarg {z}=tan~*
Shek
——-; k=1,2
Mek
v A arg (p,}
=tan7! 2
Pz
Recognizing that P,(E) = 5 Pr {Re {z}z2} <0|A0; =0} + 4 Pr {Re {ztzo} > 0|AG; =z},
evaluate a, b, and A, and show that P,(E) simplifies to
B=eg @ih) ( Ng )f (1 — p)
Ravrcest2 Ey 1
Ey
‘ a Soi: 2 exp ( No am)
Problems 501
sample at t= iT, + T,
Figure P7.4
7.4 In addition to coherent and noncoherent detection, binary FSK can also be differentially de-
tected. The receiver structures are quite similar to those for DPSK, the primary differences
being: (1) the value of delay, Tz, used to produce the delayed demodulation reference should
be chosen relative to the frequency modulation index, h; and (2) a 90° phase shift should be
applied to the delayed demodulation reference. As such, consider the suboptimum (wideband
IF, sampling) differential detector for binary FSK illustrated in Fig. 7P.4.
Assume that the IF bandwidth of the input bandpass filter is sufficiently wide that the
desired signal will pass through this filter undistorted and with negligible ISI. As such, the
output of the bandpass filter can be represented as
F(t) = V2P sin (@-t + O(t) + 8) +n, (t) cos (wet + 0) + n2(t) sin (wet + 8)
where 9 is the unknown phase introduced by the channel, 0(t) = meat is the effective phase
modulation transmitted in the i" bit interval i7, <t < (i + 1)T) with a; denoting the +1 i®
data bit, and n;(t), n2(t) are uncorrelated zero-mean Gaussian noise processes, each with
variance o* = NoB. Here B is the equivalent lowpass bandwidth of the input bandpass filter.
Also assume that the transfer function of this filter is symmetric around its center frequency, @¢,
and that the lowpass equivalent of this transfer function is denoted by H(w). Finally, we point
out that for FSK, differential encoding is not employed at the transmitter; hence, the transmitted
sequence {a;} is indeed the information data sequence.
Assume that the lowpass filter merely eliminates second harmonic terms of the carrier
and furthermore that w,.Ty = 2kz (k integer).
(a) Show that, in the absence of noise, the sample of the output lowpass filter taken at t =
iT, + Tg is given by
whTy
eta Sin
b
(b) Although the noise at the input to the threshold decision device will not be Gaussian,
one would still expect to get the lowest average bit error probability when the separation
between the signal components of z corresponding to a; = 1 and a; = —1 is a maximum.
Using the result of part (a), show that the value of delay, Tz, for which this condition occurs
is given by
(2n + 1)T,
LE caer yore iW? Oa Pc
(c) Since the delay must be restricted to lie in the interval 0 < Tz < Tp, show that for h < 0.5,
there is no suitable value of delay such that the signal components of z are maximally
separated and thus the performance of differentially detected FSK will be inferior to that of
DPSK.
7.5 Consider the receiver of Problem 7.4.
502 Differentially Coherent Communication with Waveforms Chap. 7
(a) Show that, in the presence of noise, a sample of the output of the lowpass filter taken at
t =iT» + Ty can be written in the form
1
z= 5 Re {ziza}
where
with n3(t) 4 ny(t — Ta), na(t) 4 no(t — Ta), and nj, n2,n3, n4, 6; respectively denoting
samples of the noise processes nj(t), n2(t), n3(t), n4(t) and the phase process @(t) taken at
P=SvIES SS 16%,
(b) Show that the correlation matrix of the noise samples 11, n2, 13, 14 1s given by
0 0
1 p
[ninj]=o 0 0
SS)
SiS
= vo) Si 1
ies)
— ; ; 2
te A Wer siee: io
1 et =
p,0\07 & 5 — 21)" (22 — 22); Pz= Pez + JPsz
7.6 Using the specific results of Problem 7.5 in the general bit error probability results given in the
description of Problem 7.3:
(a) Show that for differential detection of FSK, the average bit error probability is given by
Gerilscr |
a 1 — p2
(b) Show that for p = 0, the average bit error probability reduces to
ace P
= = exp {-——
a 2 *P\” Now
which agrees with the comparable result for DPSK given in Problem 7.3.
Teil A special case of Problem 7.6 corresponding to h = 0.5 and a phase modulation that is contin-
uous in time from bit interval to interval is referred to as minimum-shift-keying (MSK). (This
modulation is discussediiymore detail in Chapter 10). In particular, the elicit phase modu-
lation transmitted in the i"bit interval iTy <t < (i+ I)T is now O(t) =a; oi + x; where x;
is a phase constant whose value is determined by the phase continuity requirement at the bit
transition instant t = i7). Consider a modulation a(t) starting at t = 0 corresponding to an in-
put data sequence ay = —1, a; = 1, a2 = 1, a3 = —1, ag = —1,as= —1,a6= 1, a7 = —1, that
Problems 503
8
is, a(t) = }° ajp (t —iT,) where p(t) denotes a unit rectangular pulse of width T,. Assum-
i=0
ing that noise is absent, zero initial phase, that is, 9(0) = 0, and a delay Ty = Tp, sketch the
following waveforms over the time interval 0 < t < 87:
(a) a(t)
(b) 6(¢)
(c) A(t — Tp)
(d) A@(t) =O) — Ot — Th)
(e) sin A@(t)
Compare sin A@(t) with a(t). How do they differ?
7.8 For maximum-likelihood block-by-block differential detection of M-PSK, the pairwise error
probability is given by (7.68) where from (7.62) we have
i=0 m=0
7 Nyx? N,—-i-2
= Ve_-n,+1 + SS Vai Exp |-;(ss so-im)t
i=0 m=0
N,—1 cs Ns—t—2 rd
ce Si Ves; exp |-;(ye sin] |
i=0 m=0
i N,—2 i N;—i-2 ne
= Veen + > Vy_—i exp |;(De sion)
i=0 m=0
(a) Compute the mean, variance, and cross-correlation of the complex variables z; and 22
defined by
( =o
Brae tien nore 1 a lee
1 — — ,
p,0102 & 5 —Z1)* (22 — 22); Pz = Pez + JPsz
Wherever possible, express your answers in terms of the parameter 6 defined in (7.70).
(b) The arguments of the Marcum Q-function in (7.68) are obtained from [see (6.65)]
a 1 |S a et SS
a DN, 1 — |p, vt
= (psi
where
1 ea peu
SA =ll2;
D
OAarg(z)=tan!
="; 4 =1,2
Mek
N, Sof =03
Using the results found in part (a), show that a and b are given by (7.69).
7.9 In evaluating the upper bound on the bit error probability performance of maximum-likelihood
block-by-block differential detection of M-PSK, the sum of Hamming distances required in
504 Differentially Coherent Communication with Waveforms Chap. 7
(7.83) must be evaluated. To accomplish this, we start by observing that 5 of (7.70) can be
written in the form
N,-1 i
S=1t Yel; oj A>) 56-N,41-1
i=l n=1
We are interested in determining the various possible solutions for the 66’s such that the maxi-
mum value of the magnitude of 4, namely, |5|max, is achieved. There are four cases to consider.
Case 1: All N, — 1 vectors e/%; i =1,2,..., Ns; — 1 must be collinear and equal to efen/M
Case 2: All Ny — 1 vectors e/@; i =1,2,..., Ns — 1 must be collinear and equal to e~/27/™.
Case 3: Any N, — 2 vectors e/“' must be collinear and equal to 1 and the remaining vector
must be equal to e/?7/™.
Case 4: Any N, — 2 vectors e/“' must be collinear and equal to 1 and the remaining vector
must be equal to e~/27/™,
(a) Show that the solutions for the 50’s to Cases 1-4 are as follows:
baal oO) Peraeye Lav tee3 Liste aor estes eee IN tees
dl yen
20 (
sealSeales = OP b= 1 2y es Ne
Open
Case 4:
20 oA
JbOk—N, 4141 = Pg? SOk—N, +142 = i’ 504—N, +i =0;
eae
et On a CS aoe Ns ee eee
20 :
Ses ay a OO k= et eh == eee ee
(b — For the solutions of part (a) show that for a Gray code bit to symbol assignment, the
accumulated Hamming distances are as follows:
Case 1: w (u, a) = 1
Case 2: w (u, a) = 1
Case 3: w (u, a) = 2(N, — 2) +1
Case 4: w (u, @) = 2(N, — 2) +1
(c) From the results of part (b), obtain the result in (7.89) corresponding to M > 2.
7.10 For the special case of M = 2, show that Cases 1 and 2 of Problem 7.9 are one and the same
and likewise for Cases 3 and 4, that is, only half the number of solutions exist. As such, show
that (7.84) is now the appropriate solution for the accumulated Hamming distance.
7.11 For maximum-likelihood block-by-block differential detection of M-PSK, the conditional like-
lihood ratio is given by (7.58). In (7.60), the average-likelihood ratio was found by averaging
(7.58) over the pdf of the unknown phase, 6. The decision statistic was then found by max-
imizing this average-likelihood ratio over the signal sequence. Suppose now that we instead
References 505
first find the maximum-likelihood estimate, bu L, Of @, then substitute this value into (7.58) to
produce a maximum-likelihood ratio, and finally maximize this ratio over the signal sequence.
(a) Find the maximum-likelihood estimate of 6 defined by
quence § is identical to that obtained by maximizing the average likelihood ratio fy (v is)
of (7.60) over this same sequence.
7.12 Consider a DPSK system where the signal has a residual carrier component. As such, it can be
written in the form
where 6, < 90° is the modulation index, 6 is the unknown phase introduced by the channel,
p(t) is a unit rectangular pulse, and {b,} is a differentially encoded data sequence, that is,
bn = anbn—| with {a,} the true input data sequence. The signal is sent over an AWGN channel
and the receiver is as illustrated in Fig. 7.2b.
(a) Find the mean of the sample Z; under the hypothesis a; = 1 and a; = —1, that is, how does
the presence of the residual carrier affect these results relative to the case where it is absent
(Om 902)7
(b) Is it reasonable to expect that the receiver will work even in the presence of the residual
carrier? If so, what might you expect to choose for the decision threshold?
(c) How might you expect the value of 6,, to affect the system performance, that is, what is the
best choice of 6,,?
7.13 For the equivalent nonoverlapped signal set defined in (7.103), show that the signals have the
normalized correlation matrix given by (4.116). Hint: Write 5;(t) as 5,(t) = 5, (t) — 5/(t) in the
interval 0 <t < (#4) T; where
todff
K(t); O<t< (7) ii
5, (tf) =
06 (gh)nsts(Wh)t
and make use of the orthogonality properties of the 5;(t)’s over this same time interval.
REFERENCES
1. Fleck, J. T. and E. A. Trabka, “Error Probabilities of Multiple-State Differentially Coherent
Phase-Shift-Keyed Systems in the Presence of White, Gaussian Noise,” in Investigation of Digital
Data Communication Systems, J. G. Lawton, ed., Buffalo, NY: Cornell Aeronautical Laboratory,
Inc., January 1961. Available as ASTIA Document No. AD 256 584.
2. Park, J. H., Jr., “On Binary DPSK Detection,’ JEEE Transactions on Communications, vol.
COM-26, no. 4, April 1978, pp. 484-86.
3. Simon, M. K., “Comments on ‘On Binary DPSK Detection’,” JEEE Transactions on Communi-
cations, vol. COM-26, no. 10, October 1978, pp. 1477-78.
4. Pawula, R. F,, S. O. Rice and J. H. Roberts, “Distribution of the Phase Angle between Two
Vectors Perturbed by Random Noise,” /EEE Transactions on Communications, vol. COM-30,
no. 8, August 1982, pp. 1828-41.
506 Differentially Coherent Communication with Waveforms Chap. 7
2 Arthurs, E. and H. Dym, “On the Optimum Detection of Digital Signals in the Presence of White
Gaussian Noise—A Geometric Interpretation and a Study of Three Basic Data Transmission
Systems,” IRE Transactions on Communication Systems, vol. CS-10, no. 4, December 1962, pp.
336-72. Also TM 3250, Bedford, MA: Mitre Corporation.
. Bussgang, J. J. and M. Leiter, “Error Rate Approximations for Differential Phase-Shift-Keying,”
IEEE Transactions on Communication Systems, vol. CS-12, no. 1, March 1964, pp. 18-27. Also
SR-103, Bedford, MA: Mitre Corporation.
. Lawton, J. G., “Comparison of Binary Data Transmission Systems,” Proceedings of the National
Convention on Military Electronics, 1958, pp. 54-61.
. Cahn, C. R., “Performance of Digital Phase-Modulation Communication Systems,’ IRE Trans-
actions on Communications Systems, vol. CS-7, no. 1, May 1959, pp. 3-6.
. Weinstein, F. S., “A Table of the Cumulative Probability Distribution of the Phase of a Sine
Wave in Narrow-Band Normal Noise,” IEEE Transactions on Information Theory, vol. IT-23,
September 1977, pp. 640-43.
10. Lindsey, W. C. and M. K. Simon, Telecommunication Systems Engineering, Englewood Cliffs,
NJ: Prentice-Hall, 1973. Reprinted by Dover Press, New York, NY, 1991.
ike Lee, P. J., “Computation of the Bit Error Rate of Coherent M-ary PSK with Gray Code Bit
Mapping,’ JEEE Transactions on Communications, vol. COM-34, no. 5, May 1986, pp. 488-91.
. Korn, I. and L. M. Hii, “Relation between Bit and Symbol Error Probabilities for DPSK and FSK
with Differential Phase Detection,” to appear in the JEEE Transactions on Communications.
. Pawula, R. F. and J. H. Roberts, “The Effects of Noise Correlation and Power Imbalance on
Terrestrial and Satellite DPSK Channels,’ JEEE Transactions on Communications, vol. COM-31,
no. 6, June 1983, pp. 750-55.
. Rice, S. O., “Statistical Properties of a Sine Wave Plus Random Noise,” Bell System Technical
Journal, vol. 27, January 1948, pp. 109-57.
. Roberts, J. H., Angle Modulation, England: Peregrinus, 1977.
. Lee, J. S. and L. E. Miller, “On the Binary Communication Systems in Correlated Gaussian
Noise,” [EEE Transactions on Communications, vol. COM-23, no. 2, February 1975, pp. 255—
59.
If. Huff, R. J., “The Imperfectly-Timed Demodulation of Differential Phase Shift Keyed Signals,”
Report 2738-1, ElectroScience Laboratory, Department of Electrical Engineering, Ohio State
University, June 1969.
. Simon, M. K., “A Simple Evaluation of DPSK Error Probability Performance in the Presence of
Bit Timing Error,’ to appear in the February 1994 issue of the JEEE Transactions on Communi-
cations.
. Salz, J. and B. R. Saltzberg, “Double Error Rates in Differentially Coherent Phase Systems,”
IEEE Transactions on Communication Systems, vol. COM-12, June 1964, pp. 202-05.
. Oberst, J. F. and D. L. Schilling, “Double Error Probability in Differential PSK,” Proceedings of
the IEEE, vol. 56, June 1968, pp. 1099-1100.
. Stein, S., “Unified Analysis of Certain Coherent and Noncoherent Binary Communications Sys-
tems,” IEEE Transactions on Information Theory, vol. IT-10, no. 1, January 1964, pp. 43-51.
. Divsalar, D. and M. K. Simon, “Multiple-Symbol Differential Detection of MPSK,” JEEE Trans-
actions on Communications, vol. COM-38, no. 3, March 1990, pp. 300-08.
. Wilson, S. G., J. Freebersyser and C. Marshall, “Multi-Symbol Detection of M-DPSK,’” GLOBE-
COM °89 Conference Record, Dallas, TX, November 27-30, 1989, pp. 47.3.1-47.3.6.
- Helstrom, C. W., Statistical Theory of Signal Detection, England: Pergamon Press, 1960, p. 164.
References 507
25: Mackenthun, K., “A Fast Algorithm for Multiple-Symbol Differential Detection of MPSK,” sub-
mitted for publication in the IEEE Transactions on Communications.
26. Schwartz, M., W. R. Bennett and S. Stein, Communication Systems and Techniques, New York:
McGraw-Hill, 1966.
Dale Carlson, A. B., Communications Systems, 2nd ed., New York: McGraw-Hill, 1975, pp. 396-404.
28. Downing, J. J., Modulation, Systems and Noise, Englewood Cliffs, NJ: Prentice-Hall, 1964, pp.
184-87.
Pas): Park, J. H., Jr., “Effects of Band Limiting on the Detection of Binary Signals,” IEEE Transactions
on Aerospace and Electronic Systems, vol. AES-5, September 1969, pp. 868-70.
30. Abramowitz, M. and I. A. Stegun, eds., Handbook of Mathematical Functions with Formulas,
Graphs, and Mathematical Tables, New York: Dover Press, 1965 (originally published by the
National Bureau of Standards in 1964).
Sile Leib, H. and S. Pasupathy, “The Phase of a Vector Perturbed by Gaussian Noise and Differentially
Coherent Receivers,’ JEEE Transactions on Information Theory, vol. IT-34, no. 6, November
1988, pp. 1491-1501.
a2r Edbauer, F., “Bit Error Rate of Binary and Quaternary DPSK Signals with Multiple Decision
Feedback Detection,’ JEEE Transactions on Communications, vol. COM-40, no. 3, March 1992,
pp. 457-60.
335 Winters, J. H., “Differential Detection with Intersymbol Interference and Frequency Uncertainty,”
IEEE Transactions on Communications, vol. COM-32, No. 1, January 1984, pp. 25-33.
34, Chethik, F., “Analysis and Tests of Quadrature Advance/Retard Keying Modulation and Detec-
tion,’ Aeronutronic Ford Corp. (internal technical memorandum), August 1974.
35. Alexovich, J., P. Kossin and T. Schonhoff, “Bandpass-Limited Symmetric DPSK: A Bandwidth-
Efficient Modulation for Nonlinear Channels,’ MJLCOM’92 Conference Record, San Diego, CA,
October 11-14, 1992, pp. 35.4.1-35.4.5.
36. Davarian, F., M. K. Simon and J. Sumida, “DMSK—a Practical 2400 bps Receiver for the Mobile .
Satellite Service,’ Pasadena, CA: Jet Propulsion Laboratory, JPL Publication 85-51, June 1985. ~
Sie Simon, M. K. and C. C. Wang, “Two-Bit Differential Detection of MSK,’ GLOBECOM’S4
Conference Record, December 1984, Atlanta, GA, pp. 22.3.1—22.3.6.
38. Kaleh, G. K., “Differentially Coherent Detection of MSK,” JCC’88 Conference Record, Philadel-
phia, PA, June 12-15, 1988, pp. 23.7.1-23.7.5.
Shep Kaleh, G.K., “A Differentially Coherent Receiver for Minimum Shift Keying Signals,” JEEE
Journal on Selected Areas in Communications, vol. SAC-7, no. 1, January 1989, pp. 99-106.
Chapter 8
Double Differentially Coherent
Communication with
Waveforms
1. Herein, we shall refer to this form of differential detection as first-order differential detection (recall that the
input information phases are encoded as a first-order difference of two successive transmitted phases) to distinguish
it from second-order and higher-order differential detection schemes, whose definitions will be introduced shortly.
Sec. 8.1 Definition of Higher Order Phase Difference Modulation (PDM) 509
it insensitive to this offset, thereby maintaining the previously discussed advantage of imple-
mentation simplicity. The means by which this is achieved lies in encoding the information
data phases as a second-order difference phase process” and using a two-stage differential
detection process. The class of schemes that accomplish this goal are herein referred to as
second-order differential detection or more simply double differential detection [2,3]. When
specifically applied to phase-shift-keying, we shall refer to this class of schemes as double
differential phase-shift-keying (DDPSK or D*PSK).
From a historical perspective, the Russian literature [4-12] beginning around 1972
was first to introduce the notion of using higher order phase differences for modulation and
detection when the channel introduced higher order phase variations. Later in the 1970s,
Pent [13] contributed much additional insight to these ideas and indeed produced a result
that allowed for analysis of several extensions of the theory given in [2] and a comparison
with the work of [3]. Since much of the work in [4-13] is not commonly known in this
country, in fact the book (in Russian) by Okunev [2] is, to the authors’ knowledge, the
only textbook that even mentions higher-order differential detection, we shall make every
effort in this chapter to expose the reader to this body of literature, giving due credit where
appropriate.
namely, the phase difference’ between the current and previous transmitted signal phases.
Turning this relation around, we express the signal phase transmitted in the 7th transmission
interval as
2. If A26; denotes the data information phase to be communicated in the i transmission interval defined by
iT; <t < (i+ 1)T,, then the actual phase 6; transmitted over the channel in this same interval is related to A?6; by
the second-order difference A26; = (6; — 0;-1) — (0:1 — 6;-2) = 9; — 20;—-1 + 6;—2. This will be expanded upon
later in the chapter. For now, it is sufficient to understand this simple definition of the term second-order difference
modulation.
3. Unless otherwise noted, it is assumed that the result of first and higher order difference and addition operations
performed on phase angles are evaluated modulo 277.
510 Double Differentially Coherent Communication with Waveforms Chap. 8
exp{ j6;_2}
Figure 8.1b Equivalent transmitter to
Figure 8.1la for binary PSK
where the latter equality in (8.3) again holds because the allowable values for 6; are only
0 and z. Figure 8.1b is an illustration in complex form of (8.4).
Sec. 8.2 Structures Motivated by the Maximum-Likelihood (ML) Approach 511
Clearly the above ideas can be extended to higher (than two) order phase differences.
In particular, define the n“" order phase difference modulation (PDM-n ) by
n
ah ps a aos 8.6
@E- ess Se
is the binomial coefficient. The modulation of (8.5) is implemented by including n back-
to-back differential encoders between the information source and the phase modulator and
would be of interest in situations where one is attempting to mitigate the effects of an
(n — 1)-order phase derivative introduced by the channel. When applied to binary PSK, the
acronym D” PSK is used.
What is important in the above is that PDM-n is a modulation with memory in that
the current transmitted phase depends on the immediate n previous transmitted phases. This
property will be exploited later on in the chapter when consideration is given to optimum
detection schemes for such modulations.
Spel! (8.7)
and the corresponding received complex signal is
Fj es Ge O27 ATK) + ii; A piel Cine ren Af ti+ni) (8.8)
where nj; is a sample of complex Gaussian noise. From Fig. 8.2, we have that the output of
the differential detector in the same transmission interval is given by
512 Double Differentially Coherent Communication with Waveforms Chap. 8
For large SNR, we can neglect the term njn;_, and thus Ni can be approximately modeled
as a Gaussian random variable. Assuming that the frequency offset is constant but unknown,
then the phase © = 2x Af T, can be modeled as being uniformly distributed in the interval
(—z, 2). Hence, we can look at the combination of differential encoder, channel, and dif-
ferential detector in Fig. 8.2 as an “ equivalent” channel with additive white Gaussian noise
N, and random (but time-invariant) phase ©. This is depicted by the illustration of Fig. 8.3.
The problem of finding the optimum receiver for this equivalent channel with input e/4%
is now the same as that previously treated in Chapter 7, namely, apply a differential encod-
ing/detection scheme identical to that used for an AWGN channel with zero frequency offset
and random (uniformly distributed) time-invariant phase. For an N;-symbol observation,
the solution is again the maximum-likelihood sequence receiver derived in Section 7.2.1.
For N; = 2, we have the classical differential detector based on symbol-by-symbol deci-
sions which requires first-order differential encoding at the transmitter. This simplest case
is illustrated in Fig. 8.4. Finally, putting Figures 8.2—8.4 together, we see that for large
SNR, the optimum transmitter/receiver configuration equates to back-to-back differential
encoders at the transmitter (see Fig. 8.1a) and back-to-back differential detectors at the re-
ceiver (Fig. 8.5). The output of the receiver in Fig. 8.5 is from (8.8) and (8.9) given by
[ |
PiPi-1 EXP) jl 9; — O-1 + 22ST, +7; — Ni |
A@ i An;
Differential Differential
Detector Detector
[ |
PiP;-1Pi-2 ia} AG, - A0,-, +An; An;
A* 6; An; |
Figure 8.5 Optimum (large SNR) receiver for second-order phase difference modulation
depending on the specific implementation, the final detector output will, in the absence of
noise, either be entirely independent of the input frequency (the detection scheme is abso-
lutely frequency-invariant) or vary as a function of frequency within some specified small
frequency range (the detection scheme is relatively frequency-invariant). It is important to
emphasize that even though a detector is absolutely frequency-invariant according to the
above classification, its error probability performance will in most cases depend on the fre-
quency offset. Typically (as we shall see shortly), this comes about because of the depen-
dence of the signal x noise terms in the detector ouput on the frequency offset.
514 Double Differentially Coherent Communication with Waveforms Chap. 8
n(t)
jae (at
iT,
We begin with an examination of well-known PDM-1 receivers (at least one of which
was presented in Chapter 7) with emphasis on their behavior and performance in the pres-
ence of carrier frequency offset. The primary purpose of this discourse is to expose the
configurations that will later represent the first stage of serial-type receivers for PDM-2
modulation.
where w, is the radian carrier frequency, 6; is the differentially encoded information phase
in that interval taking on values 0, 2, and @ is the unknown phase introduced by the
channel (assumed to be constant over an interval of duration 27), sec). The sum of s(t)
and white Gaussian noise n(t) (with single-sided power spectral density No) are passed
through a bandpass filter (BPF) of single-sided bandwidth W Hz (B = W/2 is the equiv-
alent lowpass bandwidth) producing the signal x(t). Assuming, for convenience, that the
center radian frequency, w-o, of the bandpass filter is an integer multiple of the data rate,
that is, wo = MQ; 2 = 27 / Tp, m integer, then the actual carrier radian frequency in (8.12)
can be expressed as mw, = m&2 + Aw where Aw = 27 Af represents the radian frequency
uncertainty about the nominal carrier frequency. Note that the bandwidth of the BPF is
chosen commensurate with the maximum value of Aw. Assuming that the signal s(t) re-
mains undistorted after passing through the BPF (this is equivalent to assuming a wide
enough bandwidth W relative to the data rate 1/7) that intersymbol interference (ISI) can
be ignored), then expressing the filtered noise in a truncated Fourier series* in the interval
iT, <t < Gi +1)Tp, we have
4. This form of the noise expansion was used in [10] and in the context of this problem was first introduced
in [5].
Sec. 8.3 Autocorrelation Demodulation (ACD) of Binary PDM-1 Signals 515
ky . (8.13)
+ So (ns1,¢ sin KQt + Awt +) + n¢1,4C08 (Kt + Aot + 6))
k=k,
where the number of harmonics in the sum is given by kz — kj + 1 = 2BTy, and the carrier
frequency integer, m, lies in the range kj < m < kp. Thus, for the results that follow, 2B 7;
is assumed to be an integer.
Similarly, in the interval iT, < t < (i + 1)Tp, the BPF output corresponding to the
previous interval is given by
ka (8.14)
+ S> (12,4 sin (kQt + Act + A) + 1¢2,¢ 008 (KQt + Act + 6))
k=k
The quantities 51,4, M52,k, Nc1,k, %c2,k are independent Gaussian random variables with zero
mean and variance No/ Tp.
The ACD algorithm for the DPSK signal is to make the decision A6; on the dif-
ferentially encoded data phase AO; = 6; — 6;_; transmitted in the ith transmission interval
according to
ns (i+1)T,
d; = sgn / r(t)r(t — T,)dat (8.15)
iT,
where, as in previous chapters, sgn {-} denotes the signum (sign) function of its argument.
Using (8.13) and (8.14), the integral in (8.15) can be written as
AT,
ar Msn cos 6;_; cos AwTp ;
8.16
—_ a Nelm cos 6;-; sin AwT,
k2
Tp
ae > a (ns1,KMs2,k + Nel ,kNc2,k)
Substituting 51,4 = Vsi,k + Vs2,k, Ms2,k = Vstk — Vs2,k> Nelk = Velk + Ve2,k, and Ne2,k =
Vel.k — Ve2,k into (8.16) allows it to be expressed as the sum of the squares of Gaussian-
distributed independent random variables (the v,’s and the v,’s). Arbitrarily assuming that
the data phase A6; = 0, then an error occurs when the integral in (8.16) is negative and the
error probability is given as
516 Double Differentially Coherent Communication with Waveforms Chap. 8
AoT,
Py CB) Pr {(m + Acos6;— 1 cos? aah
kz
Asin cl 2 2
+
Vel,m — ee Sue ys (214 vet)
k=k
k#m
kz
Asin Sot 2
“hE
Ve2,m + see) 48 SS (v4 + v2.4)
=Pr{aj <5}
where A; and Az have noncentral x2 distributions with 4B7, degrees of freedom. The
details of the evaluation of (8.18) are carried out in [12] with the following results. Letting
a = AwTh, that is, the ratio of radian frequency offset to data rate, then
1 E
PEC) = FBT, exp |- a (2sin? S + 00s? S)|
ie
x, (2fsin?g0
n! Me ro
1, (Roots) 0
!
m
(
kL
Bd
k—m )
(8.18)
where Ep = A*T;/2 is the energy in the signal and : is the combinatorial coefficient
defined in (8.6). When the frequency error is absent, that is, a = 0, then (8.18) simplifies to
a result previously obtained in [4, 11], namely
i
2BT,—1 (£2) 257-1
1 E, b (#) ay lk PE ORT Ea
PE) = saa ex? |-5"| > aries Ds a’ soar ) (8.19)
| E LE
P(E) |\2a7%,=2 = Ds
= exp (-<)
No
1usZ| (8.21a)
5. Strictly speaking, for 2B7;, = 1, the assumption that the signal s(t) remains undistorted after passing through
the BPF is not valid. Thus, (8.20) is optimistic for the configuration of Fig. 8.6.
Sec. 8.3 Autocorrelation Demodulation (ACD) of Binary PDM-1 Signals 517
ey +30 (#)
+ exp (-3) (f) ve
3127 (8.2 1c)
which shows the degradation of performance caused by widening the input BPF bandwidth.
Again, these results are somewhat optimistic because of the zero ISI assumption, but im-
prove in their accuracy as 2B7), increases.
Figure 8.7 illustrates Pp(E) as computed from (8.19) versus E,/No in dB for values
of 2BT) equal to 1, 2, 4, 8, 16, and 32. Also illustrated are the approximate values obtained
from (7.118) of Chapter 7 corresponding to the analysis of this detector assuming Gaussian:
statistics at the output of the [&D. We observe that for small values of 2B7),, the Gaussian
approximation analysis yields very poor results relative to (8.19) over the range of P,(F)
between 107! and 10~°. As 2B7; increases, the Gaussian approximation analysis becomes
better and the range of values of E,/No values over which (7.118) tends to agree with the
exact (assuming zero ISI) result in (8.19) is increased. However, for any fixed and finite
value of 2BTp, it is clear from Fig. 8.7 that for sufficiently large E,/No, the results in
(7.118) and (8.19) deviate significantly from each other and thus it is inappropriate to use
(7.118) to predict asymptotic (E,/No — oo) performance.
An approximate account of the effect of frequency error on the noise immunity of the
ACD is suggested in [4] and is based upon replacing E,/No by Ep/No cos* a in (8.19) (the
result for zero frequency error), which yields
The advantage of this approximate expression is that, for computational purposes, one
does not have to compute an infinite number of terms. The degree of accuracy to which
(8.22) approximates (8.18) was investigated in [4] with the following results. For 2B7p <
10, Epy/No = 4 (6 dB), and a = AwT) < 45°, there is less than | dB inaccuracy in the
assessment of the E,/No requirement for fixed P,(£) using the approximate result in (8.22)
relative to that in (8.18).
518 Double Differentially Coherent Communication with Waveforms Chap. 8
10°
10°
ioe
5
Figure 8.7 Bit error probability performance of ACD of PDM-1! signals based on exact
(zero ISI) and Gaussian approximation analyses
J, Ode
(i+1)7,
Hy
(i+1)7, (i+1)T,
vie rind =A | cos (Awt +04 6;)dt+ N;
iTh iTp
(8:25a)*
sin Aol 1
and
(i+1)Tp (i+1)T)
Voi ih r(t)re(t)dt = A | sin (Awt + 6 + 6;) dt + N3
iTp iTp
(8.25b)
sin AwTp 1
Here N and N2 are independent zero-mean Gaussian random variables with variance No7p.
Analogous to (8.15), the decision algorithm here is
520 Double Differentially Coherent Communication with Waveforms Chap. 8
Din HORN 7
2 sin 5
= sen {A 7, = cosa cos AG; + Neg¢
Z
where again a = Aw7) and we have furthermore made the observation that, for binary
phase modulation, sin Ad; = 0. Also, Neg is the effective noise consisting of signal x noise
and noise x noise components as follows:
sin 5 \ | saa
Neff = ATy a me COS > {(2i = l)a ia 26; =F 20)
2
1
+ N; cos (5
2 [(2i — 3) + 26,1 + 20]); (8.27)
1 1
+N4 sin (5[(2i — l)a + 26; + 201)+ N3 sin ( [(2i — 3)a + 20;-; + 201)
+ Ni N2 + N3N4
Here again; N2, N4, which represent the noises in V, ;-; and V.;—1, respectively, are inde-
pendent zero-mean Gaussian random variables with variance No7) and are independent of
N, N3. The effective noise in (8.27) can be written in a more compact form, viz.
sin 5
Nett = ATp | —~ }Nsxn + Nuxn (8.28)
D
where Nsxy iS a zero-mean Gaussian random variable with variance 2No7, and Nyx =
N,N2 + N3Nz4 is a non-Gaussian random variable that is uncorrelated with Nsyy.
Following the approach taken by Stein [14], the error probability of the decision
algorithm given in (8.26) can be shown to be
|
P(E) = 5 [1- 0(vb, va) +0 (va, vb) (8.29)
where
2E! 2E!
a= a sinta: b= on cos? a (8.30)
with
sin % \?
B, 8 Bs ( =) (8.31)
)
and Q(a, b) is as before the Marcum Q-function. This result was originally obtained by
Henry [15].
Sec. 8.4 |-Q Demodulation of Binary PDM-1 Signals 521
Note that for zero frequency error (a = 0), we have from (8.30) and (8.31) that a = 0
and b= 2E>»/No. Then, using the special cases of the Marcum Q-function as in (7.76),
it is easy to see that (8.29) reduces to (8.20) as it should. On the other hand, there are
many values of a, the smallest of these being a = 2/2 (Af = 1/47,), for which the error
probability goes to one-half. Even at Af = 1/67>, there is already a loss of 3 dB in SNR
relative to the case of no frequency error. This rather rapid deterioration of performance
with @ is a direct consequence of using I&D filters that are not matched to the inphase and
quadrature output signals when frequency error is present.
The error probability performance of the I-Q demodulator in Fig. 8.8 can also be
obtained in several other equivalent forms. Using a geometric approach [3] to be discussed
later in this chapter in connection with the demodulation of higher order PDM, the authors
obtained the following integral form which is identical to (8.29)-(8.31), namely
(8.32)
ih e Ey ( Eos a
x F € No
ANS n Or COs n ex exp SSS
No n} an
Also, replacing the error function and exponential functions by their expansions in terms of
modified Bessel functions of the first kind, namely
ape) [o-e) 2
erf(zcos¢) = a EY ye (—1)‘en ly (=)
4 k=0 Z
Ka (8.33a)
cos [(2k + 1)¢] _ COS [(2k — 1)¢]
ip eorex 2k —1
and
CO
PAE | E,, =)
p(E) = > 2No exp No
> (8.34)
2 E’ E; cos [(2k + l)a]
—1)* pawl: |fal PE |e(iseaeSeas
. 2 | ” E (52) Se (=*)| 2k+1
Miller [16] obtained a similar and equivalent form to (8.32), which is given by
522 Double Differentially Coherent Communication with Waveforms Chap. 8
perpee E, cos” a
PAB) Se Beato) |elk Mie alae )|67)
nie =) exp | St (oe
1 m/2 E!
—.,/E, erf > cos ncosa
ON No (8.35)
2E; Bam
— erf —® cos (n+ a) cos n exp (-= sin? ")dn
No No
It is interesting to study the limit of (8.29) in the limit of large E»/No. In particu-
lar, using the asymptotic forms of the Marcum-Q function given in (7.76), we obtain for
cosa > 0 (e.g., 0 < |a| < 2/2)
= me else |e E,
ie )
ee (jsina)) (1 — [sine exp(
~ mei. 36b
0O- cosa = 0
Ep/No>co le Ccosa ra) (8.37)
that the results of (8.21) would exactly apply. In [1], this was verified for the case 2BT; =
2 where a 100% excess bandwidth raised cosine transfer function was employed for the
overall (transmitter plus receiver) Nyquist filter response. Since, for an AWGN channel with
no ISI, the average error probability is minimized (with respect to the noise) by splitting the
total filtering (transmit filter, receive filter, and matched filter) equally between transmitter
and receiver (see Chapter 9), then for the above this can be accomplished as follows.
The transmit filter (see Fig. 8.9a) is a root raised cosine filter with transfer function
Ty. 2
Pay = V2T, cos“; lol < 7 (8.38)
0; otherwise
The input receive filter is a brick wall filter whose equivalent lowpass version has a transfer
function magnitude Pr(w) given by
Pr(w) = |
AE lol < F (8.39)
0; otherwise
ols. 2x
P(o) = Pr(w) Paw) = {0;Trees a otherwise
(OS 7, (8.40)
The impulse response of such an overall characteristic has the property that two of its
samples taken at intervals of 7/2 sec are equal and nonzero with the remaining samples
all having zero value, that is,
BG) Pla)
Th Tp ; (8.41)
T;
p (+ok +1) 2) =0; k=1,2,...
The matched filter (assuming zero frequency offset) for the pulse shape of (8.41) has
a transfer function whose magnitude is given by (see Fig. 8.9)
(62)Tp
Thus, we see that the cascade of (8.40) and (8.42) achieves the Nyquist pulse shaping
corresponding to a 100% excess bandwidth raised cosine transfer function and also the equal
split (except for a scale factor) of the total filtering between transmitter and receiver.
Since zero ISI is achieved by this implementation, the optimum detector (for zero
frequency error) would be to perform the matched filtering prior to the differential detection
524
1e ojdureg
=] ae
Ao
cle
—-1)9"¢
(yu
osu
(1)u
0}
[enusraszI
D}
10}09}0p
uy.
4 (1)= (4yu—1)!d"0"{
coo—=u
ye o[dures
€
=] Loe
be)
(see Fig. 8.9b). When this is done, the error probability performance is indeed given by
(8.20). If on the other hand, differential detection is performed prior to matched filtering
(see Fig. 8.9c), then it can be shown [17] that the error probability is exactly given by
P,(E) loBT%=2 of (8.21) which corresponds to the case B = 1/Tp.
: iTy (i+1)T,
d; = sgn i r(‘)r(t — Tat [ r(t)r(t — Tp)dt
(i~1)Tp iT
where f(t), 7(t — T,) are modeled as in (8.13) and (8.14) and roo9(t — Tp) is the Hilbert
transform of F(t — Ty), then the I&D outputs V..;, Ve,i-1, Vs,i, and Vs ;-1 are given by
V.,; = E,cos AG; cosa + Ni; Ve,i-1 = Epcos AGj-1 cosa + N2;
(8.44)
V; ;= Epcos Aé; sina + N3; Vs,i-1 = E,cos AG;-1sina + N4
tu (1)
=
cou
: —1)4d""¢
(4qu
aaNnSIy
qg6"g jUSTeAIND| [eoNeUIYJeUI
Jopow
JO OY) Wd}sAs
Ul QINSIY26°8
ojdures
Je
JO poyney
|
jenuaiayjiq
1019919
|
-1)4d*v"{
(*yu
OM
4 (1)
=
9p
+
JOI[Y
ANSI
I¢6"g
payoieul
10}99]
JOLId
YM
[BNUSIAJJIP
YOO[G
WLISLIP
JO
Oy]
0}
W}sks
Cru—1)9"¢
527
528
SIL] I8¥JS | puodasa3e}S
f m),
UC1+1)
aans1yq
OT'g
= uonejauosoiny
Jow[npowlsp
Jo ASdzC (7-Wwdd)speusis payuriday)
wor[Z] © (AAAI
Sec. 8.6 Serial (Absolutely Frequency Invariant) Structures 529
AT, b AT)
i a 752.m cos6; + a Mtstam cos 6;_; cosa
AT; Th
oy ee Bel,m<08 6;—1 sina + > Sy (1s1,4752,k + Me1,kNc2,k)
k=ky
ATp ATp
No = a Ts3im cos 6-1 + a [ts2,m cos 6-2 COs a
ATp The
= a tem cos 6; 2 sina + > de (ns2,4Ns3,k si Nc2,kNc3,k)
k=k,
(8.45)
AT) AT» 4
N3 = ek cos 6; + a Msi.m cos 6;_1 sina
ATp lip ka
+ —-Nei,mcos6j—1 cosa + >2 }7 (nsi,krter,k — Nel, bs2,k)
k=k,
AT» AT), ’
N4= a Me3.m cos 6;-; + a s2m cos 6;_2 sina
ATs Tiek
Bite a Tem Cos 6;-2 cos a + oy ye (152,403,k oe Nc2,kN53,k)
k=ky
The noise random variables N;; i = 1, 2, 3, 4 are all zero mean with variance [10]
On OE
= ING |it oe (8.46)
é
M et E»/No
Furthermore, N; and N3 are uncorrelated and likewise for Nz and N4. However, the other
pairs are correlated with normalized correlation coefficients®
on on 2 [1ca ft]
(8.47)
, NiNa — N2N3 _ sina cos (A76;)
Oe Gre me ay em aoe PRAY Fe
on oN 2 [i¥ rote |
For 2BT, > 1 and E;/No not too large relative to 2BTp,, Ni, N2, N3, and Nq can be approx-
imated by Gaussian random variables. Substituting (8.44) into (8.43) gives
d; = sgn{ E; cos AG; cos A6;-; + Ni Ep, cos AO;—1 cosa + N2E, cos AG; cosa
+ N3Ep cos AG sina + N4Ep cos AG; sina + NiN2 + N3N4} (8.48)
6. This result was first pointed out by Shouxing Qu while at Technical University of Nova Scotia in a private
communication to the authors.
530 Double Differentially Coherent Communication with Waveforms Chap. 8
where Nsyy is a zero mean Gaussian random variable with variance 2az and Nvnxn =
N\N> + N3N4. The error probability may be evaluated using Stein’s approach [14]. In
particular, it is straightforward to show that
P(E) = ;[1- o(vb, va) + 0 (Vb, va)] - 5exp (-“=*) Ip (ab) (8.49)
where
1 BT aye
F (i 5 sin «(1 ar a)
1 1 Bisa, E, Bin v=
Sova ee ®iy Bore I 8.51
foes |! ncaa |e |a ne) ren
The result in (8.51) differs from that given in [8], namely
=
nie) = 5000 2No
se (14 E»/No
ele ) (8.52)
since in (8.52) the correlations of (8.47) were not taken into account. It is felt that this
difference between (8.51) and (8.52) will be small over the region of parameter values where
the Gaussian assumption is valid.
It is difficult if not impossible to obtain an exact (even assuming zero ISI) expression
for P,(E), analogous to (8.18) and (8.19), for the ACD scheme of Fig. 8.10. Thus, we
cannot check the absolute validity of the Gaussian approximation results obtained in (8.49)—
(8.51). However, based on the results for PDM-1 in Fig. 8.7, we intuitively expect that for
sufficiently large BT,, over a range of P,(E) values of practical interest, the above results
can be used with a fair degree of accuracy. As a measure of some confidence, the authors in
[10] do compare the Gaussian analysis results (not accounting for the correlations of (8.47),
however) to those obtained from a computer simulation and show reasonable agreement (see
Table 8.1).
The PDM-2 demodulator of Fig. 8.10 can be generalized to apply to M-ary (M > 2)
D*PSK (M-D?PSK) with information phase angles f,, as defined in Section 8.1. The result-
ing configuration is illustrated in Fig. 8.11. Letting Vi = Vc; + jVs,i, then the appropriate
decision rule is:
P,(E)
FT, E,/ No (dB) Simulation Results Eq. (8.49)
with
Ve = Re {&%} = Re {ViVive |
(8.53b)
= (VeiVei—1 + VsiiVs,i-1) COS Be + Wei Vs,i-1 — Vs,iWe,i—1) Sin By
Since the exact symbol error probability performance of differentially detected PDM-1
modulation with M signals and no frequency offset is given by (7.7), then by analogy with
the relation of (8.53) to the exact bit error probability when M = 2, that is, (8.20), the
symbol error probability performance of Fig. 8.11 in the presence of frequency offset can
be approximated by
-1
Es
(1re te)
BTs
[1 — cos 4,cos p|
ree Bees ties gis
sin 7/2 st {-2
P,(E) = —* Be (8.54)
Es od 1 — cos 47 cos
In (8.54), E; denotes symbol energy and 7; denotes symbol time. Furthermore, assuming a
Gray code bit-to-symbol mapping, the bit error probability, P,(E), would approximately be
given in terms of the symbol error probability by (4.135).
Finally, in Chapter 7 we also presented an easy-to-evaluate asymptotic (large E/No)
approximation to the symbol error probability performance of M-DPSK, namely, (7.12). For
the PDM-2 demodulator of Fig. 8.11, this would become
btcey 1) 1 Es BT, i. 1
eae ME ect i (1 = cos =
Be eaeoe N3 a E,/No a) - M>3(8.55
ee
532
AN AU SaJAN
I-ro,19is I-1's
ef *IK1+1) POL
.
wm),
“11+1)
| tad a7 at
te Sg
d Ror Bs I-W
/ Pees W gf_ bry Coeebarby
ainsi
["g
= Joye[np
uonrjauosoiny Jo owep WASdzd-W (7-Wdd)sjeusis payuuida
wouy
y)[Z]
© (AaaI
Sec. 8.6 Serial (Absolutely Frequency Invariant) Structures 533
aga €a ep ( 4a
[o-e) »
Gp? Ep
Teese Ve Ep
meet 8.56
ca {oe malo Cay
aa 2m+1/2 EP
2No ae 2m+1+4+1/2 Eb
2No
It is also shown in [8] that (8.56) represents a good lower bound on the performance of the
ACD of Fig. 8.11 in the presence of frequency offset when 2 BT; is small.
For M-DPSK (M > 2), an exact expression cannot be found. However, a union bound
on symbol error probability P;(£) in the form of (8.56) can be obtained with the result [13]
y oe ied
K i {im( Te)+ Int (=)| (8.57 )
m=0
lie Es xn f{M—2
x |lom+1/2 2No + Iym4+141/2 2No cos ik a Pama
The bit error probability upper bound can then be obtained from (4.140) and (8.57).
The exact error probability performance of Fig. 8.13 (specialized to the binary D?PSK
case) in the presence of frequency offset was first carried out in [3] using a geometric
approach analogous to that taken by Arthurs and Dym [20] in evaluating the performance
of DPSK. The result obtained in [3] is in the form of a double integral and is given by
7. Figures 8.12 and 8.13 are equivalent (produce the same decision algorithm) in the same sense that their
counterparts for PDM-1 modulation, namely, Figs. 7.1 and 7.2a, are equivalent [19].
8. The actual configuration analyzed in [8] and [13] replaced the I&D’s with distortion-free (zero ISI) lowpass
filters whose outputs were sampled once per bit. However, due to the zero ISI assumption, the error probability
performance results obtained there as a function of the lowpass filter output SNR, 7 = A?/o? apply directly to
Fig. 8.13 if 7 is replaced by E,/No.
9. There is a discrepancy between the error probability results obtained in [8] and [13]. Based on a derivation of
(8.56) supplied to the authors by Mario Pent, we have concluded that the result in [13] is indeed the correct one
and this is the one presented here in (8.56).
534
n="A
a" nat eeACA AAS
Tees: [I-12 2,12 I's,
i
"i141
140)
ele
oi
g
Co
5W
i,
bys
=u
l=W
Solo
cebay
ainsi
ZI{"g
AANA
ee
O-] Joyw[npowsp
Jo WSdzd-W (7-Wwdd)speusis pojuriday)
wor[Z] © (AAAI
Bai
535
poywtidoy)
(ASI
wosy
[¢]
©
sjeusis (7-Wd) ASdzC-W Jo Jowjnpowep O-] Jo uonRjuswedul aeusay ET" g ainsi]
-Wo TTS
bry Gy 6
tomes
‘ W
I
536 Double Differentially Coherent Communication with Waveforms Chap. 8
ae 3
5 ONT
E! E!
with E defined as before in (8.31). Upon careful examination of the mathematical model
used in arriving at (8.56) in the absence of frequency offset and the characterization of the
I&D outputs used to obtain (8.58) in the presence of frequency offset, it becomes readily
apparent that the two results can be made equivalent if in (8.56) E, is replaced by E;,, that
is,
2
3 E;, (—1)™ E\, (E\,
-=— In |—— I —— (8.59)
aad ae Ie m=0
seh Ng oN
x 9 Il2m+1/2 Ey
oar + I12m+1+1/2 ae
2No
Indeed, the equivalence between (8.58) and (8.59) can be formally derived. Thus, we con-
clude that (8.59) is also an exact result for the error probability performance of Fig. 8.12 or
Fig. 8.13 specialized to the binary (M = 2) case.
Figure 8.14 illustrates the error probability as given by either (8.56) or (8.58) for the
special case of no frequency error (E ae Ep). Also plotted in this figure for the purpose
of comparison are the corresponding results for coherent BPSK and DPSK as given by
(4.55) and (8.20) respectively. We observe that using D*PSK instead of DPSK requires
approximately 4 dB more E,/No and thus a sizable penalty is payed for using a double
differential scheme when indeed no frequency offset is present. Figure 8.15 illustrates (8.58)
or (8.59) for various normalized frequency offsets A f7). Also shown for comparison are the
corresponding results for DPSK with frequency offset as given by any of the forms (8.29)—
(8.31), (8.32), (8.34), or (8.35). Here we see that the D?PSK scheme is quite insensitive to
frequency offsets over a large range of values, whereas the DPSK performance degrades
rapidly. Thus, it is natural to ask: At what value of Af 7), is D°PSK preferable to DPSK,
that is, where do the performance curves of the two intersect? The answer to this question
was obtained in [3] and is illustrated in Fig. 8.16 where the values of Af 7) at which
P,(E)|ppsk = Pb(E)|p2psx are plotted versus E,/No in dB. The region above the curve
corresponds to the situation where D?PSK would be preferable over DPSK and vice versa
for the region below the curve. Thus, for a given E;/No, Fig. 8.16 offers one an idea of
how large Af must be relative to the data rate 1/7, before having to resort to a double
differentially coherent detection scheme such as Fig. 8.12 or 8.13.
Sec. 8.6 Serial (Absolutely Frequency Invariant) Structures 537
-4
109.49,
(E)
-5
-6
0 Pras 6 8 10 12 1416
E,/No,dB
Figure 8.14 A comparison of the error probability performances of BPSK, DPSK, and
D?PSK for the case of zero frequency offset (Reprinted from [3] © IEEE)
x |] f (i 10) ani
(el
where A”nn+1 is the n™ order differential operator applied to n, +1 as defined in (8.5) and
538 Double Differentially Coherent Communication with Waveforms Chap. 8
0.0
-0.5
: AER
\,
-1.0
(E)
10949,
-2.0
25 \
2 0 2 4 6 8 10 12
Ey/No, dB
Figure 8.15 A comparison of the error probability performances of DPSK and D°PSK
in the presence of frequency offset (Reprinted from [3] © IEEE)
10 a
1 ENS VE rae
ete, = Bes ;- —_——
F (10) 5 exp ( ae = cos (1 2 «)
ober 1-1
x exp Sate acres (8.61)
< et
|E Di)
e rer i) wi
= c0s(n
re 5 «)
is the pdf of n; given that the message selected for transmission corresponds to zero phase.
This result can be evaluated numerically for any n and reduces to (8.32) for n = 1 and (8.57)
join = 72.
0.17
2
DPSK HALF-PLANE
0.16
AfTy
0.15
0.14
0.13
DPSK HALF-PLANE
0.11 _
sf Onde 2804 56 671 38 NON AODTIHID
19! 1401546
E,/No, dB
Figure 8.16 Frequency offsets that yield P,(E)|ppsk = Pp(E)|p2psx (Reprinted from [3] © IEEE)
this metric reduced to the classical differential detection algorithm whose performance is
given by (7.14) for M = 2 or (7.7) for M > 2.
It is of interest here to extend these notions to PDM-2 signals. Since the first stage
of a PDM-2 demodulator (Figures 8.10—8.13) in effect converts the frequency error Af to
an unknown phase shift 27 Af 7, for processing by the second stage, then, in principle, one
should be able to improve performance by replacing the second stage of the detector with
a multiple symbol detection algorithm analogous to that considered for PDM-1 signals in
Chapter 7. It is to emphasize that the resulting total configuration is not truly maximum-
likelihood; however, conditioned on the first stage being implemented in the classical
way for PDM-1 signals, the remaining portion of the receiver is maximum-likelihood as
before.
Mathematically speaking, applying multiple symbol differential detection to the ACD
in Fig. 8.11, the decision rule of (8.53) generalizes as follows. We assume that the in-
put phase A76;_ 1,41 now acts as a phase reference for the sequence A?@ = (A76;_y,+2,
A76;_ jets Wesco A76;_1, A’6;). Then, second order differentially encoding these phases for
transmission over the channel, the appropriate sequence decision rule is:
where A76 is the particular vector, with components ranging over the allowable set {Bm =
27m/M;m=0,1,2,..., M — 1}, that maximizes
No k— 2 2
Ns~2 -j DD A°G-e—m
n=|Vien+i+ >) Vine (8.62b)
k=0
F A a A235 2
n Ss V5 Vie F(A +4761) Ae Vai a
ss (8.63a)
= |Vi-2l? + 1Vi-1? + Vi? + 2Re {ViVjt ge H+ "4-0
[ViiVjige 1AM +2Re {Viviaje fa)
+2Re
or, equivalently
@), V sf:
PE ae eee ( v3] E exp {~(<2) i (8.65)
No eff
where again Ns — | denotes the number of simultaneous decisions being made. As N, >
00, (8.65) approaches
| E
(Oe one
Pp(E) S eee eS {-(=) (8.66)
No }ett
which, using the asymptotic expansion of the complementary error function [see (7.87)],
541
(¢ = SN) UoNsa\ap JoquiAs o[dnynuw
yum speusis (7-Wdd) ASdzA-W JO Joye[npowlap uoHe[aUO90Ne JO WO} xgfduioy Lg ans
soe]
{“‘S}oy
St
asooyD
(i)
542 Double Differentially Coherent Communication with Waveforms Chap. 8
becomes
E
P,(E) < erfe (=) (8.67)
No eff
Thus, in the limit as the observation interval (decision memory) approaches infinity, the
large 2BT}, large SNR behavior of the performance of an ACD demodulator employing
multiple symbol differential detection of PDM-2 signals in the presence of frequency offset
approaches that of ideal coherent detection of BPSK (with differential encoding) with an
effective E,/No given by (8.64).
The above can be extended to M-ary PSK, again using the results of Chapter 7 to-
gether with (8.64). Analogous to the relation between (8.65) and (4.206) as N; + o9, it
is straightforward to show that in the limit as the observation interval (decision memory)
approaches infinity, the large 2BT), large SNR performance of an ACD demodulator em-
ploying multiple symbol differential detection of M-ary D’PSK signals in the presence of
frequency offset approaches that of ideal coherent detection of M-PSK (with differential
encoding) with an effective E,/No given in terms of E;/No by a relation identical to (8.64).
The conceptual double differential detection scheme illustrated in Figures 8.la and 8.5 can
be enhanced (i.e., its performance improved) by choosing the delay in the second stage of
the transmitter and correspondingly the delay in the first stage of the receiver to be 27, rather
than 7; (see Figures 8.18 and 8.19). The reason for doing this is to eliminate the high noise
correlation in the receiver output caused by the sharing of a common noise sample between
the two stages of the receiver as in Fig. 8.5. This can be noted in Fig. 8.19 where the receiver
output signal consists of four different noise samples each spaced T;-sec apart. Despite the
fact that the transmitter no longer implements a true second-order difference operation, the
receiver output is still independent of the frequency offset present in the received input
and as such Fig. 8.19 generically represents an implementation of an absolutely frequency
invariant demodulator.
To see the amount of improvement in performance gained by the above modification,
consider Fig. 8.12 or 8.13 (assuming binary modulation, i.e., M = 2, T; = Ty) where the
Tp-sec delays in the first stage are now 27)-sec in duration. The exact performance for this
modified implementation of Fig. 8.12 can be obtained by the method employed in [13] with
the result
Sec. 8.6 Serial (Absolutely Frequency Invariant) Structures 543
Differential Differential
Detector Detector
Pik)
Ve LE" See
2E'
a 2.8 Ge) exp ( aa
(8.68)
lo) / ! 4
(-—1)” E, E,
if — I,
ésm=0
D3 2m + 1 , (2 et (+)
Alternately, using the geometric approach taken in [3], the bit error probability can be
expressed in integral form as
x exp — Ey (2 ni + sin“
<2(sin <2) n2 + sin
Pay) ns) 1 + erf E,
ene dnidn2dn3
0 0
Figure 8.20 plots P,(E) of (8.59) and (8.68) versus E;,/No in dB. The large improvement
of the configuration using a combination of 7,-sec and 27)-sec delays over that using T;-
sec delays in both stages is clearly evident. The same kind of improvement in performance
can be obtained with an ACD, for example, modifying the autocorrelator (first stage) of
Fig. 8.11 to have a 27)-sec delay. Figure 8.21 illustrates simulation results for the error prob-
ability performance of Fig. 8.11 and its modification as per the above. Over a wide range of
values of 2BT), the advantage of a 7,, 27) delay combination is strongly demonstrated.
Finally, for M > 2, a union bound on the symbol error probability of the modified
version of Fig. 8.12 described above can be obtained in a form analogous to (8.57) with the
result
n (E‘\* PEON 2 aye E' ee 2he
Ds ASE (Balt pa a ASE ig a Og Ey s
bee 4 (5) exp ( No )2s 2m + | |‘i(A paint (ae)
Way (8.70)
, nz M-—-2
x 008 {2m 4 3 (=F )
The structures of Figures 8.18 and 8.19 are actually special cases of a form of D?PSK
544 Double Differentially Coherent Communication with Waveforms Chap. 8
Te Tips DELAY,
COMBINATION -
EQ. (8.59)
P(E)
referred to as nonadjacent D?PSK (NA-D2PSK) [22]. To see this, we recognize that, anal-
ogous to (8.3), the input, A26;, of the first differential encoder in Fig. 8.18 can be written
in terms of the transmitted phase, 6;, as AiO; = (0; — 0;-2) — (0;-1 — 9-3) = (6; — 8-1) —
(6;-2 — 6;-3) = AO; — AO;—2, which is recognized as the difference of two first-order phase
differences that are separated from each other by a single symbol interval. In the more gen-
eral case, if one replaces the 27; delays in Figures 8.18 and 8.19 by LT, delays (L > 2),
then the input, Me of the first differential encoder and the transmitted phase, 6;, are
related by AGO: = (6; — 0;-1) — (6;-L — 9;-L-1) = AO; — AG;_,. Here the two first or-
der phase differences are separated by (L — 1) symbol intervals. With reference to Fig. 8.13
ig oeriately modified by replacing the first 7,-sec delay by an LT;-sec delay), a deci-
sion on Bee ,9; would be made by comparing the phase difference AD ini= (Ni — Ni-1) —
(ni-L — Ni-L—1) = Ani — Ani—z with all possible transmitted phases around the circle and
choosing the one that is closest. Since for any L > 2, the random variables An; and Anj—
are statistically dependent, then from the standpoint of error probability performance on the
AWGN, there is nothing to be gained!? by choosing a value of L larger than 2; specifi-
cally, the bit error probability performance of NA-D?PSK is given by (8.68) or (8.69) for
all L > 2. Moreover, Qu [22] shows that asymptotically as E,/No becomes large, the dif-
ference in error probability performance between D*PSK (L = 1) and NA-D2PSK (L > 2)
10. For fast fading channels, it is shown in [22] that the correlation of the fading process can be utilized to
improve error probability performance by choosing L larger than 2. In fact, depending on the fading channel model,
e.g., satellite-aircraft, land-mobile, etc., an optimum (in the sense of minimum error probability) value of L can be
determined.
Sec. 8.6 Serial (Absolutely Frequency Invariant) Structures 545
RE)
is attributed entirely to the removal by the latter of the correlation between the two first-: .
order phase differences making up the decision variable. Specifically, the asymptotic E,/No
degradation of D?PSK relative to NA-DPSK is given by
where
py E {Ani Ani-v}
ae (8.72)
[An
is the normalized correlation between the two first order phase differences making up the
decision variable. Since for NA-D?PSK (L > 2), PAn = 0, whereas for D2PSK (L = 1)
2
E{(ni — ni-1) (ni-1 — mi-2)}_ —E{-njy} 1
eis 2a? mB gI in ae ot
then from (8.71) we find that AE,/No (dB) = 10 log; (1.5) = 1.77 dB, which is in full
agreement with the numerical results of Fig. 8.20. The asymptotic E,/No degradation of
D?PSK relative to NA-D?PSK as given by (8.71) is also valid for M > 2 [22].
546 Double Differentially Coherent Communication with Waveforms Chap. 8
switch is closed and as such the /'" ACD output, namely, a” (which represents a hard
decision on the first-order phase difference A6; = 6; — 0;—1) is passed on to the differential
decoder which performs the second stage of differential detection. Note that the range of
values for the L phase shifters in Fig. 8.22 only covers the interval (0, zr) [rather than (0, 277)
as one might first expect]. The reason for this is that a frequency error-induced phase shift
of @ in the region (zr, 277) will result in a 180° phase shift at the I&D output of the ACD
“matched” to the phase shift a — 2. However, since the “matched” ACD is determined
by the square of the I&D output, the ACD corresponding to the phase shift a — 7 will
ideally be selected. Furthermore, the 180° phase shift at the I&D output of this “matched”
ACD, which results in an inversion of the hard decisions at the “sgn” device output, will be
resolved by the differential decoder.
Intuitively, one would anticipate that the above scheme would yield an improvement
in noise immunity over the absolutely invariant type of ACD since: (1) first stage demodula-
tion is achieved by only a single ACD that is “matched” (within the resolution of the number
of ACD’s used in the bank) to the input frequency error; and (2) the second phase difference
is calculated from the adjacent hard decision outputs of the first stage as is done in coherent
detection with differentially encoded input symbols. As such, we would expect that as the
number, L, of ACD’s in the bank increases and the probability of correctly selecting the one
that matches the input approaches unity, the noise immunity of Fig. 8.22 should approach
(within approximately a factor of two due to the additional differential encoding operation)
that of an ACD of PDM-1 signals with no frequency error (Fig. 8.6). Practically speaking, it
has been shown [7] that a bank of L = 6 ACD’s gives performance quite close to the above
limit. This and other conclusions will be demonstrated by the results of the analysis that
follows.
Assume that the signal of (8.12) is received (in noise) where 9; is now the double
differentially encoded information phase in the i transmission interval. The output of the
multipler in the /th ACD is (in the absence of noise)
e)(t) = A* cos (AwT) + 6; — 6;—1 — 61) = A’ cos (A6;) cos (aw — Bi) (8.74)
11. Equally, one could replace the square-law devices with absolute-value devices (full-wave rectifiers). The
choice between the two would depend on the SNR as is typical in situations of this type.
547
s[eusis (Z-Wdd) 3Sdzd-W $0 Jowjnpowsep uonejasosone oseydnjnu y 77"g aaNSLy
,S09
‘@V
1
548 Double Differentially Coherent Communication with Waveforms Chap. 8
where we have again set « = Aw7}, and recognized the fact that for binary modulation,
Ad; = 6; — 9; only takes on values 0 and 7. After passing through the I&D, we get the
(noise-free) decision variable
V.° = AT; cos (AG;) cos (a — By) = Epcos (A6;) cos 1 Gre)
where ©; = a — f; denotes the phase error associated with the mismatch of the i“ ACD
to the input signal. Assuming that based on a comparison of the squared values of Vo the
correct ACD (the jth corresponding to the one whose value of ; is minimum) is always
chosen,!* then the hard decision outputs from the correct ACD are described by
where Ne? is a noise variable that accounts for S x N and N x N degradations. More
important, however, is the fact that the correct choice among L ACD’s reduces the range
of the phase error, ®;, to the interval (—2/2L,7/2L). As such, the 1&D output signal for
the correctly selected ACD will dip by no more than a factor of cos 1/2L relative to that of
the same ACD of PDM-1 signals in the absence of frequency error. For example, for L = 6,
this dip is less than 4%.
The error probability associated with the hard decisions a is obtained by noting that
(8.76) represents the decisions in Fig. 8.6 if the energy E, is replaced by Ep cos ®;. As such,
this error probability, now denoted by P,)(E |®;) to indicate its dependence on ®;, is given
by (8.19) with E,/No replaced by (E,/No) cos? ®;. For example, for the special case of
2 BT, = 2, we would have
1 Ja es 1Ep 4
Poi(E |®;) = 5 XP (-2 Cos ®i) af aN cos” ®; (8.77)
The relation between the error probability, P,;(E |®; ), associated with the hard de-
cisions d; on the first-order phase difference A@; and the error probability, P,(E |®;),
associated with the ultimate hard decisions on the second-order phase difference A206; =
AG;
— AG;—1 18
Pp(E |) = 2Ppi(E
|®;) (1— Poi
(E|®;)) (8.78)
Finally, the average probability of error of the receiver of Fig. 8.22 is given by
n/2L
P,(E) = / —7/2L
Pr(@;) f(@))d®; (8.79)
where f(®;) is the pdf of De.
Assuming that the frequency error is completely unknown and as such the equivalent
phase a = AwT) is uniformly distributed in (—z, 7), then the pdf of ®; is given by
IE 4 4
@:) led et Opa
f(%)) {a otherwise Sey
12. Practical conditions under which this assumption is valid are discussed in [4, 7].
Sec. 8.7 Parallel Structures—Multiphase ACD 549
RE)
E,/N,, dB
Figure 8.23 Bit error probability performance of the multiphase ACD in Figure 8.22
with the number of ACD’s in the bank as a parameter (Reprinted from [2] © IEEE)
Thus, for example, the average error probability performance of the receiver of Fig. 8.22 for
2BT} = 2 would be found from!
A ee Ep ts a et
P,(E) = — exp cee ®; MN ees ®;
HM J—n/2L
(8.81)
x Wes
— —exp |——Sree cos* co 9; lie
——Eeeeeaae
cos* el
®; aw ‘
2 P No l 4 No l l
13. References [4, 7] incorrectly evaluate this error probability since they first average (8.77) over the pdf of ©;
and then use a relation analogous to (8.78) to convert this result to the final error probability. Also, Figure 1 in [7]
is incorrectly drawn. Figure 4.14 in [4] is the correct figure.
550 Double Differentially Coherent Communication with Waveforms Chap. 8
this simulation for the case 2B7), = 2 and L = 6 lie within 0.1 dB of the corresponding
theoretical curve in Fig. 8.23.
Finally, we point out that the configuration of Fig. 8.22 can be clearly generalized to
apply to M-DPSK (PDM-2) signals with M > 2 in the same manner that Fig. 8.6 generalizes
to the first stage of Fig. 8.11. The details are left to the reader.
PROBLEMS
8.1 A binary phase information source whose phases take on values (0, 77) is to be communicated
using a second-order phase difference modulation (PDM-2).
(a) If the input information sequence {A76;} is 2,7, 7,0, 0, 7,0, 7, find the sequence of
complex baseband signals {e/%} transmitted in the intervals (0, 7;), (Ts, 2Ts), -.., (9Ts,
107,). Assume that the first two transmitted phases are given by 6) = 6; = 7 and since the
source is binary, 7; = Tp.
(b) If now the information source is 4-ary with phases that take on values (0, 1/2, 7, 32/2),
repeat part (a) for the information sequence 1/2, 7, 2, 0, 32/2, 1/2, 0, 37/2.
8.2 A binary phase information source whose phases take on values (0, 77) is to be communicated
using an n'*-order phase difference modulation (PDM-n). Show that the appropriate encoding
scheme relating the information phases {A”6;} to the transmitted phases {6;} is given by
APO iO) = On
8.3 A ternary phase information source whose phases take on values (0, 27/3, 47/3) is to be
communicated using a third-order phase difference modulation (PDM-3).
(a) Show that the appropriate encoding scheme relating the information phases {A”6;} to the
transmitted phases {6;} is given by
A36; = 6; — 6-3
(b) For the ternary source, does the relation found in Problem 8.2 also hold, in general, for
values of n greater than 3? Explain.
8.4 A binary amplitude (+1) information source is to be communicated using a second-order
phase difference modulation (PDM-2). Show that the two encoders illustrated in Fig. P8.4
are equivalent, that is, show that the relation between the information sequence {a;} and the
transmitted sequence {c;} is the same for both.
8.5 Consider the communication system of Problem 8.4 wherein the data source {a;} is char-
acterized as an independent, identically distributed (i.i.d.) sequence with Pr {a; = —1} =
Figure P8.4
Problems 551
Figure P8.6
p, Pr{a;=1}=1-— p. Assuming that the first two bits in the output sequence are co =
1, cCjh= ie
(a) Evaluate the statistical mean of the output sequence {c;} for all i.
(b) Evaluate the autocorrelation function of the output sequence R,(m) A E {cjcj+m}. Is the
sequence te;} in general, wide sense stationary (WSS)? Explain.
(c) If p = 1/2, that is, {a;} is a purely random data sequence, is the sequence {c;} WSS? Is it
an uncorrelated sequence, that is, R. {m} = 0 form #40?
8.6 A binary amplitude (+1) information source is to be communicated using a nonadjacent
second-order phase difference modulation (NA-PDM-2). The encoder for such a system is
illustrated in Fig. 8.18 and is repeated in Fig. P8.6 for convenience. The data source {a;} is
characterized as an independent, identically distributed (i.i.d.) sequence with Pr {a; = —1} =
p, Pr{a; =1}=1-— p. Assuming that the first three bits in the output sequence are co =
jig (5) AN O=l,
(a) Evaluate the statistical mean of the output sequence {c;} for all i.
(b) Evaluate the autocorrelation function of the output sequence R,(m) A E {cjcj+m}. Is the
sequence {c;} in general, wide sense stationary (WSS)? Explain.
(c) If p = 1/2, that is, {a;} is a purely random data sequence, is the sequence {c;} WSS? Is it
an uncorrelated sequence, that is, R, {m} = 0 form 4 0?
8.7 A binary phase information source whose phases take on values (0, 7) is to be communicated
using a second-order phase difference modulation (PDM-2). Based on Equations (8.3) and
(8.4), an equivalent transmitter implementation is obtained by replacing the pair of back-to-
back differential encoders in Fig. 8.l1a each having delay 7, by a single differential encoder
having delay 27, (also see Problem 8.4). Is it possible to make the same type of replacement
in the receiver of Fig. 8.5, that is, can the two differential detectors each having delay 7, be
replaced by a single differential detector having delay 27,;? Answer the question separately for
the case of Af = 0 and Af #0.
8.8 Consider a differential PSK system with I-Q detection that transmits an all-zero phase angle
sequence of length five, that is, (9 = 0; = 62 = 03 = 64 = 0. Assume that there is a channel fre-
quency offset of Af 7, = 1/87), and a channel phase offset 6 = 7/5. Also assume for simplicity
that there is no additive noise.
(a) Find the sequence of measured angles {n;};i =0,1,..., 4 seen by the phase measurement
portion of the I-Q detector as shown in Figures 7.1 and 8.13.
(b) If the system under consideration is binary DPSK (Fig. 7.1), find the receiver’s estimate of
the sequence of first-order differences {w; = An;} fori = 1,..., 4. Also find the receiver’s
estimate of the sequence of data information angles aéi| LOT ers 4. (Assume that
4 = 0 is known.) How large can 6 and Af be before the first-order differential decoding
results in an error in A6;?
(c — If the system under consideration is binary D*PSK (Fig. 8.13), find the receiver’s estimate
of the sequence of second-order differences {A?n;} for i = 2, 3, 4. Also find the receiver's
estimate of the sequence of data information angles {476} for i = 2, 3, 4. (Assume that
> = 6; = 0 is known.) How large can 6 and Af be before the second-order differential
decoding results in an error in A6;?
8.9 Suppose that signals s2(t), s3(t), and s4(t) as given by Eq. (8.12) are transmitted and received
eee ee eee rr eee Meee re hl ete ee ee IDO “at waalacae tlihactentad in
552 Double Differentially Coherent Communication with Waveforms Chap. 8
Fig. 8.10. Trace the signals through the receiver block diagram to find the summer output (input
to the signum operator) V...4V¢.3 + Vs.4Vs.3. Assuming that the transmitted phase angles for all
three signals have been second-order differentially encoded, explain how the signum operator
can be used to make the decision for the binary signal alphabet (0, 77).
8.10 For a D*PSK system, suppose that, in addition to the frequency offset Af and phase offset
6, there is a higher-order offset between the received signal carrier and the receiver refer-
ence signals given by at?, that is, the complex channel rotation shown in Fig. 8.2 becomes
exp{j[0+22Aft; + at?]}.
(a) In the absence of noise, find the output of the summer (input to the signum operator)
Voi Ve.i-1 + VsiVs,;-1 in the ACD of Fig. 8.10. Also find A?n; in the I-Q demodulator
of Fig. 8.13.
(b) Design a third-order differential encoder and a corresponding demodulator with third-order
differential decoding that would, in the absence of noise, eliminate the effect of this higher-
order offset.
8.11 It is desired to evaluate the channel capacity (see Chapter 1) of a second-order phase difference
modulation (PDM-2) system using a transmitter as in Fig. 8.la, that is, a double differential
encoder, and an I-Q receiver as in Fig. 8.13. Such a system can be modeled as an M-ary
symmetric channel (see Chapter 3) for which the channel capacity (in bps) is given by
M-1 , M-1
os rr D> Pen (E |m) logs Pen (E |m) + logyM
m=0 n=0
where P,,, (E |m) is the conditional probability of choosing the n'" symbol (phase) when the
m" symbol (phase) was transmitted, that is
i Qn
Pn (E \m) & Pr {A76; = By esis Bn = =
Note that Ps, (E |m) is actually the conditional probability of a correct decision given that the
mh symbol was transmitted. Since the £;’s are equiprobable, P;,, (E |m) is also the average
probability of correct symbol detection.
With reference to Figures 8.la and 8.13, the pdf of the second-order phase difference
Ax; & A?n; — A?6; (see Problem 8.12) is denoted by Fax; (Axi), which is independent of i
and the transmitted information phase A’6;, and is distributed in the interval —7 5S ANG Sok
(a) Evaluate the channel capacity of the PDM-2 system. Express your answer in terms of the
set of probabilities
(21I-M+1)2/M
ee) idx CAN Ne ene ee M-1
(QI—M—1)x/M
—(M—1)x/M +
Po =i)
=
fax, (Axi) dAXi +f fax; (Axi) dAxXi
(M—1)x/M
(b) Suppose now that we have a PDM-1 system using a single differential encoder as a trans-
mitter and a receiver as in Fig. 8.8 (or Fig. 7.1). Here, the conditional pdf of x; 4 An; —
A6;, which is denoted by fx, (x;), is independent of i and the transmitted information phase
A6;, and is distributed in the interval —z < x; < 2. Also, denote the probabilities corre-
sponding to the P;’s of part (a) by
(2l-M+1)x/M
a= | PaCOehden US MNees se. 54 ll
Ql-M-1)x/M
ag
OV= fy. (x) dx; + [ fy. (y;) dy;
References 553
In principle, how would your answer found in part (a) change, if at all; that is, express the
channel capacity in terms of the Q)’s.
8.12 Consider a second-order phase difference modulation (PDM-2) system composed of the trans-
mitter of Fig. 8.la and the I-Q receiver of Fig. 8.13. The pdf of the second-order phase
difference Ax; 4 A?n; — A6; is denoted by fax, (Axi) which is independent of i and
the transmitted information phase A*6; , and is distributed in the interval —7 SING GS He,
In terms of the equivalent first-order phase differences x; 4 An; — AO;, we have Ax; 4
An; — Anj—; — 46; — AO;-; = xi — xi-1. The pdf of Ax; can, in principle, be obtained if
———P3 OO”
A?ni A6;
the joint pdf of x,;-; and x;, namely fx,_,.x;(Xi-1, Xi), is available. In particular
8
Fax(Axi) = i at at AX ail
Say
Although the joint pdf fx,_,,x;(Xi-1, Xi) can be found and then f,x,(Ax;) evaluated as above,
the resulting expression becomes quite complicated and is in the form of multiple integrals.
In the high SNR case, the matter can be simplified by approximating fx,_,x,(Xi-1, Xi) by a
two-dimensional Gaussian form. In particular
where p = —1/2 is the correlation between the first-order phase differences in the two adjacent
symbol intervals and o* = No/ST; = R~' denotes the inverse of the symbol SNR.
(a) Show that the second-order phase difference approximate (high SNR) pdf is given by
Ffax(Axi) = Sa
ZONE
|
Ot a
= (Ax; + 2am)?
20
| UE SING GSE
Hint: In the case where x;_; and x; are not restricted to values in the interval (—z, z)
as in the approximate pdf above, one can first compute the pdf of Ax;g 4 x; — x;-1, the
“actual” phase difference without the modulo 27 condition by applying the above integral
with limits —oo and oo instead of —z and z. Then, the pdf of the modulo 27 reduced Ay;
is given by
AROS De FAXig
(AXi + 20m)
m>=—CO
REFERENCES
i Divsalar, D. and M. K. Simon, “Doppler-Corrected Differential Detection of MPSK,” JEEE
Transactions on Communications, vol. 37, no, 2, February 1989, pp. 99-109.
. Divsalar, D. and M. K. Simon, “On the Implementation and Performance of Single and Double
Differential Detection,’ IEEE Transactions on Communications, vol. 40, no. 2, February 1992,
pp. 278-91.
. VanAlphen, D. K. and W. C. Lindsey, “Higher-Order Differential Phase Shift Keyed Modulation,”
to appear in the JEEE Transactions on Communications.
. Okuney, Yu. B., Theory of Phase-Difference Modulation, Moscow: Svyaz Press, 1979.
. Fink, L. M. , Theory of Digital Data Transmission, Moscow: Sovetskoye Radio Press, 1970.
554 Double Differentially Coherent Communication with Waveforms Chap. 8
6. Korzhik, V. I., L. M. Fink and K. N. Shchelkunov, Calculation of the Noise Immunity of Digital
Data Transmission Systems (A Handbook), Moscow: Radio i Svyaz Press, 1981.
. Okunev, Yu. B., V. A. Pisarev and V. K. Reshemkin, “The Design and Noise-Immunity of Mul-
tiphase Autocorrelation Demodulators of Second-Order DPSK Signals,” Radiotekhnika, vol. 34,
no. 6, 1979 [Telecomm. Radio Engng., Part 2, vol. 34, no. 6, 1979, pp. 60-63].
. Okunev, Yu. B. and L. M. Fink, “Noise Immunity of Various Receiving Methods. for Binary
Systems with Second-Order Phase-Difference Modulation,” Radiotekhnika, vol. 39, no. 8, 1984
[Telecomm. Radio Engng., vol. 39, no. 8, 1984, pp. 51-56].
. Okunevy, Yu. B., N. M. Sidorov and L. M. Fink, “Noise Immunity of Incoherent Reception with
Single Phase-Difference Modulation,” Radiotekhnika, no. 11, 1985 [Telecomm. Radio Engng.,
no. 11, 1985, pp. 103-106.
10. Okunev, Yu. B. and N. M. Sidorov, “Noise Immunity of a Carrier-Frequency-Invariant Demodu-
lator of DPSK-2 Signals,” Radiotekhnika, no. 6, 1986 [Telecomm. Radio Engng., no. 6, 1986, pp.
81-83].
iit, Gut, R. E., “Noise Immunity of Autocorrelation Reception of Single PSK Signals,” Radiotekhni-
ka, no. 10, 1972 [Telecomm. Radio Engng., no. 10, 1972, pp. 120-122].
. Gut, R. E., Yu. B. Okuney and N. M. Sidorov, “Effect of Carrier Detuning on the Noise Immunity
of Autocorrelation Reception of DPSK Signals,’ Radiotekhnika, no. 7, 1986 [Telecomm. Radio
Engng., no. 7, 1986, pp. 96-98].
13. Pent, M., “Double Differential PSK Scheme in the Presence of Doppler Shift,’ Digital Commu-
nications in Avionics, AGARD Proceedings no. 239, 1978, pp. 43-1-43-11.
14. Stein, S., “Unified Analysis of Certain Coherent and Noncoherent Binary Communication Sys-
tems,” [EEE Transactions on Information Theory, January 1964, pp. 43-51.
. Henry, J. C., Ill, “DPSK versus FSK with Frequency Uncertainty,’ JEEE Transactions on Com-
munications Technology, vol. COM-18, December 1970, pp. 814-817.
. Miller, T. W., “Imperfect Differential Detection of a Biphase Modulated Signal—An Experi-
mental and Analytical Study,’ Ohio State University, Technical Report No. RADC-TR-72-16,
February 1972.
. Divsalar, D. and M. Shahshahani, “The Bit Error Rate (BER) of DPSK with Differential Detector
(DD) Followed by Matched Filter (MF),” private correspondence.
. Pawula, R. F., S. O. Rice and J. H. Roberts, “Distribution of the Phase Angle Between the Vectors
Perturbed by Gaussian Noise,” JEEE Transactions on Communications, vol. COM-30, no. 8, pp.
1828-1841.
. Simon, M. K., “Comments on ‘On Binary DPSK Detection’,” JEEE Transactions on Communi-
cations, vol. COM-26, no. 10, October 1978, pp. 1477-1478.
20. Arthurs, E. and H. Dym, “On the Optimum Detection of Digital Signals in the Presence of White
Gaussian Noise—A Geometric Interpretation and a Study of Three Basic Transmission Systems,”
IRE Transactions on Communications Systems, December 1962, pp. 336-372.
Ne Divsalar, D. and M. K. Simon, “ Multiple Symbol Differential Detection of MPSK,” JEEE Trans-
actions on Communications, vol. 38, no. 3, March 1990, pp. 300-308. Also see “Multiple Symbol
Differential Detection of Uncoded and Trellis Coded MPSK,” Pasadena, CA: Jet Propulsion Lab-
oratory, JPL Publication 89-38, November 15, 1989.
yp), Qu, S., “New Modulation Schemes: Principles and Performance Analyses,” Ph.D. Dissertation,
Technical University of Nova Scotia, 1992. Also see S. Qu and S. Fleisher, “Double Differential
MPSK on the Fast Rician Fading Channel,” submitted for publication in the IEEE Transactions
on Communications.
Chapter 9
Communication
over Bandlimited Channels
Except in a few instances, all of the communication scenarios we have discussed thus far
have corresponded to an AWGN channel with unconstrained bandwidth. As such, the only
effect the channel was assumed to have on the transmitted signal (aside from the addition
of Gaussian noise) was attenuation of its amplitude and the addition of a phase shift to the
carrier, the latter being assumed constant in time over the duration of the data symbol or
longer. In reality, the communication channel is never infinite in bandwidth; however, in
certain applications, it is considerably wideband when compared to the bandwidth of the
modulation being communicated over it. Thus, for all practical purposes, one can in these
situations still model the channel’s effect on the transmitted waveform as multiplication by
a fixed gain (actually an attenuation) and the addition of a constant carrier phase shift. Al-
ternately, the equivalent baseband signal when passed through the channel is multiplied by
a fixed gain and delayed by an amount that is constant for all frequencies in the waveform.
For all practical purposes, this is the case that we have treated thus far in the book.
In this chapter, we discuss the design of modulation/demodulation schemes for send-
ing digital information over a bandlimited channel whose bandwidth is on the same order as
that of the modulation. Our treatment will be brief by comparison with other fine textbooks
devoted to this subject [1, 2, 3], and will only focus on the key issues. We shall consider
only linear channels, that is, those modeled as a linear filter with bandpass transfer function
F.( f), and linear modulations of the type corresponding to a baseband pulse stream
modulated onto a carrier where A is the signal amplitude and c; = a; + jb; is, in general, a
complex discrete random variable representing the data symbol to be communicated in the
i transmission interval. The key issue here is the design of the pulse shape p(t) to allow
communication at the highest possible data rate while achieving the best trade-off between
signal power and noise power. As we shall see, this pulse shape design is dictated by the
characteristics of F.(f) over its bandwidth occupancy. In particular, if F.(f) is an ideal
(rectangular transfer characteristic) filter of IF bandwidth W Hz (lowpass bandwidth B =
W/2), then, at least theoretically, p(t) can be designed to allow distortion-free transmission
at a rate equal to W bps with an SNR (E»/No) equivalent to that of the unconstrained
bandwidth channel. If the channel is not ideal over its bandwidth occupancy, then signaling
556 Communication over Bandlimited Channels Chap. 9
channel characteristics. As such, the process is called adaptive equalization. A large body
of literature has been developed that is devoted to the design and convergence of algorithms
for adjusting the equalizer’s parameters. Since this is not the major focus of this textbook,
we shall be relatively brief in the coverage of this subject.
We begin the chapter with a brief characterization of what is meant by a bandlimited
channel and a practical application where it is encountered.
= T)dt =) P(f)Ge(fye!*f'df
p(t) =! P(T)gclt
65 5x
p(t) = Gop(t — Ta). Note that since P(f) is bandlimited, p(t) cannot be timelimited (the
uncertainty principle [8] prevents a signal from being limited in both time and frequency)
and likewise for p’(t). Thus, the pulses in m’/(t) must, by necessity, overlap. This does not,
however, mean that ISI is present at the receiver since by proper design of p(t), it may still
be possible to separate out the various pulses in the data stream. What is important here is
that if p(t) is designed to produce zero ISI at the receiver, then the distortion-free channel
above does not change that condition, that is, p’(t) also yields zero ISI.
In the more general case where |G-(f)| is not constant and 6,(f) is not a linear func-
tion of frequency over (—B, B), the channel will introduce amplitude and phase distortion
into the transmitted pulse shape. It is common to define the group delay, Ty(f), by the neg-
ative of a scaled derivative of the channel phase characteristic, 6.(f), that is,
Lf) =
1 d6.(f) (9.4)
DT ats
This definition is consistent with the above discussion for the distortion-free channel, where-
upon the group delay becomes a constant (fixed) delay, Ty, independent of frequency. For
the channel that introduces phase distortion, Tg(f) will be a function of frequency. The ef-
fect of amplitude and phase distortion on a transmitted pulse stream modulation, m(f), is to
smear the pulses in the received modulation, m’(t), so that they cannot be separated by the
receiver. It is indeed this additional smearing of the pulses that accounts for what is referred
to as intersymbol interference.
Perhaps the most common bandlimited channel that, in addition to many other impair-
ments (e.g., impulse noise, phase jitter, frequency offset, etc.), exhibits both amplitude and
phase distortion over its frequency range of definition is the telephone channel, whose us-
able band is typically in the range of 300 Hz to 3 KHz. A detailed discussion and illustration
of the telephone channel characteristics are given in [2]. As an example, an average medium
range telephone channel of the switched telecommunications network [9] has an impulse re-
sponse of about 10 ms; thus, for communication over this channel at a rate on the order of
2.5 Ksps (a symbol duration of 0.4 ms), the ISI would extend over 25 symbols.
1. Since the noise process 7(t) is, in general, complex, its PSD is defined as Sjax(f) 4 FE {a(t)a*(t + T)}}.
However, to keep the notation simple, we shall define it with only a single subscript, namely, S;(f), as is done for
the PSD’s of real random processes.
=()u
v (L1-)d'2"0
oo-=t
(1)u
=
dSd
(f)"s
1+ "74 =11e gJdues
aang
[6 WY YOOIG WeIBeIP
JO& PoyUUI[pueQ UONROTUNWIWIOS
W9}SAS
559
560 Communication over Bandlimited Channels Chap. 9
is to design G;(f) and G,(f) [for a given G,(f)] to minimize the error probability of the
receiver.”
The output y(t) of the receive filter can be expressed as
where fo is selected to match the delay of the channel (more about this in a moment). In
shorthand notation, (9.6) can be rewritten as
[o,@) CO
a! E (cleat » a Ago
P,(E) = 5)erfc [ital = 3 erfc |
ae (9.8)
2. Although, in a real communication system, the modulation pulse shaping and transmit filtering are separate
entities, from a mathematical standpoint, their cascaded effect on the transmitted waveform is represented by the
product of their respective transfer functions P(f) and G;(f). Thus, for analysis purposes, we shall denote the
product P(f)G;,(f) by simply G;(f) with the understanding that in the physical system G;(f) must be partitioned
between the modulation pulse s..ape and the transmit filter.
Sec. 9.2 Optimum Pulse Shape Design for Digital Signaling 561
Also, as noted above, the desired signal sample go depends on G;(f), G-(f), and G;(f),
ViZ.
We will return to a consideration of the selection of G;(f) and G,(f) to minimize P(E) of
(9.8), that is, maximize go/oy, after first deriving the criteria necessary to achieve zero ISI,
that is, the Nyquist criterion. For simplicity, we shall let go = 1 since an arbitrary gain can
be included by scaling G(f) later on. Also, we shall continue to consider the case of real
(but not necessarily binary) data, that is, ¢; = a;, until further notice.
We start out by recalling the well-known sampling theorem, namely, if X (f) = F {x (t)} is
bandlimited to (—B, B), then x(t) is uniquely determined by samples taken at 1/2B, viz.
0; otherwise
The lowpass bandwidth B is referred to as the Nyquist bandwidth and is equal to one-half
the sampling rate.
From (9.7), for zero ISI, we desire a pulse waveform g(t) such that its samples taken
at a rate 1/7; (the data symbol rate) satisfy?
We can rewrite the above integral in terms of a partition of adjacent Nyquist bandwidths of
width 2B = 1/T,, viz.
oo (2k+1)/2T; es
&n= i G( fel" df =, (9.14)
k=—o0o (2k—1)/2Ts
3. For the purpose of the discussion here, we shall assume without any loss in generality that the sampling epoch
to equals zero. Whatever conditions we determine on the overall pulse transfer function G(f) can later be modified
by replacing G(f) with G(f)e/?7/".
562 Communication over Bandlimited Channels Chap. 9
W275
2n= = /; G (vay =)el 2A Utk/Ts)nTs qyy
a er 3
ine) 1/2T; k ; x
=e i G €+ =)eltmunts dy (9.15)
pene ek 8
1/21; [ & Peer
=i) ) G (vSe =) eltmunts dy
ml TN pee Ts
Define the equivalent Nyquist channel characteristic
that is, the sum of all of the translates of G(f) folded into the interval (—1/2T7;, 1/2T;).
Substituting (9.16) into (9.15) gives
1/2T,
8n = i Galpe ay (9.17)
—1/2T;
But the inverse Fourier transform of Ge,(f) is by definition
00 1/2T, 4
8eq(t) = / Gel felt dfi= Geg(fyel*"f'df (9.18)
=e0 =W2re
with samples
1/2T;
8eqn = if GH (pe ar Pap (9.19)
S/T:
Thus, from (9.18) and (9.19), we see that the Nyquist rate samples of g(t) are also the
Nyquist rate samples of geq(t), that is, g¢g, = &n-
Since, by the definition of (9.16), Geg(f) is a strictly bandlimited function on the
interval (—1/2T7;, 1/2T7;), it can be expressed in the form of (9.12), namely
However, for zero ISI we require that g, satisfy (9.13). Thus, using (9.13) in (9.20) gives
= k 1
oy o(s+5) =n: lar (9.22)
k=—0o
Sec. 9.2 Optimum Pulse Shape Design for Digital Signaling 563
that is, the superposition of all of the translates of G(f) in the interval (—1/2T,, 1/2T;)
must yield a flat response.
It can also be shown (Problem 9.1) that the superposition of all of the translates of
G(f) in the interval ((2n — 1) /27; < f < (2n + 1) /2T;) must also yield a flat response
for any integer n. Thus, the zero ISI condition on G(f) as specified by (9.22) generalizes to
Note that the criterion for zero ISI does not uniquely specify the pulse spectrum G(f)
unless its bandwidth is limited to (—1/2T7;, 1/27;), in which case it must be flat since the
sum in (9.23) reduces to one term (k = 0). In this special case, G(f) itself has a rectangular
transfer function in the interval (—1/27;, 1/2T;), or equivalently
sin wt /T;
(9.24)
g(t) sca aiidbk
The interpretation of this result is that for a digital modulation of the form in (9.1), the
minimum one-sided lowpass bandwidth needed to transmit R; = 1/7; symbols per second
(sps) without ISI, is R;/2 = 1/27; Hz. When the data symbols {a,} are independent and the
noise samples (spaced 7;-sec apart) are uncorrelated, then each symbol can be recovered
without resorting to past history of the waveform, that is, a zero memory system.
Since in the above system R; sps are transmitted without ISI over a bandwidth R,/2
Hz, then the throughput (efficiency of transmission) is Rs sps/(R;/2) Hz = 2 sps/Hz. To
achieve this efficiency, one must generate the sin x/x pulse shape in (9.24). Unfortunately,
this pulse shape is impractical because its very slowly decreasing tail will cause excessive
ISI if any perturbations from the ideal sampling instants occur.
To reduce this sensitivity to timing (sampling instant) offset and at the same time
ease the filter design problem in the transmitter and receiver, we use more practical shapes
for g(t) whose Fourier transform G(f) has smoother transition at the band edges yet still
satisfies the zero ISI criterion of (9.23). One such class of channels that has achieved a
good deal of practical application is the set of raised-cosine channels defined by the overall
transmission characteristic
where 0 < a@ <1 is defined as the excess bandwidth factor. The special case of a = 0
corresponds to the minimum Nyquist bandwidth brick wall (rectangular) filter as discussed
above. Values of a > 0 trade increased bandwidth (and hence reduced throughput for a fixed
data rate) for a reduction in the rate of roll-off of the frequency response. In fact, for the class
of systems defined by (9.25), the throughput is reduced to Rs/[(Rs/2)(1 + @)] =2/(. + @)
sps/Hz. Figure 9.2a is a plot of G(f) for several values of a. The corresponding overall
system impulse response is
iG (Se) ( cosamt/Ts ) (9.26)
Ae L=—\(2at/ T.)7
———— SS
go(t) 8a(t)
564 Communication over Bandlimited Channels Chap. 9
1/2T, VTs
(b)
1.0 sin (xV/T,) cos (zat/T;)
(nT) 1 —(2a/T,)?
Figure 9.2 The raised cosine pulse: (a) frequency function, (b) time function
which is illustrated in Fig. 9.2b for the same values of a as in Fig. 9.2a. For a = 0, we get
the time response for the minimum bandwidth brick wall filter as given in (9.24). For values
of a > 0, the tails of the response decay more rapidly than for the sin x /x response of (9.24).
Note that (9.26) is written in the form of a product of go(t) (the overall system impulse
response corresponding to the minimum bandwidth brick wall Nyquist filter) and another
response, ga(t). In fact, (9.26) can be generalized to include any function gg(t) that is
bounded at the zero crossings of go(t) since under this condition, the zero crossings of g(t)
will be the combined zero crossings of go(t) and gq(t) and thus the zero ISI condition is
maintained. Stated in the frequency domain, the convolution of the minimum bandwidth
brick wall filter Go( f) = F {go(t)} and an arbitrary filter Gg(f) whose inverse Fourier
transform gq(t) is bounded at the zero crossings of go(t) is a Nyquist filter.
Having just seen how to design the overall system frequency response G(f) to achieve
zero ISI, the next issue concerns how to optimally apportion this response between the
transmitter and receiver. By optimum, we mean that the apportionment should result in
Sec. 9.2 Optimum Pulse Shape Design for Digital Signaling 565
minimum error probability or equivalently maximum detected sample SNR. From (9.8), we
see that this is equivalent to maximizing the ratio*
CONE oe
ys |ee
ace ( Sa(f) Lane
G ee ent ee Soe
eee a
Pie ( KIGAA)F
Since S,(f) is real, then we must have arg {G(f)} = e~/7ft in order for (9.29) to have
any meaning. Moreover, since go = 1, then from (9.27)-(9.29), the constant is evaluated as
a
K=(f°2, IGP PIGCAPMSaAaf) .
If we now assume that n(t) corresponds to white Gaussian noise with PSD S,(f) =
No/2, then we get
V2
G,(f)|
IG-(f)| = (2K ee — ieee!
| eG ryiGu Ald
IG(f)|
No
(9.30)
1/2 Pe 1/2
IGe(AIIGAA)| = (pre i (in | GNP GAP af)
which implies that, except for a scaling constant, the overall pulse response should be
split equally between the combined transmit/channel and receive filters. Furthermore, the
maximum SNR achieved by this apportionment is from (9.27) given by
where P, is the signal power at the input of the receiver. For binary amplitude modulation,
4. Note that for multilevel (M-ary) amplitude modulation, the symbol error probability is given by P,(E) =
aot erfe V Wa
M=1 “|
V0, (see Chapter 4) and thus the optimum apportionment is still obtained by minimizing
(9.27).
566 Communication over Bandlimited Channels Chap. 9
1 heed i
P»(E)|min = =5 erfe ,/ No
— 9.32
(9.32)
where we have also substituted 7, for T,. Comparing (9.32) with the comparable result for
the unrestricted bandwidth case (that is, rectangular pulses transmitted over an infinite band-
width channel) as discussed in Chapter 4, we conclude that the two are completely equiv-
alent if we associate P,7; with the energy per bit Ey. For M-level amplitude modulation
with a Gray code bit-to-symbol mapping, the result would be approximately given by (9.32)
with a factor of ,/(M2 — 1)/3 included in the argument of the erfc function. Similarly, the
throughput for M-level raised cosine channels would now be 2 log, M/(1 + @) bps/Hz.
It is of interest now to compare the throughput of the various Nyquist systems as
a function of the SNR in the Nyquist bandwidth, P,/NoB required to achieve, say, a
fixed symbol error rate equal to 10~>. For M-ary amplitude modulation, we have approx-
imately (ignoring the factor of (M — 1)/M in front of the erfc function) that [3/(M@ Pi
1)] P,T;/No = 9.12 (9.6 dB). Thus
P, Pets R; Pare Rp 1
NoB =( No )e 3 (No ee log,M
(9:33)
mmc (Mirek) (M*=1) 18.24
3 (lta) 3 (l+qa)
Figure 9.3 is a plot of throughput R,/B = 2log, M/(1 + a) bps/Hz versus SNR in the
Nyquist bandwidth, P,/NoB for Ps(E) = 10-> with a and M as parameters. Also indicated
on this plot is the throughput corresponding to channel capacity [see Chapter 1, Eq. 1.30)],
which is repeated here (letting C = Rp) as
Rp | PS
p 7 082 (:te i) (9.34)
A SSAA
AN)
EETSAAAATD gO en ARTGSASEEEFTTT
DISTORTION OF
SENSITIVITY TO ZERO CROSSINGS
TIMING ERROR NOISE MARGIN
DISTORTION AT
SAMPLING TIME
raised cosine filter, but the required P,/NoB would now be 18.24(5/2) = 45.6 (16.6 dB).
Thus, the penalty in P,/NoB for using higher level modulation with practical filtering
is 4 dB.
A useful tool for qualitatively assessing the severity of ISI degradation is the so-called eye
pattern. To see how this comes about, we proceed as follows. Consider the bandlimited
communication system of Fig. 9.1 wherein the transmitted modulation is of the form in
(9.1) and a received signal is of the form in (9.3). For a specific data symbol sequence
{ay}, m(t) takes on a particular waveform. Suppose we consider the superposition of all
m(t) waveforms corresponding to all possible sequences {a,}. The resulting pattern when
viewed on an oscilloscope is called an eye pattern, so-called because of its resemblance
with the shape and characteristics of the human eye (see Fig. 9.4 for a qualitative example
corresponding to binary amplitude modulation). Since the above superposition considers
all possible sequences (each theoretically of infinite length), the resulting waveform is the
same in every 7;-sec interval, that is, the eye pattern is periodic with period T,. Hence, it is
sufficient to examine the eye pattern and its properties only in a 7;-sec interval.
The two most important properties of an eye pattern are its maximum opening (the
largest unobstructed interior vertical distance at any sample time instant) and the broadness
(horizontally) of the opening in the neighborhood of its maximum. The significance of these
two eye pattern measures is as follows.
The maximum eye opening is a measure of the maximum degradation of the signal
component of the detected sample due to ISI. Equivalently, it is an indication of the ability
of the system to protect itself against bit errors when noise is present since, indeed, half the
maximum eye opening is the size of the smallest output noise sample that will cause an error
(assuming that sampling is performed at the point of maximum eye opening). Clearly, the
smaller the eye opening, the more ISI is present and the smaller the noise sample need be to
cause a bit error.
The broadness of the eye in the neighborhood of the maximum eye opening is an in-
dication of how timing jitter tolerant the system is. Clearly, the steeper the eye pattern is
568 Communication over Bandlimited Channels Chap. 9
re
Was; os Cs
8 INTO,
Ny, LBEOSSS
C4 RSX
ESEZESN\
= ZZ, S—S>
<> . —_FA
Se ———, - SS
Sioa F Figure 9.5b An eye P pattern for multilevel
s amplitude modulation (Reprinted from [3])
in this vicinity, the quicker the eye closes with any perturbation around the desired sample
point. Thus, a broad eye is desirable so that an offset in the sampling instant will produce
minimum eye closure (relative to its maximum value) and minimum error probability degra-
dation. Based on this consideration, it is now possible to clarify a statement made earlier
with regard to the reduction in sensitivity to timing offset of the raised cosine channel pulse
response relative to the sinx/x response of the brick wall channel. For the latter, the eye
will be much narrower in the vicinity of the peak because of the slowly decreasing tail.
A more realistic eye pattern diagram that results from a digital stream with pulses
having a shape that extends over many 7)-sec intervals is illustrated in Fig. 9.5a. We observe
that at the point of maximum opening, the eye is “well-focused” with little degradation
due to ISI. Figure 9.5b illustrates an eye pattern for multilevel (MM = 4) pulse amplitude
modulation. Here the maximum opening of any of the M — | = 3 eyes is a measure of the
resistance against making an error (due to noise) by choosing a symbol adjacent to the one
transmitted.
In many transmission channels, bandwidth is at a premium and thus the question naturally
arises: how can we increase the throughput of the system, for example, for a fixed band-
width, how can we increase the bit rate? If we restrict ourselves to zero ISI systems, that
is, channels that satisfy the Nyquist criterion of Section 9.2.1, then the only solution is to
increase the number of modulation levels by going to a larger signaling alphabet. While this
Sec. 9.2 Optimum Pulse Shape Design for Digital Signaling 569
does indeed produce a higher bit rate for a given symbol rate (or equivalently, a given band-
width), the penalty paid is an increase in the required transmitter power. On the other hand,
if we remove the zero ISI constraint and allow a finite and controlled amount of ISI, then
it is possible to achieve a more efficient trade-off between bandwidth and power than that
achievable by the above solution.
Lender [6] was the first to show that using binary modulation but introducing depen-
dencies or correlations between the amplitudes (that is, nonzero ISI) and suitably changing
the detection procedure, one could still achieve the maximum efficiency of 2 bps/Hz with,
however, realizable and perturbation-tolerant filters. The method was called duobinary en-
coded modulation, the prefix duo referring to the fact that the overall system time response
g(t) has two equal nonzero sample values. As such there exists 100% ISI which stems,
however, from only one adjacent symbol. Duobinary is the simplest form of controlled ISI
signaling; hence, we start our discussion with a description of this correlative encoding
technique.
The duobinary encoding technique. Consider the block diagram of Fig. 9.1
where the block labeled “modulation source” that generates (9.1) is modeled as a binary
data source represented by a uniform train of impulse functions at the data rate 1/7, fed to
a pulse shaping filter P(f). Consider now encoding the binary data source prior to pulse
shaping with a simple digital filter as shown in Fig. 9.6. The digital filter delays the input
and then adds it to itself. Thus, the n'" impulse into the pulse shaping filter has a strength
proportional to the algebraic sum of the n“ (current) input bit and the (n — 1)** (preceding)
input bit. As such, the effective encoded bit stream {a/,} = {ay + ay,_1} at the channel input
has dependencies among its members. Note that because of the algebraic addition, the
symbol sequence {a/,} is now ternary (rather than binary) in accordance with Table 9.1.
Note that when a, = —1, there are two possible levels for a}, namely, 0, —2, depend-
ing on the value of a,—1. Similarly, when a, = 1, there are again two possible levels for a,,,
namely, 0, 2, also depending on the value of a,_). We can think of this dependence on the
previous symbol as introducing a controlled amount of ISI. Thus, whereas Nyquist (zero
memory) signaling is based on eliminating ISI, signaling with duobinary or correlative en-
coding schemes is based on allowing a controlled amount (finite memory) of ISI, that is, a
decision on ap based on a threshold test performed on the sum of aj, plus noise depends on
Gn-A-
We can write the signal component of the receive filter output y(t) as
[e,¢)
(oe) (oe)
Thus, as far as the modulation is concerned, the combined transmitter/receiver of Fig. 9.6
can be represented as in Fig. 9.7 where
570
(1)u
(Lt-ad'v’¢v Sd= (f)'s
Ke (LH y odures
1 %7y 14+
ons
ed Dg. *11—
< Ng’ cy tga x yrursuely, 5
pee are Jay UOIST
(f)'9
wyo3Ty
(L0+2)-Dgo<
co-=!
aansIq
9°6 YOo[g WeIZeIP
JO payIWI[PULG UONRITUNUILOD
W}sksYM AreuIgONppepoousaoInos
Sec. 9.2 Optimum Pulse Shape Design for Digital Signaling 571
ane ‘ape a
| =
zSoLS
A ¥ a,(t—i7,)
sample at kT, + %
Decision
Algorithm
Figure 9.7 Equivalent block diagram to
Figure 9.6 for modulation
= 2G(f)cosaf
Ty exp (— jf Ty) (9.36)
IAP) = |poimretpig
:
:
oie mie
otherwise
(9.38)
We can implement H(f) directly and thus no separate digital filter is required in front of
P(f) as in Fig. 9.6. Rather, as we shall see shortly, we merely split H(f) (equally for
additive white noise) into H,;(f) and H,(f). Note that since no excess (over the mini-
mum Nyquist) bandwidth is used, the efficiency of transmission is 2 bps/Hz. Thus, we are
able to achieve the maximum efficiency for binary transmission using, however, practical
(perturbation-tolerant) filters.
The impulse response corresponding to (9.38) is
h(t) = (9.39)
and is illustrated in Fig. 9.8. Note that h(t) has two equal nonzero samples, that is, one at
t = 0 and one at t = Tp. Thus, as previously stated, the detected samples will have 100% ISI
due to one adjacent pulse.
572 Communication over Bandlimited Channels Chap. 9
h(t)
Receive
Filter
G,(f)
If we could always make perfect decisions, we could, in principle, subtract out the
ISI and use a binary decision scheme with a single threshold. The appropriate receiver for
doing this would be as shown in Fig. 9.9. If no error is made on the decision for a,_, that
is, Gk—1 = ax—1, then z_ = yy — AGg—| = Aag + fix and the probability of error for ax would
be 4 erfc (4/V20,) which is the identical result as that for binary Nyquist signaling. On
the other hand, if an error was made on the decision for a,—1, that is, @,_] = —ax_}, then
Zk = Vk — AGg—1 = A(ax + 2ax—1) + nz. Thus, depending on the relative polarities of a,x
and az—1, the double ISI can help or hurt. The problem with this detection scheme then is
that once an error occurs, it tends to propagate further errors, that is, if a,—, is in error, then
its effect is incorrectly compensated for when deciding ax; consequently, a, is likely to be
in error also.
To get around the problem of error propagation, Lender [6] suggested a means of
avoiding the necessity for ISI compensation (subtracting out the single ISI) at the receiver.
His scheme suggested that we precode the data {a;} at the transmitter into the sequence {d;}
as follows. If we think of the binary data as 0’s and 1’s, then the precoder is described by
dn = an ® dyn— 1where the symbol @ denotes modulo-2 addition. Note that a, now indicates
the change in the output symbol d,. If a, = 0, then there is no change in d,. If a, = 1, then
there is a change in d,. This simple encoding scheme is identical to differential encoding
described in Chapter 7 in connection with differential detection of BPSK modulation. The
precoder can also be reinterpreted in terms of “1”’s and “—1’’s for the input bits {ay}.
In particular, the modulo-2 summer is replaced by a multiplier resulting in the relation
dn = an X dy—). Note that d, is still a binary random sequence with equiprobable 1’s and
—1’s.
Using the +1 representation for the a,’s, we obtain the implementation for the duobi-
nary system illustrated in Fig. 9.10. The mean of yz, the sample of y(t) at t = kT, + to,
is now equal to A[dy + dy_]. Table 9.2 gives the values for this mean in relation to the
possible values of the input (information) binary symbol ax.
Thus, we see that when the information bit is a “—1”, then the detected signal sample
is “O” (independent of ay_;). When the information bit is a “+1”, then the detected signal
Sec. 9.2 Optimum Pulse Shape Design for Digital Signaling 573
Table 9.2 Possible Values for the Mean of the Detected Sample
sample is +2A with the polarity depending on az_. If we examine the magnitude of the
detected signal sample, then, independent of ax_1, the detected output sample will be 2A
when the information bit is a “+1”. Thus, in the absence of noise, we can decode perfectly
according to the following rule:
By choosing a threshold? halfway between “0” and “2A,” then in the presence of noise
we have the decision rule a, = sgn {|yg| — A}. This decision rule involves rectification of
yz followed by a single threshold binary decision. Since no knowledge of a sample other
than yy is involved (that is, the decision rule requires no knowledge of previous symbol
decisions), error propagation cannot occur. The threshold comparison and decision logic
described above can be implemented as in Fig. 9.11.
What remains is to compute the probability of error for the duobinary scheme with
precoding. For a flat AWGN channel, we saw in Section 9.2.2 that the total system response
[in this case H(f)] should be split equally between the transmitter and receiver. Thus, we
choose the two filters H;(f) and H,(f) to be identical and equal to the square root of H(f).
Under these circumstances, the input power to the receiver is
1 (2T))~
5 iP df=a? if =
27, cosnf Thdf
ae [o,2)
P, = A*d? i =(2T, —1 Tp
—0o0
pau! (9.40)
-i(f
Me Tp
5. This choice of threshold can be shown (see Prob. 9.5) to minimize the average bit error probability in the limit
of large SNR.
574 Communication over Bandlimited Channels Chap. 9
sample at kT, +t
1
Pp(E) = 7 Pr {—A < ye < Alay = 15k = 2A)]
]
Hey Pr{—A < yx < Alag = 1(ye = —2A)}
|
iy EEE = A Or ye < —Alde
=— Wye OY)
=-—
| Ip 1
exp
(y — 2A)? o
eles /2n0? 203
(9.42)
piper ty y42A)*
oF if exp Ot 5 ) d
Todd Phiters 20;
PY ee as 2
=f / exp
2A lon ae 2a5
3 2 P.T; 2
Pi Eyeene) | eat SEED
4 16 No 4 16 No
(9.43)
3 jihad aed
= 4 erfc ae i for large SNR
Thus, when compared with binary amplitude modulation using ideal (brick wall) fil-
ters to achieve 2 bps/Hz, the SNR penalty (ignoring the factor of 3/4 versus 1/2) is
—10log,y(77/16) = 2.1 dB.
Generalization of duobinary to raised cosine Nyquist channels. Suppose
now that instead of G(f) being a flat channel as in (9.37), G(f) is the raised cosine channel
characterized by (9.25). Then, the magnitude response of the equivalent channel H(/) is
given by the product of (9.25) and 2 cos 2 f Ty. Since g(t) has a single nonzero sample when
sampled at 1/7), then h(t) will once again have two nonzero samples when sampled at the
same rate. Hence, the implementation of Fig. 9.10 and the detection scheme of Fig. 9.11 still
apply here. In fact, an evaluation of the input power, P,, and variance of the detected sample,
eS, produce results identical to (9.40) and (9.41) respectively (see Problem 9.6). Thus, the
probability of error for the duobinary raised cosine channel is also given by (9.42).
What is different then about the raised cosine channel from the flat channel is the
trade-off between throughput and signal-to-noise ratio. Ignoring the factor of 3/4 in (9.42),
then for P,(E) = 10~>, we require (77/16) P,Tp/No = 9.12 (9.6 dB). The throughput is,
as before, R,/B = 2/(1 + a). Thus
Fy. P,Ty / Rp 16 2 16 —)
= pe Nae (9:12 — ne 9.44
NoB No ($) \ (3)(<5) m2 \1l+a )
Once again, we see the trade-off between throughput and SNR as a function of the excess
fractional bandwidth qa. In fact, a plot of R»/B versus P,/NoB with a as a parameter
would be identical to that determined from (9.33) evaluated for M = 2, except it would be
displaced horizontally (along the P,/NoB axis) by 10 log;9(16/27) aed Bille
| deal BopFe 3
oy ( ras \ 16 (No Nie ) eee
Thus, ignoring the (1 — 1/M7) factor, the SNR degradation for any M relative to M-ary
amplitude modulation with Nyquist pulses is again 10 log (16/77) See) OS:
Aside from extension of duobinary to multiple levels, other generalizations involve
correlative encoding techniques that make use of overall pulse responses with more than
two nonzero samples. These techniques fall into what is commonly referred to as partial
response signaling [5, 10, 11]. A generalized partial response digital filter, T(f), has the
576 ~ Communication over Bandlimited Channels Chap. 9
KaseOpe Ei SCD
Number 0}
Output Leve
T(D) H(f) for |f| < 1/2Ts h(t) (M-ary Inpt
1+D 2T, cos mf Ts Aer a 2M —1
iD j2T, sin xf T; alts eS 2M —1
f= Dp? j2T, sin 2x f T, 2 a 2M —1
(1+ D)2(1—D) 47, sin2x fT, cos fT; Aa: Gon pS) 4M —3
Table 9.3 gives expressions for the minimum bandwidth H(f) and A(t) corresponding to
the above examples as well as some others due to Kretzmer [5].°
6. In this table, for mathematical convenience, we have shifted the time origin to the center of the nonzero
samples, that is, t = (N — 1)T;/2 is the new time origin. Thus, for example, the duobinary frequency response
H(f) is given by (9.36) multiplied by e~/*/7 and h(t) corresponds to (9.39) with t replaced by t + T,/2.
Sec. 9.3 Optimum Demodulation of Digital Signals 579
8c(t)] being bandlimited distorts the pulse shape g;(t) and introduces ISI.’ The received
pulse shape is given by h(t) = g;(t) * g(t). Thus, the received signal plus noise is expressed
as
[e.@)
(9.50)
lo @)
Gln ie
n=—CO
with
le 2) CO
The n,;’s are in general independent complex zero-mean Gaussian variables with equal
variance (because of the white noise assumption all eigenvalues are equal)
= ————;
1 exp
Kes ae ae
~ 353 2 Viel
The optimum (in the sense of maximizing the probability of correct detection) decision rule
is obtained from maximizing the a posteriori probability of the transmitted sequence given
the observable, namely, f (¢|rx). Using Bayes rule (see Chapter 3), this probability can be
expressed in terms of the probability f (rx |c) of (9.53). For independent data, maximizing
7. As before, we again assume that G,(f) includes the modulation pulse shaping P(f), and thus g;(t) is, in
reality, the convolution of g;(t) with p(t).
8. Herein it shall be convenient to renormalize the problem such that the signal amplitude A = 1. Thus, the PSD
of n(t) now becomes No/A?.
580 Communication over Bandlimited Channels Chap. 9
Since the first term of (9.54) is independent of ¢, then equivalently we can maximize the
function
K
Fx@ = -» rk — 3 eel =-) axl?
ee (9.55)
k= n=—OCoO k=1
In the limit as K — ov, we get
ib iePdr=f limDae oa
Aj oe(t)dt=limDe (9.56)
The maximum-likelihood estimate of the data sequence ¢ is the particular sequence that
maximizes F(¢) of (9.57). We can simplify this result still further by noting that the first
term in (9.57) is independent of ¢; hence, it can be discarded with regard to the maximiza-
tion operation. For the second term in (9.57), we define
where
where
CO
(9.61)
= if h(t’ —nT, — t)h(t' — nT,)dt’
10,9)
We point out that we are not particularly concerned with the noncausality of h(t) since, in
practice, we can introduce a sufficiently large delay to make it causal.
Note that for g(t) corresponding to a Nyquist channel (that is, g(n7,) = 0 for n £0),
(9.62) simplifies to .
m=—oo
lo.@) (oe)
is the detected Gaussian noise sample at time k7; with complex correlation function
The brute force method for determining the optimum sequence ¢ is to exhaustively try
all possible sequences!® in (9.62) and then select the one that yields the largest value of
Fo(@). A more efficient method for processing the sampled complex matched filter outputs
{Sn} is the so-called Viterbi algorithm (VA) [7] which was originally proposed for optimum
decoding of convolutional codes. This algorithm performs maximum-likelihood estimation
of a digital sequence and thus is often called a maximum-likelihood decoding algorithm
[14]. Note again that the optimum decision rule in the presence of IS] and AWGN makes a
decision on a sequence € rather than symbol-by-symbol detection, which is a suboptimum
decision rule. The MLSE technique is capable of producing performance approaching that
of an optimum symbol-by-symbol demodulator operating in the presence of AWGN alone.
The VA may be viewed as a solution to the problem of maximum a posteriori (MAP)
estimation of the state sequence of a finite-state discrete-time Markov process observed in
memoryless noise. Before starting a formal explanation of the VA, we must first cast the
MLSE in the presence of ISI and AWGN in this form. The formulation for accomplishing
this is due to Forney [13, 15]. For the purpose of our discussion, we shall assume that the ISI
exists over a finite time interval corresponding to v discrete times on either side of go, that
is, g; = 0 for |/| > v. Although g(t) is bandlimited and thus cannot also be timelimited, the
assumption of a finite length is a reasonable approximation in most practical applications.
We begin by first writing the detected sample yz of (9.58) as
The discrete-time process {yx} is called a shift register process since it can be modeled by a
shift register of length 2v whose input is the sequence ¢ (see Fig. 9.13a). The problem with
the representation of Fig. 9.13a is the correlation of the noise samples [see (9.65)] which
makes the performance analysis of the MLSE difficult. To circumvent this problem, we shall
introduce a noise whitening discrete filter (see Chapter 4) after the sampling operation in
Fig. 9.1 so that the detected samples will take the form of (9.66) with, however, uncorrelated
Gaussian noise samples.'! The noise whitening filter is determined as follows.
Since from (9.60) and (9.61), 2m—n = 8n—m, then gx can be expressed as the discrete
convolution
v—k
where f; = 0 fori <0 andi > v. Note that 9, = 0 for |k| > v as required above. In terms of
the z-transform of these sequences, namely
10. This, of course, only has practical meaning if the sequence length and thus the number of possible sequences
is finite. We shall be more specific about this shortly.
11. The introduction of a noise whitening filter is made merely for convenience but is not necessarily required
to implement the Viterbi algorithm. A description of the operation of the Viterbi algorithm for combined ISI and
AWGN channels without using a noise whitening filter is given in [17].
Sec. 9.3 Optimum Demodulation of Digital Signals 583
Figure 9.13a Shift register representation of bandlimited system in the presence of ISI
and AWGN
where ji, is an element of a zero mean white Gaussian noise sequence with variance
E ||; ‘|= 20°. In terms of the sequence {5}, the decision rule of (9.62) can be rewritten
as
Choose the sequence ¢ that maximizes
(o @) (o@)
9)
v
De oie
ni=—CcoO i=0 n=—OO i=0
or equivalently minimizes
CO [o.2) v
2
The j,’s form a set of sufficient statistics for estimation of the input data sequence ¢ [15].
584 Communication over Bandlimited Channels Chap. 9
Figure 9.13b Shift register representation of bandlimited system in the presence of ISI
and AWGN (noise whitening filter included)
Figure 9.13b is a model of the shift register process corresponding to (9.69). !2 Note that the
length of the register is now equal to v in accordance with the duration of the f; sequence.
Also, {5x} is now a v" order M-ary Markov process.
The state, x, of the process is defined by
12. Since herein we shall be dealing with the respresentation of (9.69) rather than (9.66), for simplicity of
notation, we shall drop the prime on jy, and ji; .
13. Despite the fact that the state and transition are defined by sequences, it is customary not to denote them with
a boldface typeface.
Sec. 9.3 Optimum Demodulation of Digital Signals 585
10
11
decision rule, which minimizes the error probability in detecting the whole sequence, is thus
optimum in this sense.
Associated with a discrete-time finite-state Markov process is a state diagram in
which nodes represent states and branches represent transitions. As discrete time progresses,
the process traces a path from state to state in this diagram. An example of such a state dia-
gram is illustrated in Fig. 9.14a for a binary!* system with ISI memory v = 2. Since M = 2,
there are M” = 4 states and a total of M’*+! = 8 transitions among them. Note that from
any given state, there are only two allowable transitions to the next state. For example, if the
current (at time k) state is “Ol” corresponding to a,_; = 0 and ay_—2 = 1, then the next state
will be “OO” if a, = 0 or “10” if ax = 1. Note further that the structure of the state diagram
depends only on the input sequence a and the memory of the ISI, not the values of the ISI
(shift register tap weights) themselves.
An alternate and more redundant description of this same process is a trellis diagram
[14], so-called because of its resemblance to a garden trellis. Figure 9.14b is the trellis
diagram corresponding to the state diagram of Fig. 9.14a. The trellis diagram illustrates
the progression of the process with time, a property that is lacking in the state diagram
representation. Each node in the trellis diagram corresponds to a possible state of the state
diagram at that particular instant of discrete time. Thus, the maximum number of states at
14. For this example, C, is real, that is, ¢, = a, and takes on values 0 or 1 with equal probability.
586 Communication over Bandlimited Channels Chap. 9
any particular time in the trellis is equal to the number of states in the state diagram. In fact,
the only time the trellis will have fewer than this maximum number of states is for the first
few time instants at the beginning and at the end of the input sequence. Each branch in the
trellis diagram represents a transition from a particular state at a given time to a new state
at the next instant of time. The inital state (I) of the trellis corresponds to Xo and the final
state (F) corresponds to Xx, both of which will be “00” if the initial contents of the shift
register was empty. Associated with each sequence of states, X, is a unique path through the
trellis and vice versa. Since the most probable sequence is the one for which f (x|y) [or
equivalently f (x, y) = f (xly) f (¥) is maximum, if we assign to each path x the quantity
—In f (X, y), then since the negative logarithm is a monotonically decreasing function of its
argument, the most probable sequence is also the one for which — In f (x, y) is minimum.
We shall see in a moment that we can attach to the quantity — In f (x, y) the significance of
path “length.”
Since the process y is observed in memoryless noise, then
K-1
f 91%) = F (HE) = [TF (Sul&e) = TT] f Celie, 9.71)
k=0 k=0
Furthermore, since the states obey a first-order Markov behavior, that is,
Taking the natural logarithm of (9.74) converts the products into sums; thus
K-1 K=1
=In fH) =— Yn f Geile) +10 f Guléev, 40) = 2 (E) 0.75)
k=0 k=0
where the weight A (&) is assigned to each branch (transition) and represents its length.
Hence, the sum in (9.75) represents the length of the entire path.
Finding the shortest length path through the trellis [path with minimum — In f (X, ¥)]
in an efficient way is what the VA accomplishes. Consider a segment of the state sequence
X = (x0, %1,..., XK) Consisting of the states from discrete time zero to discrete time k, that
iS, xf = (Xo, X1,..., Xx). In the trellis diagram, xh corresponds to a path from state x9 to
state x,. For any of the M” (or fewer) nodes at discrete time k, there will be in general
several such path segments entering that node, each with a length
that depends on the particular transitions €;; i=0,1,...,k—1 along that path segment.
One of these path segments, x (X;), entering the node will have the shortest length T (x,) A
min A (x6) = (x(%)). This particular shortest length path segment is referred to as the
survivor corresponding to that particular node. The reason for this terminology stems from
the following observation:
The shortest state sequence x (the one we are looking for) must at any discrete time k along its
path correspond to a survivor path segment.
If this were not true, then at any node along X we could replace the path segment from
state xg to that node by the survivor and thus achieve a smaller length which leads to a
contradiction that x is the shortest length path. The result of all this is that at discrete time
0 <k < K, it is only necessary to save (keep track of) the survivor paths and their lengths,
one each for the M” (or fewer) nodes at that time instant. In proceeding to discrete time k +
1, one merely appends the survivors with the branches leading to the next set of states and
increments the survivor lengths with values related to the corresponding transitions. At time
k + 1, one again chooses the M” survivors and the process continues recursively. Finally,
at time K, the trellis converges to a single node with one survivor that must correspond to
the shortest state sequence x being sought after. The length, I" (x) ule (x(x x), of this
shortest path must from (9.76) be equal to the minimum value of (9.75) as desired.
Figure 9.15 illustrates the above recursive evaluation of the shortest path for the trellis
diagram of Fig. 9.14b and a value of K = 5. The values assigned to the trellis branches are
not intended to be typical of any particular application but are merely selected for simplicity
of the distance evaluations. Since M = 2 and v = 2, there are 4 (or fewer) nodes and 8 (or
fewer) transitions at each discrete time. At each of these nodes, the solid lines entering
the node indicate the survivor path while the dashed lines indicate the nonsurvivor paths.
Finally, the shortest path through the trellis has length equal to 3.
Although the VA as described above will theoretically find the shortest length path
through the trellis, the presence of very long or infinite state sequences requires a buffer
memory beyond practical limits. Hence, in practice, one must invoke a modification of the
algorithm wherein survivors are prematurely truncated to a manageable length, say 5. That
is, at time k the algorithm makes a final decision on the first k — 5 branches of the shortest
path (that is, nodes up to time k — 4) and retains only the length and path information
on the most recent 6 branches (the truncation depth) of the survivors. The ability to make
such a premature decision relies heavily on the extent to which the time-& survivors share a
common path for the first k — 5 nodes. For example, in the example of Fig. 9.15 we observe
that at k = 4, the two surviving paths shared a common path up to time k = 2. Since one
of these two survivors will ultimately be selected as the shortest path, nothing is lost by
making a final decision at time k = 4 on the first two branches (5 = 2) and truncating the
survivors to two branches. Clearly, for sufficiently large 5, there is a high probability that all
survivors at time k will pass through the same set of nodes for the first k — 6 time instants.
In the rare cases where survivors disagree in their first k — 6 branches, one must adopt a
time-(k — 5) decision strategy that has negligible effect on the ultimate error probability
performance. One such possible strategy is to choose the time-(k — 6) node corresponding
to the shortest time-k survivor hoping, of course, that this survivor will eventually wind up
588 Communication over Bandlimited Channels Chap. 9
I(x, =1) =2
T(x, =2) =3
I(x, =3)=2
I(x, =2)=2
I(x, =3) =3
P(% =4)=2
I'(x,=1)=2
I(x, =2) =3
FG; =)=I)=35
Figure 9.15 Recursive determination of shortest length path for four-state trellis diagram with K = 5
Sec. 9.3 Optimum Demodulation of Digital Signals 589
to be the shortest length path. Another possibility is to choose the time-(k — 5) node by a
majority vote, that is, the one through which most of the time-k survivors pass.
From a computational point of view, as time k gets large, the values of survivor path
length, T (x,), for each of the M” states also become large and thus from time to time it is
necessary to subtract a constant from all of them.
A few words about complexity are in order. Since the algorithm requires M” storage
locations, one for each state (after the initial transient) capable of storing a survivor path
and its length, then from a memory standpoint, the complexity grows exponentially with the
length v of the shift register. The complexity associated with the computation of the incre-
mental lengths, (&), is assessed as follows. Since from (9.74), one component of A (E.)
is equal to — In f (X,+1|x;), then since for a shift register process f (%4+1|X,x) is either equal
to 1/M or 0 depending upon whether state x; +) is allowable after the occurrence of state x;,
all allowable transitions have the same value of —In f (x,+1|X,). Thus, this component of
Xr (&) may be ignored. The computation of the remaining component of A (&), namely,
In f (Silke) depends on the data sequence. However, from Fig. 9.13b and (9.69), we ob-
serve that, in general, many different values of & produce the same value of shift register
output wz. Since yy, = we + fx, then for a given yx, all of these values of & produce the
same value of Gaussian probability f (5xl&) and thus — In f (nlx) need be computed
only once. Because of the above, the complexity associated with the computation of the
A (&)’s tends to be insignificant relative to that associated with the memory requirement.
Example 9.1
To illustrate the application of the Viterbi algorithm specifically to the ISI problem, consider
a binary PAM (cx = a, = +1) system with memory v = 2. The appropriate representation of
the system is given in Fig. 9.13b where the shift register now has only two stages. The state,
the transition, and the shift register output are real and are given by x, = (ag—1,.Ax—2), &k =
(Ap, Ak—1, Ak—2), and wy = foag + fiax—1 + foax—2, respectively. The state transition diagram
for this shift register process is identical in form to Fig. 9.14a. We redraw it here in Fig. 9.16a
where we have now labeled the branches with the input (a,)/output (w;) corresponding to each
transition. The conditional pdf f (yz |&%) is Gaussian and given by
l (ye — we)?
f (ye 1E&&) == F OK we) = a exp p 4————_——_
No
u (9.77)
ete ale [yx — (foak + fiar—-1 + frax—2)]
% /m No P No
OUT 6
ly.- fl’ Yi :
State , k=2 vans se AER So ha oI k=3
0,0 eocee
0,1 eoce
1,0 eoce
iLJt eoce
ae Di- Got ft Ay
D2-Goth }
Da-G+h) %,-(¢ +4
Figure 9.16b Trellis diagram corresponding to state diagram of Figure 9.16a
Sec. 9.3 Optimum Demodulation of Digital Signals 591
|
keL- | [re(5 (y= 4)"}
:Gs = |p
ie | (9.79)
|
Note the similarity between (9.79) and the metric used in the decision rule of (9.70).
To evaluate (9.79), let
and variance
k+L-1 7 aS k+L—1 5
ot =E (d> Re {iin(ton — tn) ) =o? > |tin—tn| (9.82)
n=k n=k
a mee . i =
P(K—> 1® =Pr{X>y|é}=~erfe| ” (9.83)
2 202
VCE
where
y kt! >
+ ~ 12
ya-2 ({i, — |Wn| ) (9.84)
n=k
nlp Rlaaa D0 hg
yi Dy X=Ces E.5 2 pee |)
Ve Se
AG? (9.85)
Sec. 9.3 Optimum Demodulation of Digital Signals 593
a 1
(9.86)
kK+L—-1| v - Z 2
Sa ph eF ees = cri)
= ae n=k_ l|i=0
D, 802
Since the correct and incorrect paths share the same set of states for discrete times less
than or equal to k and greater than or equal to k + L, that is, x; = ti BSS kao
then from the definition of the system state in terms of the input data sequence, we have that
C; —c;=0; i<k—1, i>k+L -—v. Then, in terms of the definition of the actual ISI
coefficients given in (9.67), we can rewrite (9.86) as
n=k
Aid} ie 2 (9.88)
ae _ Yatfi (Ga
(Cnt = n Cri )
2 80?
This result (except for the factor of 1/2) can also be obtained by applying a Chernoff bound
directly to (9.79).
594 Communication over Bandlimited Channels Chap. 9
The number of errors in &,, that is, the number of nonzero elements in &,, is defined as the
Hamming weight wy (€) of the error event.
Following Forney [13], we consider a particular error event € as being composed of
three subevents, €,, €2, and &3, all of which must occur in order for € to occur. Subevent €
corresponds to the requirement that both the incorrect and correct path must have the same
state at time k, that is, x, = Xz. Subevent 2 requires that for 0 <i < L — v — 1, the input
sequence must be such that, for the particular error sequence assumed, ¢; = 2€,,; + cj is an
allowable error sequence, that is, it takes on values from the set +1, +3,...+(M — 1). Fi-
nally, subevent €3 corresponds to having noise terms nj; k <i <k +L —1 such that the
likelihood of the segment of X over this interval is greater than for the corresponding seg-
ment of x. The probability of €3 occurring, Pr {3}, is precisely the pairwise error probability
discussed in the previous section applied now to the specific case where the incorrect path
corresponds to an error event.
Based on the above partitioning of € into €), €2, and €3, we can write the probability
of the particular event € occurring as
15. Actually, €- has a true dimension equal to L — v in terms of the number of elements that may or may not be
zero. However, it is convenient to define the extended sequence &, as we have done.
Sec. 9.3 Optimum Demodulation of Digital Signals 595
equations gives
Bs ] é
Pr {€3} = P(x > x|e,) = — erfc au)
2 202
and
v Ds
er 2 elo ( fitcn-i)
Pr {3} = P(x > Sie) < 5 Ul exp
1 |4 -»- 1 i=0
262 a ae 202
v
1 ate 80ES n +2 d 8iEc.nEcn—i
i=
52 Jide a (9.92)
n=k 20
where we have now changed the conditioning event to the error event sequence. It is con-
venient to normalize (9.91) and (9.92) in terms of the received signal-to-noise ratio. From
(9.60), we have
3 3 T.
Since o* = No/2A?, then solving for A? from (9.94) and substituting into (9.91) and (9.92)
gives
M2—1) No
and
: k+L—1 3 Per 5 2
(9.96)
1 ktLe! 3 pT, ;
I] ee M2=] No eon anid Me 8iEc nEcn—i
i 9
n= =I
the number of input data sequences that produce a particular error sequence with Hamming
weight wy (E) is M'~’-“#©) TT (M — |ec,;|). Since the total number of possible input
i€lwy
data sequences of length L — v is M/~”, then the probability of subevent €2 occurring is
(9.97)
TT (SE) (Ge
k+L—v-1 M — |ec.i| Reo M — |e,i|
i=k r=K
Finally, computing the conditional probability Pr {€)| €3} is difficult because of the
possible dependence of €; on the noise terms involved in €3. However, the unconditional
probability that €; does not occur, namely, 1 — Pr {€)} is on the order of the symbol error
probability. Thus, in the normal operating region, the conditional probability Pr {€;| €3}
is closely approximated as well as overbounded by 1. Substituting this upper bound on
Pr {€)| €3} together with (9.97) and (9.95) [or (9.96)] into (9.90b) gives the desired upper
bound on the probability of the particular event €, namely
_
1 k+L—1 M — lec 3 BTA
I ( M oa es (szcP :) No Eon + 2 ~- 8iEcnEc,n—i
F 2
For binary PAM where e,.; can only take on values 0, +1, (9.97) simplifies to
1 k+L—1 1
where wy(Ec,n) is the Hamming weight of e€.,,, that is, wH(€on) = 1 if en 40 and
WH(Ec.n) = 0 if &¢., = 0. Thus, (9.98) becomes
j tee 1 1 2,
(9.100)
Ves is =
= I] SACHS ass ene (35fron
r=
pRticl 4 PT» | 5 r
cs) I] wu (Ec,n) Dn No Een +2 y scat
n= i=1
Since the probability of a union of events is less than or equal to the sum of their
individual probabilities, if E denotes the set of all possible error events € starting at time k,
Sec. 9.3 Optimum Demodulation of Digital Signals 597
then the probability of any error event starting at time k is union bounded by
Al k+L—1 Abt f
P(E) < ye 5 erfic(Gea)7
PTs ©] leh (“5
EEE 1=k
1 k+L—1 Me een 3 2
= 2.5 I] (“f= exp Ga) NoE (55fee )
SIGS ~ leen|
= 5 nt (“ie (9.101)
x exp |-(@-)
Ne 9oeeal es
me 2,+
2
2 2 Sitenton-)
Similarly, the average probability of a symbol error is obtained by weighting each error
event probability by the number of decision errors wy (E) resulting in the upper bound
y KtLel M om ae
2
3 Pe tuhes
x exp = (Ga—3) No (>:
fist (9.102)
k+L-1 tee
= wn); I] (* a
ECE n=k
s eles -
x exp | (Gz a7 :) oa le,2 Ptieatent|
hell
(ORO O) thats
Bree
ay,j= exp {-(@—) ele
7a 2, | (9.103)
There are 2(M — 1) branches entering the final all zeros state (at time k + L). To each of
these branches, we assign the weight | since the input error €¢,,4z—1 that causes the tran-
sition to this final state is zero (thus (M — |éc,x4z—1|) /M = 1), J is absent (since the next
state is the all zeros state), and the exponential term in the product of (9.102) correspond-
ing to €¢.44,—-1 = 0 is unity. To the remaining branches between error states, we assign the
weight an—«,j1 (M — |écn|) /M where &¢,n = j is the input error that causes the transition
and aj, is again the exponential term in the product of (9.102) corresponding to &¢,n = j
and the current state ¢, = (GRSEi, yee pea ay that is,
3 P,T,
Tso: -
An—k, j= €XP |-(@—) No le,+2 De tren (9.104)
i=
P,Th
ay = exp {- (9.105)
No
There are two branches entering the all zeros final state each of which is assigned the weight
1. The branches between other pairs of states are assigned weights a_//2 where
as = exp |
Tpie
- hi (1+ 2¢| (9.106)
0
where the plus sign corresponds to a transition from a given state to the same state, that is,
1 to 1 or —1 to —1, and the minus sign corresponds to a transition from a given state to the
opposite state, that is, | to —1 or —1 to 1. The complete error-state diagram is illustrated in
Fig. 9.17.
To evaluate the error event error probability in (9.101) or symbol error probability in
(9.102), we must compute the generating function of the error-state diagram, which can also
be regarded as the transfer function of a signal flow graph with unity input. This function can
be most directly computed by simultaneous solution of the state equations obtained from
the error-state diagram itself. In particular, from Fig. 9.17 letting &,, &, and &4 represent
dummy variables for the partial paths to the intermediate nodes, and T (ao, a_, a4, 1) = €g
the desired generating function, then with €, = 1 (since the graph has unity input) the set of
state equations is given by
I I I
Sb = a0; gee nh tte ee (9.107)
I I I
ge a0 + 4458 oh 4-5 5b
Sec. 9.3 Optimum Demodulation of Digital Signals 599
agl/2
Paneer Vee ayer
aa (9.108) )
IEGita Gh) =
|Beee |(2.07Se
toe)Oy4
Finally, the desired upper bounds on P.(E) and P;(E) are obtained from T (ao, a_, a4, 1)
as [16]
ag/2
P.(E) < T (ao, a_-, a4, I|j-1 = 1G
= Gi aan
Pas
too A
b—exp {57 cosh (24z21)
ag/2 (9.109)
7
P,(E)< 1 Al)
b as =
of i=1 [1—-(@,+a_) /2}°
5 exp {—|
namely, P;,(E) = (1/2) erfe (/P,T,/No), is given by the numerator of the upper bound on
P,(E) in (9.109). Thus, we see that as long as the “eye” is open, that is, gj < g9 = 1, the
upper bound on the performance of the Viterbi algorithm for binary PAM with ISI of memory
] asymptotically approaches the no ISI bit-by-bit detection result as P,7),/No gets large.
Clearly, the larger the value of g;, the slower will be the approach to the asymptotic result.
Figure 9.18 is an illustration of this behavior where the upper bound on P,(£) as given in
(9.109) is plotted versus P,7)/No in dB with g as a parameter.
600 Communication over Bandlimited Channels Chap. 9
107k
10-2
10°3
(E)
F
ON
BOUND
UPPER
104
D = unit delay
eee eee
ece eee
(X) - eS és = (x) : eS -
the probability of error. Since however, the ISI is not Gaussian distributed, designing an
equalizer to satisfy this criterion is a difficult task and is typically not attempted. Instead, one
considers optimization criteria that lead to analytically tractable yet practically meaningful
results. In discussing the various equalizer configurations, we shall not place any constraint
on their implementation complexity.
The simplest (but not necessarily the best in performance) class of equalizers are the
so-called linear equalizers, which have a z-transform of the form
K
16. We note that, from the standpoint of a mathematical characterization, the tapped delay line representation of
a digital filter as in Fig. 9.19 is no different than the shift register representation such as that in Fig. 9.13.
602 Communication over Bandlimited Channels Chap. 9
Decision
Algorithm
1
Q(z) = ——
G@ (pny
fa )
De = Ce + Vy (9.112)
where {5x is a zero-mean complex Gaussian noise sequence with discrete power spectral
density
oo 2
S3(2)= > R5(m)z-" = EES PN ERIHK ahs {5e5f.m} O.113)
m=—00 G()
Note that although we have completely removed the ISI, the noise sequence is not white.
Thus, symbol-by-symbol detection is not the optimum decision algorithm; nevertheless, it
is still the most common algorithm used in receivers employing an LE-ZF because of its
simplicity.
The mean-squared error at the equalizer output is simply the noise power in 3, or
equivalently the contour integral of (9.113) around the unit circle defined by |z| = 1. If G(z)
is Zero Over some interval, then the equalizer Q(z) becomes infinite over that same interval
and hence the mean-squared error at the equalizer output also becomes infinite. Thus, a
condition on the existence of the LE-ZF is that 1 /G(z) be integrable. From (9.110), (9.111),
and (9.113), the above mean-squared error is given by
eee Po!
ei ncar LD {ax — | = E |
9.114
= Ry tO) ogee
q
v
—
90
— Co —_—
In n/t,
———ea
G (ef°%) @ fe
17. We shall assume here the most general case where the memory v can be infinite.
Sec. 9.4 Equalization Techniques 603
Example 9.2
By way of example, we shall consider a received pulse shape, A(t), that is causal and of the
following form:
2 ee
G(z)= oan lente 5 naehs (9.118)
(ee) ili ye) 42)
The corresponding equalizer is described by
that is, a finite impulse response (FIR) filter. The corresponding mean-squared error is from
(9.114) given by
2
SLE-ZF -~ 4a? (a? + B) (1 + e787)
B2 (1 — e287) (9.120)
E {lx - el? het Cm = 5: ¢z* denote the z-transform of the transmitted data sym-
k=—0o
5 (ee)
bol sequence and M(z)= >> jinz—* the z-transform of the noise sequence at the equal-
k=—0o
izer input. Since the data symbols are independent of one another, that is, the sequence {Cx}
is a zero mean “white” sequence with variance E {{cxl| A a then since
Da ea,
Ren) = E {eect = {@-' iC25 (9.121)
Furthermore, from the correlation properties of the noise sequence {/1,} given in (9.65), the
power spectral density of this sequence becomes
CO [e.@) 2 ;
lo.@)
Finally, writing the z-transform of the equalizer output sequence V(z)= > vUmz”™
m>=—OoO
as V(z) = C(z)G(z)Q(z), then the power spectral density of the error sequence {&,} =
{Vz — Cx} is given by
2 ; Te
85 (e/0™) = 5; (eT) |G(el™ Q(el#™) — 1] + 55 (e/°™) |Oe™)
f (9.124)
; 2
= 02 |Gel™) (ei) — 1] + 207G (e/™) |O(e!*™)
and hence the mean-squared error is
es Tk m/Ts , 2,
LE MSE 9NU /J_—zx/T, jo?G(el#™) Q(el*) — 1]
(9.125)
2
+20°G (ef®) |Q(eJ@1) |e
Completing the square in (9.125) with respect to G(e/®?s) and, for simplicity of notation,
omitting the argument e//* in the various transfer functions, we get
iy
éfe_wse = 5rn] (2ean IG? nam
+2076)
2
o-G
Since G(z) is real and even (since 9, = g_,), then G(e/®@") = G*(e/@"s), that is, IG/? = G?
and (9.126) simplifies to
sa
eas ae 5AY Nae tp
BadLE-MSE
2
bl ici= eae
57 (| G(o2G+2
2
(0: a”) ) |2 wih bese
2 c
02G +202
(9.127)
207o?2 r
o2G + 20? ri
Since both terms in the integrand of (9.127) are positive, the mean-squared error is mini-
mized by choosing Q so as to force the first term equal to zero at every value of w. This is
accomplished by selecting
af 1
(9.128)
2@)= o2G(z) + 202 <a G(z) + 202/02
Sec. 9.4 Equalization Techniques 605
(9.129)
2 Ts Bits 1
= 20° — : dw
2n J_x/1, G (e/27s) ae 202/02
(ste-mse),
2
,,<te—ze (9.130)
as would be expected. At high signal-to-noise ratio as measured by a2 /o”, the LE-ZF and
the LE-MSE give approximately the same performance. Also, for the LE-MSE the noise
enhancement is finite, and hence this equalizer always exists.
For Example 9.2, the LE-MSE would be
The performance of the linear equalizers considered in the previous two sections can be im-
proved upon by introducing nonlinearity into the design. One such structure is the decision-
feedback equalizer (DFE) [17, 18] which uses decisions on the previous data symbols in an
attempt to cancel out a portion of the ISI affecting the decision on the current symbol. The
primary advantage of the DFE is that it achieves its purpose with substantially less noise
enhancement than its linear counterpart. The main disadvantage of the DFE is that it suffers
from error propagation caused by feeding back incorrect decisions. Fortunately, however,
the error propagation effect is usually far less significant than the benefit obtained from the
reduction in noise enchancement.
The structure of a DFE is composed of two parts: a forward filter, Q(z), which
typically includes one of the two linear equalizers previously discussed, and a feedback
filter, P(z), which attempts to further reduce the ISI (see Fig. 9.21). Each of these filters
has a distinct purpose in the way it handles the ISI. The forward filter acts on the precursor
ISI (the samples of the overall system response occurring prior to the present time) in a
manner similar to that of the linear equalizer discussed previously. It makes no attempt to
equalize the postcursor ISI (the samples of the overall system response occurring after the
present time). Since the feedback filter operates on past decisions, it must be causal. Thus,
it has a discrete-time representation with only positive delay terms
K
P=) pez* (9.132)
k=1
and as such aids in cancelling the postcursor ISI. Since the forward filter only attempts to
equalize the precursor ISI, it can accomplish this task with less noise enhancement that if it
tried to equalize the entire system response. On the other hand, assuming that the decision
algorithm produces virtually error-free decisions, then the feedback filter produces no noise
at its output and thus the overall effect is a reduction in total noise enhancement.
wia}sks UOTBOTUNUWIWIOS
pazienbs yorqpady uoIsIoap SuIdoJ-019z & JO WIRIZIP YOo[G SWIN-a1019SIP YW 176 ANSI
eyeq g0In0S
Joqutés
wyzOZ[y
uoIsbaq
606
Sec. 9.4 Equalization Techniques 607
DFE’s are classified in much the same way as LE’s, that is, in accordance with the
optimization criterion applied to the forward filter. In this regard, we shall discuss the
so-called zero-forcing decision feedback equalizer (DFE-ZF) and the mean-squared error
decision feedback equalizer (DFE-MSE).
H+ (9.133)
where CE has the z-transform
error-free decisions, then (9.134) simplifies to C(z) = C(z) and likewise (9.133) becomes
Dp = Ce + & (9.135)
that is, the input to the decision algorithm is once again ISI-free as predicted above. The
noise sequence a rhowever, now has the discrete power spectral density [see (9.123) and
Fig. 9.21]
Eyer
Az)
° G(z)
ieet z ) (1+ PE) (9.136)
:
where we recall that G(z) = G(z7!) since gx is real and even. Comparing (9.135) and
(9.136) with (9.112) and (9.113), respectively, we see that the DFE-ZF has an identical sig-
nal component to the LE-ZF, that is, it is independent of the feedback filter P(z); however,
the noise component of the DFE-ZF is clearly affected by this filter. As such, we should be
able to choose the causal filter P(z) to reduce the noise at the input to the decision algorithm
relative to that of the LE-ZF wherein P(z) = 0.
The mean-squared error of the DFE-ZF at the decision algorithm input, namely,
2
edenop= E {li 2 el], is from (9.135) simply given by erp7p= E {| |To de-
termine the optimum P(z), that is, the one that minimizes this mean-squared error, we
608 Communication over Bandlimited Channels Chap. 9
first reinterpret (9.136) as the power spectrum of the output t of a filter Q2(z) =1+ P(2)
whose input is 3, with power spectrum as in (9.113). Based on the orthogonality principle
[19, Chapter 13], it can be shown that the particular filter P(z) that minimizes fore has
the property that the output of Q(z) is uncorrelated with the past of its input, that is,
K
Sian Oc a a ey (9.138)
a=
AP
Rz(m) SE |50m (9.140)
where again 5(m) is the Dirac delta function. Thus, we conclude that the equalizer output
roa,
noise sequence that minimizes Dae. =E a is a “white” sequence, that is, its
00 te
discrete power spectrum is a constant given by Sz Ges R; (m)z "SE |
Ee |
m>=—CO
In our discussion of noise whitening filters in Section 9.3.1, we considered partition-
ing the overall system response G(z) into a minimum phase, physically realizable part F(z)
and a nonminimum phase, nonphysically realizable part F(z~'), the former having all its
zeros and poles inside the unit circle and the latter having all its zeros and poles outside the
unit circle. To find the optimum forward filter, we again consider this same partitioning of
G(z) (actually a normalized version of it, as we shall describe shortly) and then associate
the filter (1 + P(z)) with the normalized minimum phase, physically realizable part and the
filter (1 + P(z*)) with the normalized nonminimum phase, nonphysically realizable part.
In particular, let
T] dG = Az74) [[ GQ — Az)
Z l
GQ =FOFR 2K SS (9.141)
[Jd — Xmz7") he — XmZ)
m m
where the A;’s and the Xm’s respectively represent the zeros and poles of G(z). Since from
the definition of P(z) in (9.132), (1 + P(z))|,~o9 = 1, then the correct association referred
to above would be
T]d= Aiz7!)
1
(9.142)
Mieka iho es
Sec. 9.4 Equalization Techniques 609
and thus from (9.136), the sequence {x has the power spectrum
2 2
Sp(2) = = (9.143)
which is white in accordance with the above discussion. Finally then, the minimum mean-
squared error is
fern JM 202 i
DFE-ZF) i, OK (9.144)
Note that (9.142) also determines the feedback filter by simply subtracting 1 from the right-
hand side of this equation.
Example 9.3
Considering Example 9.2, we immediately see from (9.118) that
K= B* (1 -n’)
7a? 3 Iba (gh ee) per ; ai= entat (9.145)
145
corresponding to the case of no zeros and one pole. Thus, the forward and feedback filters of
the DFE-ZF are given by
When compared with (9.120) for the LE-ZF we see that we have achieved a reduction in mean-
squared
q error by the factor 1 + e Ts The fact that (e205
DFE-ZF) <?,,_>,
LE-ZF can be shown [3]
to be true in a more general context than just this example.
The z-transform of the signal component, &s, A cx — Cx, of the total error signal &, is then
Finally, since the signal and noise components of &% are uncorrelated, then analogous to
(9.124), the power spectral density of the total error signal is
; 2 , ae Pu ie
ce {22G(elM)Q,(e/°%) — 1] + 207G(e!°™) |O1(e/2%)
| (9.151)
ENG
x 1ae (el )
eDFE—MSE = = | Qi(e/@") = i
{02G(el@*)
nes (9.152)
i ji+ P (ei) dw
é ; 2)
+207G (el?) | 1 (e/2%)
Because of the way we have partitioned the integral in (9.152) into a part (between
braces) dependent only on Q, (e/®7) and a part dependent only on P (e/®"'), we can
perform the minimization of ene a in two steps, that is, we first choose Q; (e/ots) to
minimize at each frequency the power spectrum between braces and then choose P (e/oTs)
to minimize the total power spectrum corresponding to the error signal itself. Comparing
the power spectrum between braces in (9.152) with the integrand in (9.125), we see that
we are dealing with identical minimization problems. Thus, without further ado, we can
immediately write the desired solution for the first section of the forward filter as [see
(9.128)]
|
Q\ (z) = (9.153)
G(z) + 202/ a2
which results in the term between braces in (9.152) being equal to the integrand in (9.129),
or equivalently
oS
2 xete iei‘ 7 20 2 L 1p (it) 2
DFE—MSE ™ 97 ever. G(eleTs) aE 207/02 dw (9.154)
The second step involves a spectral factorization similar to that performed for the DFE-ZF,
namely, we factor G’(z) 4 G(z) + 257 Ja? into
TT (i —Aj271)
l
1+ P(z)=————— 9.156
Td —x,2-5
m
Ce
from which we can determine the feedback filter P(z). Finally, the minimum mean-squared
error for the DFE-MSE is given as
(cire—wse),.,, = gr (9.157)
2 202
Sec. 9.5 Further Discussion 611
Example 9.4
Considering again Example 9.2, we have that
eae) gee
2 (a? + 62) (1—nz)(1—nz) o2
(9.158)
ei) (uz)
(1 — z-!) (1 = 072)
where
aE
6? (1=n?
1+ p2
UtA)_ KAM
a?
1+ pe?
o
It can be shown [3] that in general Saal PR (eer) _ (that is, K’> K) and
in min
furthermore (e
DFE-MSE} )i, <— (€?
\LE-MSE) )Jin ‘
sample at t = kT, + t,
received Transversal
signal | Receive Filter Decision
ne Equalizer Algorithm
G,(f) (2)
Coefficient
Adjustment
Algorithm
Training
Sequence
will be averaged (smoothed) out by the large time constants typically built into the adjust-
ment algorithm resulting in no net change in the equalizer coefficients. Finally, when both
noise and ISI are present, the adaptive equalizer will function on the average to bring the
input and output of the symbol decision algorithm into agreement.
The error signal produced from the difference of the data symbol decisions and the
equalizer output will be successful in adapting the equalizer provided that the data symbol
estimates are reasonably accurate, which in turn requires that the ISI not be excessively
large. While this condition is typical of the steady-state (tracking) behavior of an adaptive
equalizer, it is not typical during initial acquisition when the equalizer is learning the state of
the channel. Thus, at the outset of transmission (and perhaps periodically for reacquisition
purposes), it is common to send a training data sequence (assumed to be known at the
receiver) over the channel during which time the adaptation error signal is formed as the
difference of the known training symbol sequence and the equalizer output (see Fig. 9.22).
Upon conclusion of the training interval, actual data transmission begins, whereupon the
training sequence is replaced by the decision algorithm output for the formation of the error
signal.
Since the complexity of the equalizer and thus the speed with which it adapts is
directly related to the number of coefficients in its z-transform representation [for example,
the parameter K in (9.110)], it is only meaningful to talk about adaptive equalizers in
the context of an FIR. Hence, the second idealization mentioned above, namely, the lack
of a constraint on complexity, is automatically handled as a consequence of the need for
adaptation. However, one must re-solve the problem of finding the filter coefficients under
the various optimization criteria discussed in Section 9.4 subject now to the constraint
Problems 613
PROBLEMS
9.1 Following an approach analogous to (9.14)-(9.21), show that the Nyquist condition on G(f)
can alternately be expressed as
x (+z)
k=—oo
== AR soosm fT: IflS srl
tik ges
9.4 Consider a frequency response G(f) characterized by the convolution of an arbitrary (not
necessarily bandlimited but bounded impulse response) filter Gg(f) and a strictly bandlimited
filter
Gr f) = |2T costs
Tee Ih; IFSa7,
0; otherwise
that is, G(f) = Ga(f) * Go(f). Show that using G(f) for the overall channel response of a
communication system sending data at a rate 1/7; results in zero ISI from all but adjacent
symbols.
9.5 For the duobinary system illustrated in Fig. 9.10, show that choosing a detection threshold
halfway between “0” and “2A” minimizes the average bit error probability in the limit of large
SNR.
9.6 Show that, for the raised cosine class of Nyquist channels as described by the overall system
frequency response of (9.25), the received power P, and amplitude A are related by (9.40), and
the variance of the detected sample, yx, is given by (9.41).
9.7 A modified duobinary system is illustrated in Fig. P9.7a. As in the conventional duobinary case,
this system has an equivalent, as illustrated in Fig. P9.7b.
(a) Letting P(w) = P\(w) P2(w), and H(w) = H)(w) H2(o), find H(@) in terms of P(w) such
that the above equivalence is established, that is, in the absence of noise, y(t) is the same
in both.
614 Communication over Bandlimited Channels Chap. 9
y(t)
Precoder
to threshold
comparison
and: decision
logic
Figure P9.7a
to threshold
comparison
and decision
logic
Figure P9.7b
(b) If P(@) is the ideal brick wall Nyquist filter, that is, P(@) =7T; |w| <2/T and P(w) =0;
otherwise, find the overall channel response h(t) = F-'{H()}. (Hint: make use of the
fact that p(t) = F-!{P(o)} is a sin x/x function and then use the relation between H (w)
and P(w) found in part (a)]. Sketch H(w) and h(t).
(c) What are the values of h(t) at the sampling points t = nT, that is, how many have nonzero
value and what are these values?
(d) Based on the result of part (c), in the absence of precoding, which data bit(s) contribute ISI
to the detection of ag?
(e) As in duobinary, error propagation will also occur for modified duobinary unless precoding
is used. Find the appropriate precoder and sketch its configuration.
(f) For the precoder of part (e), what is the appropriate decision rule for determining {a,,} from
the samples of y(t)?
(g) Assuming H(q) is split equally between Hj(w) and H2(w), find the error probability for
the modified duobinary system using the results found for parts (b)-(f). Assume n(t) is
white Gaussian noise with single-sided spectral density Ng Watts/Hertz.
9.8 A system is designed for a quadrature partial response (QPR) modulation, in particular,
quadrature modified duobinary, that is, modified duobinary (Problem 9.7) on quadrature car-
riers transmitted over a common channel. The transmitter for such a modulation scheme is
illustrated in Fig. P9.8a. Binary (+1) data at a rate 1/7) is split into its even and odd compo-
nents, a, and b,,, respectively, each occurring at a rate 1/T = 1/27). These two data bit streams
are precoded and formed into baseband streams with pulse shaping equal to that produced by a
modified duobinary transmit filter, that is
-( oT ua
Hiloy= JOT sate Ne): lo| <4iP
otherwise
The two baseband data bit streams are amplitude modulated onto quadrature carriers and the
sum of the two sent over the channel as the transmitted signal
Problems 615
AY 5(t-nT)
n=—0o
Binary
Source A Ss5(t—nT)
Rate 1/T, n=-0o
Figure P9.8a
sample at t= nT
Figure P9.8b
CO [o,@)
Assume an additive white Gaussian noise channel with single-sided power spectral density No
Watts/Hz. The received signal plus noise is first demodulated with quadrature carriers
and then passed through the modified duobinary receive filters H2(w) = H\(@). The outputs
of these filters are sampled at times tf = nT (n integer) and passed through full-wave recti-
fied threshold comparison devices typical of modified duobinary baseband modulation. The
detected bits are denoted by a, and Dns respectively, as shown in Fig. P9.8b.
(a) Assuming w,T >> 1 (that is, ignore all second harmonic carrier frequency terms), find the
bit error probabilities for the in-phase and quadrature data sequences, that is, P(E)|q =
616 Communication over Bandlimited Channels Chap. 9
n(t) sample at
¥,4,8(0-n2)
Figure P9.9a
Figure P9.9b
Pr {Gn # an} and P(E)|, = Pr {brSOx |.Express your answer in terms of the bit energy-
to-noise ratio E,/No of the system where Ep, = P;,,7; is the bit energy with P;, the input
power to the channel, that is, the average power of s(t).
(b) If a pair of bits a, and b, is thought of as having sent the complex symbol x, =a, +
jbn, evaluate the symbol error probability of the system, that is, P(E)|x = Pr {X, # xn}.
Express your answer in terms of the bit energy-to-noise ratio E,/ No of the system.
9.9 A pulse g;(t) = ie {G,(f)} is designed for transmission over a given channel G,(f) per-
turbed by additive white Gaussian noise so that the overall system response go(t) = F—!
{Go(f)} = F7! {Gi(f)Ge(f)G,(f)} has no ISI. In particular, let Go(f) satisfy the Nyquist
condition of Eq. (9.22). Furthermore, assume that G,(@) = (G;(w)G_-(@))* so that the er-
ror probability is minimized. Suppose now that the channel has a somewhat different transfer
function, say Gf ), than that expected in the design such that the overall system response
g(t) =F Gif} = FUNG CAG AGG} now contains ISI at the sample time instants.
It is desired to pass y(t) through a zero-forcing linear equalizer (LE-ZF) with transfer function
Q(f) to restore the zero ISI condition at a finite number of sample times in the equalizer filter’s
output pulse response h(t) = F “1G f)Q(f)}. The LE-ZF has the form of a transversal filter
with three taps as illustrated in Fig. P9.9b. Samples of the pulse shapes are illustrated in Fig.
P9.9c.
(a) Find a matrix relation that when solved yields the tap coefficients g_}, go, gq; in terms of
the samples ho = h(O), hj = h(T;), hz = h(2T,) and an appropriate set of samples of g(t)
taken at integer multiples of T;, namely, g; = g(i7;). How many and which samples of g(t)
and h(t) are needed?
(b) If the equalized pulse is to have the same nominal delay as the equalizer, that is, h(t) =
go(t — T;) at the three sample points t = 0, 7;,27;, solve the matrix relation found in
Problems 617
8 (t)
Figure P9.9c
quences are real. Define the MSE for the k" equalizer output sample by € = E {(vx — ax)*}
where a, is the desired k'" data symbol. Note that if a, is binary (+1), then oa? = |. Since € is
a concave function of the tap weights {q,}, a set of sufficient conditions for minimizing € is
(a) Show that the above set of conditions leads to the following set of relations:
where Ryy(m) A E {yx—mvx} is the cross-correlation of the input and output equalizer se-
quences and Ryq(m) A E {yx—mag} is the cross-correlation of the equalizer input sequence
with the data symbol sequence.
(b) Using the convolution relation between the input and output equalizer sequences in terms
of the equalizer tap weights, show that the set of conditions of part (a) can be expressed in
the following matrix form:
Ryq= Ryu
where
q-K Rya(—K)
q-K+1 Rya(—K + 1)
q= : » Ry= : ,
qK-\ Rya(K = 11)
qK Rya(K)
Ry(0) Ry() vee Ry(2K)
R,(-1) Ry (0) Ry(2K — 1)
R, = : : ; :
with Ry(m) & E {yx—myx} equal to the autocorrelation function of the equalizer input se-
quence.
(c) Assume now that the equalizer input sequence is as specified in (9.66), namely y, =
v
Y> giax—1 + Ux, where 2v is the number of ISI components), {a,x} is an uncorrelated data
l=—v
sequence with variance GF and {jx} is a zero-mean Gaussian correlated noise sequence
with correlation function R,,(m) = E {ux—mULk} = O'S n2 Also, assume that go = 1. Show
that the elements of the matrix Ry and the column vector Ry, are given by
Rya(m) = ae
from which the tap weight vector q can be uniquely found using the result in part (b),
that is
q— Rak.
9.11 Consider the design of an LE-MSE in accordance with the development of Problem 9.10.
Letting Ko=p'= 1, o = 1, o? =.2 and an ISI channel described bype= 220) ee
—.1, find the optimum tap weights q_1, go, 41-
9.12 Consider a zero-forcing linear equalizer (LE-ZF) for use in a system with an overall system
response of the form
p |>
81
80
as the normalized ISI contributed by each of the two adjacent pulses.
References 619
(a) Show that the necessary Fourier transform condition G (e/®") > 0 for all w implies that
lo| < 1/2.
(b) Show that G(z) can be factored in the form
¥ mare
A eS ae a)
a)
where p A — (a + aa)?
(c) Assuming without loss in generality that |a| < 1, find a partial fraction expansion for the
equalizer filter Q(z).
(d) Expanding each of the partial fraction terms in the answer of part (c) in a power series in z
or z_' l as appropriate, show that the zero" equalizer tap go can be expressed as
Je a
pgo (1 — a?)
(e) Evaluate the mean-squared error of the LE-ZF using your answer found in part (d). Express
your answer only in terms of go, a7, and p.
(f) What is the value of the mean-squared error as p > +1/2 or equivalently a > +1? What
happens to the gain of Q(z) at these limiting values and what does this result in in terms of
noise enhancement?
9.13 Consider now a mean-squared error linear equalizer (LE-MSE) for use in a system with the
same overall system response as in Problem 9.12 and the same ISI condition, namely g_; = g}.
(a) Show that the equalizer Q(z) is now given by the reciprocal of the equivalent overall
system response
G'(z) = pgo
(c-a’)(2-a'-!)
where p’ A p (1+ 22") A (a’ ally .
(b) Expressing Q(z) asa partial fraction expansion and expanding each of the partial fraction
terms in power series in z or z~! as appropriate, show that the zero" equalizer tap go can
be expressed as
qi = — SO
a /
pFO
/
REFERENCES
1. Lucky, R. W., J. Salz, and E. J. Weldon, Jr., Principles of Data Communication, New York:
McGraw-Hill, 1968.
2. Proakis, J., Digital Communications, 2nd ed., New York: McGraw-Hill, 1989.
3. Messerschmitt, D. G., and E. A. Lee, Digital Communication, \st ed., Boston: Kluwer Academic
Publishers, 1988,
4, Nyquist, H., “Certain Topics in Telegraph Transmission Theory,” A/EE Transactions, vol. 47,
1928, pp. 617-44.
5. Kretzmer, E. R., “Generalization of a Technique for Binary Data Communication,” JEEE Trans-
actions on Communication Technology, vol. COM-14, no. 1, February 1966, pp. 67-68.
620 Communication over Bandlimited Channels Chap. 9
6. Lender, A., “The Duobinary Technique for High-Speed Data Transmission,” JEEE Transactions
on Communication Electronics, vol. 82, May 1963, pp. 214-18.
. Viterbi, A. J., “Error Bounds for Convolutional Codes and an Asymptotically Optimum Decoding
Algorithm,” JEEE Transactions on Information Theory, vol. IT-13, April 1967, pp. 260-69.
. Papoulis, A., The Fourier Integral and its Applications, New York: McGraw-Hill, 1962, Chap-
ter 4.
. Duffy, F. P. and T. W. Tratcher, “Analog Transmission Performance on the Switch Telephone
Communications Network,’ Bell System Technical Journal, vol. 50, April 1971, pp. 1311-47.
. Pasupathy, S., “Correlative Encoding: A Bandwidth Efficient Signaling Scheme,’ JEEE Commu-
nications Magazine, July 1976, pp. 4-11.
. Kabal, P. and S. Pasupathy, “Partial-Response Signaling,’ JEEE Transactions on Communica-
tions, vol. COM-23, September 1975, pp. 921-34.
. Bennett, W. R., Introduction to Signal Transmission, New York: McGraw-Hill, 1970, p. 16.
. Forney, G. D., Jr., “Maximum-Likelihood Sequence Estimation of Digital Sequences in the Pres-
ence of Intersymbol Interference,’ IEEE Transactions on Information Theory, vol. IT-18, May
1972, pp. 363-78.
. Forney, G. D., Jr., “Coding System Design for Advanced Solar Missions,” Final Report on Con-
tract NAS 2-3637 for NASA Ames Research Center, Watertown, MA: Codex Corp., December
1967.
. Forney, G. D., Jr., “The Viterbi Algorithm,” Proceedings of the IEEE, vol. 61, no. 3, March 1973,
pp. 268-78.
. Viterbi, A. J., and J. K. Omura, Principles of Digital Communication and Coding, New York:
McGraw-Hill, 1979.
. Austin, M. E., Decision Feedback Equalization for Digital Communication over Dispersive
Channels, Lexington, MA: MIT Lincoln Laboratory, August 1967.
. Belfiore, C. A. and J. H. Park, “Decision Feedback Equalization,’ Proceedings of the IEEE, vol.
67, no. 8, August 1979, pp. 1143-56.
. Papoulis, A:., Probability, Random Variables, and Stochastic Processes, 2nd ed., New York:
McGraw-Hill, 1984.
. Salz, J., “Optimum Mean-Square Decision Feedback Equalization,” Bell System Technical Jour-
nal, vol. 52, October 1973, pp. 1341-73.
. Lucky, R. W., “Automatic Equalization for Digital Communications,” Bell System Technical
Journal, no. 44, April 1965, pp. 547-88.
. Widrow, B. and S. D. Stearns, Adaptive Signal Processing, Englewood Cliffs, NJ: Prentice-Hall,
1985.
- Qureshi, S. U. H., “Adaptive Equalization,” Proceedings of the IEEE, vol. 73, no. 9, September
1985, pp. 1349-87.
Chapter 10
Demodulation
and Detection of Other
Digital Modulations
Thus far in the book, we have focused our attention on the characterization of traditional
digital modulation schemes and their performance over the AWGN channel. By traditional
modulations we mean those that are treated in most textbooks on the subject of digital com-
munications, although not perhaps in the depth to which they are covered here. While in
many of today’s applications these schemes still provide adequate performance, there ex-
ist other applications for which modulations having greater flexibility and sophistication are
absolutely essential. An example of an application requiring increased flexibility might be a
satellite channel over which it is desired to simultaneously send two independent and gener-
ically different types of information. In this instance, we would need a modulation scheme
capable of handling two asynchronous data streams with, in general, different rates, dif-
ferent powers, and different formats. Both the Tracking and Data Relay Satellite System
(TDRSS) and Space Shuttle make widespread use of such modulation schemes for com-
municating various combinations of operational, scientific, and telemetry data. An example
of an application requiring increased sophistication of design is today’s commercial mo-
dem market where baud rates as high as 19.2 Kbps are being transmitted over a nominal 5
KHz bandwidth telephone channel. Here the focus is on modulation schemes that achieve
a high degree of bandwidth conservation with as little expenditure of additional power as
possible.
For the most part, all other digital modulations that meet the above requirements fall
into two primary categories: inphase-quadrature (I-Q) modulations and continuous phase
modulations (CPM), the latter most often being represented by a constant envelope signal
with digital frequency modulation. In certain special cases, CPM also has an I-Q represen-
tation and thus the two categories are not always divorced from one another. Nevertheless,
it is convenient to separate them this way since the generic implementation for conven-
tional CPM is in the form of a single carrier which is frequency modulated by a waveform
that maintains phase continuity from symbol interval to symbol interval. More about this
later on.
622 Demodulation and Detection of Other Digital Modulations Chap. 10
pn
ig std
v2 cosa,t
Matched
Filter a;
(P, (t))
v2 cos @,t + 8)
r(t)
where p;(t) and p2(t) are unit power pulses of duration 7| and 7, respectively, {a;} and {b;}
are +1 independent sequences at rates 1/7 and 1/7, respectively, and v1,v2 are random
independent timing epochs to allow for asynchronization, then after passing through the
I and Q matched filters and assuming 1/7), 1/72 « B, we obtain the output samples (at
times t = (i + 1)7; and t = (i + 1)7)
Pyi(E)
pI — feDace
oewoi eeL
No en
=a 2,) (E) =— 5| orferic |E22
No (10.8)
.
where Ep; = Pi 7; and Ey2 = P2T> are the respective energies per bit on the I and Q chan-
nels. Thus, in the ideal phase coherent case, the bit error probability performance per chan-
nel for UQPSK is identical in form to that of balanced QPSK.
We now consider the case where a carrier phase error exists between the received
signal and the I and Q demodulation reference signals. In particular, assume now that the
624 Demodulation and Detection of Other Digital Modulations Chap. 10
€9(t) = J Prdx(t) cos@ + J Pid (t) sing +ns(t) cos6 — n-(t) sin 0
where ¢ 46 — 6 is the carrier phase error assumed to be constant” over the observation
interval pertinent to making decisions on d\(t) and d2(t).
Consider representing d)(t) and d(t) as in (10.6). After passing through the normal-
ized I and Q matched filters, we obtain the output samples [at time ¢ = (i + 1)T]
where
1 (i+1)7T) +, ;
Hi=— dy(t)pi(t — iT, — v)dt
Ti Jit, +)
1 (i+1)T2+v2
i= s- d\(t) p2(t — iT2 — v2)dt
P T2 JiT+v2
| (10.12)
1 (i+1)T+v~
Nek = => Ne(t) pelt — iT; — vy)dt; k=1,2
Tk UTK+V
1 (+1) Ty +v%
Ns k => ns(t)px(t
—iT; —vyjdt; k=1,2
T Th +V%
Here N, and N>2 are zero-mean independent noise samples with variances Oe = No/2T; and
os = No/27?, respectively. We note from (10.11) that, as for balanced QPSK, the presence
of carrier phase error introduces crosstalk into the detection process, that is, the data in one
channel affects the detection of the data in the other channel. The nature of the crosstalk,
however, now depends on both the powers in the I and Q channels as well as the I and Q
data formats through the definitions of d),2 and d2,; in (10.12). To assess the effect of this
crosstalk on uncoded bit error probability performance, we proceed as follows.
2. In Volume II of the book, we shall introduce other phase error variation models wherein the phase error can
vary in time over the duration of a data decision. For now, we consider only the so-called slowly varying phase
error case.
Sec. 10.1 |-Q Modulations 625
2 1 Pris PoT,
oe a eet Fi sno
\ No \ No
(10.14a)
i _ [Pe
ae qerfe
ote| pastue
ese— |
nd
anne]
and
: #bi|o} = je
Po( E1) = Pr {6; 1 PoT cos
[7 Pi
+ Feaiasn|
(10.14b)
1 | P2T> | P;T> :
+ —erf
Te | a No cos @ — .|—d
No 1,.2sing
where the overbar denotes averaging over the data sequences and uniformly distributed
random epochs v, and v2 in d; 2 and d2 1.
In principle, (10.14a) and (10.14b) can be evaluated exactly for given pulse shapes
pi(t) and p2(t), that is, data formats on the I and Q channels. This, however becomes a
tedious calculation when, for example, the ratio of data rates R2/7R, is large? since the
statistical average over the data sequence required in Pp( E| d) of (10.14a) will then require
evaluation of dz; for each possible sequence d2(t) in the interval nT) + vj <t <(n+
1)T; + v;. Furthermore, in order to ultimately calculate the I-channel average probability
of error [by averaging (10.14a) over the probability density function f(@) of @], these
statistical averages over d2(t) must be computed for each value of @ in the domain of f(@).
To get around this computational bottleneck, a scheme was proposed in [1] that ex-
pands Pp;(E| ); i = 1, 2 into a power series in @ whereupon maintaining only the first few
terms, the average error probability P,; for the I and Q channels is easily evaluated as a func-
tion of E4;/No, E2/No, the mean-squared crosstalks dS and rae and the mean-squared
phase error, O%. For small values of OG, as is typically the case of interest, this scheme gives
credible results. We now proceed to describe this simplified evaluation technique.
From (10.14a) and (10.14b), we note that P,;(E| @); i = 1, 2 is of the form
———— en
PACE oC) = gore [Aj cosf + &; sing] + gone [Aj cos@ — & sing] (10.15)
where A; is a constant and &; is a random variable in the sense that it is both data sequence
3. With no loss in generality, we shall assume that the data rate 722 = 1/7} is arbitrarily chosen to be greater
than or equal to the data rate R, = 1/7}.
626 Demodulation and Detection of Other Digital Modulations Chap. 10
and random epoch dependent. Differentiating P,;( |) twice with respect to @ and evalu-
ating the results at ¢ = O gives
APHCEIO) | we Ae aye
oe a gion (-A7) =°
; (10.16)
a>2 Pyi(El~)|
Py; | _ Aj Aiexp(~Aj)
— A; [1+287]
as
de> ey 35 Ja
Furthermore, from (10.15)
1
Pri( E| ’)|g=0 = moe Aj ONT)
which corresponds to the bit error probability performance in the absence of carrier phase
error. Thus, expanding (10.15) in a Taylor series about ¢ and keeping only the first three
terms gives, upon substitution of (10.16) and (10.17)
d et——)
Pyi(E o a
51 d* Py; E Pe
= Pr
E| @)SS
Pyi( b)|g-0 +——
Eli(
es (10.18)
=a _, Aiexp(—A 4)
rie AGS 25a
Comparing (10.15) with (10.14) and identifying A; and &; in terms of the receiver parame-
ters, (10.18) becomes
Ep, E oe E a=
Pri (E| 9) = erfe |= + sy nen )
Bt) f+ 2S yeh |
: (10.19)
(10.21)
Pia(E) % 5erfe Epba E
Fb2 ie2) rey Ene
eee o3
Noa No No yr
Note that (10.21) consists of the sum of a term representing perfect carrier synchronization
performance [see (10.8)] and a term proportional to the mean-squared crosstalk and mean-
squared phase error. What remains is to evaluate the mean-squared crosstalks ds , and a. 2
Sec. 10.1 |-Q Modulations 627
for various combinations of data formats in the I and Q channels. From the definitions of
dz,; and d;,2 in (10.12) and the approach taken in Chapter 2, Section 2.1, it is straightfor-
ward to show that
zz (10.22)
where Sini(f) and Sm2(f) are the power spectral densities of the modulations d)(t) and
dz(t), respectively. In [1], closed form expressions for a > are evaluated for all combina-
tions of Manchester and NRZ coding on the I and Q channels. The results are summarized
in Table 10.1.
In (10.24), p(t) is a unit power pulse of duration 7;, 7; is the symbol time which is related
to the bit time by 7; = 2m7), and the data amplitudes {a;} and {b;} range over the set of
values +1, +3,...,+/M — 1. Note that for M = 4, QAM is identical to QPSK.
4. In the special case where the I and Q amplitude modulations correspond to amplitude-shift-keying (ASK), the
quadrature multiplexed form is often referred to as quadrature amplitude-shift-keying (QASK).
628 Demodulation and Detection of Other Digital Modulations Chap. 10
Assume now that (10.26) is transmitted over an AWGN as characterized by (10.2) and
in addition the channel introduces a constant phase into s(t) producing
s(t, 0) = A [d,(t) cos (Wet + 8) — ds(t) sin (wet + 8)] (10.26)
Thus, the received signal is the sum of (10.2) and (10.26), namely, r(t) = s(t, 0) + n(t).
Coherent detection. In the case of ideal coherent detection, the received signal
is first demodulated (see Fig. 10.3) by the pair of quadrature carriers in (10.4) producing the
lowpass signals
E-(t) = Ad-(t) + n-(t) cos @ + ns(t) sind
(10.27)
&s5(t) = Ad, (t) +ns(t) cos @ — n-(t) sin@
After passing through the I and Q matched filters and assuming 1/7; < B, we obtain the
output samples [at time t = (i + 1)T;]
Ad, (t)
2 COS @,t
S(t)
Ad,(t)
Figure 10.3a A QAM modulator
Matched
Filter &
(p(t)
v2 cox @,t + A)
r(t)
Pya(C) = Pr {Aq — 1) < Voi < AC + 1); AG — 1) < Ves < AG + D)}
=Pr{-A< Noi <A;—-A< Ng < A}
2
1 vf AZT; 1 f AZT;
= =5 eric a No —=—s—5 CLL No (10.29)
2
A2T.
Lo lithcveete =} | =('—2P;(B))*
630 Demodulation and Detection of Other Digital Modulations Chap. 10
Pyg(C) = Pr {A(ai — 1) < Voi < 00; Ai — 1) < Vs,5 < Ai + ‘DY
=Pr{—A< Noi <00; -A<Nyi < A}
AT. 1 AT, 1 55 Wer .(10.30a)
=| —5 erfc eee
No =5 erfc
erfi No 5 e c | ,/——
No
= (1 — Ps (E)) i — 2Ps1(E))
or
P,p(C) = Pr {A(ai — 1) < Voi < A(ai + 1); Ai — 1) S Vs,i < 00}
— Pr {—A SING SAP SANG oo}
P,(E) =1 — P,(C)
=1—— | {4 — P(E)? 2 +4 (VM 2)d= P(E) = 2P (BD)
1 3 E,
( ra) (cara) #)
=) >) ——— F =
Se ee
where we have substituted for A? in terms of P from (10.25) and as in previous chapters
E;/No & PT;/No denotes the average symbol energy-to-noise ratio.
It is interesting to observe that (10.32) can also be expressed in terms of the average
error probability performance of M-AM , that is, the individual modulations on the I and
Q channels. Since a correct QAM decision is made only when a correct decision is made on
both the I and Q channels, then
or, equivalently
2
P;(E)|m—qam = 1 — [1= FAC ae
(10.34)
1
=2P;(E)| az_am (1~ 5 PAE apna)
which, using the result for the symbol error probability of M-AM [see (4. 167)}° and replac-
ing M by VM and E, by E,/2, demonstrates the equivalence.
To obtain the average bit error probability of QAM from the average symbol error
probability of (10.32), we observe that QAM can be assigned a perfect 2-dimensional Gray
code [2], that is, for each symbol error in the set, all adjacent symbol errors cause only a
single bit error. Thus, since the number of bits/symbol is log, M, then for large SNR we
have the simple approximate result that
P,(E) = 5
Go) US) (10.35)
og,M sm
where in addition the symbol energy-to-noise ratio E;/No is equal to 2m times the bit
energy-to-noise ratio E,/No. The values of E»/No corresponding to P, = 10-> are tabu-
lated in Table 10.2.
An exact evaluation of bit error probability can be obtained by considering the number
of bit errors that are associated with the QAM symbols having each of the symbol error
probability types, P;4(E), Psg(E), and Psc(E) and then averaging over all possible cases.
For arbitrary M, it is tedious to express the result in closed form. However, for any particular
value of M, an answer can be obtained. For example, for M = 16 and M = 64 and a Gray
code mapping (see Chapter 4), it can be shown that
where we have furthermore replaced Ey, by Es for consistency with the usage in this chapter.
632 Demodulation and Detection of Other Digital Modulations Chap. 10
Table 10.2 Bandwidth and Power Efficiencies of Various Modulation Schemes (P;, = 10-5,
Values of E,/No are in dB)
I Ne Re a GS SS iad =A a Sl naeLt Bat 2
Constant Envelope
QFPM CPFPM N-FSK/
QFPM (N-FSK/QPSK) CPQFPM (MFMSK) BPSK QAM _ M-PSK
Se Aa Ce MR aie E E; E
ee E
LIEy
41.0 9.60 1.0 9.60 0.67 9.60 0.67 9.60 0.76 9.60 1.0 9.6 1.0 9.6
Gree, ae ee kG 8.31 2. "0.60. 83 0830831. = salem
16 1.33 9.60 0.8 7.36 0.80 9.60 0.44 7.36 0.71 7.36 2.0 13.4 2.0 17.4
BOG ae yt RKONSSE 67! = (Sv"0.39" 6.700.5206.7 1 =o = es
6401222831 035. rele 0.67 8.31 0.18 6.16 0.34 6.16 3.0 17.8 3.0 27.5
256 0.89 7.36 0.12 5.36 047936 ee ee
1024 0.59 6.71 0.04 4.80 0:30. 6.71 = See 5 RTS es Onion
4096 0.36 6.16 0.01 4.40 O18 6.16) =) —) 22260. 57.61a0NGes
and
Ey 1 Ep
P,(E) = — erfc — }+ —erfc {|3,/ —
lB) = 74 (i, 4 ( i,
(10.36b)
where Noi and Nii are independent zero-mean Gaussian noise samples with variances
No/2Ts.
6. This observation was made in [5] for the special case M = 16.
Sec. 10.1 |-Q Modulations 633
To compute the conditional probability of symbol error we again consider the proba-
bility of a correct decision (conditioned on @) for each of the three different types of points
in the first quadrant of the constellation. Defining
P,(C|j,k, @) =Pr {AG —1)< Voi < AG H+ 1); AK—-1) < Vi < AK +)
=Pr{A[—1+ j(1 —cos$) +ksing < N,; <1+ j(1—cos¢) +ksing];
1 | A2T, A‘ ;
= 5 erfc No 1411-6) +4sne
1 AZT; :
— 5 erf No (1+ jC —cos¢)
+k sing) (10.39)
1 AZT;
(—1+k(1 —cos¢) — j sing)
» No
1 A2T;
(1+k(1 —cos@¢) — j sing)
2 No
k=1,3,...,(VM -3)
634 Demodulation and Detection of Other Digital Modulations Chap. 10
1071
10-2
10-3
10-5
0 2 4 6 See t0M 124 Sal eS ae20 Figure 10.4 Exact bit error rate of 4, 16,
and 64 QAM
or
P,(Clj,k, 6) =Pr {AG 1) < Voi SAG +); AWM - 2) < Vs, < 00]
2 ferfe JAE (-1+ id ~ cos) + (VM ~ 1)sind)
1 AZT;
x | —erfc (—1+ (VM — 1)(1 ~ cos) — j sing) ;
2) No
j=1,3;...,(¥M-3)
Finally, for the single corner point, the conditional probability of a correct decision is
Sec. 10.1 |-OQ Modulations 635
P,(C| j,k.) =Pr {ACM - 2) < Voi <00; A(WM — 2) <V,; < 00]
From the symmetry of the signal set and the decision regions, the following properties
hold:
Ps (C| j,k,
b) = Ps (C| — j, —k,)
Ps (C| j,k, 6) = Ps (C| j, —k, —¢) (10.42)
P;(C| j,k, 6) = Ps (Clk, j, —?)
Bs, VM-\1/M-1
us j=1,3,..k=1,3....
isk
Gh,
10.45
Deter a= (O ( )
Substituting (10.40a), (10.41), and (10.42) into (10.44) and combining like terms gives the
final form of the desired result, namely [3]
636 Demodulation and Detection of Other Digital Modulations Chap. 10
; +(/M-1) +(VM— 2) 4
P.(E|d) = 1 — P(C1$) = = ps oe
j=
x ate Gren. e+ oe
oo — 1 (10.46)
M hale A lOc Dyers 2
a 5 a
«ete (<a)
la
Ne aa
No (k+(U—k)cos¢+
’
(1 vt ind)
1 3 :
< fete Gaa)e d+ —1)cos¢ — sing)
The conditional error probability of (10.46) is plotted in Fig. 10.5 for M = 16 with A=
- wo i a as a parameter. The limiting cases of (10.46), namely, A — 0 and A — ov, are
also illustrated in this figure and are mathematically described by
1
Ae P,(E|¢) =1— v (10.47)
+(VM-1) +(/M-2)
lim pws) ee Sy [5 588m + (100s ~ jsind)|
Aes ee Mee iat ea eee
+(VM-2) +(/M-2)
4 1 1
— — Ly [p- pen +0 Heose-+d— Dsing)|
M 0, 50,44,...1=0,4044,.
Note that the limiting curve for A — oo exhibits discontinuities at values of phase error
at which the four signal points in the first quadrant move across a decision boundary (see
Fig. 10.6), namely, @; = 16°53’, 2 = 18°26’, 3 = 20°48’, and ¢5 = 45°. The absence of a
discontinuity at 4 = 32°20’ is explained as follows. For a phase error of value 3, the signal
point at (3A, A) moves across a decision boundary and into the decision region of the signal
at (3A, 3A), thus resulting in an error in its detection (see Fig. 10.6). When the phase error
increases to #4, the signal point at (3A, A) moves across another decision boundary and into
the decision region of the signal at (A, 3A). However, this still results in an error insofar as
detecting the transmission of the signal point at (3A, A); thus, there is no discontinuity in
the conditional error probability at dg.
In order to calculate the average error probability performance of QAM in the pres-
ence of carrier phase error, we must characterize the pdf f(@) of this phase error. Since this
characterization in turn depends on the particular carrier synchronization loop used, we shall
delay our discussion of average error probability performance until Volume II of this book,
Sec. 10.1 |-Q Modulations 637
Se Soe ee
(E\9)
Pg
0 0.2 0.4 0.6 0.8 1.0 Figure 10.5 Conditional error probability
4 d performance of 16-ary QAM (Reprinted
Tt from [3] © IEEE)
where we shall present carrier synchronization loops for QAM and their associated f()’s.
At that time we shall be in a position to average (10.44) over these pdf’s.
Vi A Vek + UVs.k = A (ae + fbx) cosO + A (—b_ + jax) sind + Neg a ING,
(10.48)
A Acpe*el® + N;
1 1 =19
fy\s,o V\|s,(Vs, 6)
@) = ———-
Q DN, exp ——|\VEV
5 7 | | (10.50)
Using the definition of V, the argument of the exponential in (10.50) can be expressed as
Im {Vs*"}
of Atay ee 10.52
Re {Vs*"} ( )
Substituting (10.51) into (10.50) and averaging over the uniform pdf of 6, we get
4
(10.53)
= a
Qo)"
x {515
20;
(IMP? (05-2 vs"? )
+I?)|10aes
7. For notational convenience here, we number the elements of an N,-length sequence vector by 0,1,...,N;—1
rather than by 1, 2,..., Ns. A similar notation applies in two dimensions to matrices.
Sec. 10.1 |-Q Modulations 639
where as before /o(x) is the zero order modified Bessel function of the first kind. Since the
natural logarithm is a monotonically increasing function of its argument, then maximizing
fvjs(V|s) over s is equivalent to maximizing In fyjs(V|s) over s. Hence, eliminating all
irrelevant terms (that is, those that are independent of the transmitted data), the following
decision rule results: Choose the sequence s that maximizes the metric
n= saa?
e +In Io (o,,*)vstT
1 N;-1 Ns—1
(10.54)
Fane? pure, 4 int (00 DOV pcicni
On i=0 i=0
Unfortunately, as was true for the analogous result for M-PSK derived in Chapter 7, the met-
ric of (10.54) is ambiguous in phase in the sense that the argument of the Bessel function
is unchanged by a fixed phase rotation. Thus, even in the absence of noise, it is possible to
make an erroneous sequence decision. The simplest case of this to envision occurs where the
phase rotation is a multiple, say J, of 90° in which case a particular sequence of QAM sym-
bols is mistaken for another sequence whose symbols correspond to a rotation of the true
symbols by J quadrants. To resolve this ambiguity, we again propose the use of differential
encoding at the transmitter. Thus, if AQ, now denotes the phase of the k" transmitted com-
plex QAM symbol, that is, s, = cye/4%, then the actual phase transmitted over the channel
in this same interval is given by the usual differential encoding relationship
Note that unlike M-PSK, the phases {6,} do not necessarily take on values corresponding
to QAM symbols. This, however, is of no consequence since indeed (10.54) holds for
any AM-PM constellation. Finally, then, substituting (10.55) into (10.54) gives the desired
unambiguous metric
Nea
Ns =i) 3 AO—i—m
(10.56)
1 iba
Using the approximation (valid for large arguments, that is, high SNR) In Jo(x) = x, then
after normalization by 0,7, (10.56) becomes
Ns—i-2
| Ns—1 j >y A®k—i-m
n=—5 Dy cgay eS Vi-ice-ie —"™=9 (10.57)
ji i=0
Analogous to Fig. 7.7, a direct implementation of a receiver that employs a decision rule
based on the metric in (10.57) is illustrated in Fig. 10.7. Note that the receiver uses one
extra complex amplitude, namely, cx—n, + 1, to make its decision on cy—j, A@-i} 1 =
Ge SINae:
Analysis of the average bit error probability performance of (10.57) is difficult. Thus,
we present in Fig. 10.8 computer simulation results typical of this performance for the
case of 16-ary DQAM. Also illustrated for comparison is the performance of ideal coherent
detection of QAM as given by (10.36). We see that despite the use of a maximum-likelihood
a[dure
13s= “11
peyoie,y 7) Aeylaq "yy Aejaq
49A!4 : l Aepoq myPr keoq Ny
Zi
640
SSoI
= ppYCal
‘ boned “IN
Y w—1-¥
Ip= x4 9
;
WOU |
———
edojaaua
10}99}9p
oredwo5
pue
199T9S
91
WN sapako D,
JadN ‘L aqeue kyafoA9 out)
Aeyap
SOV. a VV
7
v A
v
v
vv
10-1
10-2},
10-4
receiver, the cost (in E,/No) of using DQAM relative to coherent QAM is quite significant,
although one can achieve a good reduction in this cost by increasing the observation interval
from two to three symbols times. Further increase in the observation interval at the expense
of increased complexity of the receiver processing brings one closer and closer to the ideal
coherent performance.
from modulating the orthogonal carriers cos w-t and sin @-t with N-FSK/BPSK modula-
tions. Here the minimum frequency spacing required for orthogonality corresponds to se-
lecting h = 1(faT; = 0.5) for each of the two N-FSK/BPSK modulations and the resulting
modulation is referred to as quadrature frequency/phase modulation (QF PM) [7]. A math-
ematical representation of a QFPM signal and its bit error probability performance on an
AWGN are presented in the following section.
bs.i, for example, ac¢(t)= >> dei p(t —iTs), then a QFPM signal is characterized by
1=—0O
s(t) = VP {ac(t) cos [(we + be(t)wa) t] + as(t) sin [(e + bs (t)@a) t]}
(10.58)
= (Soh) Sst)
where P is the average signal power and wg = 27 fq is the radian frequency deviation about
the carrier radian frequency w,. Note that (10.58) represents a constant energy (E; = PT;)
but nonconstant envelope signal. To see this, we rewrite (10.58) as
s(t)
s(t) = Pa¢(t)
V {cos(en se baea) te cos [we Bey Aya aI
ac(t)
be(t) + bs (t) T
X COS OEE TTS ens
where the upper signs correspond to a;(t)/a¢(t) = 1 and the lower signs to as(t)/a¢(t) =
—1. The cosine factor in (10.59) having a radian frequency [(b¢(t) — bs(t)) /2] wa rep-
resents the envelope of s(t), whereas the second cosine factor of radian frequency w, +
[(be(t) + bs(t)) /2] wa represents the carrier. Clearly, the envelope is constant only when
b-(t) = bs(t), that is, the inphase and quadrature component frequencies in (10.58) are iden-
tical. In this special case, the (2N)*-ary QFPM signal reduces to a 4N-ary N-FSK/QPSK
signal.
8. The coding rule used to map the data sequence {d,,;} into the binary amplitude and N-ary frequency sequences
{ac,i} and {b-,;}, and likewise the data sequence {d,;} into the binary amplitude and N-ary frequency sequences
{as,;} and {bs,;}, is identical to that given in Chapter 4 for realizing a biorthogonal signal set with an N-FSK/BPSK
modulation (see Example 4.6).
Sec. 10.1 |-O Modulations 643
a, (t)
eM a
4 FM f=120.0N ©)
ile Encoder Modulator
J
Demultiplexer
sin @,t
0, =@,-[N-(2n-l]@,; n=iorj
Block diagrams of modulators for (2N)*-ary QFPM and 4N-ary N-FSK/QPSK are
illustrated in Figures 10.9a and 10.9b, respectively. Since for the former, the quadrature
components of s(t), namely, s-(t) and s;(t), are orthogonal over the signaling interval T,,
they can be demodulated and detected independently. As such, an optimum coherent re-
ceiver for a QFPM signal (see Fig. 10.10) is a parallel combination of two optimum coherent
receivers for biorthogonal signals of the type illustrated in Fig. 4.17. Exact bit error proba-
bility results for VN-FSK/BPSK biorthogonal modulation are given by Eq. (4.107) together
with (4.102), (4.105), and (4.106). Since the two constituent N-FSK/BPSK signals in the
M-ary QFPM signal of (10.58) are demodulated and detected independently, then the bit
error probability performance of the latter is also given by Eq. (4.107), keeping in mind that
now N = /M/2 and thus E, = (1 + log, N) Ep = (3logy M) Ep. This equivalence be-
tween the average bit error probability of the resulting modulation and that of its individual
inphase and quadrature components is typical of quadrature multiplexed modulations, for
example, QPSK and QAM.
An upper bound on the bit error probability performance of QFPM that is simple to
evaluate can be obtained from the same upper bound on the performance of N-FSK/BPSK
and the equivalences noted above. In particular, since the probability of error of the first
kind, P;;(E), of (4.105) can be upper bounded by the error probability for antipodal signal-
ing, that is, 4 erfc ./E;/No, then from (4.107) together with (4.103) and (4.106) we have
1 |
P(E) = 5 Psi(E) te 5 Ps(B)
Es
te
ae |No
— er "|] [(— 9)) arte
erfc
Es
Dignity
Eig
fe.( |=4 (10.60)
ea)
]
ree ) ctleerfc it
E;
ane
1
oe Ccert |No Ee
where M = 2N. For (2N )?-ary QFPM with the same number of frequencies as the N-
FSK/BPSK scheme, the bit error probability has the same upper bound as in (10.60) which,
644 Demodulation and Detection of Other Digital Modulations Chap. 10
a, (t)
tH
5)
oO
=—
a.
om
—
:O
a)
@,= @,-(N-(2j-Nlo,
Figure 10.9b A block diagram of a modulator for 4N-ary N-FSK/QPSK
COs @,t
b=12 5-0. ,N
Optimum
Biorthogonal Decoder
Detector
Output
Received Data
Signal
Convert
Optimum
Biorthogonal Decoder
Detector
j =1,2,...,N
E
p(E) <<
Pp(E) erfc ,/(1+ log, N) se
No +<erfe
2 /(1+log)N)—No (10.61)
For the constant envelope 4N-ary N-FSK/QPSK modulation, the receiver in Fig.
10.10 is not optimum. The reason for this is that the two biorthogonal components s,(t)
and ss(t) are no longer independent of each other (they share the same frequency) and
thus should not be detected as if they were. More specifically, in Fig. 10.10, the common
frequency of s-(t) and s;(t) is redundantly detected twice (once in each biorthogonal detec-
tor) each time using only half the total power available. Rather, an optimum receiver (see
Fig. 10.11) for this special case of QFPM should coherently combine both quadrature sig-
nal components prior to detection of the common frequency thus utilizing the entire signal
Sec. 10.1 |-Q Modulations
645
92 sin @,t
y(t)
d, (t)
2 cos @yt
power for this purpose. The idea behind such a configuration is generically the same as that
used in constructing the optimum biorthogonal (V-FSK/BPSK) receiver, namely, remove
the phase modulation for detecting the frequency, and then use the frequency decision to
determine which of the N parallel channels the phase modulation should be extracted from.
Removal of the QPSK modulation in Fig. 10.11 is accomplished at a given frequency by co-
herently adding and subtracting the I&D output samples (projections of the signal onto the
quadrature coordinate axes) and choosing the larger in absolute value of these two quanti-
ties. The maximum of the N largest absolute values so produced then provides the desired
frequency decision, and a binary hard decision on the amplitude in each of the quadrature
components corresponding to this frequency channel provides the QPSK phase decision.
To compute an upper bound on the bit error probability performance of Fig. 10.11,
646 Demodulation and Detection of Other Digital Modulations Chap. 10
we observe that for the N-FSK/QPSK signal of (10.58) with b(t) = b-(t) = b(t), ihe.sum
and difference I&D outputs of the i" frequency channel (at frequency @;), namely, V,; and
V._.,
n,i?
corresponding to an input signal at the same frequency, that is, wp + b(t)wa = @j, are
given by
+ EC ~ -
a 5 (45,0 + den) + (Men + 75.0)
(10.62)
Pe
=! iP oe ay.
Vie = 9 (den wt as.n) * (Ee 5 ne)
(n+1)T
nes = Vi| n-(t) sin [(@j — @e)t +2 /4] dt;
nT
(n+1)T
oe = v3| ns(t) sin [(@j — We) t + 72/4] dt
i (n+1)T
(10.63)
Non = v3| Nc(t) sin [(@j — we) t— 1/4] dt;
nT
(n+1)T
Non= vi| Ns (t) sin [(@j — We) t— 1/4] dt
Tiel
In (10.62), the noise components n* and n~ are zero mean and uncorrelated (also indepen-
dent since they are Gaussian) and have equal variance o,2 = No/T. The sum and difference
I&D outputs, V,"; and V,,;, of the j™ frequency channel (j 47), have zero signal com-
ponents (due to the orthogonality of the signaling frequencies over the symbol interval) and
noise components analogous to those in (10.62). Note that depending on whether as, = dan
OF ds. n FAc.n, the signal component of Vee CU ae)will have a magnitude equal to /2P.
Thus, the effective SNR of the signal used for frequency detection is 2PT,/No, which is a
3 dB improvement over the equivalent signal in Fig. 10.10. Once again, this improvement
comes about because of the coherent combining of the I and Q signals prior to frequency
detection which is possible when the two quadrature components of s(t) share a common
frequency.
An upper union bound on the symbol error probability performance of the receiver
in Fig. 10.11 can be obtained by observing that a symbol error results when any of the
following occur: a wrong frequency is detected, the sign of a¢(t) is erroneously detected,
the sign of a(t) is erroneously detected. As such
ie
|
Wo4)erfc i./(2+log,
5
B N ) Fes —ete
Ep 1
erf (2+
2+ log,oe N)
) ——
ao
P(E) Ep
< = (M —
(10.64)
1 Ep Ep
+ = éric _/ (2+log, N) == ON— I
2 i 62 ) ONG ( ye erkoy 2+ ine?
log, N) ——
) ON
Sec. 10.1 l-O Modulations 647
The first term of the first line in (10.64) corresponds to the detection of an incorrect fre-
quency, while the second and third terms correspond to erroneous detection of the signs of
as(t) and a,(t).
An upper bound on the bit error probability performance is obtained by employing
reasoning analogous to that used in relating bit error probability to symbol error probabil-
ity for biorthogonal (in particular, N-FSK/BPSK) signals. In particular, with reference to
Table 4.4, if a frequency error occurs, then on the average, half of the bits are in error. If an
error occurs in detecting the sign of as(t), then all but one [that corresponding to a¢(t)] of
the total number of bits are in error. Finally, if an error occurs in detecting the sign of a,(t),
then only one bit is in error. Since the total number of bits per symbol is 2 + log, N, then
from (10.64) and the above arguments, the average bit error probability has the upper bound
PoE)
ane < =)AS(=) =(M= (Mt —4)erfc
—4)erfe, |(2 + logy
fe) N) 22
52°") 3No
1+1 Iss
abses (2 + logy N
2+ log, N oe
(10.65)
] eh
2+ log, N
+ (Scher )ee= B2 ) aN
=5ON SDNate I(
(2 + logy N
which is simply one-half of the symbol error probability upper bound in (10.64). Finally,
we shall see later in this chapter that the structure of the receiver in Fig. 10.11 forms the
basis of the optimum receiver for continuous phase FPM, which is constant envelope by
construction.
Using the methods presented in Chapter 2, it is straightforward to show that the
equivalent lowpass two-sided power spectral density (PSD) of QFPM is given by [7]
N-—1
FS sine? (sr +i BS ~~)
Ss(f) = —_
He)
(10.66)
2 (1 +log, N) PT, X Ne
= * De sitte? (21 + log, N) fTp+i-~—)
i=0
which is illustrated in Fig. 10.12 for various values of N. Note that since QFPM corresponds
to a modulation index h = 1, its PSD is not? simply a frequency scaled (by a factor of two)
version of the PSD for N-FSK/BPSK (h = 0.5). This is readily seen by comparing (10.66)
with the comparable result in Example 4.6 and Fig. 10.12 with the comparable figure in that
example, that is, Fig. 4.45.
The null-to-null bandwidth, Wyn, (Ban = Wnn/2 is the single-sided lowpass null
bandwidth) of QFPM is easily obtained from (10.66) by setting the argument of the sinc
9. The exception to this statement occurs for the trivial case of N = | whereupon QFPM becomes QPSK and
N-FSK/BPSK becomes simply BPSK.
648 Demodulation and Detection of Other Digital Modulations Chap. 10
dB
S,(/PTp,
dB
S,(f)/PTp,
function equal to unity when i = 0 and letting f = W,,,/2 with the result that
1 New 1
Weg Ne) Se (10.67)
Ts; 2(1+ log, N) T
Table 10.2 illustrates the bandwidth efficiency 5 = 1/WnnTp = 2(1 + logy N)/(N + 1) of
QFPM for various values of M = (2N)?. Also tabulated are the exact values of Ep/No
(in dB) for a bit error probability P, = 10~-° as determined from the exact N-FSK/BPSK
results of Chapter 4, that is, Eq. (4.107). We observe that in comparison with QAM and
MPSK, the bandwidth efficiency of QFPM is significantly lower for M > 16 as one might
expect; however, the power efficiency is significantly better. In particular, QFPM yields a
power savings of 3.8 dB for M = 16, 9.5 dB for M = 64, and 15.1 dB for M = 256 when
compared with QAM. When compared with M-PSK, the power savings are even larger;
however, the reader is reminded that M-PSK is a constant envelope modulation whereas,
in general, QFPM is not. For the special case of constant envelope QFPM (N-FSK/QPSK)
as discussed above, Table 10.2 also gives the exact values of E,/No (in dB) required for a
bit error probability P, = 107°.
If instead of using null bandwidth, we were to use a fractional out-of-band power defi-
nition of bandwidth (see Chapter 1) for computing bandwidth efficiency, then QFPM shapes
Sec. 10.1 |-Q Modulations 649
Hon
i
z22e2
onnM—
ie NEG
dB
n,
up considerably better when compared with QAM and M-PSK. Figure 10.13 illustrates the
fractional out-of-band power [see (2.125)] 7 = (1 — x/100) in dB of QFPM versus normal-
ized bandwidth W,7; where x is the percentage of the total power that lies in the band-
width W,. Tables 10.3 and 10.4 [7] present the comparable data to the column labeled 6 in
Table 10.2 for 90% and 99% power containments, that is, 6 = 1/Wo97} and 6 = 1/ Woo97)
where
available signal dimensions than conventional two-dimensional signal sets such as QPSK.!°
The modulation scheme dubbed quadrature-quadrature phase-shift-keying (QPSK) places
two QPSK signals in quadrature to form a four-dimensional signal set that, in a given
bandwidth, ideally allows for an increase of the bit rate by a factor of two (i.e., twice the
throughput) relative to QPSK. Mathematically speaking, a Q?PSK signal takes the form
CO
where T; = 47; is the signaling interval, p;(t) and p2(t) are unit power real pulse shapes
of duration 7; and orthogonal to one another over the interval 0 <t < 7;, and 6, ;;n =
1,2 denotes the n' data phase (constant over the n" signaling interval nT, <t < (n+
1)T;) from each of the two QPSK signals with values ranging over the set 6», = (2m +
— 1)m/4; m=0, 1, 2,3. A particular case of interest is where p)(t) = J/2cos (zt/T;) and
Prt) = J2 sin (rt /T;), whereupon (10.69) becomes (in the zero” signaling interval)
wt t
s(t) =V2P co (F) Cos (wet + 61,0) + sin (+) cos (wet + 620)|
ut t
=/P jarcos (+) COS Wet — by.9 cos (=) SiN Wet (10.70)
See Wy Cine
+az2,9 sin Tr COS Wt — b2,9 sin TT SIN @ct
with dno, bn,o; n = 1,2 representing the equivalent +1 inphase and quadrature pairs of
data bits corresponding to the data phases 6,9; n = 1, 2, that is, an,9 = J/2cos On.0. Ono =
J/2 sin 6n,0; N= 1,2. The structure of (10.70) clearly identifies this particular form of
Q’PSK as the sum of two sinusoidally pulse-shaped QPSK signals that are in quadrature
(with respect to their pulse shaping) to each other.
10. By “ QPSK” here and in the discussion that follows, we mean the more generic case that allows for arbitrary
(rather than rectangular) pulse shaping.
Sec. 10.2 Continuous Phase Modulation (CPM) 651
(10.71)
+ /P Eo cos (+) Sin Wet + as(t)bs(t) sin (=) cos vt |
Since a-(t), ac(t)bc(t), as(t), and as(t)bs(t) are all binary waveforms, that is, rectan-
gular pulse data streams of rate 1/7; = 1/47), then letting a,(t) = a(t), ac(t)b-(t) =
b2(t), as(t) = —bi(t), and as(t)bs(t) = az(t), we obtain the desired equivalence with
(10.70). Thus, Q?PSK of (10.70) can be generated, demodulated, and detected as a QFPM
signal with N = 2 in accordance with Figures 10.9 and 10.10.
where a, denotes the n'® transmitted data symbol which takes on values from the M-ary
alphabet {A, =-M+(2k+1); k=0,1,2,...,M—1}, hn is the modulation index,
which may vary from symbol interval to symbol interval, and q(t) is the phase pulse, which
has length LT, (L integer) and is expressed as the integral of the normalized frequency pulse
g(t) by
t LTs 1 woes
q(t) =i) g(t)dt; q(LTs) =} g(dt= 5 (10.73)
0 0
CPEM schemes characterized by (10.72) fall into several categories depending on the vari-
ation of h and the shape and duration of g(t). Typical shapes for g(t) are rectangular and
raised cosine which lead, respectively, to the modulation types LREC and LRC. Multi-h
modulations refer to the class of CPFM’s where h,, varies from interval to interval (typically
the variation is cyclic through a set of indices). Full response CPFM is one where the fre-
quency pulse duration is one symbol, that is, L = 1. Otherwise the modulation is referred to
as partial response CPFM. Continuous phase FSK (CPFSK) is the acronym associated with
a full response CPM having a rectangular frequency pulse, that is, 1REC. For our purposes
in this chapter, we shall consider only the case where /; is independent of i, that is, the mod-
ulation index is constant over the duration of the transmitted signal, and also L = 1. For this
case, (10.72) reduces to
n—1
S(t) =v 2P.cos ee+ 2mhang(t —nTs;) +h a a
1=—CO (10.74)
=vV2Pcos[wt+O6(t,@)]; nI;<t<(+1)T;
The first term in 6(t, @) represents the contribution to the total phase in the n™ signaling
interval due solely to the n'” data symbol a,. The second term represents the contribution
due to all previous data symbols a,j; i = 1, 2,..., 00 and as such is responsible for the
memory associated with this form of modulation.
Analogous to (10.74), the excess phase in the (n — 1)*' transmission interval (n —
LQTS 1 AS CIVEDY
M2
A(t, @) = 2thonig(t
—(n-1)T.) +h Day; (10.75)
i=—oo
Comparing (10.75) with (10.74), we observe that @(t, a) is continuous at the symbol transi-
tion instant t = nT; [since q(0) = 0] which is required for CPFM.
Transmitting s(t) over a constant phase AWGN channel results in the received signal
r(t) =s(t, 0) +n(t) = Acos (wet + O(t, @) + 8) +n(t) where as in previous sections n(t)
denotes the additive white Gaussian noise process.
Since, as previously mentioned, the continuity of phase from symbol interval to sym-
bol interval introduces memory into the modulation, the optimum receiver (coherent or non-
coherent) should base its decisions on an observation interval, say N,7;, commensurate with
the memory duration. In this regard, two types of optimum receivers have been considered
in the literature. One type calculates an estimate of the data sequence in the observation in-
terval and performs block-by-block (Ns-symbol) detection. The second type decides on only
one symbol (e.g., the first one) per N,-symbol observation by averaging over the remaining
Sec. 10.2 Continuous Phase Modulation (CPM) 653
1
ee eyae,
9 OST,
=f 10.76)
k
Bt) {2 otherwise
Thus from (10.73), the phase response pulse, q(t), is a linear function of t over its T,-sec
duration and likewise the excess phase in the n" transmission interval nTy <t<(n+1)Ty
is also linear, that is,
ha,(t — nT,
iia) = wee +h) a; (10.77)
11. Unless otherwise noted, we shall consider only the case of binary FSK and thus the terminology “symbol”
is replaced by “bit.” Also, the generalization of the results in this section to continuous phase M-FSK is given by
Schonhoff in [13].
654 Demodulation and Detection of Other Digital Modulations Chap. 10
|
is (r |, a ) = Tee exp {7 [
Lope
[r(t) — s(t, a0, @ )} at
(10.78)
9) NsTp
= C exp (= | r(t)s(t, ao, @’) ar)
where C is a constant that is independent of s(t, QQ, a’). Averaging (10.78) over the pdf of
a’, namely
5c No=1
fla) = I] f(@)= I] Faces b+ 58a +0]
1 I
(10.79)
i=1
Sec. 10.2 Continuous Phase Modulation (CPM) 655
where a’; i =1,2,...,2'~! represents the set of possible sequences for @’. A decision
between a = | and ag = —1 is then made by comparing the likelihood ratio
QNs-l ; A QNs-l
ae > exp (# Jo r()s(t, 1, a) de) my, X1,i
ae ee Se Ot
oy 1 2Nse! aNss!
ES ae X exp (i Ost, —1,a') dt) Dees
i= i=l
with unity, that is, if A > 1 then decide ap = 1, otherwise decide ag = —1. Figure 10.15
is an illustration of a receiver that implements this decision rule. Unfortunately, the exact
error probability performance of this receiver cannot be evaluated analytically. However, it
is possible to approximate and bound this performance at low and high SNR.
At low SNR, we may use the approximation e* = | + x, in which case the decision
rule dictated by (10.81) and Fig. 10.15 becomes
QNs-!
2 NsTp
ao = sgn ss (:+ a, r(t)s(t, I.a{)dr)
2 NsTp
-)> (:+o f rids(t.~1a{) dt)
No Jo
NsTp :
= sgn / r(t) 5 s(t, 1,a@;) — »s, s(t, —1,a;) |dt
0
NsTp
a sen|| r(t) [s(t, b 5-1
0
Here, s(t, 1) and S(t, —1) denote the average transmitted waveform given that the first bit
is a “1” or a “—1”, respectively. A block diagram implementation of (10.82) is illustrated
in Fig. 10.16, which takes the form of a conventional matched filter receiver for binary
communication over the AWGN. This implementation is referred to as an average matched
filter (AMF) receiver in [11]. Other variations of the AMF are also discussed in Chapter 8
of [11]. Using the results of Chapter 4 and recognizing the symmetry of the present system,
656 Demodulation and Detection of Other Digital Modulations Chap. 10
D
— s(t,1,@,)
No
we immediately have that the average bit error probability performance for low SNR is
QNs-!
]
P(E) =aa 2 Poi(E) (10.83)
i}
where
NgT,
Ppi(E) = serte| r ua a, LejEG Ese ila:
J20 0
(10.84)
20No
Oo Rr
pt =
[s(t, 1) —
=
s(t,
2
—1)] dt
0
Sec. 10.2 Continuous Phase Modulation (CPM) 657
S(t,1) —S(t,-1)
r(t)
2 = Note
NoTi
5
; 1 — cos 2h)
[— sinc 2h + SES
(1 + sinc 2h) (cos?%s~2
TOUUE Se Coen
rh — 1
ter) (10.85a)
2 (cos?
xh — 1)
and
Soares
; ale 1 — sinc 2h + sinzth s (cost * rh)
k=2
: (10.85b)
=
x sinOx, + te (cos Ox — cos (27hai,x + e))|}; Q,A S 5whoi,j
j=l
where
s) NsTp
= Pr {exp(=f r(t)s(t, Ia) dr)
0/0
Z NsTp /
< exp (=f r(t)s(t, -1,a))dr)|s(t, 1, o)/ ae,
NsTp NsTp
=Pr{ f r(t)s(t, 1, a5) dt < | r(t)s(t,
—1, o;) dt |s(¢, Ia)
0 0
1 N; Ep
= 5 erfc 7; (1 — pi,;)
658 Demodulation and Detection of Other Digital Modulations Chap. 10
aon / /
pi, & cE
I
f s(t, —1, a) s(t, 1, a) dt
1 : Ns h (aii, = j,k)
= N, [sscos7h + ms (Aas (10.88)
mh (ai,~ — j,k) kl
x COS [em + ma: (ai. — @;,1)
A lower bound on average bit error probability for all SNR can be obtained by suppos-
ing that for each transmitted sequence the receiver needs only decide between that sequence
and its nearest neighbor, that is, the sequence most correlated with it. Then
QNs—l
1
P(E) > ot ye cue (10.89)
r=)
where
N; Ep
Pi,jmax = = erfc s (1 aia max); Pi, jmax = Max Pj,j (10.90)
Z No j
The lower bound of (10.89) and the upper bound of (10.86) asymptotically approach
each other at high SNR and thus represent the true bit error probability performance in that
region. Thus, recalling that (10.83)-(10.85) represents an approximation of the true per-
formance at low SNR, then a composite curve formed from (10.83)—(10.85) and (10.86)—
(10.88) can be looked upon as an approximation of the performance for all SNR. Fig-
ure 10.17 is an illustration of the behavior of these various bit error probability bounds and
the composite approximation for an observation interval of five bits (Ns = 5) and a modula-
tion index h = 0.715. This value of modulation index was chosen since in [10] it was shown
that it achieves the maximum value of the minimum distance between ali pairs of signals
s(t, 1,q@‘) and s(t, —1, a’), i, j =1,2,..., 16 which dominates performance at high SNR.
By distance between signals s(t, 1,q@) and s(t, —1, a’), we mean
57h le
dij = / [ste 1,ai) —s(t, -1,¢")| dt (10.91)
0
It can also be shown [12] that little or no further improvment in performance can be achieved
by extending the observation interval beyond five symbols. This behavior appears to be
typical of CPFSK systems independent of modulation index.
Noncoherent Detection. _Here we assume that the channel phase 6 is unknown and
uniformly distributed in the interval (—z, 2). Also, in contrast with the coherent detec-
tion case, we shall be making decisions here on the middle bit of the observed sequence
as explained in [12]. Thus, to simplify notation, the length of the observed sequence N,
shall be equated to 2n, + 1 and the decision will be made on the bit Qn,. If a’ now de-
. notes the transmitted sequence excluding the bit a,,, and s(t) is notationally replaced by
Sec. 10.2 Continuous Phase Modulation (CPM) 659
10-2 COMPOSITE
APPROXIMATION
10-4
OBSERVATION
INTERVAL (N,) =5 BITS
S(t, &p,,@’) and s(t, 0) by s(t, &p,, @’, 0), then analogous to (10.81), the likelihood ratio is
now
Q2ns
> 10 (21,1)
_ isl
a, Q2ns
Bla (z=1,1)
1=1
where
660 Demodulation and Detection of Other Digital Modulations Chap. 10
2
, 2 (2ns+1)Tp F
LE ha af r(t)s(t, 1, @;,0) dt
2
2) (2ns+1)Tp
ae =| r(t)s(t, 1,a;, 2/2) dt
No Jo
(10.93)
2 p2nst1)Tr 2
)
21,1 = =f s(a,
r(t)—1, 0)dt
t,
9 pQns+l)Ts 2
a val r(t)s(t,
—1, o;, 2/2) dt
0 Jo
A block diagram of a receiver that implements the decision on ap, based on (10.92) is
illustrated in Fig. 10.18. As was true for the coherent receiver of Fig. 10.15, the exact error
probability performance of Fig. 10.18 cannot be evaluated analytically. Thus, once again we
resort to suitable approximations and bounds for high and low SNR.
At low SNR, the Bessel functions in (10.92) can be approximated by Jp(x) = 1+
x?/4. Making this approximation in (10.92) results in the simplified decision rule
22ns
SiG An,» a’, 0) af js(t, Qn,» a’, 1/2) BS S (é Qn,» a’) = A exp G [wet ot A(t, a)]) (10.95)
Q2n 5
92n s
_ Then
Sec. 10.2 Continuous Phase Modulation (CPM) 661
Bei
No Lan0
(t,1, 1»Y)
2
Feta
Oe ,0)
2
yp het) ae 2° -\ hp One ge
=| r(t)S(t, 1) dt} — x | r(t)S(t, 1) dt
0 No Jo
2
(2ns+1)Tp
= | r(t)S(t, 1, wt) dt
No Jo
2
5) (2ns+1)Tp
=| r(t)S(t, —1,
a) dt
0 JO
22ns Q2ns
2 enstlTp ;
r(t)S(t, 1, @;) dt (10.98)
as 2uk xh
I= J=
if]
(2 ostna)a)|
(2ns+1)T,
For any given sequence a; in the double summations of (10.98), the complement (negative)
of that sequence, say a; = —a@j;, also exists. Hence, it can be shown that these double
summation terms cancel each other and what is left (that is, the single summation) is the
argument of the sgn function in (10.97). Finally, we conclude that the decision rule of
(10.97) can alternately be written as (eliminating the 2/ No scaling)
2
OnvttT, (mani 2
= |= sgn
gAn, il r(t)S(t, 1) dt} — it r(t) S(t, —1) dt
0 0
(10.99)
S$(t,1)
S(t,-1)
with
| 2)
Se = 5 ELAN: Kaa, 2
o? Dy =atteSE
1
(iZ1-at E (Zi)?og |ae)=58
1
{lZ2—~ EZ)?2 |
1 (10.102)
p= x58 {(Z1 ~ E (Zi) (Za ~ E (Za))"}
In (10.102), the values of the various parameters are computed conditioned on the trans-
mitted middle bit, a»,, being a +1 and the remaining bits belonging to the vector'@;. The
evaluation of these parameters is carried out in [12] under the assumption that all signals
have zero phase at the beginning of the middle bit interval. The results are as follows:
664 Demodulation and Detection of Other Digital Modulations Chap. 10
NoAZT) 1 — cos?" th
o? =~?
4
41.4 (1 + sinc 2h) ||
1 —cos* zh
oe NoA2T;, 1 : 1 —cos”" rh FZ ,
{5 + sine 2m) (1+ -7**) | }+ 7 sinc |
eed , 1 — cos? 2h
(10.103)
AGT:ie
ATT |
Si ao ye ith inch + e7 jah a, At cee 2 |Ay +14 Aol?
where
= ;So (-
jah ye fines)(cos!
= th) (1+ e IThAins—l sinc h)
f=!
a (10.104)
1 Ns ia 5
=~ ; Son —jmh re
ELS nat) (cos’! rh) )((1 + € eI 7 hist! sinch)
Finally, the average error probability is found from (10.83) with N, = 2n; + 1.
As a simple example of the above, consider the case of N; = 1(n; = 0). For this case
A‘TZ A‘T2
AlE—vA |) te Z sinc’ h: Y= Us
8 8
NoA2T, NoA2Th _; 10.105
ae po’ = ae sinc h ( )
ae al
A7T; y V1 sine?) = E
5%[1 yi=sine?| (10.106)
0
0
for all sequences a@;. Combining (10.106) with (10.100) gives the well-known result [14, p.
207] for the error probability performance of noncoherent FSK with modulation index h. If
in addition the signals are orthogonal, that is, h = 1, then this result simplifies to
1 Acre Atre
At=Ao=-;
I Sool Sj=
1 7 Le S} ee ean
5
and
ale Eb J3
i = No [+ + (10.109)
for all sequences @;. Combining (10.109) with (10.100) results in an improved error proba-
bility performance over the case N; = 1.
For high SNR, we may use the approximation
Q2n s
|
Pein aca
(2ns+1)Ty
2.
/ r(t)S(t, 1, @;)dt
No 0 (10.113)
|
9) (2ns+1)Tp
a, —
/ rns(t, =1,¢”;) dt
No 0
The pairwise error probability of (10.113) can be expressed in the form of (10.100)
where
. IE /
Di,j 2No
with
2ns+1 k—]
1
Pl,m ne pe exp (isLd,(ev o-n))
(10.115)
h h
x exp GF (aK.,1 _ 7) sinc (5(ax, _ aun)
where a, is the v'h bit in the sequence a; whose middle bit is a —1 and a@),m is the v" bit
in the sequence @,, whose middle bit is a +1. Note that for Ns = 1, (10.1 14) together with
(10.115) reduces to (10.106) and hence the high SNR upper bound of (10.86) agrees with
the low SNR approximation, both of which yield the exact performance.
Figure 10.20 is an illustration of the behavior of the composite performance approx-
imation made up of a combination of the low SNR performance of (10.83) together with
666 Demodulation and Detection of Other Digital Modulations Chap. 10
10°
BINARY
ORTHOGONAL
1071
3 BIT OBSERVATION
10-2
COHERENT PSK ————>
(E)
P
10°3
10-4
5 BIT OBSERVATION
(10.100) and the high SNR approximate bound of (10.86) together with (10.113). The nu-
merical results are plotted for 3-bit (ns; = 1) and 5-bit (ns; = 2) observations and h = .715.
Also, plotted for the purpose of comparison is the performance of noncoherent orthogo-
nal (h = 1) FSK [Eq. (10.107)] and coherent PSK. We observe that for E,/No greater than
about 7.3 dB, noncoherent CPFSK with 5-bit observation outperforms coherent PSK. For
bit error probabilities below 10-3, the performance of noncoherent CPFSK with 3-bit ob-
servation is within | dB of that corresponding to coherent PSK.
a sequence detection standpoint. Hence, we shall consider only the noncoherent case and
furthermore show that considerable simplification of the generic implementation with 2s
correlators can be achieved. The results to be presented here for CPFSK are taken from [15]
where the more general case of full response CPM is considered.
where as before we have incorporated into the notation the arbitrary (uniformly distributed)
phase @ introduced by the channel. For the CPFSK signal of (10.74) together with (10.77),
the correlation in the argument of the exponential in (10.116) can be expressed as
NsTp Ns Tp
/ r(t)s (t,@,0)dt= (| r(t)s (40) cos 0
0 0
(10.117)
NsTp
_ (| r(t)s (.a,1/2)a) sin@
0
NsTp NsTp
if TA) SEN CORO aL — E / r(t) COS (Wet + A(t, a)) ar|cos
0 0
NsTh ne LOSELS)
= i r(t) sin (@ct + O(t,
@)) dt |sind
0
NsTp
seal r(t)exp[—j(@ct + O(t, a))] dt (10.119)
0
Tb r(the
b * sf ;
Jet e—Imhoot/ Todt 2. (ends) /2Tp (tye Jeet eI hat —1)/ Th dt
0 Tb
(10.120)
Ns—2 NsTb F
4. and + il e I mhax / nitye Ite ithon,-10-Ws-D)15)/ Thge
rae, (Ns—1)Tp
Substituting (10.119) into (10.116), averaging over 6, and taking the natural logarithm gives
668 Demodulation and Detection of Other Digital Modulations Chap. 10
Note that any signal phase (due to prior data bits) that is present at the start of the obser-
vation interval can be included in (added to) 6 and because of the averaging performed in
(10.121) will have no bearing on the decision rule. This affirms a previous statement that for
noncoherent detection, the structure of the optimum receiver does not require knowledge of
the signal phase at the beginning of each block.
Since In Jo(x) is a monotonic function of x, then maximizing In f(r |@) leads to the
appropriate decision rule as follows. For each finite sequence (block) @; = (@;,0, @i,1,.--,
Qj,N,—1), the optimum receiver computes
Ns-1
Bo Poa ceri nent Cie (10.122)
k=0
where
(Ns—m)Tp ‘
Tr} n= i; r(te Jet e—Jtht—mT)/ To gy
5 (N,—1—m)Tp
(10.123)
(Ns—m)T), ;
Css = if r(the Jeet eI ™ht—mT5)/To dy. m=0,1,..., Nei
(Ns—1—m)T,
and the complex constants C;,, are defined recursively by
10.21. We note that only two correlators are required, each of which is followed by a tapped
delay line and suitable processing to make the sequence decision.
To evaluate the performance of the receiver in Fig. 10.21, we make use of the tech-
nique used in the discussion of multiple symbol detection of MPSK in Chapters 6 and 7
to obtain an upper bound on the bit error probability, P,(£). In particular, we use a union
bound analogous to that used for upper bounding the performance of error correction coded
systems. This bound is expressed as the sum of the pairwise error probabilities associated
with each N;-bit error sequence. Exact evaluation of the pairwise error probabilities is again
made possible by the results in [14] analogous to (10.100)-(10.102).
Mathematically speaking, let @ = (ao, a1, ..., @y,—1) denote the sequence of N, in-
ot bits and @ = (a, @, ..., &y,—1) be the corresponding sequence of detected bits.
Then
4)
yl
a
(1)4 aouenbas yore I0j
"ny= 0't)
6 2699
e+ (TO'N'gg
*/(41-"N)
3urpuodsai09
0}
12s00yD
fo|xeu
669
670 Demodulation and Detection of Other Digital Modulations Chap. 10
where w(@,@) denotes the Hamming distance between @ and @ and Pr ||B| > || a|
denotes the pairwise probability that & is incorrectly chosen when indeed @ was sent.
The decision statistic || is defined in (10.120) and the corresponding error statistic |B|
is identical to (10.120) with each a, replaced by a. Unlike M-PSK, CPFSK is not a
symmetric signaling set and thus it does satisfy a uniform error probability (UEP) criterion
where the probability of error is independent of which input phase sequence @ is chosen as
the correct sequence. Hence, without any additional characterization of Pr ||B| > |Blal,
further simplification of (10.125) is not possible. Later we shall show that Pr |B| > |B |a|
depends only on the error sequence ¢ = @ — @ and thus further simplification is possible.
To compute Pr ||BI > || al,we use the approach taken in [14] for evaluating the
probability that one Rician random variable exceeds another. In particular, analogous to
(10.100)
1
Pr{ IAI > IBl|@} == [1- ovo, va) + Ova, Vb)| (10.126)
where
b| _= Fo |v.+ niane
2} | (10.127)
with
n=0 2
Before proceeding to the evaluation of (10.126)-(10.130) for specific cases, we first present
an expression for the pairwise error probability under the conditions a, b >> 1 which from
(10.127) corresponds to the case of large E,/No. Using the appropriate asymptotic expres-
sions for the Marcum Q-function and the complementary error function, (10.126) can be
Sec. 10.2 Continuous Phase Modulation (CPM) 671
written as
P No lope |
P{iAl> laa] = ( a )se |set} (10.131)
‘ m {°|
a) — a ao
— aQ
a
(1+/1- l+e/7
ansinch
'
0 2 Ep (ae 1 ei
2
}
2 =2 Ep
(1 pe jsinc h|?)
The first two cases in the above tabulation correspond to error sequences having a Hamming
distance w (a, &) = 1 while the third and fourth cases have w (a, &) = 2. We note further
that the values of a and b for cases 1 and 2 are identical. Finally, computing the number
of times that each entry in the above table occurs corresponding to the 4 x 3 = 12 possible
cases for which a # @, then from (10.125)-(10.127) and the above tabulation we arrive at
the desired upper bound on bit error probability given by
1
PoC) = Pi Ais (Parte) (10;732)
where
|b3 |= Ep 2
(ityi- |cos 7th sinc) oa (1 ae
l sine h?)
a3 No
Figure 10.22 is a plot of the upper bound on P;(E) [as given by (10.132)] versus E,/No in
dB for various values of h between 0 and 1. Figure 10.23 is the associated plot which shows
the optimization of P,(E) as a function of h with E,/No in dB now as the parameter. We
note that the optimum value of h [corresponding to minimum P)(£)] is quite insensitive to
Ep/ No over the large range of values considered.
672 Demodulation and Detection of Other Digital Modulations Chap. 10
10-2
10°3
10-4
PF
ON
(E)
BOUND
UPPER
10°5
106
bj |_ 3Eb
fe ate (10.136)
and
Sec. 10.2 Continuous Phase Modulation (CPM) 673
10°
E,/Ny = 6 dB
E,/Ny = 8 dB
10-2
E,/No = 10 dB
10°3
10 -4 rau
10°5
MINIMUM
OF
BOUND
UPPER
ON
(E)
F,
10°6
E,,/Ng = 14 dB ———>
Figure 10.23 Minimum of upper bound
on bit error probability versus modulation
index with E),/Npo as a parameter; binary
10°8 modulation with rectangular frequency
0.1 02 03 04 05 06 07 0.8 0.9 1.0 pulses; two-symbol observation (Reprinted
h from [15] © IEEE)
i Id
Figure 10.24 illustrates the analogous results to Fig. 10.22 where the upper bound on Pp(E)
is now computed from (10.135) together with (10.133), (10.136), and the above table.
Comparing these two figures, one can assess the improvement in performance in going from
a two-symbol to a three-symbol observation interval.
674 Demodulation and Detection of Other Digital Modulations Chap. 10
100
1071
10-
10°3
10-4
(E)
P
ON
BOUND
UPPER
10°5
10°6
Figure 10.24 Upper bound on bit error
probability versus E,/No with modulation
5 index as a parameter; binary modulation
10° ith rectangular frequency
S10. Ae a2 46 ae tig 8 i il pulses; we three-
: : y E ny dB symbol observation (Reprinted from [15] ©
fay (0) IEEE)
To extend these results to values of N; > 3, we make use of the fact that from
(10.130), |5| is only a function of the error sequence ¢ = @ — @ rather than @ and @ indi-
vidually. Thus, from (10.126) and (10.127), Pr |\B| Ss i6||a| is likewise dependent only on
the error sequence ¢ = a — @. Hence, analogous to [16, Eq. (4.9.21)], we can write (10.125)
in the simpler form
(minimum value of equivalent squared distance N, — |5|) provided that the other values of
|5| generated by the remaining sequences are sufficiently smaller than |5|max. For arbitrary
h, one cannot show that this always occurs. In fact, for certain values of h, the values of
|5| may indeed be bunched around |5|max. However, for h = 1, the values of 65 are all real
and integer and there is sufficient separation between them that one can evaluate the upper
bound on error probability based on the above approximation. Thus, we proceed with an
analysis of the behavior for this special case.
For h = 1 (orthogonal CPFSK), it is easy to show that the maximum value of |4| oc-
curs for each error sequence & having Ns — 1 elements equal to the correct sequence @ and
one element in error. From (10.130), |6|max is easily shown to have value N, — 1. Since the
element in error can occur in any one of the N; positions of the sequence, then the total
number of pairs of sequences yielding |5|max is Ns x 2s. For each of these sequence pairs,
the Hamming distance is w (a, @) = 1. Thus, from (10.125)-(10.127), an approximate up-
per bound on the performance of orthogonal CPFSK with N,-symbol observation (N, > 2)
is given by
Po(E) <Pr |1B| > (Bllar}= [1— O6/bmaxs Vaimax) + Q(/aimax> VPmax)| (10.138)
where
For large SNR, we can use (10.131) instead of (10.126), which gives the asymptotic upper
bound
IN; +|dlmax \ 1 Ep
Pp( bE)
) (—————
Dine ] -—erf
5° c ——(N,
No: Ss — ||6 lmax)
(10.140)
2N;-—1)\ 1 Ep
= ; — erf —}; Ns >2
(4) Sie (a :
As Ns — oo, (10.140) reduces to
1
PRE) 5 erfc si (10.141)
2No
which is identical to the performance of orthogonal FSK with coherent detection (see Chap-
ter 4).
t—nT
a(t, a) = SR th eetSa =a (=)
mt
+ Xp (10.142)
2Tp
where x, is a phase constant required to keep the phase continuous at the transition points
t =nT» andt = (n + 1)T7p. In the previous interval, (n — 1)Tp, < t < nT), the excess phase is
given by
aa
Tanae
(t — (nHamma
— 1)Tp)
Bega
— mt
als ais +Xn—-1 (10.143)
1 1
Qn (= (nTp) + Xn = Qn-1 (=) (nT) + Xn-1 (10.144)
or
wn
Xn =Xn-1 + in — An) (10.145)
Equation (10.145) is a recursive relation for x, in terms of the data in the n'® and the n — 1°
intervals.
We observe that (@,—1; — a@,)/2 takes on values 0, +1. Thus, for a, = a,_1, we have
Xn = Xn—1 while for a, = —dy,_1, we have x, = x,—-1 + nz. If we arbitrarily set x9 equal to
zero, then we see that x, takes on values of 0 or z (when reduced modulo 277). Note from
(10.142) that x, represents the y-intercept (reduced modulo 277) of the line representing
O(t,@) in the interval nT, <t < (n + 1)Tp. This is easily verified by letting h = .5 in the
trellis diagram of Fig. 10.14.
Let us now apply trigonometric identities to (10.74) with @(t,@) defined as in
(10.142). In particular, we rewrite (10.74) as
S(t) = V2P [cos wet cos O(t, @) — sin wet sin O(t, a)] (10.146)
Now
wt se cule
cos O(t, a) = cos oT, COS Xn — Gy Sin oT, SiN Xp (10.147)
Recalling that x, only takes on values 0 or 2(modulo 271), then sinx, = 0 and (10.147)
simplifies to
t
cos O(t, @) = dy COS ae =O} = Gell (10.148)
Qi,
Similarly
; : wt RELA A TUUAS .
sin 6(t, @) = sin (ane ae ‘|= My, Sin oT, COS Xn + COS aT SIN Xp
(10.149)
A URS
= Pa sin 57 by = @y COS An = Ondn = =k
Sec. 10.2 Continuous Phase Modulation (CPM) 677
Finally then
S(t) = V2P [anC(t) cos wet — by S(t) sin@et]; nTy <t <(n+1)T,
An =COSXn; by =SiINX, =Andn => effective I&Q binary data sequences (10.150)
wt 5 CULE
C(t) =cos ——; S(t)=sin— = effective 1&Q waveshapes
Dilip dip
(10.151)
S(t) sin wet =--— cos aF a || Se ze
oc ae) OTR Bee BN BITS he
QO, —— _ —- —
ons (olla)
](85%)[le-2)
Thus
1
ay = OaOn = —1) => S(t) = Aa, cos |
(o- =| |
Note that in (10.150), C(t) is S(t) shifted by 7, seconds. Thus, it appears that s(t)
might be in the form of offset QPSK with half sinusoidal pulse shaping. To justify that
this is indeed the case, we must examine more carefully the effective data sequences {ay}
and {b,}. Since the actual data sequence a, can change every 7) seconds, it might appear
that a, and b, can also change every 7, seconds. To the contrary, it can be shown that as
a result of the phase continuity constraint [Eq. (10.145)], an = cos x, can change only at
the zero crossings of C(t) and by = a, cos x, can change only at the zero crossings of S(t).
Since the zero crossings of C(t) and S(t) are each spaced T; = 27) apart, then a, and by are
constant over 27p-sec intervals (see Fig. 10.25 for an illustrative example). Noting that C(t)
and S(t) alternate in sign every 27) sec, we can strengthen the analogy with sinusoidally
pulse-shaped offset QPSK by defining the basic pulse shape
j&)
singe:
2Tp’
OSSey f= 27)
PE 10.153
pw) |0; : otherwise )
where
with
678 Demodulation and Detection of Other Digital Modulations Chap. 10
0 O0<t<T,
1 T, Sts 21,
1 | — PYhe SEES BIW.
0 3T, <t $47,
— 0 4T, st SST,
0 =
ee
el
oe
oe
el
Ln 5T, <t $67,
0 -1 67, <t< 77,
1 -1 -1 TBIPSES ESOL:
FWY
UN
nA
OS
NY
oN 4 -1 1 87, <¢ <9T7,
Figure 10.25 An example of the equivalent I&Q data sequences represented as rectangular pulse streams
To complete the analogy, we point out that the sequences {a2,—} and {b2,} are the odd/even
split of a sequence {v,} which is the differentially encoded version of {a}. This is easily
verified (see Fig. 10.26 for an illustrative example) by recalling that for differential encod-
INZ, Vp = An Vp_— and then noting from (10.150) that b, = a@,a,. The alternate representation
of MSK as in (10.152) can also be expressed in terms of the differentially encoded bits, vy.
In particular
Sec. 10.2 Continuous Phase Modulation (CPM) 679
Figure 10.26 Equivalence between differentially encoded input bits and effective 1&Q bits
Forn odd
+(25)
5 c ( - oT,
os || We = )]| 10.157a)
(10.157a
For n even
*
s(52)=[-8)])
Un—1 + Un ry,
mom
‘an VP ( 2 Joo (o. a ie,(
Un—1 — Un IU
680 Demodulation and Detection of Other Digital Modulations Chap. 10
MSK
Modulator
1)" cos@,t
7) a-1
Pulse
Shaping
(p(2)
Data
Splitter
Pulse
Shaping
(p(t)
(=1)" sin@,t
(10.158)
tee Wa (252) COs |
(vx— ) |}SWAT SSE IMG
b
Finally, the I-Q implementation of MSK is as shown in Fig. 10.27. We observe that this
figure resembles a transmitter for offset QPSK except that here the pulse shaping is half-
sinusoidal [see (10.153)] rather than rectangular and in addition a differential encoder is
applied to the input data sequence prior to pulse shaping and modulation onto the quadrature
carriers.
ee
(2n-1)T, (e)dt
Carrier
Recovery Data
r(t) Diff.
Circuit Combiner Decoder
Unity Transmission
S(t) . MSK
Modulato
offset QPSK (or QPSK or BPSK). Thus, the often quoted statement that “optimal coherent
detection of MSK and optimal coherent detection of BPSK yield the same bit error probabil-
ity performance” is, strictly speaking, incorrect. Again we remind the reader that the reason
behind this inaccuracy is that MSK is defined as a frequency modulation as per (10.74) to-
gether with (10.142) and its equivalent I-Q representation requires a differential encoder at
the transmitter and hence a differential decoder at the receiver.
It is possible, however, to modify MSK so that the above statement becomes valid.
Since a differential encoder followed by a differential decoder is equivalent to a unity trans-
mission device, that is, input = output, then preceding the MSK modulator with this combi-
nation (see Fig. 10.29) still produces s(t) as an output. Comparing Figures 10.27 and 10.29,
we then have the equivalence illustrated in Fig. 10.30, namely, that preceding an MSK mod-
ulator with a differential decoder has an I-Q representation identically equivalent to that of
offset QPSK with, of course, half-sinusoidal pulse shaping. The modulation implemented in
Fig. 10.30 is referred to as precoded MSK and as such its receiver requires no differential
decoder following the data detection. /t is this modulation that has the exact same bit error
probability performance as BPSK, QPSK, or offset QPSK. While all of this may seem like a
moot point, it is nonetheless important to point out the discrepancy. In this text, we shall be
careful to make the distinction between MSK, and precoded MSK both of which have the
identical power spectral density given in Chapter 2.
682 Demodulation and Detection of Other Digital Modulations Chap. 10
Differential MSK
Decoder Modulator
Yc, At-(2n-1T)
(-1)" COs @.t
2) n-1
Pulse
Shaping
(p(t))
Data
Splitter
Pulse
Shaping
(p(t))
» 4,At- 2n7,)
(4s)
(1)4
3utpuodsai09
0}
[xews
Id
rot-? -? uf 172(%-1) fe
eansig[¢E-O, uy uonejuswedui
Jok Yoo]q 10j99}ap
JO} Wasayoouou
YS paseg
UO& [oquiks-om}
uonearasqo paquday)
wos [$1]© aI (Aa
683
684 Demodulation and Detection of Other Digital Modulations Chap. 10
Tp 27, gon
Ba = (| r(tye
[ete Jom dr — if rite
eg a dt
0 Ty
Taking the squared!? magnitudes of the four complex quantities in (10.152) gives terms of
the form:
|Bil? = (arRs+ aaIe +.a3Rer + aalsr)” + (a5Re + 461s +a7Rst t+agler)? (10.161)
where
2T, 2T)
Rs =I) P(t) COS @ct SIN@m (t — Th) dt, Re = / r(£) COS wel COS Wm (t — Ty) dt
Tp Th
2T, 2Tp
Noe / r(t) sin @ct sin@p, (t — Tp) dt, I-= / r(t) SiN Wet COS @m (t — Tp) at
3 ci (10.162)
Tp Th
Rsr =|! r(t) COS Wet SiIN@mtdt, Rer =! r(t) COS @ct COS Wmt dt
0 0
Tp Th
[37 a r(t)Sin@ct sin@,t dt, Icr =i] r(t) Sin @ct COS Wmt dt
0 0
and the a;’s take on values of +1 depending on the value of i. For example
13. Note that a decision based on the maximum |8;| is equivalent to one based on the maximum |Bi\*.
0}
UOISIDap
s01Aap
685
686 Demodulation and Detection of Other Digital Modulations Chap. 10
Since
p*
1
PoE) <5 [1= O(/b1, fai) + OL/ai: vb]
E; (10.166 )
+5[1- O(vbs, Jaa) + (Vas, Vou)|+ 7 exp (-=)
1
2+ /71+4/n?1 =. (
Ppl) Soh terpenes
3 2/1+4/n2 2 (2No
(10.167)
=F
[2+4/x
8/7
1 | E, L, Gee,
jet si0~4/x)) + exo =)
What we see from (10.167) is that the bit error probability performance (at least in terms of
its asymptotic upper bound) of noncoherent MSK with two symbol observation and block-
by-block detection is a scaled mixture of the performance of DPSK and coherent PSK’s
with effective E/No’s determined from the arguments of the erfc functions. Figure 10.33
illustrates the upper bound of (10.167) as well as the accompanying results for N, > 2
as determined from (10.137). From this figure, one can see the dramatic improvement in
performance obtainable with moderate values of Ns.
As mentioned at the beginning of Section 10.2.1, another common pulse shape for full
response CPM is the raised cosine (RC) pulse, which is a special case of LRC corresponding
to L = 1. In particular, the frequency pulse g(t) has the form
| a Qnt\.
et) =4 EF (1 COS Sr i Ostet (10.168)
0; otherwise
For h = .5, this modulation scheme has independently been referred to as sinusoidal
frequency-shift-keying by Amoroso [17]. The primary advantage of this form of CPM over,
say, MSK is its increased rate of spectral roll-off as discussed in Chapter 2. It also possesses
an I-Q representation in the form of offset QPSK with I and Q pulse shapes as in (2.122).
As such, coherent and noncoherent detection techniques are quite similar to those for MSK
and will not be discussed here in detail. As an example, for noncoherent block-by-block de-
tection, the upper bounds on bit error probability given by (10.132) [together with (10.133)]
Sec. 10.2 Continuous Phase Modulation (CPM) 687
10-1f
10-2
10°3
10-4
RP
UPPER
ON
BOUND
(E)
10°5
10°6
for Ns = 2, (10.135) [together with (10.136)] for N, = 3, and (10.137) for arbitrary N, still
apply with 6 now given by [15]
n—\
Ns~1 i% Dd (i-4)
Ri Sh E,e i=0 (10.169)
n=0
where
Ie h , h i h me
&j = 5 Deedat (5(a; = x) c (5(a; = a) + 2)=H A (5(a; fae ai) — 2)
and
Also in (10.170), ex is again the Neumann factor, that is, e, = 1 ifk =O and e, = 2 ink 505
and Jx(x) is again the k"* order Bessel function of the first kind with argument x. As such
the values of a and 3, that is, the arguments of the Marcum-Q function, given previously for
CPFSK still apply to SFSK if sinc h is replaced by sinch x f(h) where
= h so h(2k — 1)
fies Dderau(h) lean |—2 2 Jnx—1(h) eed (10.172)
€ mt
+a;(t) sin (=)sin [(@e + bs(t)wa) n} (10.173)
AY
= Sc(t) + Ss(t)
where the even- and odd-numbered data waveforms d,(t) and d,(t) that generate the
amplitude-frequency pairs a;(t), b-(t) and as(t), bs(t), respectively, are offset by 7;/2 in
their relative alignments.'* This staggered alignment between a;,(t) and a;(t), and simi-
larly between b,(t) and b;(t) together with the sinusoidal pulse shaping by cos (xt/T;)
and sin (7t/T;) results in the phase of s(t) at data transition instants t = (2i — 1)T, /2 and
t = iT; being continuous. To see this, we rewrite (10.173) as follows:
14. The data streams a,(t) and b,(t) have data transition instants at t = (2i — 1)T, /2 and the data streams a, (t)
and b;(t) have data transition instants at t = iT,.
Sec. 10.2 Continuous Phase Modulation (CPM) 689
IP
s(t) = [eat {cos|
(«.+ be(t)@g + =)|+ cos |
(x+ b-(t)@g — x)|
as(t) IU
Brass |
(o.+ bs (thog + =)|
(10.174)
= V2Pa¢(t) {cos(OO out)
bs(t) + be(t
xX COS |
(2Ar OO 4 a =)|
: bs a be < S Cc
+ sin eo
5 Put sin |
(e+ bs(t) :belt) @g + x)|}
where the upper signs correspond to a,(t)/ac¢(t) = 1 and the lower signs correspond to
as(t)/ac(t) = —1. Evaluating (10.174) at t= (2i — 1)T,/2, we get
Since a,;(t) and b,(t) cannot make a transition at t = (2i — 1)7;/2, then s(t) will be contin-
uous at these time instants. Similarly, at t = i7;, (10.174) evaluates to
SW) = J2P factcos (+) COS Wet + as(t) sin (=) sin vat| (10.177)
AY AY
which is in the form of an offset QPSK signal with sinusoidal pulse shaping, or equivalently,
precoded MSK.
A block diagram of a modulator for (2N )?-ary CPQFPM is obtained from the block
diagram of a (2N )?-ary QFPM modulator (Fig. 10.9a) by introducing a 7; /2 offset between
the alignments of d,(t) and d,(t) and adding sinusoidal pulse shaping to the s,(t) and s,(t)
signals prior to their summation. The result is illustrated in Fig. 10.34a. Similarly, the block
690 Demodulation and Detection of Other Digital Modulations Chap. 10
Demultiplexer
ane!
@, =@,-[N-(2n-])]@,; n=iorj sin
(|
Bile FM
Demultiplexer
sin @,t
@,= @, —[N-(2j—)]@,
diagram of a modulator for 4N-ary CPFPM is the same modification of the modulator for
4N-ary N-FSK/QPSK (Fig. 10.9b). The result is illustrated in Fig. 10.34b. The optimum
receiver for CPQFPM is a modification of the optimum receiver for QFPM (Fig. 10.10)
wherein the I and Q integrate-and-dump filters of the optimum biorthogonal detector blocks
are replaced by matched filters, that is, amplitude weighting by cos (zt/T;) and sin (zt/T;)
followed by I&D’s with integration limits as well as the sampling instants staggered by T, /2
with respect to one another (see Fig. 10.35). It should be clear from Fig.10.35 that the ideal
coherent error probability performance of CPQFPM on the AWGN is identical to that of
QFPM (see Table 10.2).
Using the methods presented in Chapter 2, it is straightforward to show that the
equivalent lowpass two-sided PSD of CPQFPM is given by [7]
Sec. 10.2 Continuous Phase Modulation (CPM) 691
oni tcoess
i eenT:
i=1,2,...,N
Optimum
Biorthogonal
Detector
Received
Signal
Optimum
Biorthogonal
Detector
sinossin=
J=1,2,...,N
Ss(f) = 5
T. alae
Sf i oe 3\72
1+ lo 1
Aesoes (1 + logy N) fT+2i-—-N+
8)PTp D[sine(2( 5) (10.178)
3\72
+ sinc (2(1+ log, N) fT) +2i-N +3)
which is illustrated in Fig. 10.36 for various values of N. The curve labeled N = 1 is
identical to the PSD of MSK. Comparing Fig. 10.36 with Fig. 10.12 we observe that,
analogous to the PSD’s of QPSK and MSK, the main lobe of the CPQFPM spectrum is
wider than that of QFPM; however, the PSD of CPQFPM falls off at a faster rate than that of
QFPM. This can also be seen by comparing the fractional out-of-band behavior of CPQFPM
(Fig. 10.37) with that of QFPM (Fig. 10.13). Finally, the power and bandwidth efficiencies
of CPQFPM are included in Tables 10.2—10.4 along with the comparable data for the other
I-Q modulations previously discussed.
As previously mentioned, CPFPM is a special case of CPQFPM wherein the I and Q
signal component frequencies are identical. With reference to (10.173), it is mathematically
modeled as a 4N-ary constant envelope signal of the form!>
15. We remind the reader that, as for N-FSK/QPSK, the number of bits per symbol, i.e., 7;/ 7p, is now logy 4N =
2 + log, N since two bits are required for the pair of binary amplitudes a,(t) and a;(t) and log, N bits to determine
the frequency common to both channels.
692 Demodulation and Detection of Other Digital Modulations Chap. 10
+ b(t)wa) t] (10.179)
+ a;(t) sin (+) sin [(@e
AY
= Sc(t) + S5(t)
Here a,(t) and a;(t) are offset in their relative alignment by 7,/2, that is, a;(t) has transi-
tions at t = i7; and a,(t) has transitions at t = (2i — 1)T;/2. However, b(t) being common
to both I and Q signal components now has the same transitions as b,(t) of CPQFPM, that
is, att =iT;. Also, s-(t) and s;(t) are orthogonal over the interval (i7;/2, (i + 1)T,/2) of
duration T;/2 sec, which allows for demodulating and detecting them separately.
An equivalent representation to (10.179) is
Sec. 10.2 Continuous Phase Modulation (CPM) 693
dB
n,
which clearly elucidates the constant envelope nature of the CPFPM signal. Moreover, the
phase continuity of s(t) is easily identified since a transition in b(t) and/or a;(t) at t = iT, or
a transition in a,(t) at t = (2i — 1)T;/2 produces a phase change equal to 277 (modulo 277).
An optimum receiver for coherently demodulating and detecting CPFPM is given in
[18] and is illustrated here in Fig. 10.38. Its structure resembles the optimum receiver for
constant envelope QFPM (see Fig. 10.11) . Because of the coincident alignment of b(t) and
a;(t), that is, both transition at integer multiples of 7;, the bit stream, d;(t), corresponding
to the signal frequency is estimated only in the receiver block, Ds, associated with s,(t).
This estimation, however, is aided by the receiver block, D,, associated with s,(t) so as to
improve performance. This will be explained shortly. Thus, at the output of block Ds, as(t)
and d,(t) [or equivalently, b(t)] are recovered.
In the block D,, only the binary amplitude data stream a,(t) (or a delayed ver-
sion of it) needs to be recovered since as mentioned above no frequency decisions are
made in this block. The particular correlator in D,, that is, the ith, from which a,(t)
is estimated should be chosen in accordance with the frequency w; decided on in the
D; block. However, since the time at which this frequency estimate is made, namely,
t =(n+1)T;, occurs in the middle of the pulse a,(t)cosmt/T; associated with s,(t),
then it is possible that a frequency transition may occur at this very same time instant.
Stated another way, the first half of the pulse a,(t)cosat/T; may correspond to one
frequency while the second half of this pulse may correspond to another frequency. Be-
cause of this, the integration interval of the I&D’s associated with the correlators in block
D, must be limited to a duration of 7;/2 sec so that the appropriate correlator for the
VisinE sin@,t
OL
(at)T,
(a+),
i, Coat
Pets Ler
x2 sin—sinayt
ty
y(t)
sample at
t=(n+ DT,
Ci
Delay
D, T,/2
a,(t—T, /2)
a2 cos COS Wyt
’
Figure 10.38 Block diagram of the optimum receiver for 4N-ary CPFPM (Reprinted from [18]
© IEEE)
694
Sec. 10.3 Further Discussion 695
first and second halves of the symbol a,(t) can be separately chosen in accordance with
the frequency decisions at t = (n + 1)T, and t = (n+ 2)T,, respectively. For example,
in order to detect a,(t) [the sign of s,(t)] in the interval T, /2 <t <3T,/2, the particular
correlator corresponding to I&D integration in the interval T,/2 < t < T, should be selected
by the frequency decision at t = T, whereas the particular correlator corresponding to I&D
integration in the interval T; < t < 37,/2 should be selected by the frequency decision at
t = 2T;. To effectively utilitize both of these components in making the decision on a,(t),
the first-half symbol correlator output should be delayed by 7; sec and then added to the
second-half symbol correlator output (see the lower portion of block D, in Figure 10.38).
By taking the sign of this sum, we obtain an estimate of a,(t — T,/2).
As mentioned above, an improvement in detection of the signal frequency can be
obtained by utilizing the correlator outputs corresponding to the quadrature signal compo-
nent s-(t). Since midway in the detection interval for s,(t) a transition in a,(t) may occur,
we must separately add and subtract the first- and second-half symbol correlator outputs
corresponding to s,(t) (this is accomplished in the D, block of Fig. 10.38 by the blocks
Ci;i = 1,2,..., N) to obtain pairs of signals that can be used to aid in the formation of the
block D, output magnitudes, w;;; i=1,2,...,N, j=1,2,3,4. As in Fig. 10.11, the
pair of signals from each of the blocks C;;i = 1,2,..., N is coherently combined (added
and subtracted) with the corresponding I&D output of the D, block to produce a 3 dB im-
provement in SNR relative to the case where the frequency estimate is obtained without
aiding from the D, block.
In determining upper bounds on the symbol and bit error probability performances of
the optimum 4N-ary CPFPM receiver of Fig. 10.38, we follow the same arguments as used
to arrive at upper bounds on the performance of the optimum M-ary receiver for 4N-ary N-
FSK/QPSK (see Fig. 10.11). In particular, it is straightforward to show that the symbol and
bit error probabilities of Fig. 10.38 are upper bounded by (10.64) and (10.65), respectively.
Figure 10.39 is an illustration of the upper bound of (10.65) for values of N =
1,2,4,8, and 16. We observe from these numerical results that the power efficiency of
CPFPM is a monotonically increasing function of N and that relative to MSK (N = 1) sig-
nificant power savings (on the order of 1.5 dB at Pp(E) = 10->) can be achieved with only
N =2. A more detailed numerical comparison is made in Table 10.2.
P(E)
ON
BOUND
UPPER
E,,/No, dB
Figure 10.39 Upper bound on the bit error probability performance of CPQFPM with
N as a parameter (Reprinted from [18] © IEEE)
where
1 Tote Ven
p(t) = (1 — cos Hi,5 Ostet (10.182)
: otherwise
Note that for QORC, the amplitude A is related to the average power of the modulation P
by A = /4P73, whereas for QPSK the relationship is A = /P and for precoded MSK,
A = V2P. While the power performance of QORC is quite comparable to that of QPSK
and precoded MSK, the real advantage of QORC lies in its spectral occupancy. Indeed it
can be shown [19] that, except for a normalization constant, the equivalent lowpass PSD
of QORC is the product of the equivalent lowpass PSD for QPSK and MSK (or precoded
MSK), that is,
sinin 27
2xffTy||” Fe DfT,
nf 4 (10.183)
Ss(f) = ZPD
2nfTp 1— (4fT,)°
Thus, asymptotically the PSD of QORC rolls off at a rate f~° as compared with that of
MSK, which rolls off as f~*, and that of QPSK, which rolls off as f~*. Offset (staggered)
versions of QORC have also been considered for application on nonlinear amplification
channels [20].
Another I-Q technique that has merit on nonlinear amplification channels is a variation
of conventional QPSK referred to as 7/4-QPSK and was initially described by Baker [21].
Sec. 10.3 Further Discussion 697
The idea here is to reduce the maximum 100% envelope fluctuations due to possible 180°
phase reversals in QPSK without resorting to an offset form of modulation. While indeed
offset QPSK reduces the envelope fluctuation by limiting carrier phase changes to 90°, its
performance suffers relative to conventional QPSK when differential detection is employed.
By limiting the maximum phase change to 135°, 2 /4-QPSK achieves a compromise be-
tween QPSK and offset QPSK in the spectral spread caused by nonlinear amplification. The
following brief description of 7/4-QPSK explains how this reduction in maximum phase
change is accomplished without offsetting the I and Q channels.
The basic difference between 2 /4-QPSK and conventional QPSK is the set of values
assigned to the phases representing the four possible two-bit information sequences. In
conventional QPSK, the information phases used are 0, +7/2, 2 whereas in 1/4-QPSK
we use +77/4, +37/4. Since with differential encoding (which is required for differential
detection) the information phases represent the phase changes between adjacent transmitted
symbols, then we see that whereas with QPSK a maximum phase change of z radians is
possible, with 7/4-QPSK the magnitude of the maximum phase change is limited to 37/4
radians. In terms of the actual phases transmitted over the channel, with QPSK the allowable
set remains the same in every transmission interval, namely, 0, +2 /2, 2 whereas with 7 /4-
QPSK it alternates between 0, +7/2, a and +7/4, +37/4.
Finally, we bring the reader’s attention to multi-amplitude minimum-shift-keying
(MAMSK), a generalization of MSK introduced in the late 1970’s [22, 23]. As the name
of the modulation indicates, MAMSK is a multilevel version of MSK in that the effec-
tive inphase and quadrature data sequences are multilevel rather than binary. Specifically,
MAMSK assumes the form of (10.154) where the data sequences c, and d, take on values
B=) esi See +./M — 1. An alternate representation of MAMSK is in the form of the sum
of n = log, (log, M )MSK signals whose powers are distributed such that the i" component
of the sum, i = 1, 2,..., has an amplitude that is 2‘! times that of the first component
[23]. For example, for M = 16, the signal
wt wt
s(t) = Acos (ou+ Ay (=) te *»)+2Acos (ou+ Bn (=) oF mn):
2T, 2Tp (10.184)
nlp <t <(n+1)Tp
is identically equivalent to
mt
S(t) = ACOS wet Cos a7, [cos Xn + 2 COS yn]
b
, eee
— Asin wet sin aT, [Qn COS Xn + 2B, COS Yn] (10.185)
b
wt . a EES
= ACy COS Wet COS rT, — Ad, Sin Wet sin oT,
where the effective inphase and quadrature data sequences cy and d, take on values'E1, +3.
The probability of error performance of precoded MAMSK is identical to that of
QASK with the same value of M in the same sense that the probability of error performance
of MSK is identical to that of PSK. The advantage of the former in both cases is, of course,
the faster spectral roll-off achieved by using I and Q half-sinusoidal pulse shaping.
698 Demodulation and Detection of Other Digital Modulations Chap. 10
Table 10.5 A Table of Various Modulation Types and Their Properties, continued
ee
EE aE a ee a ee ae Se
We conclude this chapter by presenting, for quick reference purposes, a table listing
the various modulations considered in this chapter together with their acronyms, relation to
other modulation types, and key properties such as discontinuous versus continuous phase
and constant versus nonconstant envelope. These results are presented in Table 10.5.
PROBLEMS
10.1 Consider a digital communication system transmitting an offset QPSK modulation. As such,
the received signal is of the form
de(t)=.)) deip(t—iTs)
i=—0o0
— 1
d;(t) — ye dsi p (:
7 (:
at ;)r.)
i=—0o
with p(t) a unit amplitude rectangular pulse of duration 7; sec. The two data streams {d.;} and
{d,;} are independent, identically distributed (iid) +1 sequences.
The carrier reference signals generated at the receiver (see Fig. P10.1) are not ideal in
that their phase differs from that of the received signal by ¢ radians. That is
(a) Find the conditional error probability, P;(E£; 6), associated with detecting the sequence
{d.i} in the presence of the phase error $. For simplicity, consider only the detection of the
data bit do and ignore any second harmonics of the carrier frequency.
700 Demodulation and Detection of Other Digital Modulations Chap. 10
Carrier Parallel
r(t) Recovery to Serial
Circuit Convert
(n+1/2)T,
Figure P10.1
(b) If the I and Q data sequences {d,;} and {ds;} were originally obtained as the odd and even
bits of an iid +1 data sequence {d;} occurring at a bit rate of 1/7, = 1/27;, what is the con-
ditional bit error probability, P,(E; ¢) |ogpsk ; of the sequence {d;}? Express your answer
in terms of the conditional bit error probability of QPSK (non-offset), P,(E; ¢) |Qpsk »
and the conditional bit error probability of PSK, P,(E; ¢) |psx, in the presence of the
same phase error, ¢. That is, write P,(E; ¢) |ogpsk as a function of P,(E; ) |apsk and
P,(E; &) Ipsk-
(c) For any ¢ £0, how does P,(E; ¢) |ogpsk compare with P,(E; ¢) |psx? Is it better or
worse? How does P,(E; ¢) |ogpsk compare with P,(E; ¢) lapsk ? Is it better or worse?
How do these conditional bit error probability performances compare at ¢ = 0?
10.2 Consider an unbalanced QPSK receiver with imperfect carrier phase synchronization. Because
of the presence of a carrier phase error, @, between the received modulation s(t, 9) of (10.3) and
the I and Q demodulation reference signals, r.(t) and r;(t) of (10.9), the I and Q integrate-and-
dump outputs will each contain a term due to crosstalk [see (10.11)]. Normalized versions of
these crosstalk terms denoted as d,2 and d2,; are defined in (10.12). Using the approach taken
in Chapter 2, Section 2.1 to derive the PSD of a random pulse train, show that the mean-squared
values of d;,2 and dz; can be expressed in terms of the PSDs of the I and Q data streams as in
(10.22).
10.3 Using (10.22), derive the mean-squared crosstalk de, for the case where d(t) is an NRZ
waveform of rate 1/7; and d2(t) is an NRZ waveform of rate 1/7.
10.4 Using (10.22), derive the mean-squared crosstalk a for the case where dj (t) is a Manchester
waveform of rate 1/R, and d2(t) is an NRZ waveform of rate 1/R2.
10.5 (Gray Code for QAM) In Chapter 4 we defined the notion of a Gray code as applied to M-
PSK modulation. In particular, a bit assignment was made to each symbol (signal point on
the circular constellation) in such a way that adjacent symbol errors resulted in only a single
bit error. Since each point on the circle has only two adjacent neighbors, when viewed in
polar (amplitude-phase) coordinates, the Gray code for M-PSK is one-dimensional, that is,
it relates only to the phase coordinate. For QAM modulation, a two-dimensional Gray code
can be assigned in such a way that for each of the two rectangular coordinates, an adjacent
symbol error results in only a single bit error. Construct such a 2-dimensional Gray code for
Problems 701
Acos(@,t + 0)
PSK
Modulator
Figure P10.6
16-QAM, that is, show the appropriate bit to symbol assignments for each of the 16 points in
the constellation.
10.6 (Serial MSK) Consider the transmitter implementation illustrated in Fig. P10.6. Binary (+1)
data {dx} is PSK modulated onto a carrier Acos (wt + 6) to produce the signal x(t) =
co
A >> dkp(t —kT) cos (wt + 6) where p(t) is a unit rectangular pulse shape and @ is an
k=—oo
arbitrary phase. The PSK output is then passed through a linear filter H7(@) with finite impulse
response
Zsin@t; O<t<T
mys yi ; aie
re) {3 otherwise
Suppose that we choose the carrier frequency w2 = w, + Aw/2 and the filter impulse response
frequency w; = w- — Aw/2 where AwT = 72.
(a) Show that the output signal s(t) in the k transmission interval kT <t <(k+1)T depends
only on the input bits d, and d,_;, that is, write s(t) as a convolution of x(t) with h7(t)
and express the result as two integrals with appropriate limits involving only dy and dy_.
Ignore double harmonic terms.
(b) Evaluate your result of part (a) and show that it can be written in the form of an MSK-type
signal, that is
s(t) = A {ug
cos wzt + (—1)* vx cos wt} het Ska T
where {u,} and {vz} are (—1, 0, +1)-valued sequences that depend on dx and dy_; such
that uz, = 0 when v; = +1 and vice versa. Indicate your choice of 0.
10.7 In addition to coherent detection, MSK can be differentially detected as illustrated in Fig. P10.7.
Assume that n(t) is white Gaussian noise with PSD No/2, and s(t) is the usual MSK input, that
is
t
s(t) = Acos(wet + 6(t)); @(t) = deo xy KT SPER EDT
The MSK signal plus noise is multiplied by itself delayed one bit and phase shifted 90°.
The resulting product is passed through a lowpass filter H(f) which, for simplicity, we shall
assume does nothing more than remove second harmonics of the carrier frequency terms, that
is, assume that H(f) is wide enough not to introduce any distortion into the signal s(t). Also
assume that the carrier frequency and data rate are related such that w-T = 2kz (k an integer).
(a) Write expressions for A@(t) = 0(t) — 0(t — T) and the statistical mean of y(t) in terms of
the appropriate data bits {d,} that are valid over the interval kT <t < (k + 1)T.
(b) Show that a single sample of y(t) taken at t = (k + 1)T yields a suitable decision variable
702 Demodulation and Detection of Other Digital Modulations Chap. 10
n(t)
Threshold
Vi Compare
y(k+1)T)
Figure P10.7
for detecting d;, that is, express the statistical mean of y((k + 1)T) in terms of d; and give
its values for d, = —1 and dj = 1.
(c) Consider an input data sequence dp = —1, d) = 1, dz = 1,43 = —1,d4 = —1,d5 = —1,
dg = 1, and d7 = —1. Letting p(t) denote a unit rectangular pulse of width 7, and assuming
that noise is absent, sketch, over the interval 0 < t < 87, the following waveforms:
[o,2)
6. d(t —T)= D> y(kK+1)T) p(t — (k + 1)T) [the sample and held values of y(t)].
k=—00
You may assume A = | for the purpose of illustration. Assume an initial phase value
xo = 0. Compare d(t) and de) and justify your previous results.
10.8 Consider an offset QPSK signal with half sinusoidal pulse shaping, that is, precoded MSK. The
conventional receiver for this modulation (see Fig. 10.28 without the differential encoder) uses
I and Q matched filters that are implemented as follows. The I and Q demodulated signals are
multiplied by the half sinusoidal pulse shape p(t) = sin 7t/27) and then passed through I&D
filters spanning the appropriate 7;-sec time intervals. An approximation to the above matched
filter design that avoids the multiplication by p(t) is a windowed I&D filter, namely, one whose
integration time is shortened by an amount ¢ at the beginning and at the end of the integration
interval. In particular, with reference to Fig. 10.28, the multiplications by p(t) in the I and
Q channels are removed and the remaining I&D filters have integration intervals respectively
given by ((2n — 1)7T, + €, (2n + 1)T) — €) and (2nT, + «€, (2n + 2)T; — €).
(a) Evaluate the error probability of the resulting (suboptimum) receiver in terms of the rela-
tive gate parameter &/ Tp.
(b) Find the value of ¢/ 7), that minimizes the error probability of part (a).
10.9 Redo part (a) of Problem 10.8 for the case where the I and Q carrier reference signals are not
ideal and are given as in Problem 10.1. Express your answer in terms of the carrier phase error,
od, and €/ Tp.
10.10 An MFMSK signal can be viewed as an MSK signal with carrier frequency which may vary
from symbol to symbol and can take on one of N equiprobable values. Consequently, the
PSD of the MFMSK signal can be derived as the average of the PSD’s of N MSK signals
with carrier frequencies @; = w, + [2i-—(N +1)]a/T;, i=1,2,...,N. Starting with the
equivalent lowpass PSD for an MSK signal as given by (2.148), derive the equivalent lowpass
PSD for an MFMSK signal in a form analogous to (10.178), namely
&
Problems 703
Figure P10.12
catpyey N N\72
i=0
=
N Dp
10.11 For MAMSK modulation with M = 16, show the equivalence between the I-Q representation
of (10.185) and the multiple MSK representation of (10.184).
10.12 Consider a 16-signal amplitude-phase-shift-keyed (APSK) modulation whose signal point con-
stellation is composed of two circles (one having radius R; and one having radius Ry) each
containing eight uniformly spaced points (see Fig. P10.12). The phase angles associated with
the points on each circle are identical, that is, for each of the eight possible phase angles,
B; = (i — 1)7/4, 1 =1,2,..., 8, the signal can be at one of two amplitude levels. As such,
it is possible, in principle, to independently detect the amplitude and phase of the modulation.
It is desired to detect this modulation differentially both in amplitude and phase, hence
the acronym 16DAPSK (sometimes called 16 star QAM). Assume an ISI-free environment, for
example, Nyquist filtering equally shared between transmitter and receiver. Then the complex
samples of the receive filter output (taken once every 7; sec) are given by rz = 5ye/° + A, where
5. = V2Paze!™ A /2PS; is the transmitted amplitude-phase signal in the k'" signaling inter-
val, 9 is the unknown channel phase, and nz is the AWGN sample with variance a. Writing sx
in differential form, namely 5, = Sx—1 (+) exp {j (0% — O—1)} = 8-1 Aag exp (j Ady), then
the differential phase Ad, ranges over the set of values {6; = (i — 1)w/4, i1=1,2,...,8}
and the differential amplitude, Aax, ranges over the set of values {y~', 1, y} where y =
Ry/Rz > | denotes the ratio of the circle radii.
704 Demodulation and Detection of Other Digital Modulations Chap. 10
(a) Show that if P = S Y} \5,|* denotes the average power of the signal set, then the radii
2 2y2
Ri ey ee
‘ 1+y? é 1+y?
(b) Show that the conditional pdf of 7, given Aa,, Ady, Sp_1, 6 is expressed by
F S . 2
lik— Acne (V2PS.-12!*)|
vy (7 |Aax, Agx, 5,-1.6) ae
2no2 exp y— 20;2
(c) It would be desirable to detect the information pair Aaz, Ad, using maximum-likelihood
detection, namely, select that particular pair of values, Ad, Ady, that maximize the pdf
in part (a). However, S,_; and @ are both unknown. Thus, to obtain an implementable
but suboptimum solution, the term /2P S;_,e/® in the argument of the exponential is first
replaced by an estimate, cps (see Problem 10.13 for the characterization of the estimate)
and then the maximization of the pdf is performed. With this replacement, show that the
joint decision can be decomposed as follows:
. es TERRA yaa
Aa,
= min ~' |Aa, — Re {———~e-/ 4%
Aa, s 2
rk-1
Note that the decision rule for Adx follows the form of that corresponding to pure M-
DPSK, whereas the decision rule for Ady depends on the decision Adx.
10.13 In Problem 10.12 the unknown received signal in the k — 1® interval, 5,e/9 = /2P S;e/®, is
replaced by an estimate, r~—1, before maximum-likelihood detection of Aa, and Ady is per-
formed. One possibility for rh | is the maximum-likelihood estimate of 5,e/° using the set of L
past samples r,_1, 7x2, ..-, Tk Wherein the past decisions Adx-1, SGee alee ere
arrived at by the decision rules in Problem 10.12 are substituted for the true past values
IN Spy INGE), TES NE wk 1G
(a) Show that S,_) and S,—1 are related by
I-1
Spor = Se ht Aa, e-i4e-
f=l1
(b) Using the relation in part (a), and defining the past received complex sample vector r 4
Geetha rk—L), Show that the joint conditional pdf of F given Aa, Ad¢,, Seay 6 is
expressed by
1
f (F|Aa, Ad,, Sk 1,6) =
(2102)”
t=
of part (b), show that the maximum-likelihood estimate of 5,e/9 = /2P S,e/®, that is, the
value that maximizes the joint pdf is given by
Upon substitution of this result in the decision rules of part (c) of Problem 10.12, one
obtains a subeptimum detection scheme for 16DAPSK based on decision feedback.
10.14 An approximation (asymptotically tight for high SNR) to the average symbol error probability
performance of an M-ary two-dimensional (amplitude-phase) modulation scheme transmitted
over an AWGN channel is given as
1 M-1
where P7;,,,, denotes the maximum of the pairwise error probabilities between signals s;(t) and
sj(t), j =0,1,2,...,M—1, i j which is expressed in terms of the phasor representation
of these signals by
fh | eee
E, is the average symbol energy, and N7,,,, is the number of signals that result in P;;,,,,Tmax for each
value of i, that is, the number of signals at the minimum distance from signal point 7.
Evaluate the approximate average symbol error probability for the 8-ary signal constel-
lations illustrated in Fig. P10.14.
10.15 A quadrature overlapped raised cosine (QORC) signaling scheme is illustrated and defined as
shown in Fig. P10.15 where p; (t) is a unit rectangular pulse of duration 7; and p(t) is a unit
rectangular pulse of duration 27;. The output can be expressed as:
S(t) =A (a anp(t— 2m) COS Wet — (pe byp (t — ar) sin oa]
n=—oo n=—OoO
Note that the transmission rate on the in-phase and quadrature channels is 1/27) but the pulse
width is 47}, that is, the adjacent pulses overlap in time; hence the name of the modulation
QORC.
(a) If we consider the pulse shape p (t) as being generated by passing a rectangular unit pulse
pu (ty Ostsct,
P2\') =) 0; otherwise
through a filter H(f), find the impulse response h(t) = F{H (f)} of the filter.
(b) Show that the average power P of s(t) is related to the amplitude A by P = 3A/4.
706 Demodulation and Detection of Other Digital Modulations Chap. 10
Figure P10.14
(c) From your result of part (a), which choice, when filled in the blank space, correctly com-
pletes the following sentence: A raised cosine pulse of duration 47) is obtained by con-
volving a square pulse of duration 27; with the pulse shape of modulation.
(a) PSK (b) QPSK (c) MSK (d) OQPSK (e) SFSK
(d) Using the results of parts (a) and (c), derive the two-sided equivalent low-pass PSD S;,(f)
of s(t) as given by Eq. (10.183).
(e) If the modulation is now changed to an offset QORC scheme, which is now characterized
by
l
ais (dj+di-1), agi= an)
Laie (d; — dj_1)
Problems 707
¥ 4,7,0-207)
A>-0o
d(t)= )'d,p,(t—nT,)
n=—2
¥
b,p, (t-2nT,)
a>
¥
bp
n=-co
2nf,) Asin @,t
Figure P10.15
{a;;}
Figure P10.16a
{d;}
{ao; }
Figure P10.16b
(a) Show that for each i, either a7; = 0 or ag; = 0, but not both; that is, in each bit interval,
one of the two transducer outputs is zero while the other takes on value +1 or —1.
(b) Show that the transducer illustrated in Fig. P10.16c, whose outputs satisfy the relations
1 l / / / I
an = 5 (di +di-1), agi = 5 (di +i), di = (-1)'*"d;
708 Demodulation and Detection of Other Digital Modulations Chap. 10
{d;}
{29; }
Figure P10.16c
and thus have identical statistics, is equivalent to the transducer shown in Fig. 10.16b.
The inputs C;(t) and Cg(t) to the RF selector are carrier waveforms that are as-
signed to the modulator output s(t) in accordance with the following rule:
CMutivan— a
—C,(t) if aj; =—1
s(t) = » ifnst<@+)T
Co(t) if agi = ]
= Cot) if agi = —]
(c ~ Show that the phase of s(t) is continuous at the potential carrier switching transitions, that
is, at times t = iT, for all i and for all choices of the data sequence {d;}. Consider even and
odd values of i separately.
(d) With the choice of carrier waveforms as in part (c), show that the modulator realizes
differentially encoded MSK, that is, show that the output signal, s(t), is equivalent to what
one would get from an MSK modulator preceded by a differential encoder.
10.17 For the modulator of Problem 10.16, assume that the carrier waveforms are selected as
(a) Show that the phase of s(t) can never have a transition of z radians. Equivalently, show
that s(t) can never make a transition from C;(t) to —C;(t), or Cg(t) to —Cg(t).
(b) Show that the modulator realizes differentially encoded offset QPSK, that is, show that the
output signal, s(t), is equivalent to what one would get from an offset QPSK modulator
preceded by a differential encoder.
REFERENCES
1. Simon, M. K., “Error Probability Performance of Unbalanced QPSK Receivers,” JEEE Transac-
tions on Communications, vol. COM-26, no. 9, September 1978, pp. 1390-97.
2. Weber, W. J., Ill, “Differential Encoding for Multiple Amplitude and Phase Shift Keying Sys-
tems,” [EEE Transactions on Communications, vol. COM-26, no. 3, March 1978, pp. 385-91.
References 709
. Simon, M. K. and J. G. Smith, “Carrier Synchronization and Detection of QASK Signal Sets,”
IEEE Transactions on Communications, vol. COM-22, no. 2, February 1974, pp. 98-106.
. Simon, M. K., G. K. Huth and A. Polydoros, “Differentially Coherent Detection of QASK for
Frequency-Hopping Systems—Part I: Performance in the Presence of a Gaussian Noise En-
vironment,” JEEE Transactions on Communications, vol. COM-30, no. 1, January 1982, pp.
158-164.
. Simon, M. K. and A. Polydoros, “Coherent Detection of Frequency-Hopped Quadrature Modu-
lations in the Presence of Jamming—Part 1: QPSK and QASK Modulations,” JEEE Transactions
on Communications, vol. COM-29. no. 11, November 1981, pp. 1644-60.
. Divsalar, D. and M. K. Simon, “Maximum-Likelihood Differential Detection of Uncoded and
Trellis-Coded Amplitude-Phase Modulation over AWGN and Fading Channels—Metrics and
Performance,’ JEEE Transactions on Communications, vol. 42, no. 1, January 1994, pp. 76-89.
. Fleisher, S. and S. Qu, “Quadrature Frequency/Phase Modulation,” submitted for publication
in the JEEE Transactions on Communications. Also see S. Fleisher, S. Qu and S. Periyalwar,
“Quadrature Biorthogonal Modulation,’ Canadian Conference on Electrical and Computer Engi-
neering, September 25—27, 1992, Quebec, Canada, pp. 61.2.1-61.2.4.
. Saha, D., “Quadrature-Quadrature Phase Shift Keying,” U.S. Patent 4,680,777, July 1987.
. Saha, D. and T. G. Birdsall, “Quadrature-Quadrature Phase Shift Keying,” EEE Transactions on
Communications, vol. 37, no. 5, May 1989, pp. 437-48.
. Pelchat, M. G., R. C. Davis and M. B. Luntz, “Coherent Demodulation of Continuous Phase -
Binary FSK Signals,” Proceedings of the International Telemetry Conference, Washington, D.C.,
OTe
. Anderson, J. B., T. Aulin and C.-E. Sundberg, Digital Phase Modulation, New York and London:
Plenum Press, 1986.
. Osborne, W. P. and M. B. Luntz, “Coherent and Noncoherent Detection of CPFSK,” JEEE Trans-
actions on Communications, vol. COM-22, no. 8, August 1974, pp. 1023-36.
. Schonhoff, T. A, “Symbol Error Probabilities for M-ary CPFSK: Coherent and Noncoher-
ent Detection,’ IEEE Transactions on Communications, vol. COM-24, no. 6, June 1976, pp.
64452.
. Proakis, J., Digital Communications, New York: McGraw-Hill, 1983.
. Simon, M. K. and D. Divsalar, “Maximum-Likelihood Sequence Detection of Noncoherent Con-
tinuous Phase Modulation,” /EEE Transactions on Communications, vol. 41, no. 1, January 1993,
pp. 90-98.
16. Viterbi, A. J. and J. K. Omura, Principles of Digital Communications and Coding, New York:
McGraw-Hill, 1979.
We Amoroso, F., “Pulse and Spectrum Manipulation in the Minimum (Frequency) Shift Keying
(MSK) Format,” JEEE Transactions on Communications, vol. COM-24, no. 3, March 1976, pp.
381-84.
. Fleisher, S. and S. Qu, “Multifrequency Minimum Shift Keying,’ JEEE Journal on Selected Ar-
eas in Communications, vol. SAC-10, no. 8, October 1992, pp. 243-53. Also see S. Qu and S.
Fleisher, “Continuous Phase Frequency/Phase Modulation,” AIAA 14th International Communi-
cations Satellite Systems Conference, Washington, D.C., March 22-24, 1992, pp. 500-06.
. Austin, M. C. and M. U. Chang, “Quadrature Overlapped Raised-Cosine Modulation,” JCC’80
Conference Record, Seattle, WA, June 1980, pp. 26.7.1—26.7.5.
20. Simon, M. K., J. K. Omura and D. Divsalar, “Performance of Staggered Quadrature Modulations
over Nonlinear Satellite Channels with Uplink Noise and Intersymbol Interference,’ GLOBE-
COM’82 Conference Record, Miami, FL, December 1982, pp. A7.1.1—A7.1.8.
710 Demodulation and Detection of Other Digital Modulations Chap. 10
21. Baker, P. A., “Phase-Modulation Data Sets for Serial Transmission at 2000 and 2400 bits per
Second, Part I,” AEE Transactions on Communication and Electronics, no. 61, July 1962, pp.
166-71.
22. Weber, W. J., II, P. H. Stanton and J. Sumida, “A Bandwidth Compressive Modulation System
Using Multi-Amplitude Minimum Shift Keying (MAMSK),” NTC’77 Conference Record, Los
Angeles, CA, December 1977, pp. 05:5-1—05:5-7.
23. Simon, M. K., “An MSK Approach to Offset QASK,” JEEE Transactions on Communications,
vol. COM-24, no. 8, August 1976, pp. 921-23.
Chapter 11
Coded Digital Communications
1. Did you know? The first code to be used in digital communications was the Morse Code. The Morse Code
System lasted as the primary communication code until the end of the nineteenth century. The insightful Samuel
B. Morse used what is called a “variable-length code”; however, the “variable-length bit pattern” of the Morse
Code was a problem to interpret so a new code was developed. This was done by the brilliant Frenchman Jean
Maurice Baudot. Baudot’s code was based on five bits of information with each alphabet character represented by
combinations of binary digits “1” and “0”; there are 32 such combinations (codewords). It is from Baudot’s name
that the speed (information rate) of information transmission called the baud rate evolved. It is frequently used in
computer communications. Donald A. Huffman was apparently the first to qualitatively investigate variable-length
codes (see Proc. IRE, 40, 1952).
712 Coded Digital Communications Chap. 11
Now the K,-bit binary sequence which represents message m,, can be represented by the
“carrier waveform”
K.-1
sn@)= Yo pibokhh)y ma 0aMe (11.2)
k=0
where 7 is a random variable taking on values “0” and “1” with equal probability and
pi(t), i =0, 1, are the corresponding pulse waveforms. When viewed as vectors in a space
of K, dimensions, the M- points represent the vertices of a hypercube. Assuming the use of
the most efficient modulation-demodulation and detection system (BPSK) operating in the
2. For coded digital communications, we use M, as the number of messages to be coded and reserve M for the
number of distinct waveforms the modulator generates. This is consistent with the notation used in the first ten
chapters.
Sec. 11.2 Coded Digital Communication System Architectures 713
face of no bandwidth constraint, we have then that po(t) = — Pit) = /2E,/T; sin wet for
0 <t < 7; and zero elsewhere. With equal energy signals given by
Tp Tp
E, = PT, = if s*(t)dt = / p;(t)dt joules/bit (11.3)
0 0
2E
niey=0( 7") =0( am) a (11.4)
where we have assumed the AWGN channel and P is the average received power. Since the
errors in the K,-bit message occur independently, then the probability of correct symbol or
message detection is
Ps(C)
=(1— pp)*¢
=(1— pp)? (11.5)
Therefore, the message or symbol error probability (MEP) is
P,(E)
=1 —(1 — pp)* (11.6)
Since pp < 1/2 for all values of E,/No, then from (11.4) and (11.6), we have
which says that the MEP tends to one as the amount of information transmitted (K, bits)
becomes large. The reason for this behavior is that the Euclidean distance ./2E;, between
nearest neighbors remains fixed as K, increases, whereas the number of nearest neighbors
and the number of dimensions occupied by the signal set increase linearly with K,. The
probability that at least one of the K, relevant noise components will carry the received
vector closer to a neighbor than the transmitted signal vector becomes large as K; increases,
that is, there are K, chances for this to happen. On the other hand for a fixed T and No, the
MEP can be made small by either decreasing the data rate R, or by increasing the received
power of P Watts. This fate of an uncoded communication system performance can be
alleviated by employing encoding and decoding techniques at the transmitter and receiver
respectively; we now investigate methodologies of practical interest.
Chapter 1. In particular, level 3 and level 5 coded-system architectures will now be used to
further our investigation from a systems point of view (see Fig. 1.6).
Figure 11.1 illustrates the level 3 system architecture discussed in Chapter 1; this architec-
ture will lead us to our mathematical model of coded information flow between the informa-
tion source and information sink (user). As discussed earlier, we will not treat the problem
of source encoding and decoding? from an information redundancy perspective, that is, the
techniques associated with data compression should not be confused with channel coding
which is used for error control.
For most practical purposes, the channel encoder (see Figs. 11.1a and 11.1b) is char-
acterized by the three parameters M,, N-, and r-. Here, M, serves to characterize the number
of codewords in the codeword dictionary. These M, words are frequently called the code.
The parameter N; serves to characterize the dimensionality (codeword length) of the code
while r¢ specifies the code rate defined as r. = log M-/N¢ bits/dimension. Assuming that
the source encoder output is characterized by a Q-ary alphabet and that the messages that
are generated from this alphabet contain K, letters, then we require that M, = QO. This as-
sures that the source encoder and channel encoder* are matched with respect to the informa-
tion to be transmitted. Clearly, each message contains log, M, = K; log, Q bits. In the bi-
nary case, a case of great practical interest, Q = 2 and M, = 2*< so there are K, = logy M-
bits per message. In general, we will consider the class of (M., N-) Q-ary codes and, in par-
ticular, the class of (N., K-) binary codes.> In the binary case, the code rate is re = K-/Ne
bits/dimension. As we shall later see, r,' serves as the parameter which characterizes the
bandwidth expansion due to coding; both r, and r- | play key roles in what follows.
In the case where we are working with Q-ary codes, we will find it convenient to
refer to the codeword elements as channel or coded symbols. When we talk about binary
codes, we will refer to the NV. elements which make up the codewords as coded message bits,
message bits, or codeword symbols (there seems to be no standard in the literature). Thus,
the log, M, information bits, each of duration Tp sec, inputted into the encoder are converted
into N, coded message bits, message bits, or codeword symbols. The exact prescription for
this mapping defines the encoding algorithm (the type of code) and constitutes the subject
of Coding Theory. Later, in Section 11.6, the classification for the error correction and
detection coding techniques of greatest practical interest is presented.
When time is introduced, we require that each codeword duration T sec satisfy the
relationship T = T, log, M. = T,K- logy Q sec. For a binary code, this means that T =
K-Tp = N-T,- where T, is the duration of the coded message bit or code symbol. From this
relationship, we see that the message bit duration is related to the bit duration via the code
rate, viz., Tc = KeTp/Nc =1rcTp» sec. Thus, the channel symbol rate R = 1/T: = Rp/Tre is
r,' times the data rate Ry. This is why the inverse code rate rz! is referred to as the
bandwidth expansion. For example, a binary block encoder accepts data in blocks of K;
3. The term codec used in Fig. 11.1 was defined in Chapter 1 and refers to the channel coder-decoder pair.
Similarly, the term modem refers to the channel modulator-demodulator pair.
4. In general, N. > K. and the encoder adds redundancy in the data stream for error correction purposes. When
Nc < Ke, the data is compressed to remove redundancy in the information and thus, the term data compression.
For N. = Ke, the encoder acts like a scrambler for secure communications.
5. We use two different notations for binary and Q-ary codes to emphasize the different alphabets.
Sec. 11.2 Coded Digital Communication System Architectures 715
| M
Code Rate r, = dae = a bits/dimension
Nc c
AAA eneanneneeennesensenannassaseeesuseeseesseeneseneaeeuegenanessusneuansnannsasansunsusasesnacsseaacsnsansenensesenpeasnsenseneeesusnesescessanssseesenensseestensnsansacensegensensesensnsusinansensnteaensenentstensensenensenerenes “
bits and outputs coded message bits in blocks of N, coded bits (see Chapter 12). Since
redundancy r = N, — K- bits is added, N, will be larger than K,. On the other hand, we
shall see in Chapter 13 that binary convolutional coders can also be described as providing
N, output coded message bits for each group of K, input bits. The major difference is that
any channel symbol output consisting of N; coded message bits depends, not only on the
last set of K, input bits, but also on several preceding sets of input bits. Furthermore, with
binary convolutional encoders, K, and N, are usually much smaller than their block code
counterparts (see Chapters 12 and 13). In either case, we will see that low-rate codes have
716 Coded Digital Communications Chap. 11
Output Sequence
Message Input Sequence (Coded Bits)
mo 00 00000
my 01 01010
m2 10 10100
m3 11 11111
This one-to-one mapping between input and output channel encoder sequences defines the
code. Note that the above mapping produces a nonlinear code, that is, the modulo-2 sum of
two output sequences (codewords) does not necessarily generate another possible codeword.
Example 11.2
Consider a 16-ary message source (M, = 16) whose output is to be encoded using a 4-ary (OQ =
4) rate 2/3 code (K, = logg M. = 2, N- = 3). The 16 possible messages mo, m1, m2, ..., 15
are respectively represented by the 16 possible 4-ary sequences 00, 01, 02, 03, 10, 11, 12,
13, 20, 21, 22, 23, 30, 31, 32, 33 which serve as the input to the channel encoder in each
T-sec interval. Of the Q%: = 43 = 64 possible 4-ary sequences of length 3, we choose the
6. The codes used in the examples in this chapter are for illustration purposes only and are not necessarily “good”
codes.
Sec. 11.2 Coded Digital Communication System Architectures 717
following 16 sequences to act as channel encoder outputs corresponding to the 16 possible input
sequences.
Output Sequence
Message Input Sequence (Coded Bits)
mo 00 000
my 01 010
m2 02 202
m3 03 030
m4 10 102
ms 11 020
m6 12 302
m7 13 110
mg 20 002
mo 21 300
m0 22 133
m\\ 23 210
m2 30 321
M13 31 111
mi4 32 322
m15 33 213
Once again the above mapping represents a nonlinear code since the modulo-4 sum of any two
output sequences (codewords) does not necessarily result in another codeword.
Based on the coded system architecture discussed thus far, we observe that, as N,
gets large, the decision regarding the message is delayed for T = N.T, = K-Tp log, Q sec.
This delay is called coding delay and represents a constraint which must be considered by
the system engineer. Moreover, since M, = Q*« grows exponentially with K, we would
expect the complexity of the encoder and decoder to grow rapidly; this represents another
issue to be faced by the system engineer. In practice, there are other issues which affect the
system engineer’s choice of a particular coding-decoding technique, for example, it is driven
by implementation complexity, performance requirements, and cost. We will not be able to
address these issues from a theoretical point of view.
Furthermore for the AWGN channel, if codeword x, is transmitted, then x, = Xmn and
1] (Yn — sony]
Xmn) = ——=—= exp |-—————_ (11.9)
Level 5 of our information transportation system architecture (see Fig. 1.6), serves as the
coded system model when a communications carrier of intermediate frequency (IF) f,
Hz is introduced to transport the information (see Fig. 11.2). Introducing the carrier into
the system picture requires the use of a Modulator and a Demodulator and this requires
separate attention. Unlike the uncoded system, for analysis and synthesis purposes only, the
modulator function in a coded system can be viewed as incorporating the two functions of
signal generation and modulation. Similarly the Demodulator in a coded system can also be
viewed as incorporating the two functions of demodulation and detection (see Fig. 11.3). We
adopt this point of view here.’ In practice, the demodulator projects the received IF signal
to baseband; this projection is accomplished by a carrier synchronization circuit.
In one simple form, the modulator can transform, via a known technique, the code-
word x from the channel encoder output into the transmitted signal vector S,, [waveform
Sm(t)] for presentation to the IF channel. In yet another simple form, the modulator can
transform each coded bit or symbol from the codeword into the transmitted waveform. In
between, the modulator can transform a group of coded bits or symbols into the transmitted
waveform. In general, this process is tedious and is best accomplished in two steps using a
signal mapper which we now discuss.
7. Demodulation involves four major tasks: (1) synchronization of the local communication carrier with the
received carrier component used at the modulator, (2) use of this reference to convert the received modulation
to baseband, (3) extraction of channel symbol timing so that the “matched filtered” version of this baseband signal
can be sampled and the recovered clock synchronization maintained with the transmit clock, and (4) detection of
the demodulated channel symbols and decoding them in accordance with the appropriate decoding algorithm.
Sec. 11.2 Coded Digital Communication System Architectures 719
Prttcsceeneeenessenscenncnececnsacccncenerscececsncccccscccccensencccccacececs.
Information|
Source He
i™1 Source
Encoder
Channel
Encoder
:Xm| Modulator Salt
(M,, No, T,)
farhdZ
and zero elsewhere; we will refer to s,,(t) as the transmitted signal when codeword Xj,
is to be transmitted if and only if N = N, and M = M,. When this is not the case, we
will consider mapping methods to provide codeword transmission. From the above repre-
sentation, it is clear that the waveform set {s,,(t),m =0,---,M — 1} can be represented
by the N-dimensional vector 8 of its coefficients, that is, Sm = (Sm1, Sm2,°°°,SmN),m =
0, ---, M — 1. In other words, there exists a one-to-one correspondence between the wave-
form s(t) and the vectors s,,. AS we shall see, s(t) = Sm(t) if and only if M = M, and
N= Ne.
It is convenient to identify Ty 4 7;/N as the time used per dimension by the mod-
ulator and D = 1/Ty as the number of dimensions used per second by the modulator. In
Chapter 12, we will need to introduce the notion of bandwidth as it relates to signal dimen-
sionality. The interconnect will be made (Chapter 12) via the Shannon-Nyquist sampling
rate defined as B = 1/(2Ty) Hz; there, we identify dimensionality D = W = 2B Hz as the
required channel bandwidth.
There are two modulator rate parameters of practical interest. They are defined using
720 Coded Digital Communications Chap. 11
eeneceencnsessacenoscscncecesenss
eeeceveeseeesvesesescenecccssccscccsceaccscncscsecccscecesececsscsssesascsacassecssescseseassvaconsssevoesenenscusscsscasssocsnooaussssnoasooonavasseassesgseeeseeeey
Encoder ( Sm(t) §
(Mg No,T)
To
Decoder
the parameters which characterize a digital modulator. These are the channel rates defined
by R A log, M/T; bits/sec or by Ry A logy M/N bits/dimension. Thus, the modulator
output rate is R = Ry/Ty = DRy bits/sec. These two rates will play a key role in what
follows. We will be interested in coded digital systems which are implemented using the
(M., Nc, Tc) code and the (M, N, T,) modulator. Thus, the modulation/demodulation tech-
niques studied in the earlier chapters are applicable here. In the coded digital communica-
tion system architecture of Fig. 1.6, the signal generator and signal detector serve as bridges
which respectively connect the channel encoder to the modulator and the demodulator to
the decoder. We will describe these bridges next and present combined coding-modulation
waveform designs which can support the transmission of binary and multilevel codes using
the modulation-demodulation techniques (BPSK, QPSK, M-PSK, M-FSK, QASK, QAM,
and so on) presented in Chapters 3 through 10. In this chapter, we deal with rectangular
baseband pulses; generalization to arbitrary pulse shapes is straightforward (see Chapter 9).
Finally, the energy in the m'” waveform output of the (M, N, T;) modulator is
Ts
E,, = / s*(t)dt (11.11)
0
for m=0,---, M — 1. For equal energy signals, E;,,= Es and the energy generated per
dimension is Ey = E;/N joules. Since the time per dimension is Ty = T,/N, then E, =
PT, = PNTy = NEv joules. If the (M, N, 7;) modulator is used k times per codeword
transmission and we denote this composite waveform by S,,(t), then the energy required per
Sec. 11.2 Coded Digital Communication System Architectures 721
codeword is E = PT = kE, =kPT, = NcEn, = (Ke logy Q) PT, = (Ke logy Q)Ep Joules.
These various energy relationships will be useful in what follows.
Example 11.3
Consider mapping the channel encoder output into a ternary (M = 3) signal set of dimension-
ality N = 2. Such a signal set can be generated by an M-PSK modulation with normalized
signal points located on the perimeter of a unit radius circle at 9 = 0, 27/3, and 47/3 radians.
The orthonormal basis functions in (11.10) are defined by o1(t) = /2/T,cosw.t, d2(t) =
V/2/T; sin wet and the coefficients, 5.,, for each signal are tabulated below:
Signal 7] Coefficients
So(t) 0 Soi=1, so =0
51 (t) 21/3 Stl /2s 512 = V3/2
Kr x m2 Xm3 < Fr
XmNo =
{Sy(t)}
(1 of Mo)
- T
(d) Coded Waveform
the information rate as required. In this sense, the (M;., N., Ts) code is mapped onto the
(M, N, T) modulator.
More generally, Table 11.1 defines all time intervals, rates, energies, bandwidths,
and so on. Table 11.2 summarizes the coder-modulator time-interval and rate relation-
ships, while Fig. 11.4 demonstrates the time scale mapping and summarizes the encoder-
modulator mapping for some examples which follow. Some examples will clarify the role
of this mapper (see Table 11.2)
Coded BPSK: Consider first the (N., K;) binary code and the (2, 1) BPSK modulator. Here
M =2 and N =1 s0 thatk = N., with T = N.Ty and Ty = T,. In principle the n!” coded
message bit is used to modulate the basis function g(t) = /2/Ty cos wet, (n — 1)Ty <
t <nTy, which is used serially in time for all n = 1, 2,---, N-. In other words, the set of
Sec. 11.2 Coded Digital Communication System Architectures 723
j
Sm(t)
= /2/Tn > Smncoswct; O<t<T=NTy (11.12)
nae
724 Coded Digital Communications Chap. 11
Table 11.2 The (M-, Nc, T-) Code Mapped into the (M, N, T,) Modulator at a
Bit Rate of R, = 1/7), bits/sec
and Sinn = /EnXmn for (n — 1)Ty < t <nTw and zero elsewhere. The parameter Ey rep-
resents the energy per dimension. This coded waveform is generated using the (2, 1) BPSK
modulator k= N- times. If one now identifies the coded message bits x.» with the voltage
levels +1 or —1, we have a waveform carrying N,. coded message bits (see Table 11.2 and
Fig. 11.4). Figure 11.5(a) demonstrates the coded waveform mapping. Needless to say, bi-
nary phase-shift-keying combined with coherent detection can always be used to transport
coded information. As we shall see shortly, QPSK can also be used.
Coded QPSK: An alternate approach to transmitting the (N., K;) code is to use the (4, 2)
QPSK modulator. Here M = 4, N = 2 and one uses the modulator k= N-/ log, M = N,/2
times to generate the coded waveform S,,(t) corresponding to one codeword. In this case we
introduce an I-channel and a Q-channel which supports the transmission of the coded mes-
sage bits. Thus, T = kT, ss 2kT = N.T, and T; = 2Ty = 2T,. In order to map the code-
words onto the (4, 2) QPSK modulator, we demultiplex the coded message bits out of the
encoder for codeword x,, into the | and Q channel components Xmz = (Xm2, Xm4, +++, XmN,)
and Xm@Q = (Xm1, Xm3,***,XmN.—1). The orthonormal basis functions are $7 (t) = d2n(t) =
J/2/Ty cos wet and Pg(t) = d2n-1(t) = /2/Ty sin wet with w, a multiple of 27/Ty and
(n—))ivest snine n= 1,206, k. The resulting QPSK coded waveform model be-
comes (for N, even)
A
k k
Sin(t)= /2/Tw pe, Smn
COS Wet +) SmnSinwct | O<t<T (11.13)
n odd
1 channel Q channel
where Sn A / ENXmn for QPSK and xm» takes on value +1 or —1 volts. Figure 11.5 illus-
Sec. 11.2 Coded Digital Communication System Architectures 725
l- or Q-Channel py (t) = V2/Ty COS wet; (Nn -1) Ty StS nTy and zero elsewhere
Sint :
J Yi “
Yy A yj 4,
(NZ
if SLE,
T=
I-Channel Pon-4(t) = V2/Ty COS wt; (N -1) Ty <t<nTy and zero elsewhere
i. «
Sint
0F Og
Y Z
a Lia LLL
a(t) ‘7
7 Ot he
T; T = kT, = 2kTy
(b) For Binary Coded QPSK k = N,/2, 8-PSK k = N,/3, and M-PSK k = N,/ogoM
Figure 11.5 Codeword transmission using a set of time-orthogonal functions
mo 00 000
mM, 01 001
m2 10 010
m3 11 111
Table 11.3 Coded Modulation Waveform Mappers for Binary and Multilevel Codes
Table 11.4 Coded Modulation Waveforms for Transmission of Binary and Multilevel Codes
(2, 1) BPSK) %p< Smif) = op) Sn (Gn) for See Table 11.3,
or(M,
1) M-| m=0,1,...,Me—1;
k= Ne columns
2 and 3_
ASK
k k
(4, 2) QPSK, Xb) — Sy Sin P2n—1() + De SmnP2n(t) See Table 11.3,
(M, 2) n odd n even columns 2 and 3
ASK, M-
fee M-FSK I-channel _ Q-channel
Or =O, Nn bsog ie lle Kei
The expansions assume unit energy rectangular pulse shapes; if other baseband pulse shapes
are used these models can be readily modified (see Chapters 9 and 10) to accommodate
arbitrary I and Q pulse shapes
Here the modulator is used k = N-/ log. M = 3/3 = | times to generate the coded waveform
Sm(t) corresponding to each codeword. That is, all three of the N. = 3 coded bits per code-
word are assigned to one 8-PSK waveform. These waveforms are transmitted sequentially
over the channel every T; = 2Ty = T sec. Also note that 7, = Ty = T, /2. Letting So(t) =
/2/T; cos wet be assigned to the coded bit sequence “000,” S)(t) = /2/T, cos (wet + 1/4)
= J/1/T, coswt — /1/T; sinw-t be assigned to the coded bit sequence “001,” S(t) =
J2/T; cos (wet + 1/2) = —./2/T,
sinwet be assigned to the coded bit sequence “010,” etc.,
then the mapping from a typical message sequence to the transmitted 8-PSK modulation would
appear as below:
S3(t)
si (t), 81(t), so(t), So(t),
SSeS ee
s1(t), So(t), ---
So(t)
This mapping is identical to that in part (d) of this example. The reason for this is that the
code to waveform mapping is only dependent on the values of M and N associated with the
modulation and not on the actual form of the signal constellation, that is, the values of s,,,, in its
orthonormal representation.
Example 11.7
In this example, we consider the use of a ternary M-PSK (M = 3, N = 2) modulator in combi-
nation with binary (Q = 2) and ternary (Q = 3) rate 2/3 codes.
(a) Binary Code (Q = 2)
Since each symbol of the ternary PSK modulator would have to be assigned to logy M = 1.585
coded bits, the use of a binary code does not lend itself naturally to this sort of mapping since
the conversion of coded symbols to coded bits (in accordance with BCD) has an integer rela-
tionship.
(b) Ternary Code (Q = 3)
Here it is not necessary to convert the coded symbols to coded bits since the ternary code al-
phabet represents a natural match to the ternary modulator alphabet. In particular, each coded
symbol of the ternary codeword, regardless of the value of N,, would be assigned to one of the
three signals in the modulator of Example 11.3.
(Sm1+ Sm2s°**+ Smn,) Of its coefficients in this expansion for 0 < ¢ < T. The demodulation-
detection process in Fig. 11.3 reduces r(t) to its projection r = (71,72, °°, ry,); this process
is covered in Chapter 3 and this chapter. The vector r characterizes the output of the demod-
ulator (demodulation processor followed by matched filter detection). The channel decoder
then transforms r into the codeword output v = (v1, v2, *--, UK.) from the codeword dic-
tionary. The source decoder, based on the rules of source encoding, transforms v into an
estimate m of the information source output. Source decoding (data compression) tech-
niques are not covered herein.
The IF channel model consists of modeling numerous subsystems; the subsystems
which make up the model are demonstrated in levels 6 through 8 of the system architecture
shown in Fig. 1.6 of Chapter 1. For the continuous IF CMC it is characterized by the known
channel transition probability density f(r|S) and, if the channel is memoryless, then
Ne
fOISm) =| |£ralsmn), (11.14)
n=!
when message m is transmitted, that is, message m — X, — Sm. For the AWGN channel
] Ca aa
Sa
(Tn\Smn) = Hor Xp P|
f rn\S No (11S)
for all m and n. Frequently, sn» = NE Near for all n and m so that the IF channel input-
output is directly related to the baseband channel input-output via the energy Ey. used per
dimension. So much for the generic baseband and IF channel models of a coded digital
communication system. They are both referred to as CMCs.
Practical demodulator outputs are defined on the continuum, that is, re R%¢ where Re de-
notes the N,-dimensional Euclidean space. However, before the demodulator output can be
processed by a digital signal processor (decoder), some form of amplitude quantization on
the observable r must be performed. The quantization process is accomplished in practice
using some form of vector quantization circuit (see Fig. 11.6). In essence, the quantizing
circuit converts the CMC into the DMC which is then an approximation to the CMC. We
will discuss the details of this probability transformation shortly. For the moment however,
we note that this circuit transforms each observable r, € r from a random variable (r.v.) of
the continuous type into a discrete r.v., say R,. Strictly speaking, we will need N, quantiz-
ing circuits; however, this can be avoided by using BPSK modulation, requiring one quan-
tizer, and using the modulator and demodulator k = Nz times to create r (see Fig..11-7);
In this case, the components of r are output serially in time every Ty sec. If M-PSK or
QAM modulation is used, we need two quantizing circuits; one for the I-channel and one
for the Q-channel. In addition, one will need to compute the tan! (ro, /r2n—1) sequentially
in time for M-PSK, M > 4. The QASK or QPSK modulator and demodulator will only
need to be used k = N, /2 times to transmit the coded message bits serially in time (see
Fig. 11.7). The quantizer converts the continuous random variable r, into the discrete ran-
dom variable R,. In particular, a J-point scalar quantizer Q) is a mapping Q| : rp, > Rp
where ry is the real tine in the n™ direction of an N--dimensional vector space and R, =
Qi(rn) € {Rno, Rni, +++, Rn,s—1} in the quantizer output set of size J forn = 1,2,---, Ne.
The output values are referred to as output levels, Associated with each J-level quantizer is
a partition of the real line r, into J cells of width A; for j=0,1,---,J—1.
Sec. 11.2 Coded Digital Communication System Architectures 733
f(rAlSon)
f(rAlS4,)
The vector quantizer of dimension N, and size J illustrated in Fig. 11.3 is a mapping
from a vector r = (r},72,---,7n,) in N--dimensional Euclidean space Rc into a finite
set R = (Rj, R2,---, Rn.) where each component takes on one of J output levels; R, €
{Rno, Rni,**-, Rn,s—1) for all n = 1,2, -++, Nc. Thus, R = Qy,(r): RNe > R.
The design, that is, selection of the number of quantization levels, of the quantizing
circuit is an important part of the demodulator design. Assume that R, takes on one of J
values (a J = 27 level quantizer), say (Rno, Rni,---, Rny—1). For example, an 8-level (¢ =
3 bit) quantizing circuit is illustrated in Fig. 11.6; this figure demonstrates the quantization
regions, and the input-output transformation law.
In order to discuss the process further, two more concepts involved in error control
system design must be presented, viz., the concept of hard decisions and soft decisions.
From the point of view of codec design, creation of the DMC from the CMC outputs is
most important. When the demodulator outputs r, and this value is quantized into two levels
(J = 2), for all n, the demodulator is said to make hard decisions and the decoding process
is termed hard-decision decoding. When the demodulation outputs are quantized into more
than two levels (J > 2), the demodulator is said to make soft decisions and the decoding
process is called soft-decision decoding. A soft-decision demodulator is more complex to
implement when compared with a hard-decision demodulator. Soft-decision demodulators
require automatic gain control in the signal path and qg = log, J bits have to be manipulated
for every channel symbol interval of duration Ty sec. Infinite bit quantization is frequently
referred to by system engineers as no quantization, that is, perfect representation of the am-
plitude voltages associated with r is retained, no quantizer is used, and no information is
lost at this point in the decoding process. We defer until later the presentation of the details
regarding codec performance as a function of the quantizer design parameter J = 27 and the
734 Coded Digital Communications Chap. 11
Quantizer
Quantizer
Quantizer
Figure 11.7 Optimum demodulation of the coded message bits: (a) BPSK and ASK
demodulator, (b) I-Q demodulator for QPSK, M-PSK, QASK, or QAM
chosen coding-decoding technique. Suffice it to say, we shall show in the next chapter that 3-
bit soft-decision demodulators improve the communication efficiency by approximately 2.2
dB for coherent detection and 4.7 dB for noncoherent detection when compared with hard-
decision demodulators. On the other hand, 3-bit quantizers loose little in efficiency when
compared with the unquantized demodulator output (see Fig. 11.6). Figure 11.8 demon-
strates the DMC model for the quantizer shown in Fig. 11.6. The confidence thresholds
illustrated in Fig. 11.6 will be useful in creating erasures in Reed-Solomon decoding (see
Chapter 12).
As we shall see later, there is only a small difference in the communication perform-
ance efficiency achievable between eight-level quantization and the unquantized case. The
eight-level quantized outputs involve one decision threshold and three pairs of confidence
Sec. 11.2 Coded Digital Communication System Architectures 735
Decoder Input
7 Ta
P 6 110
My, =0-® So, = VEN ¢ Cee 5 pomdonce
101 | Thresholds
4 100
Hard-Decision
3 011 Threshold
Via
mM, = 1->5,, = -VEy LN 010 Confidence
c BN 2
Thresholds
1 001
Channel Transition Probabilities:
LaF SR a ee ee oO 0 000
ee Od OS a Vay, Confidence Bits
qe OOn ls Nat, 2,033; KeeiN. Hard-Decision Bit
Figure 11.8 DMC for BPSK demodulation using 8-ary (3-bit) soft-decision
demodulation (for QPSK one has an I-channel and a Q-channel; however,
Pale) er IN-/D)
QAM or QASK. To accommodate multilevel codes, that is, the case where the coded mes-
sage bits x, can take on more than two values, quadrature amplitude modulation (QAM)
can be used and the coefficients s,,, can be adjusted to give the appropriate mapping. The
telephone channel is a good example of this. Here the data rate requirement is 9,600 bits/sec
and the modulator can support the transmission of 2,400 symbols/sec (bauds). Without
channel coding, a 16-point QAM modulator (Chapter 10) can be used to accommodate
the data rate. With coding and a code of rate r,=4/5, a 32-point signal structure will sup-
port the 9,600 baud data rate. In practice, the CCITT V.32 convolutional encoder is used in
combination with a QAM modulator. We consider the transmission of multilevel codes in
Chapter 12.
m=1;S,=-VEy R=1
P[0|0] R=0
m = 0; S9 = VE
m=1;8,;=-VEy_ a
mo =0—> so=./En, or mj = 1 > 5; = —/Ey,, for all coded message bits, the channel
transition probabilities are given respectively by
0
Pir <0\sol=f_f(rlsopdr = p
—0Oo
(11.17)
CO
multipath, interference, jammed, and fading channels. Telephone channels typically have
impulses on the line which last for approximately 10 ms.
We now address techniques whereby channels with memory can be engineered such
that they appear to the channel decoder as memoryless. Interleaving or scrambling is the
common technique used to make the channel appear to the decoder as memoryless when
burst noise is present. In practice, the system designer begins by specifying some maximum
noise burst length which the system must tolerate—the so-called channel memory time.
Once this is determined, the design of the interleaving system can proceed.
To mitigate channel memory, thereby taking full advantage of coding, an interleaver-
deinterleaver (I-D) system is inserted into the signal path. We will discuss the placement or
logical position of the I-D system in the functional system architecture in the next section.
For the moment, we present various interleaver types and discuss the trade-offs between
their performance measures of interleaving depth (span), interleaving (channel) delay, and
interleaving memory size. The I-D system is critical to being able to realize the performance
gain provided by coding in channels with memory. The function of the I-D system is to ef-
fectively reduce the degradation in bit error probability by randomizing the distribution of
the burst errors (or correlated bits) caused by channels with memory. The interleaving proc-
ess serves to scramble the order of the code symbols x,,, in time while the deinterleaving
process unscrambles the recovered symbol stream into the original sequence for application
to the decoder. Error bursts which would overwhelm the decoder are spread out by the dein-
terleaver over a number of channel symbols longer then the duration of the error burst. Thus
for purposes of error control, the burst or memory characteristics of the channel must be
specified.
There are at least five I-D system techniques of practical interest. They include: (1)
the periodic block I-D system, (2) the helical I-D system, (3) the periodic convolutional I-D
system, (4) the pseudorandom I-D system, and (5) the hybrid I-D system (see [2], [4], [6],
[7], and [8]).
Figure 11.10 depicts a block of length L = 4 symbols which is block interleaved to depth
D = 3. In essence, an (L, D) periodic block interleaver is characterized by the block inter-
leaver length L and interleaver depth D. Symbols moving across the channel are obtained
by scanning rows from left to right; symbols leaving the channel encoder, or entering the
decoder, are obtained by scanning succesive columns from top to bottom. The codewords
occupy the positions labeled ABCD, EFGH, IJKL, ... . The transmitted code symbols are
AEI, BFJ, CGK, DHL, and so on. Suppose a burst of noise “wipes out” the transmitted code
symbols AEI, and converts these into XXX (see Fig. 11.10) . The decoder input consists of
(XBCD, XFGH, XJKL). A single error correcting code which corrects one error will allevi-
ate this error. Also a single error detecting code will detect but not correct this error. Thus,
an interleaving depth of D units means that a noise burst must have length (D + 1)7y sec
in order to strike any codeword more than once.
Figure 11.11 demonstrates a block of length L symbols which is block-interleaved
to a depth of D units of time. We follow the convention that the block symbols x;; to
be transmitted over the channel are read into the interleaver columns and then transmitted
over the channel row by row from left to right. On the receiver side, the action is reversed.
The deinterleaver inputs are the rows of the interleaver output. The rows are read into the
738 Coded Digital Communications Chap. 11
ie vas
TIME Noise Burst
| = | b ae?) See.4
De oahloene | Diorieeee le
TIME
eld eee
| A. BaG.D xX BC D
DELAY
pes Eye Geek Xe a Gar
Figure 11.10 Block I-D system of length L = 4 and depth D = 3: (a) Interleaver action,
(b) Channel action, (c) Deinterleaver action, (d) Deinterleaver action in burst noise
columns of the deinterleaver and are read out of the deinterleaver (decoder inputs) by rows
as shown in Fig. 11.10(b). Thus, the D x L deinterleaving matrix is the transpose of the
L x D interleaving matrix.
In practice, the encoder output symbols are written into L vertically contiguous mem-
ory cells in a column and the outgoing symbols are read from D horizontally contiguous
memory cells in a row. The interleaver span is a measure of the time across which pre-
viously contiguous data is spread. It serves to characterize the ability of the interleaver to
randomize burst errors or spread out adjacent symbols over time intervals greater than the
noise burst durations. The span of the block interleaver is given by
CD NSEC (11.19)
where Ty is the time duration of each channel code symbol. The interleaving (channel)
delay introduced into the recovered code symbols is
whereas the required memory at the transmit and receive end is M, = 2LD symbols for
hard-decision demodulator/decoders. ;
Figure 11.12 demonstrates a helical interleaver for a block of length L = 4 with a depth
D = 3. The codewords are seen from this figure to be ABfa, CDbc, and EFde. As a function
of time the output of the encoder is ABfaCDbcEFdeABfa, the output of the interleaver
(channel input) is abFAcdBCefDEabFA, and the input to the decoder (deinterleaver output)
Sec. 11.3 Interleaving-Deinterleaving to Mitigate Channel Memory 739
Pee D
Ke Aeine PeX.1\4|
ieee D
Delay, L Mo eAL ees Xp {ta
1 2 D
DE OL, (ees) Ah
Time
=|
X; 1 Xo 1 eee XL 1
2 2 2
Atte Sbs2 * 2° Xr
Delay, D
e e
X(p-1)L +1
D e e e
Xo
D
is the same as the encoder output; no errors are assumed. The total interleaving delay
is seen to be 4 x 3 = 12 characters. If a burst of noise in the channel were to convert
the transmitted codeword abFA into the word XXXX, then the deinterleaver action would
scramble the X symbols such that the input to the decoder would be the codeword sequence
ABfXCDXcEXdeABfxX. Again a single error correcting code would be able to correct these
errors or a single error detecting code would be able to detect the errors.
Figure 11.13 depicts a code of length 8 which is helical-interleaved to a depth of 7.
This means that a noise burst must have a length of at least 8 in order to hit any code block
more than once. Each codeword lies in a column and the symbols are fed into the interleaver
by columns and then read out by rows. A block interleaver requires the receiver to know the
synchronization modulo LD; however, the helical interleaver requires the receiver to know
synchronization modulo L. In general, a code of length L can be helical-interleaved to a
depth of L — 1. The helical-interleaver requires a memory size about one-half the size of
the corresponding block interleaver of the same depth.
—Channel Time—>
re
A Cc d
faa orl
B Cc e
wae
f D E
la b F
Sars
Interleaver A Cc d
Time oar
B Cc e
ie eet
f D Ee
la b F
ayal
A Cc d
Sanam
B Cc e Figure 11.12 A (4,3) helical interleaver
delay lines are tens of thousands of symbols (bits) long, then they are most effectively imple-
mented with random access memory (RAM). We note from Fig. 11.14 that the convolutional
interleaver has a staircase geometry, as opposed to the rectangular geometry of the block
type. This staircase geometry may be thought of as a series of shift registers of graduated
lengths, as indicated in Fig. 11.14. Actual implementations are more likely to use a random
access memory (RAM) and stepped address pointers as opposed to shift registers.
The shift register model facilitates discussion of the operation. In contrast to the block
case where the data was stationary and a two-dimensional commutator was used, a one-
dimensional commutator is used in the convolutional interleaver and the data in each row of
the interleaver and deinterleaver is periodically shifted to the right. The commutator at the
interleaver input is stepped row by row from top to bottom inserting one bit in each row.
After the commutators reach the bottom position, each data bit in every row is shifted one
bit toward the output, and the commutators are reset to the lowest row. The output or last
bit in each row is overwritten after it is read. In contrast with the block interleaver, there is
no conflict between input and output, so long as both input and output precede shifting on a
consistent basis for all rows.
The convolutional interleaver has several advantages over the block interleaver. The
memory and delay required are lower and it lends itself to rate buffering of the data. When
inserted in the encoded symbol sequence, the system guarantees separation (of any two
symbols within LTy sec of each other in the interleaver symbol sequence) to be at least
[LD/(D — 1)]Tw sec between each other in the deinterleaved sequence. Thus, the inter-
leaver span is given by
and the interleaving delay is tq = 2t; sec. The memory required for the convolutional
interleaver is the sum of the row lengths. For a typical interleaver, it becomes
Sec. 11.4 Synchronization of I-D Systems 741
et e f g h Xx >
oe M| q r Ss
eat Pies Q | u Vv
Interleaver 15 SO Hl
ee eaten? iGo g Tr WZ
rch si ooleng
: as‘2 ee
axe
EOE a ee Ores
eS as “ erty
kat Q | u,v
Vener mMoovA: a U|
Mi DAE ne 1) (ie22)
biel
characters or symbols.
Cover . P
Sequence
oe ee
Channel
" f Cover
O 286 Sequence
29
Note: D = Delay Element; The cover sequence can be used to provide data privacy.
synchronization and resynchronization of I-D systems which accept hard or soft decisions
from the demodulator largely remain the trade secrets of organizations who design and build
error control systems.
: Transmit :
Local
)
Clock
Distribution Oscillator
Figure 11.15 Coded-digital communication system architecture for channels with memory
the system architecture. The interleaver lies between the channel encoder and the modulator;
the deinterleaver is inserted between the demodulator output and the channel decoder. Both
units frequently require substantial memory.
Suffice it to say, block decoders tend to be less complex when driven by hard deci-
sions from the demodulator; convolutional decoders can perform using either hard- or soft-
decision demodulator outputs. In order to gain the advantages of both error control meth-
ods in the design of an overall error control system, a convolutional decoder is driven by
the soft decisions from quantized demodulator outputs. This decoder then outputs hard de-
cisions to a block decoder for further error reduction. Combining the attributes of block and
convolutional codes into cascaded codes is known as concatenated coding.
Concatenated coding, introduced by Forney [5], represents an alternate technique to
mitigate noise bursts when channel bandwidth is not constrained. Figure 11.16 demonstrates
the architecture of a system that employs an inner code, an outer code, and interleavers and
deinterleavers designed to randomize the location of the burst errors in the symbol stream
seen by the channel decoders.
Typically, the inner code, that is, the code that interfaces with the channel, is designed
to correct most of the channel errors while the higher rate outer code reduces the error rate to
the desired level. The purpose of concatenated coding is to obtain a small error probability
with an end-to-end coded system implementation complexity which is less than that which
would be required by a single codec. Such coding techniques could also be applied to
improve the error probability performance of an existing coding system. One of the most
successful concatenated coded systems is one that employs a convolutional inner code and
a Reed-Solomon outer code with interleaving-deinterleaving [8]. The inner codec can take
maximum advantage of soft decisions from the demodulator output while the outer Reed-
Solomon codec can easily deal with the hard-decision output from the inner convolutional
codec.
744 Coded Digital Communications Chap. 11
In many cases, the analog signals being processed in the front end of the receiver or
the confidence symbols shown in Fig. 11.6 can be used to generate a “fade detector” or
“erasure indicator” signal (shown dotted in Fig. 11.16). When an error-correcting code is
used as the outer code, this erasure signal may enable the decoder to correct more severe
noise bursts than it could if there were no such control signal. We will discuss the error
and erasure correction capacity for RS codes in the next chapter. This will be followed by
presenting the performance of certain convolutional and concatenated coding systems.
Consider the binary concatenated coding-modulation system depicted in Fig. 11.17 which
cascades the binary codes (Nj, Kei) of code rates rej with M;j = QXci for all i = 1, 2,---,
L. At each encoder output, we must have ( let T;1 = Tp)
and since the channel rate 7? = (log, M)/T;, we can use (11.23) and (11.24) to write
L
T = (logy M)Tp || rei (11.25)
t=]
is the equivalent concatenated code rate. The bandwidth expansion of the coded system over
that of an uncoded system is r;!, that is
Ren al
—=r;! (Lean
Rp Ve
A BPSK modulator can be used to transmit the codewords one coded bit at a time, while
demultiplexing the coded message bits; a QPSK modulator can transmit them two at a
time, one in the I-channel and one in the Q-channel (see Fig. 11.5). In the next chapter,
we investigate block-coded digital communication techniques; this is followed by looking
at convolutional coded digital communication techniques.
| yiWSuUeI
je007
490/D,
J0}BII9SO
uoNNqisiqg
i Aouanbe4
pue Bulwiy Waj}skS I
jouueyD
: BAISOOY
49019 [2007
uoRNgISsIG JOJeIIIOSO
eee
SP aeaneeseeeconeecesseceucnoeoeseecescssecenasonsesenseeey
eeecesoeeees soneccononse
errrry Coe
YOS sudIsiog9q
et ferree
Secuscereceecceveeeuconensency
pocuececoncvncceoGe
aans1q
TT] gy
= Popoo-payeusjwou0
[eIISIP D UONROIUNUTWIOD
WII\sKs aM{oaqTYOIe
JO} sauULYO
YIM AJOUIOUL
745
746
18p0D
Z 48p09
7 Joyeijnpo;
9 — wg uoneing
: (yy) Pry) (Py ‘W) (S1'N
ae eeeeeeeeeasevece.
aanB1gLTT] payeuawwouog
SuIpoos
pue UoNe[Npow
waisXs Jopow
Sec. 11.6 Classification of Error Correction and Detection Codes 747
are a subset of linear block codes [9, 10]. The text book by MacWilliams and Sloane [10]
probably contains the most comprehensive discussion of block codes to date while Michel-
son and Levesque [11] gives a clear discussion of the details of codec design using block or
convolutional codes.
Bose, Chaudhuri, and Hocquenghem (BCH) codes are a subset of cyclic codes. BCH
codes can be classified as P-ary or binary (P = 2). P-ary BCH codes have codeword
symbols and codeword generator polynomial coefficients which come from Galois field
(GF) algebra, viz., GF(P) where P is prime. BCH codeword generator roots are from
GF(q)=GF(P”) where gq is the order of the field and m is an integer greater than one. Binary
BCH codes are most often used in practice; they are from GF(2). The codeword generators
of binary BCH codes are nonbinary; they come from GF(g) = GF(P™) = GF(2”). These
codes have inherent error detection capability without sacrificing any of the error-correction
capability. Berlekamp [12], Lin [13], Lin and Costello [14], Peterson [15], and Peterson
and Weldon [16] treat the Galois field algebra of linear block codes and present numerous
decoding algorithms. This important material lies outside the scope of this book.
Reed-Solomon (R-S) codes are special cases of q-ary BCH codes (see Fig. 11.18).
The code symbols of a binary R-S code are also nonbinary; they are generated from GF(q)
= GF(P”). R-S codes are very powerful burst and random error correcting codes. An ex-
cellent tutorial on R-S error correcting coding can be found in Geisel [17]. Discussions of
current applications of codes and codec design can be found in Odenwalder [18], Berlekamp
[19], and Berlekamp, Peile, and Pope [20]. Odenwalder [18] presents a clean presentation
from a system engineering and design point of view. Finally, Chapter 12 covers block coded
communication system design and performance, while Chapter 13 presents the use of con-
volutional codes and their performance.
748 Coded Digital Communications Chap. 11
PROBLEMS
11.1 (a) For the one-bit quantizer characteristic illustrated in Fig. P11.1, characterize the channel
transition probabilities when the input channel symbols are given by so = +/En, and
5; = —,/Ey, with equal probability. Assume additive white Gaussian noise and BPSK
signaling.
(b) Sketch the DMC (binary input—binary output) model and label all input to output connec-
tions.
11.2 Repeat Problem 11.1 for the quantizer characteristic illustrated in Fig. P11.2. The DMC model
for this case represents the binary erasure channel (BEC), discussed in Chapter 3. Here, M = 2
and = 3:
11.3 We are given the demodulator output random variable r to quantize. The pdf for r is given by
Apap a4
f(p)= ,Atp —-l<p<0
0 otherwise
Assuming a 2-level (1-bit) quantizer for this input signal with output levels R; = 1/2 and
Ry =—1/2
(a) Find the constant A.
(b) Find the pdf f.(o) of the quantizer error e(r) = Q(r) —r= sgn(r), where sgn(r) = 1 if
r >Oand sgn(r) = —1 ifr <0.
(c) Find the distortion D = E{e?(r)}.
(d) Now redesign the quantizer so that R; = 1/3 and Ro = —1/3. Find the new distortion
D = E{e*(r)}. Compare this to the result obtained in (c).
11.4 The span of an interleaver is a measure of the time duration over which previously continuous
bits are spread, and is a metric of the ability of the interleaver to randomize burst errors. For the
(10, 20) block interleaver, determine the span for symbol rates of 2.4, 4.8, and 9.6 kb/sec.
11.5 (a) For the block interleaver, show that the memory required at the transmitter or receiver is
given by
M, =2LDAq
where Ag = 1 for block interleavers and Ag = 1 for deinterleavers with Demodulators
using hard decisions and Ag = 3 for deinterleavers with Demodulators using 3-bit soft
decisions.
(b) For the (10, 20) block interleaver, evaluate the required transmitter and receiver memory
for both hard- and soft-decision Demodulation.
(c) Repeat (b) for the (39, 116) interleaver.
11.6 (a) Given a coded symbol rate of R = 1/Ty = 19.2 kb/sec, find the span for a block inter-
leaver of length L = 8 and depth D = 8.
(b) Determine the channel delay ty.
(c) Determine the required memory size M, for hard-decision demodulators.
(d) Using Fig. 11.11 and the numbers 1, 2, 3, ... to fill the interleaver input rows, demonstrate
the interleaver output matrix and the deinterleaver output matrix.
11.7 Repeat Problem 11.6 assuming a convolutional interleaver.
11.8 Show that the memory M, required for the convolutional interleaver is given by
Kab + De
a Be -1)
no
fr |
So) f(r \s,) quantization
5 ee ie
oe | ae Ey
matched filter output y
een) See hard
quantization
<q——_—_——- 0) 1 2 ———— 2-bit
quantization
1
REFERENCES
1. Shannon, C. E. and W. Weaver, The Mathematical Theory of Communication, Urbana, IL: Uni-
versity of Illinois Press, 1949.
2. McEliece, R. J., “Communication in the Presence of Jamming—An Information-Theoretic Ap-
proach,” Chapter 3 in Secure Digital Communications, ed. G. Longo, Wien, NY: Springer-Verlag,
1983.
3. Berlekamp, E. R., Algebraic Coding Theory, New York: McGraw-Hill, 1968.
4. Olson, D., “Hybrid Interleaver,’ International Télemetering Conference, vol. 19, San Diego, CA,
pp. 703-04, 1983.
5. Forney, G. D., “Burst-Correcting Codes for the Classic Bursty Channel,’ JEEE Transactions on
Communication Technology, vol. 19, pp. 772-81, 1971.
6. Ramsey, J. L., “Realization of Optimum Interleavers,” JEEE Transactions on Information Theory,
vol. IT-16, pp. 338-45, 1970.
7. Berlekamp, E. R. et. al., Interleaved Coding for Bursty Channels, Cyclotomics Technical Report,
April 12, 1983.
8. Berlekamp, E. R. and J. L. Ramsey, “Readable Erasures Improve the Performance of Reed-
Solomon Codes,” JEEE Transactions on Information Theory, vol. IT-24, pp. 632-33, 1978.
9. Clark, Jr., G. C. and J. B. Cain, Error-Correction Coding for Digital Communications, New York,
NY: Plenum Press, 1981.
10, MacWilliams, F. J. and N. J. A. Sloane, The Theory of Error-Correcting Codes, Amsterdam:
North Holland, 1977.
11. Michelson, A. M. and A. H. Levesque, Error-Control Techniques For Digital Communication,
New York: John Wiley and Sons, 1985.
12. Berlekamp, E. R., Algebraic Coding Theory, New York: McGraw-Hill, 1968.
References 751
13: Lin, S., An Introduction to Error-Correcting Codes, Englewood Cliffs, N.J.: Prentice Hall, 1970.
14. Lin, S. and D. J. Costello, Error Control Coding: Fundamentals and Applications, Englewood
Cliffs, N.J.: Prentice Hall, 1983.
15. Peterson, W. W., Error-Correcting Codes, New York: MIT Press, 1961.
16. Peterson, W. W. and E. J. Weldon, Error-Correcting Codes, New York: MIT Press, 1972.
iN/E Geisel, W. A., “NASA Technical Memorandum 102162,’ Houston, TX: Lyndon B. Johnson
Space Center, June 1990.
18. Bartee, T. C., Data Communications, Networks, and Systems, Chapter 10 by J. P. Odenwalder,
IN: Howard W. Sams & Co., 1985.
19) Berlekamp, E. R., “The Technology of Error-Correcting Codes,” Proceedings of the IEEE, vol.
68, no. 5, pp. 564-93, May 1980.
20. Berlekamp, E. R., R. E. Peile and S. P. Pope, “The Application of Error Control to Communica-
tions,’ IEEE Communications Magazine, vol. 25, no. 4, pp. 44-57, April 1987.
Chapter 12
Block-Coded Digital
Communications
In previous chapters, we have optimized the design of a digital communication system from
the viewpoint of the receiver; that is, we have determined optimum receiver structures and
evaluated their performance for an arbitrary set of modulation waveforms which represent
the messages in the channel. The following two chapters are primarily concerned with the
problem of optimizing the transmitter design, that is, determining the set of coded modula-
tion waveforms that minimizes the bit error probability. Optimizing the transmitter design
for a given channel characterization represents the signal design or coding and modulation
selection problem. We shall equivalently pose this problem as one of finding the best codec
and modem design which minimizes the bit error probability when the receiver used is opti-
mum. This leads us to presenting the Shannon capacity theory.
In Chapter 11, the functional architecture of a coded digital communication system
was presented. In this chapter, we shall exercise this model by developing the performance
of block-coded digital communication systems. The “Ro-criterion” for modem and codec
design will be introduced as the methodology for separating the design of the codec from
that of the modem; hence, the material presented in Chapters 3 through 10 can be used in
the modulation-demodulation design process for uncoded systems, while the theory given
in Chapters 11 through 13 can be used in the design of the coder-decoder (codec) and
Modulator-Demodulator (modem).
The analysis revolves around the fundamental parameter called the cutoff rate, Ro (“R
zero”). As we shall see, Ro (in bits/dimension) is a Shannon capacity-like quantity defined
for any CMC or DMC, whose value is less than the Shannon capacity, Cy., by approxi-
mately 3 dB in the energy-efficient region where coding is of interest. From a system engi-
neer’s perspective, Ro sets the practical limit on the rate at which information can be reliably
transmitted through the channel. Apparently, Wozencraft and Kennedy [1] were the first to
propose the use of the Ro criterion for modulation-demodulation system design; however, as
pointed out later by Massey [2], their proposal fell on deaf ears. They observed that a mo-
dem for use in a coded system should be designed to achieve the highest possible value of
Ro rather than the lowest value of post-demodulation bit error probability for the compara-
ble uncoded system. In this context, Ro is optimized not only with respect to the probability
distribution of data symbols entering the channel and the modulation-demodulation choice,
but also with respect to the quantization boundaries which define the decision regions at the
quantized demodulator output; that is, for a given number, J, of quantized demodulator out-
Sec. 12.1 Block-Coded Digital Communication System Architecture 753
Saas MIN eerie mice caSraco anos no cnucb seus soz aM Nosy aves zcaiohicoasansonXenasusueninsssnyshnyrohaqnscnscohreansatnasonsSuesneseascencosstassovsc«qbad
snaseaecouassdustessvbivexsokesbtassogncoseoterivonevastertses cavenessusvedseestoveg
Encoder
VEC IN 17)
puts (quantization intervals—see Chapter 11), the decision boundaries will be chosen so as
to maximize the value of Ro subject to a hardware complexity constraint. This criterion is
also applicable when convolutional and/or concatenated codes are used for error control. In
fact, in the development and analysis of powerful decoding algorithms for long constraint-
length convolutional codes, Ro has been given a more descriptive name—the computational
cutoff rate, sometimes denoted Rcomp. We proceed by developing an analytic understand-
ing of this important parameter with respect to the coding and modulation process.
of being precise, in this section we assume a Modulator which produces the coded modula-
tion waveform
Ne
Sm(t)
= D7 SmnPn(t), m=0,1,...,Me—1, O<t<T (12.1)
n=1
so that k = N- and T, = T;. As discussed in Chapters 3, 4, and 11, the building blocks
{@n(t)} form a known orthonormal set over the time interval T sec. Each of these coded
signals has energy
Ie
Em = PmT =i) S? (t)dt =| Sm |? (12.2)
0
for all m=0,...,M_ — 1. Notice that smn =./En.Xmn where Ey, is the peak energy
available per codeword dimension. For constant envelope signals (for example, M-PSK),
Ey, is also the average energy used per codeword dimension.
A particular mechanization of the encoder-mapper-modulator is shown in Fig. 12.1b
for the special case k = N., N = 1. Two cases of particular interest are binary coded BPSK
and Q-ary coded Q-ary ASK. We note that while the actual coded message bits assumed
in coding theory for BPSK are xm, =0, 1, we often write x, = +1, —1 corresponding
to voltage ievels realized in hardware. Thus, for binary coded BPSK, we have xm, = +1,
while for Q-ary coded Q-ary ASK we have Xmy = qj = (—1) + (2i — 2)/(Q — 1) with
probability p;, P= 1,...:,:Q;/ where Q is: even, and: Sap = ENOnn: For both of these
coded modulation techniques,
10
Wee | ees De eel
0, else
for all n.
Communication
Basis Functions (n = 1, 2,..., k)
time orthonormal y
(n—1)Tw St <nTw
(we a multiple of | gy_\(t) = {V2Tw cos wt, else
2n/Tn) 0,
Baseband channel perspective. For the baseband channel model (see Chap-
ter 11), we can equivalently pose the problem of communicating with block codes from a
geometric point of view. A block code consists of a subset of M. vectors (points or code-
words) chosen from an N,-dimensional vector space. It may be thought of as a way of
transmitting an integer m = 0 to M, — | to the receiver by sending the corresponding code-
word X. A decoding system for such a code is a partitioning of the N,-dimensional ob-
servation space into M, subsets or disjoint decision regions Aj». If the received signal plus
noise projection r belongs to A, the transmitted message is taken to be message m. See
Chapters 3 and 4 for all the details. An optimal decoding system for a code is one which
minimizes the probability of error for the code. Since the Gaussian density is monotone
Sec. 12.2 Performance of Block-Coded Communication Systems 757
decreasing about its mean, an optimal decoding system for a given code is one which de-
codes any received signal as the integer m corresponding to the geometrically nearest code-
word, assuming equally probable, equal energy signals. If there are several codewords at
the same minimum distance, any of these may be used without affecting the probability of
error. As shown in Chapter 3, a decoding system of this sort is called minimum distance
decoding or maximum-likelihood decoding; it amounts to partitioning the N--dimensional
space into N,-dimensional polyhedra or decision regions “centered” around the different
signal points. Each polyhedron is bounded by a finite number (not more than M, — 1) of
(N_ — 1)-dimensional hyperplanes. In this context, there are M. = Q*< different codewords
to be assigned to points in our N.-dimensional signal space, in which there are Q< points
available; it follows that there are QNe“ different ways (codes) to assign the M, codeword
vectors. This is the usual approach to studying coding system design; that is, the baseband
channel model is assumed, and the modem design problem does not explicitly enter into the
system model. The question of isometric mapping of the encoder output onto the modulator
output does not arise in the baseband model; the modem remains hidden in the baseband
channel and is assumed to be ideal. In this chapter, we work with the IF CMC and DMC
models which take the modem outside the channel.
1. A coded digital communication system is complex and consists of many parameters involving different time
intervals, rates, energy, energy-to-noise ratios, and bandwidth.
758 Block-Coded Digital Communications Chap. 12
us to introduce the reader to the important concept called the channel cutoff rate, Ro. As we
shall see, this rate is a simple function of the modem and the communication parameter
En,/No. In Section 3.2.7 and in (4.81), we showed that the conditional message error
probability performance of an M,-ary communication system with equally likely messages
over an AWGN channel is union bounded from above by a sum of the pairwise error
probabilities (PEP’s) discussed in Appendix 3D; that is,
Sec. 12.2 Performance of Block-Coded Communication Systems 759
PGy = WTy
= 2BTy
Processing PG,=WT,=2BT,
Gains PG=WT =2BT=N-
M.-1
c di;
From (12.3) we see that good coded modulation techniques employ coded modulation vec-
tors such that the Euclidean distance between any pair is large enough to ensure that the
channel noise will not cause the message that was sent to be confused with another message.
The Hamming distance D(x;, Xj) 4 Dj; between the two binary codewords x; and x;
is the number of positions in which the two codewords differ. For example, the Hamming
distance between (0, 0, 1, 1, 1) and (1, 0, 1, 0, 1) is 2. Considerable insight into the coded
communications problem is obtained by writing (12.3) in terms of the Hamming distance
Dj; when the coded modulation waveforms have equal energy E. In Section 12.8.1, we will
show that for binary coded BPSK modulation, the squared Euclidean distance is related to
the Hamming distance through
ely 2E ( Dij |
BES) o( 72(34) (12.5)
J=014)
760 Block-Coded Digital Communications Chap. 12
If we denote the minimum Hamming distance by Din and the minimum Euclidean distance
by dmin, then (12.5) can be used to upper bound the message error probability for a block-
coded system with equally probable messages which uses binary coded BPSK modulation.
The resulting bound is
denn 2E ( Dmin
P(E) < (Mc ~ 1) (2%.) = (Me - no( =z(N ))
(12.6)
=(M,-1
=(M,.-1)Q sine”
re ma
where we have used the facts that E = PT = P(N,T,-) = PN¢-(rcTp) = Ep(Nerc) and the
definitions
Since (4.80) cannot in general be simplified, we first illustrate key features of (12.3) by
assuming transmission of N.-dimensional, equal energy, orthogonal coded modulation vec-
tors, namely
Si = VEQ; (12.8)
with @; - ?;= 6;; for all i, j =0,1,...,M_- — 1 (these codes will be considered further
in Section 12.3.1). In this case, dj; = /2E. Thus (12.3), when averaged over equal prior
probabilities, reduces to
If we assume a binary code with Q = 2, then M, = 2*-. Since the bit rate Ry = log, M-/T
= K,/T or K- = RpT (see Table 12.2), we can write (12.10) as
From (12.12) we see that the decoded message error probability can be made to approach
zero exponentially with increasing T as long as the bit rate Ry satisfies the bound
R. <a aad 4 t e
Alternatively, from (12.13) we see that P(E) approaches zero exponentially as K, increases,
as long as the communication efficiency parameter
Ep
— >2In2~1.4dB (12215)
No
Since No = kT°, it follows that the minimum received energy per bit per degree Kelvin is
where the brackets denote the largest integer no greater than (Dmin — 1)/2. It is also true (see
Section 12.8.1) that the cross-correlation coefficient between any two binary codewords, say
x; and xj, of length N¢, is given by
where Aj; represents the number of term-by-term agreements and Dj; represents the number
of term-by-term disagreements (or the Hamming distance) between the two codewords. We
shall use these facts in what follows by considering the transmission of specific block codes
using BPSK. For this case, we note that Ry = 1 and that the coding to modulation mapping
is isometric.
Hie i AE (12.21)
Here the overbar denotes the complement of the “H” matrix, wherein the ones are replaced
by zeros and vice versa. Note that the rows of Hx, represent the M, orthogonal code-
words {Xm = (Xm1,Xm2,---,Xmn,)} (see Figures 11.3 and 12.1). Such a matrix is called
a Hadamard matrix [4] and the codeword dictionary generated by (12.20) is known as a
first-order (24°, K-) Reed-Muller code. For an orthogonal code, Aj; = Dj; so that y;j =
0, for alli # j. Thus, from (12.4b), yj; =0 = 1 — 2D;;/N¢ since Ajj = N- — Djj, so
where the superscript “7” denotes a matrix transpose. It is understood that the multiplication
and addition operations involved in the matrix multiplication are reduced modulo-2.
Example 12.1
(8,3) Reed-Muller code
o7"000
O00
Oo 1 0
artodl arom On kl tomy
G= pees
t
O0L 1 Shi0080 es | be t
me
Dea Om iO 1 Qual one : a
pe beat)
i tee
so that
Hx.-1
Bx.=|= *
ii ae
The message and bit error probabilities have been derived and are given by (4.102) and
(4.107), respectively. Here M, = 2N- and for any particular codeword x;, Djj = N-/2 for
M,. — 2 of the codewords x; (i # j), and Dj; = N- for the complementary codeword x; = Xj.
Table 12.4 Comparison of the MEP for Various Binary Coding Techniques
Uncoded BPSK
(signal points on
hypercube)
is Dmin = M-/2 with N. = M, — 1. Also note from (12.19) that y,; = —1/(M_- — 1) since
Ajj = Djj — 1 and Ne = Me — 1 = Ajj + Dij-
A comparison of the message error probability bounds, as found from the union bound
result given in (12.5), is given in Table 12.4 for various binary coding techniques. The case
of bit-by-bit signaling (uncoded BPSK) is also given for comparison purposes. For this case,
we note that r- = 1 and Dmin = | since the signal coordinates disagree in at least one place.
In Chapters | and 2, definitions and techniques for assessing the bandwidth requirements of
a communication system were presented. Although no constraint was placed on the band-
width of the binary coded waveforms discussed in the previous section, it is necessary in
practice to be able to allocate and control bandwidth. Assuming that the transmitter uses the
(2, 1, T;) binary coded BPSK modulator, then D A N/T; = 1/Ty = 1/T; = 1/T; dimen-
sions per sec; however, we know from the sampling theorem that a signal of bandwidth B
Hertz can be uniquely represented by NV, = 2BT samples taken at the Nyquist sampling rate
of 2B = 1/T, = 1/Ty = D samples per second [15]. Thus
2a a
wesconr=1] €[—[—_
Trans-_ codes BPSK orQpsk [1
Consider now the binary coding case (Q = 2) and assume that m = m,,. This implies that
the code vector is x,,, and the coded modulation vector is Sin = (Sm1,..-, Sm n,)- With BPSK
modulation we assume that Sm, takes on +,/Ey, or —/En, with equal probability for all
m=0,...,M-—1,andn=1,..., Ne. For these equal energy signals, the received signal
energy is
Japoouy Joye|NpoW |(;)1, 40VeINpOW
iy] 40/4
SQ
is} Jepooeq
‘{?x}
(a)'u .
Joye|npoW
yoo|g
SS
'{!s}
yoo|q
'f!x}
Jepoouz
Jepo0eq
'u (2) ;
((7)“c)
wn Oy apn (J) ‘ Oy 2wn
my)
*w'nO(
y90|q
Japoouy J0}e|NPOW Joyeinpowag 490/q
{ s} Oh, u ?WN Jepo0eq
*w?nO {1} >wnOf!
Oy wn (2)
eins
"ZI sjquissuq
Jo pepoo-yoo]q
[eIISIP UOTBITUNWIWOD
SUIDISAS
767
768 Block-Coded Digital Communications Chap. 12
Source
Encoder
“a, fer hf
Ensemble Modulators
Ensemble
of IF
Channels
Source
oder |
Decoder Ensemble Demodulators ll
Ne
E=PT=) $s? =N-Ey, joules (12.28)
n=1
and since T = N-T;, we have that Ey, = PT, joules. Since the calculation of the decoded
message error probability (MEP) is not feasible for an arbitrary block code, we consider
instead the problem of upper bounding the ensemble average MEP for the class of bi-
nary block codes. Examples of binary codes in this class include the Hamming and Golay
codes, while examples of Q-ary codes are the Reed-Solomon and BCH codes. Applica-
tions of these coding techniques to the error-control problem will be given later on in this
chapter.
Now the conditional MEP, P(E|{S,,}¢), for the ¢'” system in the ensemble can be
expressed using (4.80) as a function of E/No and the é'" code cross-correlation matrix
I'e = [yjx]e; however, as we have seen, it is difficult to find the global minimum of P(£)
as a function of Iz. In principle, the ensemble average MEP
M.-1
:
P(E\{Sm}e,Sm)< D> ({dink
ae}e (12.30)
kém,k=0 2No
where {dx}¢ denotes the Euclidean distance between the coded modulation vectors Sj, and
S, in the £’" codeword dictionary, £=1,...,2%ec. In the context of random coding, we
note that dmx is a random variable since the components S,,,, and sxn,n = 1,2,..., Nc, of
the vectors S,, and S, are binary random variables taking on values ,/Ew, or —\/E Nn. With
equal probability over the ensemble. To further characterize the random variable d,,x, let
Dmx be a random variable denoting the Hamming distance between S,, and S,;, taking on
value h. If S,, and S; differ from each other in h coordinates, then
Sec. 12.4 The Ensemble of Block-Coded Modulation Systems 769
1\%
PCDane= b= (4) (5) (12.33)
forh=0,1,...,N- and £=1,2,..., 2NeMe, Using (12.33) to perform the ensemble aver-
age over the Hamming distances in order to eliminate the conditioning on {S,,}¢ in (12.32)
produces
ref Y o(FE)()O
M.-! N
Nee Ma 2hE eas
=0 k=0,k xm
Ome ae)
(12.34)
Ne M--1 Ne hEy,
2 k=0,k4m h=0 ;
where the last line follows by changing the order of summation and upper bounding Q(x) by
e~*’/2. Note that the inner sum is independent of k and reduces to (1 + e~2Ne/ Noy ¢, Upper
bounding M, — 1 by M;, and summing over h and k yields the closed form expression
and Ro 4 Ro/T- = W Ro, the normalized channel cutoff rate in bits/sec. Since M;== 2Ke=
QRoT— 2"Nc, (12.36) can be rewritten as
P(E) <2 NelRo~re] — 9—KelRo/re—1] — 9-T(Ro-Ro) (12.38)
For the system engineer this is a very striking result: it says that as long as the code rate
re < Ro, the ensemble average error probability can be made arbitrarily small by taking
the block length N- (and hence message length K,) sufficiently large. Equivalently, as long
770 Block-Coded Digital Communications Chap. 12
as Ry < Ro, the MEP can be made arbitrarily small by taking the codeword duration T
sufficiently large. Since the ensemble of binary codes satisfies (12.38), it follows that at least
one code in the ensemble must have an error probability P(E) no greater than the ensemble
average, P(E). Notice from (12.37) that Ro approaches zero for small Ey,/No and unity
for large Ey./ No.
Now let us determine the form of Ro in the energy-limited region, Ey./No < 1,
where, with unlimited bandwidth, coding can be used to reduce the required signal energy
to the lowest possible level. To do this, we differentiate (12.37) with respect to x = Ey,/No
and take the limit as x approaches 0 to obtain
lim{d Ro/dx] = 1/21n2
>
P(E) = or ae oe (12.39)
by noting that e* = 2*/!"?. We observe that the bound on the MEP tends to zero with
increasing message length K; if E,/No > 21n2 = 1.4 dB.
Since T, = rceT) (see Table 11.2), then for binary coded BPSK
—Ro =_ Th Ey Ro
— Ry = —— , _= ———_
Eb/No Ro (12.40)
rc T, En, En./No
where
En./N
garni (12.42)
Ro
Thus, the bound on P(E) goes to zero with increasing K- if Ep/No > B for the
ensemble of binary coded BPSK versus E»/No > 21n2 for orthogonal modulation vectors
(signals). Notice that for small Ew,/No, where coding is most useful, we have shown that
Ro = (En,/No)/2 1n 2, so that B in (12.42) approaches 21n 2. This says that for very noisy
channels, binary coded BPSK is essentially equivalent to binary coded orthogonal signals,
which (from Chapter 4) are equivalent to binary coded transorthogonal signals for large M,
which are known to be optimum for the AWGN channel.
The binary coded BPSK result can also be obtained another way. Since Ey,/No =
rc(Ep/ No), then we can use this in (12.37) to write
Ep
ates =
Qrre— Dre (12.45)
which says that, as the code rate re approaches zero (N, approaches infinity or the band-
width expansion 1/r, approaches infinity), that
Ep
— >2ln2=1.4dB (12.46)
No
for re < Ro. Can one do better? If yes, then the question is how much better? We proceed to
answer this question.
In Section 12.4.1, we used the (2, 1, T;) BPSK modulator to generate the binary coded
BPSK waveform
Ne
Sm(t) =) 7)Smnn(t), O<t<T (12.47)
n=1
Figure 12.4 Q-ary ASK and Q?-ary QASK signal vector models: (a) Possible set of
values permitted the sy,; M = Q =8, N = 1, (b) Possible set of values permitted the
(Sinn (2); Smn(Q)); M= Q? =O Nat
rate ratio Ra /R2 = log, Q or Rg = R2 log, Q. These relationships define the throughput-
bandwidth trade-off of interest to the systems engineer.
The second concern mentioned above is addressed by obtaining a bound on the en-
semble MEP and examining the resulting Ro. Our approach is identical with that used for
the case of binary codes; that is, we determine the Ro (as a function of Q and Ey,/No)
which exponentially bounds the ensemble of codes average error probability P(E). There
are Q*< signal points available (potential codewords), and M, messages to be assigned to
the available points. Thus, the total number of ways to make the assignment, and hence the
total number of possible codeword dictionaries in the ensemble, is @%«“<. This says that the
ensemble of communication systems illustrated in Fig. 12.3 is now QNeMc systems; each
system uses different coded modulation vectors {S}¢,@=1,..., ONcMe: together with the
optimum receiver.
Notice from Fig. 12.4a that the end values gg = | and gq; = —1 have neighbors only
on one side. In the presence of noise these values have a higher noise immunity (or are
Sec. 12.4 The Ensemble of Block-Coded Modulation Systems 773
more distinguishable) and it is intuitively clear that the error probability will be smaller for
systems which use the coded modulation vectors {S,,}¢ in which the values g; near the ends
are used more frequently than the interior values.
If we let N; count the number of times that g; is used in the entire ensemble of
encoder-modulator pairs and if we assume that the message symbols occur independently,
then to the coded modulation vectors {S,,}¢ we assign the a priori probability
N
Sep, Dy ee lnc Oe (12.48)
where pj + p2+...+ po =1. Since each code yields M, coded modulation vectors S,n,
made up of N. symbols S,, then Nj + No+...+ Nog = N-M.. In order to bound the en-
semble MEP we use the union bound as before. Using the fact that O(x) < ent /2 we note
that the pairwise error probabilities in (12.3), conditioned on using the £th code, become
{dnk}e )
{P2(Sm,Si)}e = Q (
J/2No
(12.49)
-¢o (Ss - sh) E ples [-«s aa )2/4No]
J2No ahs mn n
Since the coefficients s,,, of S,, are random variables, we will need to average over all pos-
sible values in order to get the ensemble error probability, P(E). If Sinn = qi JEN, and sxp, =
qjVEn, then we have that the coordinate distances are |Smn — Skn| =|gi — aj|VEn, &
gij/En.- Moreover, the squared coordinate distance (Smn — Skn)* = ij Ey, occurs with
probability p;pj; since the levels q; and qj; are chosen independently and at random from
the Q-ary alphabet. Thus
Q
E {exp |— (mn z sin)”/4No| =) pipjexp(—)jEw./4No) (12.50)
i=) =i)
and the conditional PEP in (12.49) can be averaged over all codes in the ensemble to give
Ne
OF :
Px(Sm, Sx) < bsY= pip exp (-ahen./4m) (12.51)
i=1 j=1
Averaging the union bound over the ensemble of codes, we have
Tiss BUEN ey heat Sd ror =
P(E|Sm)< >. P2Sm, Sk) = (Mc — 1) P2(Sm, Sk) (12.52)
k=0,k#m
where the last equality holds since the right-hand-side of (12.52) is independent of k. Since
it is also independent of m, the MEP for the ensemble of codes is P(E) = P(E\S»m).
Bounding M, — 1 by M. = Q*c, and assuming equally likely messages, we obtain
RUE) (12.53)
where
Q Q 5
Ro A — logy Pi pj exp (—q;; En,/4No) (12.54)
I Let
774 Block-Coded Digital Communications Chap. 12
and where r, = K- logy Q/ Ne. Since Ro = T-Ro = Ro/ W and re = Rp/ W, we can rewrite
(12.53) as
-
P(E) < 27 Nel(Ro—Rs)/W] _ 9-T(RoRe)
Equation (12.53) demonstrates that the error probability can be forced to zero by using long
block lengths as long as Ro > re. Furthermore, there exists at least one code in the ensemble
of codes with P(E) < P(E).
If we further assume that the levels gj and qj occur with equal probability, then
pi = 1/Q@ foralli=1,..., Q, so the channel cutoff rate in (12.54) reduces to
1 Q oO ’
Ro = Ro/
W = —log, FEye) xn -ah n/N bits/sec/Hz (12.55)
i=l j=l
Lor. O=.2, ij = 4 fori # j and qj; = 0; substitution of these conditions in (12.55) leads to
(12.37) for coded BPSK as it should. Upper and lower bounds for the cutoff rate Ro can be
derived from (12.55) by noting that 4Ey./(Q — Ly qi EN < 4Ey, for all i A j. Using
these conditions, we find from (12.55) that
—Evn,/No
og: |+ (Q — l)exp || < (Ro
— log, Q)
10.0
8.0
6.0
4.0
2.0
N
=1)
v
AD
g
S +0
=So 08
fe
0.6
pe
So
0.4
0.2
0.1
-10 0 10 20 30 40 50 dB
Energy per Dimension to Noise Ratio, 10 log;y Ex, /No
Figure 12.5 Ro versus Ey,./No for equispaced coded Q-ary ASK; p; = 1/Q (Reprinted from [7])
To complete the discussion of Q-ary coded ASK, we revisit the bandwidth considera-
tions in Section 12.3.4 pertaining to use of the (2, 1, 7;) binary coded BPSK modulator. We
note that the argument that D = 1/Ty = 1/7; = 1/T; is true as long as k = Ne and N = 1.
This condition holds for the (M, 1, 7;) Q-ary coded ASK modulator of this section. Thus
(12.25) holds for Q-ary coded ASK.
Q?-ary coded QASK. The cutoff rate for Q?-ary coded QASK can be derived
easily starting with the I-Q generalization of (12.49). Figure 12.4b demonstrates the ex-
tension of the Q-ary coded ASK signal design to the Q?-ary coded QASK signal design
using the (16, 2, 7;) modulator. For the Q?-ary coded (M = Q?, 2, T;) QASK modulator,
the modulation vectors S,, are broken into in-phase, s,,(/), and quadrature, sm,(Q), com-
ponents with Sn(1) = /En,/2XmnU) and smn(Q) = J En,/2%mn(Q), where Xnn(1) and
Xmn(Q) take on values g; = (—1) + (2i — 2)/(Q — 1),i=1,..., Q for Q even, indepen-
dently and at random with probability p;. In other words, we transmit one of M = Q?
776 Block-Coded Digital Communications Chap. 12
10.0
8.0 a
64
6.0 |-
SZ
4.0 |- 16
* 8
Ro
32.0 4
ea@ 3
“we
2
=
= 10- Q=2 4
ce 0.8 1- =
M
oc 0.61 a
#o
Re ——— Upper envelope of Ry
0.4 |- curves for {p;)
optimized
02) a
0.1
-10 0 10 20 30 40 50 dB
Energy per Dimension to Noise Ratio, 10 log; Ex /No
Figure 12.6 Ro versus Ey,/No for equispaced coded Q-ary ASK; {p;} optimized (Reprinted from [7])
QASK signal vectors (Smn(Z), Smn(Q)) for each of the k= N,. codeword coordinates x,
with energy constraint |(Smn(1), Smn(Q))|* < En,
Thus, the I-Q generalization of (12.49) becomes
Ne
(P2Sm. So} < ] [exp {(smn) ~ sin VP + [5nn(Q) ~ Sin Q)P)/(—4No)|
n=1
where the expected value is computed over the ensemble. We use the above expression to
find the ensemble average PEP, P2(S,,, S,), which when substituted into the union bound
yields
2N,
Q Q
P(E|Sm) < (Mc — 1) b Pip; exp (—4;, EWN
t= 1
Q @Q
where Ro = Ro/W = —2 log, b»Y- pip exp (—47, En, ano bits/sec/Hz denotes the
i=1 j=l
cutoff rate for Q?-ary coded QASK and where r, = 2K, logy Q/Nc. We note that both r,
and Ro are twice the values obtained for Q-ary coded ASK, as we would expect.
Here Rj denotes the channel cutoff rate for a continuum of possible transmitted signal
points on or within the hypersphere of radius VE =./NcE N,- This derivation assumes
that the vector space generated by the modulator is isometric to the vector space generated
by the encoder. For small values of Ey,/No, where coding is most useful, we note that
R5 ~ (Ey,/No)/21n2 = (reEp/No)/21n2, which is the same approximation obtained for
the channel cutoff rate, Ro, for binary coded BPSK in Section 12.4.1. Operation of the
system at code rate r, = Ro in this energy-efficient region requires E,/No > 21n2.
Figure 12.6 plots Rj versus Ey,/No together with the Ro of (12.54) for the ensemble
of Q-ary codes with equally spaced amplitude levels and an optimum probability assign-
ment. We note that the richness of the continuum of possible signal points from the trans-
mitter results in a larger channel cutoff rate; that is, Ro > Ro for all O > 2. This in turn
leads to a tighter bound on the ensemble error probability, as seen by comparing (12.56) and
(12.53).
Some insight into the physics of the parameter Ey./No will be useful before we
further discuss Fig. 12.6. This can be obtained by using the bandwidth defined via the
Nyquist sampling rate of 2B = 1/T.. Thus, 2Ey,/No = P/NoB & CNR, which can be
interpreted as the receiver input signal-to-noise ratio in a single-sided Nyquist bandwidth of
778 Block-Coded Digital Communications Chap. 12
B Hz. For low input CNR’s, we see from Fig. 12.6 that Ro nearly coincides with Ro for all
Q, but for CNR’s greater than zero dB the binary codes are less desirable. In fact for CNR
> 10 dB binary codes are exceedingly undesirable; however, for high input CNR’s, coding
is generally not required to achieve the desired performance. At low input CNR’s (power
limited channels) where coding is usually applied, binary codes are essentially optimum.
The increase in achievable Ro as Q increases for large CNR’s is due to the fact that the
signal vectors (codewords) that lie inside the signal hypersphere of radius ,/N-En, can
be detected without error, while at low input CNR’s only the signal vectors (codewords)
that lie on or near the surface of this sphere can be detected without error. This behavior
tends to justify the fact that binary codes are optimum for use in channels that are power
limited, that is, those with low values of CNR. For large values of CNR, it should be clear
why Ro ~ log, Q bits/dimension can be made to approach Ro for large Q. In this region
of operation, those codewords in the Q-ary codes which lie inside the hypersphere have
enough signal-to-noise ratio to be detected with essentially the same reliability as the set
of codewords constrained only in energy. In fact, as Q approaches infinity, the discrete
constrained-energy signal points approach a continuum, and Ro approaches Rj. For large
CNR’s, the channel throughput is usually limited by bandwidth so that the application of
coding (which tends to expand bandwidth) is of lesser interest.
where
l hoe E k
Ro =Ro/W A ao log, FeSy exp(- sin? 4] bits/sec/Hz (12.59)
Sec. 12.4 The Ensemble of Block-Coded Modulation Systems 779
for M > 3. For small values of Ep/No, it is easy to show from (12.59) that
Ro& 4 =
. sm (7!) 2In2 (=) Oe
If we insist on operating the system at Ro > rc, then this requires a communications effi-
ciency of
E
ere in?) (12.61)
No
equivalent to that given for block orthogonal codes in (12.15).
It was shown in Chapter 4 that the M-PSK symbol error probability is well approxi-
mated by
2 wen ot
P(E) 220 (\vA sin ) (12.62)
for E;/No > 1. For large M and E;/No, we have, using the small-angle formula and the
fact that E,; = Ey, for this model, that
Except for the special cases of M-PSK, orthogonal, biorthogonal, and transorthogonal
coded modulations, we have seen that the problem of expressing the message error prob-
ability performance of a coherent communication system in closed form and evaluating the
expression is a formidable problem. The problem of optimum code selection centers around
being able to determine that signal cross-correlation matrix, namely I’ of equation (4.78),
which minimizes P(E). Shannon no doubt recognized this and in order to gain insight into
the coding problem proposed the astute method of bounding the MEP by using the random
coding techniques presented in the previous sections.
This method uses the pairwise error probability and the union bound to evaluate the
MEP of an ensemble of communication systems which use amplitude modulation and the
class of all possible Q-ary codes to communicate over the AWGN channel. The significant
result, an existence proof, is that this ensemble of communication systems has the ensemble
message error probability union bounded for isometric coded modulation mapping by
P(E) < q—Ne(Ro—re) — 9—Nel(Ro—Ro)/W] _. g—~PG(Ro—Ro/W) _ 9-T(Ro—Ro) (12.64)
where we have used the facts that PG = WT = N, (see Table 12.3). This says that the
ensemble MEP approaches zero as N. approaches infinity provided that Ro = Ro/W >
re = Ry/ W or as T approaches infinity provided that Ro > Ry. From this, we can conclude
that there exists at least one codeword dictionary among the ensemble of Q-ary codes with
MEP less than or equal to the average value. Indeed, such code(s) are of great practical
interest; however, it appears that nature has hidden them and the search for such codes
780 Block-Coded Digital Communications Chap. 12
has led to the development of Coding Theory. Later in this chapter, we investigate specific
block-coding techniques for error control from a system design perspective. In the next
chapter, we will investigate the use of convolutional codes as a coding alternative.
When message m is sent over an AWGN channel, the pairwise error probability has been
shown (in Appendix 3D) to be given by
dm
PGiSi=o (=) (12.65)
for all m 4k =0,1,...,M- — 1. More generally, it can be shown [8, 20] that, for the
arbitrary IF CMC
under the same assumptions, where a vector quantizer produces the quantized random vec-
tor R = (R,..., Rv.) whose coordinates are the quantized random variables R,, taking on
values Rj. Here P(R|S,,) and f(r|S;) denote, respectively, IF channel transition probabil-
ities and probability densities. The bounds in (12.66) and (12.67) are called Bhattacharyya
bounds and the negative logarithms are the Bhattacharyya distances, say By. For the BSC
with crossover probability p, we have from (12.67)
oe (12.69)
Thus, application of the total error probability formula, the union bound, and the PEP result
from the AWGN channel yields
M-—1 Mc-1
P(E) <(1/Mc) }) }> exp(—dpy/4No) (12.70)
k=0 m =0
m#k
Now let dmin denote the minimum Euclidean distance as defined in (12.7). Using
this in (12.70) clearly demonstrates the important role dmin plays in the design of error-
correcting codes, that is, P(E) < M, exp (—d?., /4No). Recall further from (12.4b) that the
Hamming distance is proportional to the square of the Euclidean distance for binary coded
BPSK.
Shannon’s Channel Coding Theorem. The rate r, of an (M;, Nc, T-) code is
log, M.
ro= a bits/dimension (12.71)
Cc
and all code rates below the channel capacity Cy, are achievable. Specifically, for Gaussian
channels and M, sufficiently large, an optimum receiver, and code rate
1 1 ; : :
re< Cy, = 5 log» ( + )= 5 logs[1 + 2r-(Ep/No)] bits/dimension (12.72)
No
there exists an (M., N-, T-) code with MEP approaching zero. The restriction re < Cy,
implies that Ep,/No > (22re — 1)/2r. for all r. < Cy,. If re > Cy, and M,¢ is sufficiently
large then P(E) approaches one for every possible (M., Nc, T-) code. It can be shown [7]
also that 5C Ne < Ro < Cn,. The code rate inverse rks 4 1/r, is frequently identified as the
bandwidth expansion factor by system engineers.
The proof of this famous theorem is given in numerous books on information theory
[7, 8, 9] and in Shannon’s original paper [5]; however, it is not our purpose to present this
important proof here since it is too lengthy to reproduce. Shannon’s approach to proving
this famous theorem included the modulator and demodulator as a part of the channel; that
is, he used the baseband channel model to prove the theorem. This is equivalent to assuming
an isometric mapping from coder output to modulator output when the IF channel model is
used. In this sense a codec and modem exist such that the theorem is true.
In Shannon’s famous paper [5], it was shown that all information sources, for example,
produced by people speaking, television cameras, telegraphy key, and so on, have a rate
(r-) associated with them which can be measured in bits/sec. Communication channels have
a capacity (Cy,) that can be measured in the same units. Shannon’s theorem states that a
codec-modem exists such that information can be transmitted essentially error-free over the
channel (baseband or IF) if and only if the rate does not exceed the channel capacity. To
782 Block-Coded Digital Communications Chap. 12
gain further insight, let us investigate the communication efficiency requirement implied by
(12.72) for the AWGN channel. Since Ey,./No = rcEp/No, we can write
1
re=< Cy, = 5 logy(1 + 2r-Ep/No) (12:73)
Thus
then, with N. = 2BT per our bandwidth discussion of Section 12.3.4, we write
2E 2PT P
NoNN SAEED
No(2BT) NoB~A CNR (12.81)
Thus, we have a physical interpretation of the parameter 2E y,./No as merely the input signal
power-to-noise power ratio! Substituting the above relationship into (12.72), we see that the
temporal channel capacity is given by
P
C ACy,/Te = B log, | 1 + —— } bits/sec (12.82)
NoB
Sec. 12.5 Shannon's Channel Coding Theorem 783
P
Sareeae
NoIn2 bit
(bits/sec)
(bits/sec)
C
P/N,
B (Hz)
Coe lm C= hin 10 ee
Sey ees B00 82 No B
(12.83)
The parameter Cy is the so-called infinite bandwidth Shannon capacity limit for the AWGN
channel. Furthermore, the maximum amount of information that can be transported error-
free in T sec is
|RR aie
Le Col = Noin2 bits (12.84)
The parameter /,o is the so-called infinite bandwidth Shannon information transport limit!
In practice, most radio communication systems are designed to operate with 10 dB-Hz
< P/No < 120 dB-Hz. Thus, with P/No = 120 dB-Hz, Io = 10!2/ In 2 bits of information
can be delivered error-free in one second; this is more information than stored in the Library
of Congress.
One question that remains as a consequence of the existence proof is to specify the
ultimate codec, modem and their implementation. Unfortunately, nature seems to have hid-
den the particular coded modulations; they remain to be the slippery eel which motivates the
further development of coding theory from the perspective of coded modulation.
784 Block-Coded Digital Communications Chap. 12
in bits/dimension where the available energy-per-dimension to noise density ratio for each
coded message bit is
P 2P P
SENS ee yw Sets ae ee es ONY (12.86)
No No NoW NoB
and W is the double-sided Nyquist bandwidth. Thus, (12.85) can also be written as
1
Oe 5 log,(1 + CNR) (12.87)
As before, converting Cy, to bits/sec requires that one normalize Cy, by T,; that is, define
C2 Cy.) dz,,so that
Cy.
CA= oe = B log,(1 + CNR) bits/sec (12.88)
SINCE. 2 Beale
Let us now relate the capacity theorem to the matched filter theory developed in
Chapter 4. We note that the input signal power to noise power ratio can be written as
22 fogs Tt |
CNR Litesbe Pa eee (12.89)
NoW No WT), No WT,
Now, with the 7p-sec processing gain defined by PG, A WT), and noting that the output
signal power-to-noise power ratio of a matched filter is pp,; = 2E,/No, then
2Ep
= (PGp) (CNR) = (PGp) pin (12.90)
0
where ~jn = CNR is the input signal-power-to-noise power ratio to the matched filter. Thus
so that in T sec, the maximum amount of information that can be transported error-free is
E
Noin2 bits!
its (12.93)
(12.94)
Sec. 12.5 Shannon’s Channel Coding Theorem 785
Ey _(2Re/®
No >
RaW
en
— 1) (12.95)
€ = tog, (1+ =)
Cc P
bits (12.97)
which says that C/B ~ (P/J)/1n2 if the jamming margin is large (J/P > 1), a case of
great practical interest.
12.5.4 The Union Bound for Binary Orthogonal Codes Written in Terms
of the Shannon Capacity Limit
Using the fact that 2* = e*!"*, (12.12) can be rewritten in terms of the Shannon capacity
limit, namely
<2 Ry)
P(E) <2 (F (12.98)
which says that, as long as 0 < Rp < Coo/2, then P(E) can be made to approach zero by
making T large. This remarkably simple expression serves to upperbound (4.80); unfortu-
nately this bound is not very tight.
Gallager has developed a tight bound on the ensemble error probability P(E) for
block codes. In particular, consider the case of the Q-ary coded (M, 1, T;) modulator used
k = Nz times per codeword to transmit the class of Q-ary codes where M = Q, as discussed
in Section 12.4.3. For this case, he has shown that
for all observables r,,n = 1,..., Nc, and all messages m =0,..., M, — 1. Here f (rn|smn
= qk JEN.) denotes the known channel transition probability density function (which char-
acterizes the channel) for the n‘ h received observable, given that message m is sent.
For the IF DMC (quantized IF CMC) model defined in Section 3.1.1, Section 11.2,
and used in Section 11.2.4
sea ge a l+p
Eo(p, Q) = — log, | 3? [A= Rate En.|' “ (12.102)
+
j=0 Lk=1
for all quantized observables Rp, n=1,..., Nc, and all messages m=0,...,M. — 1.
Here P(Ry = Rnj|Smn = gkx/En,) = prjA& P(j|k) denotes the known channel Panction
probability for the n’ h received quantized observable (that is, the quantized version of ry),
given that message m is sent.
For modem and codec design purposes, it is of interest to connect E;(r-) with the
channel cutoff rate Ro for the IF CMC model and with the quantized Ro for the IF DMC
model. This is accomplished using Gallager’s function Eo(p, Q) and setting o = 1, that
is Ro = Eo(1, Q). In regions of low Ey,/No, then E;(rc) is maximized at p = 1 [20], so
E;(re) X (Ro — re) for equally likely, independent message symbols. For the IF CMC we
have, from (12.101)
Ro = Ro/W = Eo(1, Q)
2
oo f @
= — log, if bsPka| f (Fabs ~ alex) dry | bits/sec/Hz
—0oO k=1
788 Block-Coded Digital Communications Chap. 12
Ro = Ro/W = Eo(1, Q)
2
J=1 (29
=— log, a bsDEAL. (Rr = Rnj|Smn = 4k Fx) bits/sec/Hz
j=0 Lk=1
where the channel transition probabilities for the quantized observables Ry are related to the
channel transition probability density functions f (7n|Smn = Gk./ En.) through
1 (fn — Qkx/En,)?
j
for Q-ary coded ASK and A; is the observation region which gets mapped into the observ-
able Rj for all j =0,1,..., J — 1 (see Fig. 11.6). This defines the connection between the
IF CMC and the quantized IF CMC (IF DMC).
Properties of the random coding exponent E,(r-) have been studied with much suc-
cess; from our system perspective, one important feature is that E,(r-) > 0 for allO <r, <
Cy,. Thus, (12.99) says that for appropriately chosen codes and modems, the probability
of decoding error can be made to approach zero with increasing block length N, for any
rate less than capacity. In expression (12.100), E,(r-) is a positive function of the channel
transition probability via Eo(o, Q). This function yields a fundamental characterization of
the channel for coding purposes. It brings the relationship between decoding error proba-
bility, data rate, block length, and channel conditions into focus. In fact, as we shall later
see, the question of what modulation-demodulation techniques one should use when coding
is applied to system design can be addressed on the basis of E;(r-). Moreover, since the
ensemble of block codes satisfies (12.99), it follows that at least one block code in the en-
semble must have P(E) no greater than the ensemble average P(E). From a system design
perspective, this bound is not useful in accessing design point requirements with respect
to communications efficiency (E,/No), supportable data rate (72,), or power requirements
which achieve a specified error probability. The insight provided is philosophical in the
sense that a code(s) exist(s) such that if r. < Cy, then the error probability can be made
arbitrarily small by employing large block lengths N,. Unfortunately, the complexity of the
encoder and decoder which accommodate long (or short) block lengths is not addressed by
this interesting bound. In fact, the complexity issue remains to be the important unsolved
problem in information theory. The random coding exponent versus rate curve is of inter-
est to the system engineer, especially for comparing the Ro values corresponding to specific
codec-modem pairs. We will consider an example of how such a comparison can be useful in
the design of the demodulator in Section 12.7.5, for binary coded DPSK. The general char-
acteristics of Ro are discussed extensively in Gallager. However, for our purposes, we are
particularly interested in its behavior for very noisy channels, the case of greatest practical
interest.
12.6.1 Block versus Convolutional Coding for the Very Noisy Channel
Without elaborating the details, let us consider the class of very noisy channels (low values
of E»/No, where coding is of interest) and present the random coding exponent versus rate
curve E,(r-) which results and discuss the implications from a coding viewpoint. For block
Sec. 12.7 Demodulator Design Using the Ro Criterion 789
E,(r,)
Ne Convolutional Codes
Block Codes
CN,
Figure 12.8 Random coding exponent versus code rate for the very noisy channel
Figure 12.8 demonstrates plots of the random coding exponent given in (12.103) versus
code rate. For comparison purposes we also demonstrate the random coding exponent of
convolutional codes (to be discussed in Chapter 13) found by Viterbi [10]. From a systems
perspective, we conclude from Fig. 12.8 that for a given code rate, there exist convolu-
tional codes which yield a lower error probability than that of block codes. (This statement
assumes a bandwidth unconstrained, memoryless channel.) Notice too that the maximum
value of E,(r-) predicted by this bound is Cy,,/2. The result in (12.103) was derived using
the baseband system architecture which absorbs the modulator and demodulator as parts of
the channel model.
Consider first the IF DMC with binary coded BPSK and with equally probable messages
mo = 0 and m, = 1. The rate Ro, found from (12.102) with p = 1 for soft-decision demod-
ulation, reduces to
l
Ro = — log, {5 (12.105)
qe]
x 1+ >> P (Rn = Rujlsnn = VEN.) P (Rn = Rnjlsnn = —VEn.)
j=0
Sec. 12.7 Demodulator Design Using the Ro Criterion 791
for J = 2" output quantization levels or L bits of quantization. This compares with the
capacity of the same IF DMC [8]
2e-1
Cy. = Ss [P (Rn = Rajlsmn = —VEn.) (12.106)
j=0
As we have mentioned, the Ro criterion is natural for use in the design of the modem as
it can be optimized as a function of the number J of quantization levels, allowing the
system engineer to trade off system performance with the implementation complexity of
the decoder. For example, in making this trade, the engineer would examine the quantized
Ro attainable with J = 2, 4, 8, 16, and so on, until he has obtained an Ro sufficiently close
to the unquantized (J = 00)Ro which justifies the decoder cost and complexity versus the
improved bit error probability performance. As we have already pointed out, the absolute
upper limit on the code rate r, is the channel capacity Cy,, and the upper limit for all
practical purposes (considering cost and implementation complexity) is the computational
cutoff rate Ro; the unquantized Ro appears to be approximately 3 dB smaller than Cy, for
the cases of greatest interest.
The procedure used in the early design of modems for selecting the quantization cell-
size parameter A demonstrated in Fig. 11.6 (see Sections 11.2.3, 11.2.4, and 12.6) was to
choose it to minimize the E,/No required to operate at a code rate of Ro bits/dimension.
792 Block-Coded Digital Communications Chap. 12
System operation
atr,=R ‘
System operation
atr, = Cy
€
This philosophy is intuitively justified since one is attempting to maximize the coding gain
(the reduction in E,/No required when coding is used to that required in an uncoded BPSK
system for the same BEP) when operating at rate r, near Ro. Usually computer simulations
of the coding system are required to make this trade, and the results have shown good
agreement with those determined using the theoretical considerations due to Massey [2]
which we now describe.
Consider the case of binary codes with equally likely message symbols using BPSK
modulators, that is, Sinn =./En, or —,/En,, and J output symbols; see Fig. 11.6 for
the case L = 3 and J = 8. Recall that each received observable r,, must be assigned one
of J quantization levels Rnj € {Rno, .--, Rn(y—1)}. To do this, the quantizer needs (J —
‘ 1) thresholds 7;, 7 =0,..., J — 2. Here T; denotes the threshold between two quantizer
Sec. 12.7 Demodulator Design Using the Ro Criterion 793
T) T, T,
ee bere os peed
output levels Rn; and Ry(j+1). The sketch in Fig. 12.10 illustrates the decision regions or
quantization levels for the case J = 4.
Our goal is to select these threshold levels in an optimum way; specifically, we seek
a selection that will maximize the quantized cutoff rate Ro, given by (12.105) and repeated
below for convenience:
1
Ro = — log, {> (12.109)
aE
x 1+) > P (Ru = RnjlSnn = VEN.) P (Rn = RnjlSnn = —VEn.)
j=0
Hence the demodulator design problem is to choose each threshold 7; such that
dRo Nes
0 (12.110)
ony
Applying the above equation, Massey found that the optimum threshold levels were given
by
T; = ,/ A(Rnj)A(Rn,j+1) (12.111)
where the likelihood ratio is given by
for a given output quantization level Ryj, j =0,..., J — 1. This says that each optimum
threshold is the geometric mean of the likelihood ratios for the quantizer outputs whose
decision regions the threshold value separates.
Odenwalder [17] considers the use of equally spaced thresholds at 0,+7,...,
+(24—! — 1)T, with the spacing A; =T chosen to maximize Ro. For L = 3 (3-bit quanti-
zation) and Ey,/No = 1.5 dB, we obtain threshold spacing T of roughly .5 or .6 times the ©
standard deviation of the unquantized demodulator outputs [17]. Decreasing (or increasing)
En,/ No results in only a minor decrease (or increase) in 7. Thus, a fixed value for T can
be used over the entire dynamic range of Ey./No, and an automatic gain control (AGC)
circuit can be used to estimate the noise variance No/2. System performance is not sensitive
to small variations in the AGC output, and the value of T is usually selected for the worst
case value for Ey./No or Ep/No.
794 Block-Coded Digital Communications Chap. 12
1-bit quantization
no quantization
E./No
dB
,
1 2 3 4 5 6 7 8 910
Bandwidth Expansion = 1/r,
Figure 12.11 E),/No required to operate at re = Cy,, for coded BPSK modulation (Reprinted from [17])
The results presented so far in this section are now combined to demonstrate the effects
of the number of quantization levels on demodulator performance. We assume that coher-
ent demodulation of coded BPSK or QPSK is used in the AWGN channel. First, we con-
sider the problem of demodulator design when we wish to operate at code rates re = Cy..
Equation (12.72) for infinite quantization and (12.108) for 1-bit quantization establish this
performance by relating the bandwidth expansion r- ' — 1/r, and the required communi-
cation efficiency to operate at r- = Cy. Plots of this relationship are demonstrated in Fig.
12.11. In addition, comparable results found from (12.106) for soft-decision demodulators
are demonstrated for 2- and 3-bit quantizers with equally-spaced threshold levels.
Figure 12.12 gives the corresponding curves of required E,/No versus 1/r, for system
operation at r, = Ro using (12.37), (12.105), (12.107) and equally spaced threshold levels.
From Figures 12.11 and 12.12 we see that 3-bit soft quantization is almost as good as
- infinite quantization (L = oo), and that 1-bit quantization (hard-decision) requires about
2 dB more E»/No than infinite quantization for comparable performance. It can also be
concluded by comparison of Figures 12.11 and 12.12 that for small values of required
E»/No corresponding to large bandwidth expansions, approximately 3 dB more Ep,/No is
required to operate the system at r- = Ro than is required to operate at re = Cy,.
Sec. 12.7 Demodulator Design Using the Ro Criterion 795
1-bit quantization
no quantization
E,,/No
aB,
1 Z 3 4 5 6 7 BO) 1
Bandwidth Expansion = 1/r,
Figure 12.12 £;/No required to operate at r, = Ro for coded BPSK modulation (Reprinted from [17])
For binary coded DPSK modulation, p = (1/2) exp (—r-Ep/No) is the coded message bit
error probability as given by (7.14). For binary coded double differential phase shift keying
(DDPSK), the expression for p is given by (8.58). In both cases, we have assumed optimum
demodulators and no frequency offsets. Furthermore, the Ro for both coded DPSK and
coded DDPSK with hard decisions is given by (12.107). Figure 12.13 illustrates the E,/No
required to operate at code rate r, = Ro versus bandwidth expansion 1/r, for coded BPSK,
DPSK, and DDPSK. These curves reveal the minimum value of E,/No for which coding is
useful for each of the coded modulation techniques. From the DDPSK curve, for example,
we see that if the E,/No is less than 10.45 dB, no practical code exists at any rate that will
allow us to achieve arbitrarily small bit error probabilities. Figure 12.13 also illustrates the
E,,/ No needed to achieve a bit error probability of 10~> without coding for each of the three
modulation schemes; these curves enable us to compute achievable coding gains. From the
DDPSK curve, for example, we see that if r- = 1/2, then the needed E,/No is 10.45 dB.
Thus, the maximum coding gain achievable (for BER = 10~> and hard-decision decoding)
is 3.55 dB. To achieve a greater gain we must increase the complexity of the receiver to
allow for soft decisions.
We further note from Fig. 12.13 that for both coded DPSK and DDPSK it is not the
796 Block-Coded Digital Communications Chap. 12
dB
Ey/Np,
case that decreasing the code rate (say by increasing the redundancy of the code) automat-
ically increases the achievable coding gain, as was the case for BPSK. This is because the
energy in each coded message bit is given by Ey. = r-Ep; thus, decreasing the code rate r,
decreases the channel symbol energy, effectively making the demodulation phase reference
noisier. This effect is not evident in the case of BPSK because a perfect phase reference is
assumed [17]. Both DPSK and DDPSK appear to achieve maximum coding gain with code
rates near r, = 1/2.
We now consider the case of binary coded DPSK with infinite quantization (L = oo); that
is, we assume that the demodulator output random variable is handed over to the decoder
unquantized. For this case, Odenwalder [17] has shown that
1 om ;
Ro = — log, E (:+/ Vy f(vlA! = 0) Fwle =mav)| (12.113)
‘where f (y|0! = 60) denotes the conditional probability density function for the demodula-
tor output random variable y, given that the information phase angle 6! = A"6; is equal to
69. (See Chapters 7 and 8.) Clearly, the form of the conditional probability density function
:. depends on the detector implementation, via y.
In order to illustrate our thesis that the Ro criterion should be used to design the
Sec. 12.7 Demodulator Design Using the Ro Criterion 797
modem in a coded system versus using the post-demodulation bit error probability sug-
gested for the comparable uncoded system, we shall consider two demodulation techniques,
namely the optimum I-Q demodulator and the suboptimum autocorrelation detector (ACD)
demodulator (delay received signal one coded message bit and multiply; see Fig. 7.18). In
the case of the autocorrelation detector, the demodulator output random variable w is equal
to the sampled output of the lowpass filter, and the resulting Ro is given in [17]. For the
optimum I-Q detector the random variable w; is equal to the difference of the two adjacent
transmitted phase angle measurements, nj — ni—1. The conditional probability density func-
tion f (|! = 0) is given by (7.3), while f(W|6! = 7) is simply a shift of the same density
function by z(mod 277).
In Fig. 12.14, we plot the E,/No necessary to operate at rate r, = Ro versus band-
width expansion for binary coded DPSK with infinite quantization using both the ACD and
I-Q demodulators. For comparison, we also show the curve for binary coded DPSK with
hard decisions using the I-Q demodulator. It is very intriguing to note that, while in the
absence of coding the I-Q demodulator is known to be optimum, with coding the autocor-
relation detector (ACD) actually gives better Ro results. Thus, for binary coded DPSK, it is
preferable to use an autocorrelation detector rather than an I-Q detector. This confirms the
thesis that, from a communication efficiency perspective, the Ro criterion is better to use
in the design of a modem for coded systems; the post-demodulation bit error probability is
better to use in the design of an uncoded system.
In (12.105) through (12.108), we have presented the cutoff rate Ro and capacity Cy, for soft-
and hard-decision binary coded BPSK demodulators. We know from Shannon’s theorem
that channel coding will be of no value unless r, < Cy,; we also assert that, for most
practical purposes, operation at r- < Ro will be required in order to realize the decoder
mechanization with acceptable hardware complexity.
As in Section 12.4.3, we now consider the use of the (M, 1, 7;) modulator with
k= N, to transmit M = Q-ary (M-, N-, T-) codes. This implies that the Q-ary coded ASK
modulator is used k = N, times for the transmission of a codeword. If we use (12.54) and
optimize over the a priori probabilities of the coded message symbols, we have
Q
Ro < - min log bsye PjPk exp (-aivens400) (12.114)
j=l k=l
where the minimization is taken over all possible probability distributions P = {pj}, j =
1,..., Q, on the coded message symbols. (Notice we have assumed an isometric mapping
Beaeen the coder and modulator.) Since the minimization occurs when p; = 1/Q for all /,
then one can show [2] that (12.114) reduces to
+(0 Do (3-2)
ore mae
(12.115)
fy sop O— Wor, |1
where the equality holds only for equally likely Q-ary transorthogonal (simplex) signals and
where Ey, denotes the average energy per codeword dimension:
798 Block-Coded Digital Communications Chap. 12
4B
E,/No,
1 2 3 4 5 6 7 8 9 10
Bandwidth Expansion = 1/r,
Figure 12.14 £,/No required to operate coded DPSK at r, = Ro (Courtesy of Debbie van Alphen)
hae
En.
= 5 D4
Oa
This inequality can be used to gain considerable insight into the design of coded modulation
systems for low and high values of En,/No.
Now for transorthogonal (simplex) signals, the equality in (12.115) holds and we
differentiate with respect to x = Ey,./ No to obtain
=—- =
dRo _ Q(En./No) a
Te ae) sse( x0 —1) +@Q J (12.116)
which decreases monotonically with increasing x. Thus, for small values of x, where coding
is most useful, (12.116) yields
dR at
ee ~ 1/21In2or Ro ~ (Ew, /No)/21n2 (12117)
which is independent of Q! This demonstrates that one should use binary coded BPSK
(Q = M = 2) in the design of coded modulation systems since this technique is easiest
Sec. 12.8 Linear Block Codes 799
to implement. In this sense, the optimum value for Q is Qopt= 2. From a systems engi-
neering perspective, this value for Q simultaneously implies two things: (1) it says that
M =2= Qop which implies that the (2, 1, T;) binary coded ASK or BPSK modulator is
optimum, and (2) that the binary transorthogonal (simplex) code (M, = 2K¢, N. = 2Ke — 1)
is simultaneously optimum! Further, with Q = 2 in (12.115), we observe for transorthogo-
nal codes and the (2, 1, T;) coded BPSK modulator that
ue Ro A ee = = Co/2 bits/sec
This says that 70 is approximately 3 dB less than the infinite bandwidth channel capacity,
Coo. This 3 dB backoff from channel capacity is frequently taken to be the achievable region
of operation by practicing system engineers.
For high values of Ey./No, then we readily observe from (12.115) that Ro < log, Q
and dRo/dx = 0. This serves to demonstrate why one might choose a large Q in system
designs which require a large value of Ey,./No, where (12.115) holds with near equality.
In conclusion, in this section we have demonstrated how the Ro-criterion can be
used to accomplish signal design, that is, the simultaneous choice of both the coding and
modulation techniques.
fo Ke —+|«—__—_—-N,- K,-—+|
ne
Figure 12.15 The codeword for a systematic block code
We will primarily be concerned with the class of (N-, Ke) codes.” An (Nz, Ke) block
code is a code in which each codeword contains N, symbols; K- of these are information
symbols and r = N. — K¢ are redundant parity check symbols. In this context, the dimen-
sionality of the codeword space refers to the number of encoder output symbols for each
block of K,. encoder input symbols. It is possible to encode the K,-symbol message into
a codeword such that the first K. symbols of the codeword are exactly the same as the
message block and the last N; — Ke symbols are redundant parity check symbols which
are functions of the information symbols (see Fig. 12.15). A code of this form is called a
systematic code. The redundant symbols are used to combat (control) errors which occur
during transmission. Systematic codes are of greatest practical interest and we shall con-
centrate on the use of such codes for error control in what follows. Linear block codes can
be described by a K, x N- generator matrix G. If the K, information symbols into the en-
coder are represented by u = (uj, u2,..., UK), the encoder output word is represented by
the N,-dimensional row vector x. The encoder input-output relationship is characterized by
+= uG (12.118)
so that the N,--symbol encoder output blocks (that is, the codewords) are linear combinations
of the rows of the generator matrix. For the binary case, the output blocks are bit-by-
bit modulo-2 sums of the appropriate rows of G. Usually block codes are decoded using
algebraic decoding techniques; the important and elegant details can be found by the system
engineer in many coding and information theory books [3, 11, 12, 13]. The Venn diagram
for error-correcting codes is illustrated in Fig. 11.18. In what follows we shall assume that
the mapping of code vectors x to coded modulation vectors S is isometric.
From the coding perspective it is very instructive to relate the Euclidean distance, dj;, to the
Hamming distance, D;;, and use this result in the union bound result given in (12.3) to gain
further insight into block coding. To do so we must make sure that the modulation technique
chosen to transmit the coded message bits does not destroy the Hamming distance properties
of the code; otherwise, the effectiveness of the coding scheme to minimize the required
Ep»/No is lost. As was demonstrated in Chapter 11, coded BPSK and QPSK modulators
can be used since they map the coded message bits into the modulator output waveform
space without modifying the effective distance properties of the code. In this section we
will quantify this fact for the class of binary codes.
2. In the coding literature, it has been fashionable to refer to such codes as (n, k) codes. We prefer to use (Ne, Ke)
here so as to minimize notational confusion; that is, k is Boltzmann’s constant, and n and k are summing indices.
Sec. 12.8 Linear Block Codes 801
Recall that the Hamming distance D(x;, x pi SD; j between the two binary codewords
x; and x; is the number of positions in which the two codewords differ. On the other
hand, the Hamming weight of a codeword (N-,-tuple) x is defined as the number of nonzero
components in x. Using coded BPSK, the Euclidean distance d;; between two binary coded
signal vectors (S;) and (S;) can be written in terms of the Hamming distance D; j- Now
measures the normalized cross-correlation between the two coded modulation vectors. For
binary codewords
(12.12%)
since Ey,/No =rcEp/No. This result clearly demonstrates the importance of selecting
codes with large values for the minimum Hamming distance. As we shall subsequently see,
the minimum Hamming distance Dyin plays a key role in defining the error-correcting and
detecting capability of a good block code. Notice from (12.126) that the minimum Ham-
ming distance parameter serves as a multiplier on the energy-to-noise ratio, thereby increas-
ing the effective communication efficiency E,/No. From this discussion, we see that one
802 Block-Coded Digital Communications Chap. 12
where
[Din 1G Dea JEip
Re Secs é})= aes 12.128
NcRo S In2 (5) in Ne ( )
and, as N. approaches infinity, P(E) approaches zero if Ro > re. From (12.128) we can
write the ratio
Pas| i
Ro = = 12.129
CC Gea NaN Cee
which says that the product of the bandwidth expansion factor and the cutoff rate is propor-
tional to both the relative minimum Hamming distance, Dmin/Nc, for any particular binary
code and the communication efficiency, E,/No. Thus, the communication efficiency must
be
E
a > In2/(Dinin/
No) (12.130)
in order to achieve arbitrarily small P(E) with large N.. For orthogonal codes, Dmin =
N,/2, so that (12.130) simplifies to (12.15) as it should.
Now Plotkin has shown [8] that the minimum Hamming distance is bounded by
1 NM.
D (42.131)
SMT)
for binary codes with M, codewords and block length N,. This is called the Plotkin bound.
Thus
Dyin Z M.
(12.132)
Ne ~ 2(M. — 1)
Direct substitution of (12.132) into (12.130) demonstrates the requirement that E,/No >
(21n2)(M, — 1)/M- in order for P(E) to approach zero for large N,. Moreover, for large
M_-, we see from (12.132) that Dmin < N¢/2 with the equality being achieved by orthogonal
codes.
- 12.8.3 The Role of the Minimum Hamming Distance in Specifying Good Error
Control Codes
Much insight into the problem of selecting a good error control code can be obtained by
_ the systems engineer by noting that Din is directly related to the error correcting and error
Sec. 12.8 Linear Block Codes 803
detecting capability of any block code. In particular, it can be shown [3] that a block code
with minimum Hamming distance Dyin has t error-correcting capability
This means that any occurrence of t or fewer symbol errors per codeword will be corrected.
On the other hand, it can be shown [3] that if a block code with minimum Hamming distance
Dyin is used for error detection only, the decoder can detect all binary error patterns of
(Dmin — 1) or fewer errors per codeword.
As previously discussed, codes can be characterized as systematic or nonsystematic.
For a systematic code, the encoding process takes the K,-tuple of bits and maps them
into the N.-tuple by appending the additional redundant bits for error correction and/or
detection; see Fig. 12.15. If the error-correcting capability of the code is not exceeded,
then the decoder is guaranteed to correctly decode the N,-tuple into the transmitted K,-
tuple. If the error-correction capability is exceeded, then the decoder will usually do one
of two things. It will either decode the received N,-tuple into incorrect information bits or
it will detect that the error-correction capability was exceeded. If the decoder error cannot
be detected then it is an undetectable decoder error. If the number of errors detected is in
excess of the codes error-correction capability, then when a feedback channel is available,
the decoder might be designed to send an automatic repeat request (ARQ) [12]. Otherwise,
the decoder will output the errors to the user or source decoder. With an ARQ approach, the
decoder detects errors and makes use of a feedback channel to the transmitter to have the
incorrectly received codewords (frame) repeated. Presumably these requests can be repeated
until the codewords are received correctly. This procedure has the effect of reducing the
information rate. In an ARQ system which employs error-correcting codes, a bit error occurs
when the decoder incorrectly accepts a frame of data with errors. Lin and Costello [12] give
a good summary of numerous approaches to error control design using ARQ approaches.
We now investigate the error-correcting and detecting performance characteristics of other
block codes of great practical interest.
As we have observed in Section 12.8.1 for the AWGN channel, the performance of block
codes can be estimated in terms of the minimum Hamming distance if x is mapped onto
S isometrically. Furthermore, for a fixed code rate r, and block length N., the goal is to
choose a code with a large minimum Hamming distance. Usually, block codes are decoded
using algebraic decoding techniques which require a demodulator-detector to make a hard
decision on each received channel symbol, that is, “1” or “0.” Let p denote the probability of
a channel symbol (coded message bit) error on the resulting binary symmetric channel. The
performance of block codes using hard decisions (one-bit quantization of the demodulator
outputs occurring every Ty = T; sec) on the observables rj, r2,..., Nn, can be characterized
in terms of their ¢ error-correcting ability. In particular, suppose that the decoder can correct
any combination of t or fewer channel symbol errors, and no combination of more than
t errors. Such a decoder is called a hard-decision bounded distance decoder. Thus, the
codeword error probability is
where € is a random variable representing the number of coded message bit errors in the
codeword. Since there are ts different ways of having k errors in N, detected coded
message bits, the hard-decision codeword error probability is
Ne
{eos (Xi) ofa agi (12.135)
k=t+1
On the other hand, when hard-decision decoding techniques are not used (that is, with
finer quantization of the observables r; see Fig. 11.6), the decoder implementation com-
plexity is usually much greater than that of hard-decision (one-bit) quantization. Forney
[13] has investigated such block decoder quantization techniques; they are sometimes called
erasure-and-error decoders. For infinite quantization and binary coded BPSK, (12.126) up-
per bounds the codeword error probability for the AWGN channel in terms of Dmin and
Ey/No.
The bit error probability for systematic block codes can be found from (12.135) to be
[17,18]
P pall Do
= ee(Nev ifae
° |pl —p) nee (12.136)
vEV==-
© k=t+1
where e; is the average number of channel symbol errors that remain in the corrected N--
tuple when the channel caused k symbol errors. In practice, it has been fashionable to use
ex =k for k > t, which appears to be a good approximation for the majority of codes of
greatest interest, thereby making (12.136) easy to evaluate numerically.
If the block code is designed for the sole purpose of error detection, the decoder fails
to detect an error in the observed block if and only if the channel transforms the encoder
output into another valid codeword. The undetected error probability P(E) is given by
Ne
PED= >° wp'a—py (12.137)
k=t+1
Hamming codes are perhaps the simplest class of block codes. The (N¢, Ke) Hamming
codes are characterized by the parameters defined in Table 12.6. Using this table, (12.135),
(12.136), and (12.137), one can numerically evaluate the codeword error probability, the
bit error probability, and the undetected error probability. Figures 12.16, 12.17, and 12.18
demonstrate these probabilities versus the hard-decision coded message bit error probability,
p, for m = 3,4, and 5 respectively. Figure 12.19 demonstrates the bit error probability ver-
sus E;,/No for various Hamming codes on an AWGN channel, assuming a BPSK modulator
_ so that p = Q(./2r-E,/No). This figure also demonstrates the coding gain as a function of
_ parameter m. Coding gain, G, is the difference between the E,/ No (in decibels) required for
uncoded communication and the E, /No (in decibels) required for coded communication to
achieve the same bit error probability; that is, G = (Ep/No)uncoded — (Ep/ No) coded:
Sec. 12.8 Linear Block Codes 805
Minimum distance
Golay codes are more powerful than the Hamming codes discussed in the previous section;
however, the decoder complexity and the bandwidth expansion factor are greater. The ubiq-
uitous Golay (23,12) code [3] is the only known multiple-error correcting binary perfect
code that can correct any combination of three or fewer errors per 23-bit codeword. From a
system perspective, the extended Golay code is more useful. This code is formed by adding
one parity bit to the (23,12) Golay code to form the (24,12) extended Golay code. This par-
ity bit increases the minimum Hamming distance of the code from seven to eight and yields
a code rate 1/2 which is easier to decode in hardware than the (23,12) Golay code. The code-
word, bit, and undetected error probabilities for hard-decision decoding can be determined
from (12.135), (12.136), and (12.137) to be [17,18]
P(E) =)(24)
(A )p¢ (t— p) 24-k (12.138)
1SDat (
PE) = 35 24| )p
\ ok (1 — p) 24—k (12.139)
and
107 oe,
10
10
P(E)
10
10
Figure 12.16 Codeword error probability versus channel bit error probability for block
length N. = 2” — 1 Hamming codes with m = 3, 4, and 5 (Reprinted from [17])
P-ary BCH codes. Here the codeword symbols belong to the Galois Field
G F(P). Every codeword is associated with a code polynomial with coefficients in GF (P).
The roots of the generator polynomial lie in an extension field GF (P”) of GF (P) for some
_ integer m > 1. BCH codes can either be primitive or nonprimitive; see the Venn diagram in
~ 3. An introduction to the algebra of Galois Fields (also known as finite fields) may be found in Lin and Costello
{12].
Sec. 12.8 Linear Block Codes
807
107 F
P,()
10 : : :
107 107 107 10°
Channel Bit Error Probability = p
Figure 12.17 Bit error probability versus channel bit error probability for block length
N,. = 2" — | Hamming codes with m = 3, 4, and 5 (Reprinted from [17])
Fig. 11.18. Primitive BCH codes are defined as codes whose block length is N. = P™ — 1.
Given any error-correction capability rt, there exist choices of the message length (K, sym-
bols) and block length (NV, symbols) for which a t-error correcting P-ary (N., K-) code
exists [8, 12, 18]; see Table 12.7 with Q = P.
Binary BCH codes. Binary BCH codes are a special case of P-ary-BCH codes
with P = 2. The codeword symbols (coded message bits) of binary BCH codes are from the
Galois Field GF (2). The roots of the generator polynomial of the code lie in the extension
field GF (2”) for some integer m > 1. In particular, a t-error correcting (t < 2”~') prim-
itive, binary BCH code having generator polynomial with roots in GF (2”) can be found
which achieves the parameters listed in Table 12.8 for any m > 3. There is no simple for-
mula for the system engineer to use for calculating VN; — K-; however, N. — K- is less than
or equal to mt. The parameters for all binary BCH codes of length NV. = 2" —-1 with m < 10
808 Block-Coded Digital Communications Chap. 12
10
10
10
10
Undetected
of
Error
Probability
an
10
10
Channel Bit Error Probability = p
Figure 12.18 Probability of an undetected error versus channel bit error probability for
block length N, = 2” — 1 Hamming codes with m = 3, 4, and.'5 (Reprinted from [17])
are summarized in Table 6.1 of [3] and Table 4.6 in [18]. The codeword error probability
performance versus channel symbol (coded message: bit) error probability, p, is illustrated
in Figures 12.21, 12.22 and 12.23 for several binary BCH codes. Figure 12.24 provides a bit
error probability comparison versus E»/No for various BCH codes.
These codes often have some inherent error detection capability beyond their error-
correction capability. If some error-correcting capability is sacrificed to gain some addi-
tional error detection capability, say tg, then the resultant ¢ error correcting, tg additional
error detecting (tg is an even number), primitive, binary BCH code is characterized by the
parameters summarized in Table 12.9.
Q-ary BCH codes. Q-ary BCH codes are a generalization of the P-ary codes.
The codeword symbols of a Q-ary BCH code come from GF(Q) = GF(P‘), for some
integer c > 1, while the roots of the generator polynomial lie again in an extension field
Sec. 12.8 Linear Block Codes
809
€)
P,
4 5 6 7, 8 9 10 11
E,/Noy dB
Figure 12.19 Bit error probability versus E,/No for block length N. = 2” — |
Hamming codes with m = 3, 4, and 5 on an AWGN channel (Reprinted from [17])
GF(Q”) =GF(P”™), for some m > 1. If c= 1, Q-ary codes become P-ary. There exist Q-
ary BCH codes having the code parameters listed in Table 12.7. A discussion of the theory
of finite fields underlying Q-ary BCH codes can be found in [3, 8, 11, and 18].
Reed-Solomon (RS) codes have found widespread applications beginning with the 1977
Voyager’s Deep Space Communications System [21]. At the time of Voyager’s launch,
efficient encoders existed, but accurate decoding methods were not available! Jet Propulsion
Laboratory scientists and engineers gambled that by the time Voyager II would reach Uranus
in 1986, decoding algorithms and equipment would both be available for decoding of the
810 Block-Coded Digital Communications = Chap. 12
Codeword
error
probability
Bit error
probability to
Probability
Error
Probability
of an
undetected
error
1 107 107 10
4 <4 3
Figure 12.20 Codeword, bit, and undetected error probabilities versus channel symbol
error probability with extended Golay coding (Reprinted from [17])
data transmitted back to Earth. Fortunately, they were correct! Voyager’s communications
system was able to transmit at a data rate of 21,600 bits per sec from 2 billion miles away
and successfully extract the data from a signal whose received power was 100 billion times
weaker than that generated by an ordinary wrist watch battery!
Other areas of application of RS codes include: (1) communications over telephone
systems with such applications as advanced facsimile machines or faxes (which send and
receive imaging data) and high-speed modems (which usually send and receive computer
data); (2) satellite communications with such applications as the Hubble Space Telescope,
Mobile Satellite Terminals, and the 300 megabits per sec (Mbps) return link of the Space
Station Freedom via the Tracking Data Relay Satellite System (SF/TDRSS); (3) Deep Space
Communications with such applications as Voyager, the Galileo Orbiter, the Mars Observer,
Sec. 12.8 Linear Block Codes 811
and the Cassini Titan Orbiter/Saturn Probe; (4) Space Shuttle Communications; (5) enter-
tainment, with applications such as compact disks (CDs) and CD-ROM, etc.
In general, Reed-Solomon (RS) codes may be regarded as a special subclass of Q-ary
BCH codes that results when parameter m = 1. Thus, the coefficients of the code polyno-
mials and the roots of the generator polynomial all lie in GF (P‘). A primitive RS code has
block length N, = P* — 1. For any RS code, the error-correction capability t and the mes-
sage length K, are simply related by K, = N; — 2t. For example, if Q = P© = 2° = 23 = 8,
then this can either be a (7,5) or a (7,3) or a (7,1) RS code depending on whether it is de-
sired to correct t = 1, 2, or 3 symbols respectively. Table 12.10 summarizes the achievable
paramaters for a ¢ error-correcting, primitive RS code.
Binary RS codes are most frequently used in practice. A “binary RS code” is simply
an RS code where P = 2. In the nonbinary case, P is an odd prime. Note that the code
symbols of a binary RS code are not in fact binary, but belong to the field GF (2°). The
parameters for a binary, ¢ error-correcting, primitive RS code are summarized in Table 12.10
with: P= 2.
RS codes meet the Singleton bound (Dmin < Nc — Ke + 1) because for an RS code,
Dmin = 2t + 1 = N- — Ke + 1. Therefore, RS codes are optimal with respect to this bound,
and the term maximum distance separable (MDS) is used to describe optimality in this
(127, 120)
t=1
JEXGE)
(127, 36)
t= 15
(127, 22)
t = 23
sense. It is also worth pointing out that the block length N, of an RS code over GF(Q) =
G F(P‘) can be extended to either Q or Q + | while still maintaining the MDS condition. It
can also be shown [17, 18] that the block (codeword) and 2”-ary symbol error probabilities
for RS codes are given respectively by
Qgm_} Qm fe 1 f
and
P(E)
10°
where K; is the number of bits per codeword; see Chapter 4. Figures 12.25, 12.26, 12.27,
and 12.28 demonstrate the codeword error probability (MEP) versus p profiles for various
RS codes of practical interest.
Interleaving of the channel encoder output is required when the channel errors occur
in bursts; see Section 11.3 in Chapter 11. Inserting an interleaving-deinterleaving system
into the signal path introduces another new system error control system parameter, namely
the interleaving length or depth. The value of this parameter is specified in terms of the
prespecified channel memory length (sec), say T-. Physical considerations which lead to
specifying t- are the expected burst-error duration and the multipath duration which are
frequently characterized in terms of the temporal correlation time of the channel.
RS codes also have applications in error control system designs as outer codes in
concatenated coding systems (see Section 11.5 and Chapter 13). In such designs, an inner
code is also used to provide a portion of the error control, thus the name concatenated code
iN
814 Block-Coded Digital Communications — Chap. 12
107 ‘i Fy
10°
WA,
10°3
P(E)
104
(511, 157)
t= $1
(511, 448)
(511, 211) t=7
t= 41
(511, 259)
5 {a=230
10°
(511, 340) (511, 466)
te =<20
(511, 376)
p t=ot5)
[13]. The inner code is usually a binary code which maps each symbol in the RS alphabet to
a unique binary codeword. Decoding of a concatenated code typically proceeds as follows:
First, the demodulator-detector outputs are soft quantized (soft-decision) and presented as
inputs to the inner channel decoder. The output of this decoder are hard decisions which
serve to drive the outer RS decoder. The inner code protects the RS code against random
channel errors, leaving the correction of burst errors to the outer RS code.
The performance of concatenated coding systems which use an RS outer code are
given in the next chapter. Concatenated codes were first introduced by Forney [13], and their
performance for special coding techniques will be presented in Chapter 13.
Sec. 12.8 Linear Block Codes 815
gPp
a
on>)
¥
7
N
PI
gurpoo2d
E p/Nov dB
Figure 12.24 Bit error probability versus E,/No for coherently demodulated BPSK
over a Gaussian channel using BCH codes (Reprinted from [22])
816 Block-Coded Digital Communications Chap. 12
(31, 13)
t=9
(31,9)
C= sii
P(E)
1071
10°2
10°3'
P(E)
104
10°5
P(E)
(127, 97)
1071
a Nes
\\\
Seay
ee
1072
10°3
P(E)
104
10°5
PROBLEMS
12.1 Consider the discrete memoryless channel (DMC) modeled by the diagram in Fig. P12.1. Write
an expression for the cutoff rate Ro and the channel capacity Cy, assuming equally probable
input messages. :
12.2 For the discrete memoryless channels illustrated in Fig. P12.2, write expressions for the channel
cutoff rate Ro and capacity Cy. Assume equally probable input symbols.
12.3 (a) Develop an explicit expression for the achievable Ro performance for coded QPSK, M =
adh,
(b) ie the result in (a) with that for coded BPSK and comment on your results.
12.4 Consider the M-ary signal set
N
si(t) =) sindn(t)
n=
My
l-p 1
Ry My My Ro
P
Pp
P
R, a R,
l-p l-p
Figure P12.5
probability is bounded by
Ep
P(E) < exp [-K.(= —In 2)
Calculate how large the number of messages M, must be to guarantee that P(E) < 10~° when
E,/No = 0, 3, and 6 dB.
12.7 (a) For M equilikely transorthogonal signals disturbed by additive white Gaussian noise, de-
rive a bound similar to (12.12) and (12.13) which is valid for orthogonal signals.
(b) Repeat (a) for M equiprobable biorthogonal signals.
(c) Using the bounds, discuss the relative advantages of the three signaling systems from a
systems engineering viewpoint.
Hint: See Chapter 3 for the union bound result and Table 12.4.
12.8 Consider communicating over the AWGN channel using binary codes and antipodal signals.
(a) Using the fact that 2B7,. = 1, write an expression for Ro as a function of P/NoB.
(b) Show that as B increases, Ro in bits/dimension decreases for a fixed P/No.
(c) Now consider the channel throughput Ro = Ro/ T; = 2B Ro in bits/sec and show that Ro
increases monotonically with increasing bandwidth B.
(d) Explain these results.
12.9 Show that the channel capacity C in bits/sec increases monotonically toward its maximum
value Ca. as B increases. Show that the capacity Cy. in bits/dimension decreases monotoni-
cally as B increases. Explain this phenomenon.
822 Block-Coded Digital Communications Chap. 12
so that as T approaches infinity, P(E) approaches zero if Rp < Co./2. This bound is not tight
but we know that orthogonal signals actually achieve Co. in the limit (see Chapter 4).
12.11 Derive an expression for the channel cutoff rate for the ensemble of binary codes when orthog-
onal, equal energy, equiprobable signal waveforms (BFSK) are used to transport information.
(a) Derive the channel cutoff rate for the ensemble of binary codes when equal energy,
equiprobable waveforms having cross-correlation coefficient y are used.
(b) Compare and discuss the results found in (a) and (b) with the channel cutoff rate in (12.37)
for antipodal signals (BPSK).
12.12 Consider a voice-grade telephone circuit with a bandwidth of B = 3 kHz. Assume that the
circuit can be modeled as an AWGN channel.
(a) What is the capacity of such a circuit if the CN R is 30 dB?
(b) What is the minimum CAN R required for a data rate of 4,800 bits/sec on such a voice-grade
circuit?
(c) Repeat part (b) for a data rate of 19,200 bits/sec.
12.13 Consider that a 100-kbits/sec data stream is to be transmitted on a voice-grade telephone circuit
(having a bandwidth of 3 kHz). Is it possible to achieve error-free transmission with a CNR
of 10 dB? Justify your answer. If it is not possible, suggest system modifications that might be
made.
12.14 Telephone modems operating at 19.2 kbits/sec are available in the market place.
(a) Assuming AWGN and an available E,/No = 10 dB, calculate the theoretically available
capacity in the 2,400-Hz single-sided bandwidth.
(b) What is the required E,/No that will enable a 2,400-Hz bandwidth to have a capacity of
19.2 kbits/sec?
12.15 Consider the set of J orthonormal waveforms {@,(t)}, k = 1, 2,..., J over the interval [0, T,].
These waveforms are used to construct the signal set {s;(t)} of the form [7]
N
in which the {k;} are integers between | and J. Thus, each signal in the set {s;(t)} is specified
by a vector of the form (kj, k2,..., kz).
(a) Assuming N = 5, J = 4, and
(a) With M, = 2%, show that the coding gain G, defined in Section 12.8.5, is given by
G 4 1010849 [(Eb/No)uncodea/(Eb/
No)coded]
== 10 log jo {re-Dmin — K,/[In 2(E,/No)coded]}
if one uses uncoded BPSK (antipodal) transmission. Use the asymptotic expansion for
Q(x), (4.55), and (12.126).
(b) What is the limiting value for G as E;,/No approached infinity?
(c) Assuming the use of a Hamming code defined in Table 12.6, write an expression for the
code rate r, and plot the asymptotic gain achieved with m = 2, 3,4, 5,..., 10.
12.17 Using the coding gain formula provided in Problem 12.16
(a) Develop an explicit expression for the coding gain achievable for the class of binary BCH
codes defined in Table 12.8.
(b) From Problem 12.16, we see that the asymptotic coding gain for large E;/No is given
by G = 10logjg7¢Dmin. Repeat (a) and write an expression for G in terms of the error-
correcting capability t, N., and Ke.
12.18 Repeat Problem 12.17 for the binary Reed-Solomon codes defined in Table 12.10.
12.19 Consider the coding gain formula given in Problem 12.16.
(a) For the binary primitive (15,9) RS code, write an expression for coding gain.
(b) Plot this gain as a function of E,/No.
(c) What is the maximum achievable gain using this code?
(d) Repeat (a), (b), and (c) for the (154,130) RS code.
12.20 (a) Repeat Problem 12.19 (b) and (c) for the (255,235) RS code with t = 10.
(b) Repeat Problem 12.19 (b) and (c) for the binary (255,179) BCH code with t = 10.
12.21 Figure 12.5 shows a family of channel cutoff rate (Ro) curves for the set of equispaced, equally
probable Q-ary ASK signal points as shown in Fig. 12.4a.
(a) Find the expression for the distance g;; in terms of Q,i, j, and VEN. é
(b) Use the expression for the distance q;; found in (a) to find Ro for the special case of Q = 4.
(c) Graph your result as a function of Ey,./No.
(d) Verify your answer by checking for agreements with Figure 12.5.
12.22 Consider a set of QPSK signal points assigned to the coded message symbols {xm} = {00, 01,
10, 11}.
(a) Find an isometric mapping from coded message symbols to coded modulation vectors
{Sm}. For each pair of points in your signal set, compare the Hamming distance to the
Euclidean distance.
(b) Find a nonisometric mapping from coded message symbols to coded modulation vectors
{Sm}. For each pair of points in your signal set, compare the Hamming distance to the
Euclidean distance.
12.23 Heuristically, justify use of the signal space dimensionality to bandwidth relationship N. =
WT =2BT where T is the signal duration and B (in Hz) is the single-sided bandwidth mea-
sure.
Hint: See Chapter 5 of [7], Nuttall and Amoroso, /T, July 1965, and Landau and Pollak,
Bell System Tech. Journal, 41, July 1962, pp. 1295-1336.
12.24 A customer wants to send information through a channel with single-sided bandwidth 10 kHz.
Channel noise is white and Gaussian with single-sided power spectral density No = 107?”
Watts/Hz. He/she can achieve a received signal power of | pW.
(a) What is the maximum theoretical rate at which error-free transmission is possible?
(b) If the customer decides to send at rate 100 kbps using uncoded quaternary PSK, what will
the actual symbol probability be?
824 Block-Coded Digital Communications Chap. 12
12.25 The received signal power in a coherent AWGN channel is —102 dBm and the power spectral
density of the noise is —228 dBm.
(a) What is the maximum rate (bits/sec) at which one can send information error-free?
(b) What is the corresponding value of E,/No?
(c) What is the maximum amount of error-free information that can be sent in two seconds?
12.26 For QPSK signaling and M = 4
(a) Evaluate Ro using the appropriate Q-ary coded M-PSK model from Section 12.4.5.
(b) Evaluate Ro using the appropriate QASK model from Section 12.4.3.
(c) Compare your answers from parts (a) and (b). If they agree, explain why; if they disagree,
explain why.
REFERENCES
1. Wozencraft, J. M. and R. S. Kennedy, “Modulation and Demodulation for Probabilistic Coding,”
IEEE Trans. Information Theory, IT-12, July 1966, pp. 291-97.
2. Massey, J. L., “Coding and Modulation in Digital Communications,” Zurich Seminar, vol. E2,
1974, pp. 1-4.
3. Lin, S., An Introduction to Error-Correcting Codes, Englewood Cliffs, NJ: Prentice Hall, 1970.
4. Lindsey, W. C. and M. K. Simon, Telecomunication Systems Engineering, Englewood Cliffs, NJ:
Prentice Hall, 1973. Republished by Dover Publications, 0-486-66838-x, 1991.
5. Shannon, C. E., “A Mathematical Theory of Communication,” Bell System Tech. J., 27, 1948,
pp. 379-423 (Part I), pp. 623-56 (Part II). Reprinted in book form with postscript by W. Weaver,
Urbana, IN: Univ. of Illinois Press, 1949; also see Claude E. Shannon Collected Papers, edited
by N. A. Sloane and A. D. Wyner, IEEE Press, 1992.
6. Shannon, C. E., “Probability of Error for Optimal Codes in a Gaussian Channel,” Bell System
Tech. Journal, vol. 38, May 1959, pp. 611-56.
7. Wozencraft, J. M. and I. M. Jacobs, Principles of Communication Engineering, New York: John
Wiley and Sons, 1965.
8. Gallager, R. G., Information Theory and Reliable Communication, New York: John Wiley and
Sons, 1968.
9. Cover, T. and J. A. Thomas, Elements of Information Theory, New York: John Wiley and Sons,
1991.
10. Viterbi, A. J., “Error Bounds for Convolutional Codes and an Asymptotically Optimum Decoding
Algorithm,” JEEE Trans. Information Theory, vol. IT-13, April 1967, pp. 260-69.
11. Berlekamp, E. R., Algebraic Coding Theory, New York: McGraw-Hill, 1968.
12. Lin, S. and D. J. Costello Jr., Error Control Coding: Fundamentals and Applications, Englewood
Cliffs, NJ: Prentice Hall, 1983.
13. Forney, G. D. Jr., Concatenated Codes, Cambridge, MA: MIT Press, 1966.
14. “Troposcatter Interleaver Study Report,’ CNR, Inc., Report Number RADC-TR-75-19 for Rome
Air Dev. Center, Air Force Systems Command, Griffiss AFB, NY, February 1975.
15. Nyquist, H., “Certain Topics on Telegraph Transmission Theory,” Trans. AIEEE, vol. 47, April
1928, pp. 617-44.
16. Wozencraft, J. M. and B. Reiffen, Sequential Decoding, Cambridge, MA: MIT Press, 1961.
17. Odenwalder, Joseph P., Error Control Coding Handbook, Final Report under contract No.
F44620-76-C-0056, Linkabit Corporation, July 1976. Also see Chapter 10 on Error Control in
Data Communications, Netwoks and Systems by Thomas C. Bartee, ed., Indianapolis: Howard
W. Sams, 1985.
References 825
In the previous chapter, we saw that a linear (N;, K-) block code can be used for error
control purposes by partitioning the data stream into blocks containing K, bits each. Then,
using an encoding rule to associate the r = N, — K¢ parity check bits with each K; data bits,
an N.-symbol codeword was formed for transmission. However, with convolutional codes
(originally called recurrent codes), the encoded data do not have a simple block structure,
and, as we shall see, a convolutional encoder operates on the input bit stream such that each
information bit can affect a finite number of consecutive symbols (coded message bits) in
the encoder output.
Apparently convolutional codes were first introduced by Elias in 1955 [1] as an alter-
native to block codes. In this chapter, we shall introduce convolutional codes, methods for
representing such codes, and present the maximum-likelihood (ML) decoding algorithm due
to Viterbi [2]. Detailed performance results achievable with maximum-likelihood decoding
of convolutional codes are given for a set of particular convolutional codes and key design
issues are treated. The chapter concludes by presenting the applications and performance
of concatenated codes introduced by Forney [3] for error control. Punctured convolutional
codes are also discussed.
Other decoding techniques such as threshold decoding, sequential decoding, etc., lie
outside the scope of this book; however, the details of these techniques can be found in
Gallager [4], Michelson and Levesque [5], Berlekamp [6], Anderson and Mohan [7], and
Lin and Costello [8].
role for block or convolutional codes. An important difference between binary block codes
and binary convolutional codes is that the encoder has memory. As we shall see, the Ne-
tuple output of a convolutional coder is not only a function of the input K,-tuple but is also
a function of the previous K, — | input K,-tuples. Usually K, and N, assume small integer
values and K is varied to control redundancy in the transmitted data stream.
The functional architecture of a (K-, Ne, K) convolutional encoder is mechanized
with a kK-stage linear shift register and N, modulo-2 adders (see Fig. 13.1). From this
figure, we observe that a convolutional code is defined by the number, kK, of stages in
the shift register, the number of modulo-2 adders (number of outputs), and the connections
between the shift register and the modulo-2 adders. The encoder state is defined to be the
contents of the rightmost kK — | stages in Fig. 13.1. Knowledge of the next K, input
bits is therefore necessary and sufficient to determine the next output sequence. In other
words, an input sequence of k data bits (“ones” and “zeros”) of time duration T = K-Tp
sec are shifted into the first k stages of the kK register and all bits in the register are
shifted k stages to the right, thereby establishing a new state of the kK-stage register.
The outputs of the N. modulo-2 adders are sequentially sampled to yield the N, output
channel code symbols (coded message bits) which are identified as the codeword branch
vectors X; = (Xj1, Xj2,.-.,XjN,) Of N- channel symbols. In what follows, we shall consider
the most commonly used code for which k = 1. Generalization to higher-order alphabets
can be found in [8]; however, treatment of such alphabets here lies outside the scope of
this book. The N, code symbols occurring at the encoder output during the time interval
tj+1 — tj = N-T- comprise the ih codeword branch xj; = (Xj1,...,Xjn,). Here xji(i =
1,..., Nc) is the i binary channel symbol in the j" codeword branch and is assumed
to take on values “O” or “1.” When K, = 1, the convolutional coder of rate re = 1/N¢ is
implemented with a (K — 1)-stage register and the constraint length K can be referred to in
units of bits.
The encoder of Fig. 13.2 performs two polynomial multiplications. The output orthe up-
per mod-2 adder is the product of the oe -2 aie Bi Na eee eer x? and the
input sequence u(x) = ug + ux + nx? +n. + uj ix/ +..., where uj; € {0, 1} represents
the input information bit in the j™ bit time iieral Gi- DT. < 5< jT;. The output of
the lower modulo-2 adder is the product of u(x) and g2(x) = 1+ x*. The output sequence
is represented by the two (in this case) semi-infinite polynomials X (CY)= X10, lees
(1-4)
- eBeysylus Jejsibes
I 4
828
€ L-y
‘ln=n
‘en ieahe In Sa
yndu|
jiq eouenbes
peyiys
ul °y syiq
Auane
°y 4 ‘spuoses
jy—
2 c-O|Npow
sioppe
promepon eouanba
x = s‘tyx) x («lx
OJ8UM
ly= ‘tX) aly Cae NN
(
a P1oOMAapod
youesg
= I
youelg
plomapod
aanBrgq
"ET [eUONNJoAUOD
Jopooud
Jo jUTeNsUOD
YISug]
yy pur ayes°4 = °N/?Y
x= ysul4 apog indjno
joquiAS
jo apoyo ebessay
sig
yy! youesg
=X X)
‘ Sy Goes (Caay
jig
Beg
yndu;
=a cIneacta)
2
iG
eouenbas
|
2x
Jo
apog
puosas
= yy!
joquiAg
youelg
829
830 Convolutional-Coded Digital Communications Chap. 13
xox +...+ x4 jx) + ...= g1(x)u(x) and X2(x) = x20 + x21% + x22Xx eo) haga oe
.. = g(x) w(x) which are interleaved to produce the encoder output sequence. The con-
volutional code is completely defined by the set of generator polynomials g(x) and g2(x).
These polynomials define the length of the shift register, the connections to the modulo-2
adders, and the number of modulo-2 adders. For example, the encoder of Fig. 13.2 maps
each information bit into N, = 2 coded message bits using a rule which depends on the pre-
vious two information bits. Specifically, an input 1 produces the outputs 11, 01, 00, or 10
when the previous inputs are 00, 10, 01, or 11 respectively. The state of the encoder in Fig.
13.2 corresponds to the contents in the shift register, viz., the previous two input bits. The
constraint length K of the code is one plus the number of shift registers required to im-
plement the encoder. The number of stages in the shift register required to implement the
encoder is K — | and the number of encoder states is 2X~!. The encoder of Fig. 13.2 has
constraint length K = 3 since the current output pair is a function of the current input plus
the two previous inputs. Note that different definitions of constraint length can be found in
the literature on convolutional coding. In all cases, however, constraint length is a measure
of the memory required to implement the encoder.
In general, convolutional codes produce N, output aneerel symbols for each K, input
symbols and have rate re = K-/N-. Although this would appear to make convolutional
codes identical to block codes, they are significantly different. For instance, the number of
past inputs which affect the mapping of the current K, inputs into N; outputs is an important
parameter of convolutional codes. This parameter affects convolutional code performance
and complexity much as the block length affects block code performance and complexity.
As already pointed out, consideration of the more complex (K;., Nc, K) codes lies outside
the scope of this book and the reader is referred to Lin and Costello [8] for a through
treatment. With this limitation, however, convolutional codes can be defined by a set of N,
generator polynomials {g1(x), g2(x),..., 9N.()}, where there is one polynomial for each
of the N, output channel symbols. The constraint length K of the code is equal to one plus
the degree of the highest-degree generator polynomial. The length of the shift register for
the encoder is equal to the degree of the highest-degree generator polynomial. Table 13.1
summarizes the code generator vectors gj,i = 1,..., Nc, for a set of optimum rate 1/2
and 1/3 codes. The free distance, dy, indicated in this table will be further discussed in
Section 13.4. For example, if the code generator vector g; = 111 then gj(v) =1+x+ x"
and if gy = 101 then go(x) = 1 + x? as required in the implementation of the encoder
illustrated in Fig. 13.2.
Finally, we point out that convolutional codes can be characterized as systematic
or nonsystematic depending on whether or not the information sequence appears directly
within the code sequence. For a convolutional code to be systematic, at least one of the
code generator polynomials must equal | (for example, g;(x) = 1) so that one of the N;
outputs is simply the current information bit. Unlike block codes, the performance of the
best systematic convolutional codes is not equivalent to the performance of the best nonsys-
tematic convolutional codes. For the same code performance, a systematic code will have a
longer constraint length and be more complex to generate than the performance-equivalent
nonsystematic code. For this reason, most of the convolutional codes used in practice are
nonsystematic.
Sec. 13.2 Representation of Convolutional Codes 831
101111
110101
1 1001111
1101101
1 10011111
11100101
110101111
100011101
101111
110101
111001
1001111
1010111
1101101
11101111
10011011
10101001
832 Convolutional-Coded Digital Communications Chap. 13
The disadvantage of the state diagram as a method for representing a convolutional encoder
is that it suppresses the time scale. The tree diagram, although a little more complex, adds
the time dimension to the state diagram. Figure 13.4 illustrates the tree diagram for the
encoder shown in Fig. 13.2. At each successive input bit time, the encoding procedure can
be described by traversing the diagram from left to right. Here, in Fig. 13.2, each tree branch
depicts the encoder output codeword branch x;. The approach used to generate the tree
diagram is described as follows: If the input bit is a zero, its associated codeword branch
output is found by moving to the next rightmost branch in the upward direction. If the
input bit is a one, its codeword branch output is found by moving to the next rightmost
branch in the downward direction. Assuming that the initial contents of the encoder is all
zeros, the diagram shows that if the first input bit is a zero, the output branch word is 00
and, if the first input bit is a one, the output branch word is 11. Similarly, if the first input
bit is a one and the second input bit is a zero, the second output branch word is 10. Or,
if the first input bit is a one and the second input bit is a one, the second output branch
word is a 01. Following this procedure, it is easy to see that the input data bit sequence
u = 11011 traces the heavy line illustrated in Fig. 13.4. This path corresponds to the output
codeword sequence x = (Xj, X2, X3, X4, Xs) where xj = 11, x7 = 01, x3 = 01, x4 = 00, and
x5 = 01. From this tree diagram, we notice that the tree repeats itself (period K in general)
after three input bits which corresponds to the constraint length; however, for large K one
readily sees the limitation of using the tree diagram to represent the encoder output. Noting
this periodic feature of the tree diagram suggests an alternate representation, viz., the use of
a trellis diagram first suggested by Forney [10].
Sec. 13.2 Representation of Convolutional Codes
833
00
01
Legend
Sees - - - Input
Bit 1
10
To produce the trellis diagram for the encoder shown in Fig. 13.2, we take advantage of
the fact that the tree structure repeats itself after K branchings, that is, it is periodic with
period K. We therefore begin by labeling the tree nodes in Fig. 13.4 to correspond to the
four possible code states A, B, C, and D defined by the state diagram illustrated in Fig. 13.3.
To incorporate time in the diagram, let t; denote the time epochs of the codeword branches.
In this case t;+; =1t; + 27; sec, for all j = 1. We arbitrarily set t; = 0. More generally, fora
code rate re = 1/N-, then tj+; =t; + NcTc sec. We now develop the trellis diagram shown
in Fig. 13.5 for the encoder shown in Fig. 13.2.
Assume that the encoding operation begins in state A = 00 shown on the far left of
the trellis at ¢ = t; = 0. If the first information bit is a zero, the encoder moves along the
solid line out of state A = 00 arriving again at state A = 00 at t = fy. The encoder output
is the symbol pair 00 which is labeled on the trellis branch between the two states. On the
other hand, if at t = ft; the first encoder input bit were a one, the encoder would move along
the dashed branch out of state A = 00 arriving at state B = 10 at time t = f2 = 27,. In this
case, the encoder output is 11 which is indicated on the trellis branch connecting states A
and B. The second possible branching takes place at t = f2 where the encoder is in either
834 Convolutional-Coded Digital Communications Chap. 13
Codeword
branch ‘
oo A
state A or B. If state A exists and a zero is inputted to the encoder, then it moves along the
solid branch outputting 00 and arriving in state A at t = fp; if a one is input, the encoder
moves to state B along the dashed line outputting 11 at t = &. Similar comments apply if
at t = f2 the encoder is in state B. This process of moving from left to right every bit time
through the trellis and outputting the trellis branch labels continues as long as the encoder
input is clocked with data bits. Using our representation, input information bits which equal
zero cause the encoder to move along the solid trellis branches to the next state while input
bits which equal one cause the encoder to move along the dashed branches. The codeword
branch output sequence x = (x), X2,..., Xj, ...) consists of the sequence of branch labels
Xj = (xj1,Xj2) encountered as the encoder follows the path from left to right. A typical
input output sequence and associated modulator input waveform for this convolutional code
is shown in Fig. 13.6. For L input bits into the encoder, there are 2%</ possible output
sequences that can be generated.
Since convolutional encoding is best understood using illustrative examples, we fur-
ther demonstrate the concepts of the state and trellis diagrams using a rate r, = 1/3 convolu-
Sec. 13.2 Representation of Convolutional Codes
835
State A=00
B=10
C=01
D=11
Legend
Input bit O
— Input bit 1
Figure 13.5 Trellis diagram for the rate 1/2, K = 3 convolutional encoder of Figure 13.2
tional of constraint length K = 4. Figure 13.7 illustrates the encoder mechanization for the
code generator polynomials defined by
gi(x)=lt+xtx7?
$x
g(x) =14+x4+x3 (13.1)
g3(0) z= 1 ex
For this case, three channel output symbols defined by the output branch word x; =
(xj1,%Xj2, xj3), are produced at multiples of 7, = 37, sec for each input bit. The output
from the 3 mod-2 adders in Fig. 13.7 are sampled to produce the components of x;. The
specific registers summed to produce an output are defined by the generater polynomials in
(13.1). None of the generators equal | so that the code is nonsystematic. Since all powers
of x between 0 and 3 occur in the generator g;(x), the first output symbol is the modulo-2
sum of the input bit and all three previous inputs stored in the shift register. The number of
stages of the shift register is equal to 3, which is the degree of the largest-degree generator
polynomial (the degrees of all generators are the same). The constraint length of the code
is K =4 which is 1 plus the degree of any of the generator polynomials. Thus, an input
bit influences the output due to the next three input bits. The number of states in the trellis
representation of the code is 2* ~! — 8. The state transition diagram for this convolutional
encoder is illustrated in Fig. 13.8 while the corresponding trellis diagram is shown in Fig.
13.9. The trellis is found using the procedure discussed earlier for the rate re = 1/2 code
836 Convolutional-Coded Digital Communications Chap. 13
Information
Sequence, m | m,=0 Mo= 1 M3= m,4=0 m,=1 mg=0 m7=0
sora SendMps H O n
+1
Waveform
t2
-1
t,=T, =2T,
Figure 13.6 Encoder input and output sequences and waveform for L = 7
where the trellis state is the contents of the shift-register in Fig. 13.7. The three components
of a trellis branch label represent the mod-2 sums defined by the generator polynomials or
by the operations indicated in Fig. 13.7. Since x° = 1 is a component of all three generator
polynomials, the labels on the two branches leaving the trellis state are complements of each
other. The careful reader should take time to verify the state and trellis diagrams shown in
Figs. 13.8 and 13.9.
Be
fy= Hy Ay (ffx
jndjno sjoquiAs
pepod)
aBessoyy
(sig
ai
ely
eansiq
LE] AeY ‘¢/[y p= [euolnfoAuos
Japoous
837
838 Convolutional-Coded Digital Communications Chap. 13
to 13.10 is accomplished by adding the sub or superscript “j” to binary source bits, channel
oo 299
codeword branches, modulator and demodulator outputs to denote that the j" branch in the
trellis diagram is being processed in the decoder. This branch occurs during the j th time in-
terval tj41 —t; =NcTc, j = 1, where knowledge of t; corresponds to codeword branch or
node synchronization we assume ft; = 0. Assuming further that the IF channel in Fig. 13.10
is memoryless, that is, a CMC, then the demodulator functions to create the output vec-
tor rj; = (rj1,.-.,7jN,) whose components are the projections of received waveform r ;(t)
onto the basis functions of the coded modulator output waveform Si.(t); see Chapters 3 and
4 and Fig. 13.10. Furthermore, by quantizing the demodulator outputs the modulator, IF
channel, and demodulator can be easily converted into the DMC discussed in Section 11.2
of Chapter 11. This model will be useful when we wish to operate the decoder using hard-
or soft-decision outputs from the demodulator. This important design issue will be discussed
further when we present the so-called Viterbi algorithm.
For purposes of being precise, consider now the class of binary convolutional codes of rate
rc = 1/Nc; specifically consider encoding a binary message sequence of L message bits. In
Sec. 13.3 Decoding Convolutional Codes 839
State
A=000 5 ee ee
B=001
C=010
D=011
E=100
F=101
G=110
H=111
this context, one of M, = 2! messages sequences exist and can be represented by the binary
L-tuple m = (mj, ..., my). Here m A m; is a binary random variable taking on values one
and zero. Let ui, (with m = 0 or 1) characterize the source encoder output of Fig. 13.10 dur-
ing the What known data bit interval (j — 1)T) <t< < JT =v= Ned conjucrile Otle eaeeretn OUIe re
words, if m = 0 then We= 0 and if m = 1 then ul 11, The Cee Ae be in Fig.
13.10 maps the ee bit ul, into the j'"codeword branch a — (xj; ais Xn) of the trel-
lis diagram where xj; is a binary random variable taking on values ay = 0or ee = | during
the i*° channel symbol interval of the j"” bit interval. To connect the codeword branches to
the physics of the modulator, then as veces x), are mapped one-to-one onto the coded
BPSK modulation vectors $j,= (5) Pa se “) which are equivalently represented by the
coded modulation waveform
for (j — 1)T, < t < jTp sec. Here the i coded message bit (channel symbol) coefficient sj,
takes on values a = Ses if m= 1 during the j'" codeword time interval tia. —t7 =Nele
and sd’ = —,/Ev, if m = 0 during the same time interval; Ey, represents the “energy” of
the BPSK channel symbols and Ty = 7;. Furthermore, the pulse shape pj(t — i7;) of unit
energy serves to characterize the modulator output waveform during the known i” channel
symbol time of the ies codeword branch and is zero elsewhere. Thus, the IF channel coded
modulation waveform for a sequence of L input bits (the message m) can be represented as
Uist=NebTy=LT, (13.3)
L .
2) aSe
er)
j=1
840 Convolutional-Coded Digital Communications Chap. 13
Codec Codec :
Information Source
Source Encoder
Channel IF
Clock Channel
Information
Sink Decoder Decoder modulator
where for each j, m=O or 1 with equal probability and m is one of Pligbinary message
vectors. Now Sm(t) encodes L bits into N,L binary channel symbols si; or one of 2Ne4
ae ee viz., the L trellis codeword branches containing N, channel symbols
x),= (x; x)JNey Substitution of (13.2) into (13.3) leads to the channel waveform
which is observed to be the sum of L waveforms which represent the codeword branches in
the trellis diagram. The vector equivalent to the coded modulation waveform S(t) is de-
fined by the coded modulation vectors Sm = (Soe S27; age Ss‘) where Sj, = (si) SHAKY.JNey
If r;(t) denotes the noise corrupted received waveform occurring during the aie inter-
val tj41 —tj =NcTc, j =1,...,L, then the demodulator creates the output projection
rj =(rji,.--,17jn.) of this waveform using the equivalent basis functions for the modulator
output waveform sj, (t).
We are now in a position to establish the maximum-likelihood decision function for
the j" trellis codeword branch. Assuming infinite “leepear eaeifeh m “Na 0 or 1) is
to be transmitted, then the codeword branch xj, implies Si, A (si! a JNey so that the
Sec. 13.3 Decoding Convolutional Codes 841
Na
where we have assumed the noise samples in r;j are independent and p(r; iS, yA Sr; (0; SZ, )
characterizes the conditional joint probability density function of the demodulator output,
or equivalently, the IF channel transition probabilities. If we assume the L bits transmitted
consist of the message m = (m , m2, ..., m,), then the corresponding maximum-likelihood
decision function (conditional upon m being transmitted) becomes
Since the logarithm is a monotone increasing function of an increasing argument, the de-
coder could equivalently use the decision metric In A(m). In this case, the decoder metric
(decision function) becomes
L Ne ae
In A(m) = 5 In inp Go) (13.7)
j=! j=l
and the maximum-likelihood decoder outputs the vector m = m which maximizes In A(m).
When the IF CMC is modeled as the AWGN channel, then
ij 1 (ri fal 7)
any
p (riilsm) = a exp inane (13.8)
2En. 2roEE
ee | a 13.9
for all i, 7, and m. Substitution of (13.8) into (13.7) leads to a path (decision) metric Uae
tion) which can be written equivalently in terms of the squared Euclidean distances dtjm=
rj — S/,|2, that is
j=]
1. Note again that m= 0 or 1 for each j = 1,2,..., L; hence, no reason to subscript m in the probability
densities.
842 Convolutional-Coded Digital Communications Chap. 13
This says that one selects the message sequence m which minimizes the sum of the squared
Euclidean distances, that is, set i = m if
where i is now one of the 2” binary vectors representing the binary message sequence.
This says that the maximum likelihood decoder evaluates the Euclidean distance between
the received vector r = (r},¥2,...,1%z) and all possible coded modulation vectors Sm 4
(Ss! S2 LoS) and selects that binary vector mm = m = (m1, m2,..., mz) which min-
pnived. d(m). We are using the fact that maximum likelihood decoding of convolutional
codes is equivalent to minimum distance decoding; see Chapter 3. Equivalently, for the
AWGN channel the minimum distance decoding can be shown to reduce to that of maximiz-
ing a set of 2'«“ cross-correlation operations; see Chapters 3 and 4. Further simplification of
(13.10) leads to the equivalent decision rule of selecting mm = m if and only if m maximizes
the cross-correlations
16
ngs | Dor +c (13.12)
ele
where for each value of i and j, m =0 or | and C’ is a constant bias. This says that to
perform maximum-likelihood decoding we see that the test in (13.7) requires that one eval-
uate the 2%" cross-correlations defined by (13.12). For large L this algorithmic approach to
decoding becomes too complex to implement, that is, exponential growth in decoder com-
plexity occurs. To avoid this exponential buildup in decoder complexity and still achieve
the performance offered by maximum-likelihood decoding, Viterbi [2] recognized two key
facts: first, that the tree diagram representation of a convolutional code was periodic with pe-
riod K; and second, that paths in the tree diagram actually merge and hence this fact could
be used to simplify the decoding process. Forney [10] clarified this point by introducing
the trellis diagram from which it is clear that certain codeword paths remerge as a function
of the state of the encoder. In what follows, the Viterbi algorithm for use in maximum-
likelihood decoding of convolutional codes is presented.
As the L-bit message grows in length we observed that the numbers of correlations (decoder
complexity) required for a rate r, = 1/N,- code grows exponentially. The Viterbi algorithm
(VA) is an elegant and efficient method for performing ML decoding of convolutional codes.
The algorithm was first described mathematically by Viterbi [2] in 1967. Since that time
the algorithm has been described many times by many authors; most notably Forney [10]
provides a highly readable and insightful description of the algorithm and its performance.
Recall from the previous section that the function of a ML decoder is to find the
codeword vector sequence S,,= (Se 5 An So) which was most likely to have been
transmitted given the received channel output vector sequence r = (rj,..., rz) (see Fig.
11.2). Corresponding to each Sy there is a binary message vector m = (m ,..., mz), which
corresponds to a path through the trellis that gives the largest value for the log-likelihood
function In A (m) established by (13.7). Maximizing this is equivalent to selecting the binary
vector m, say m, such that the Euclidean distance
Sec. 13.3 Decoding Convolutional Codes 843
dy (m) = |r — S| = (13.13)
is minimum at m=m. Here d;jm 2 Ej — si, | is the distance between the demodulator
output r; during the j" bit interval and the hypothesized coded modulation vector S/, =
(S745 ees Sin.) transmitted during the j" bit interval. In the context that follows, the variable
L is used to truncate the “length” of the cross-correlation operations which in practice will
be a few constraint lengths.
In order to illustrate this decoding procedure, it is simplest to consider the case of hard-
decision decoding, that is, the case where R j =(Rj1,..., Rjn,) is one of QNec binary vectors
with components taking on values Rj; = 1 or 0. The vector R; is generated from rj by
making hard decisions on each component of rj, that is, Rj; = sgn(rji); see Figs. 11.3,
11.6, and 11.7. As opposed to S$”, these hard-decision vectors are to be compared with all
possible encoder output codeword branch vectors x), = Oi, wees a iNey, generated in the
trellis. From (12.122) we know that the Euclidean distance squared is proportional to the
Hamming distance; this important fact will now be used to rewrite (13.13) and illustrate the
Viterbi algorithm. Thus, for hard-decision decoding, the ML decoding algorithm in (13.13)
uses the decoding components qd‘, =|rj- si,|? == 4Ey,Dj(m) where Dj(m)4 |Rj — x),|
for all j = 1,2,..., L with m =0 or 1. Here the distance measurement is now made in the
Hamming sense, that is, Dj(m) 4 |R; — x),| is the number of places in which the binary
vector R; differs from x},. In other words, for hard-decision decoding of coded BPSK,
the Euclidean distance measure is “equivalent” to the Hainnnng distance. For example, if
the hard- rc BPSK demodulator output during the j codeword branch is R;j=ill
and the j'4 codeword branch is x), == 00 then the distance D;(m) = |Rj — Si,f= 25 Inkthis
context, the ML decision rule for hard-decision decoding is to select that message vector
=(mj,...,mz), say m, selected such that the accumulated Hamming distance
Ib L
Dim) = )/ Dj Gn) = > IR; — Xml (13.14)
is minimum. The Hamming distances Dj(m) = |Rj — x},| are sometimes called branch
metrics and D,,(m) is called the accumulated path metric (consisting of the sum of the
branch metrics) due to transmission of data vector m. Specifically, each codeword branch
vector x), in the trellis diagram is at a Hamming distance
algorithm. We now demonstrate the Viterbi algorithm using the encoder of Fig. 13.2 and its
corresponding trellis diagram shown in Fig. 13.5.
Dj(m) = (13.16)
At t =t; we assume that the encoder is initialized to the state A = 00. In fact, (13.16)
serves to define the cumulative Hamming distance (path metric) as the sum of the branch
Hamming distances relative to the demodulator outputs up to time t = t;. For example, at
t = t7 the demodulator output is R2 = 11 and the Hamming distance between R> and the
encoder output of XG = 00(m = 0) is 2, while for x} = 11(m = 1) and the distance is 0.
These accumulated distances are labeled on the decoder branches at t = fy in Fig. 13.11,
that is, going from encoder state “A” at t = t) to state “A” at tf = tf we label the decoder
trellis branch with 2, while going from state “A” at t; to state “B” at t2 we label the decoder
trellis with an accumulated distance of 0. One continues in this manner by labeling the
accumulated distances on the decoder trellis branches as the demodulator outputs occur
Sec. 13.3 Decoding Convolutional Codes 845
Received sequence R 11 01 01 10 01
ty to
State A=00
Branch
metric
B=10 = >
ay
C=01 =
D=11 rT]
through t = f. The VA uses these accumulated Hamming distances to find the accumulated
path metric D;(m) which gives rise to the minimum distance, that is, the shortest path
through the trellis relative to the demodulator outputs R = (Rj, ..., Rs).
Viterbi made the following creative and innovative observation which in essence elim-
inates the need to create all cross-correlations required in (13.12). If any two paths in the
trellis merge to a single state at any time f = t;, one of them can always be eliminated in
the search for the shortest (minimum distance) path. For example, in Fig. 13.12 two paths
merge at t = f5. The upper path has the accumulated distance (path metric) Ds = 4 while
the lower accumulated path metric Ds = 1. Thus, the upper path need not be given further
consideration with respect to finding the shortest path through the trellis because the lower
path, which enters the same decoder state, has a lower path metric. This lower path is called
the survivor path for this state. Needless to say, this was Viterbi’s astute observation and is
primarily responsible for eliminating the need to compute, store, and accumulate all 2Nel.
correlations required in (13.12). Technically speaking, this is justified on the basis of the
Markov nature of the encoder state, that is, the present encoder state summarizes the en-
coder history in the sense that past encoder states do not affect future trellis branch outputs.
More generally, we can state that a (N., K-, K) convolutional encoder has 2*-! states in
the trellis at time t =¢;4; =1t; + NcTc, j = 1. Each of these states can be entered by one
of two paths. The VA consists of computing the accumulated branch metrics for the two
paths entering each state and eliminating one of them. This computation is done for each of
846 Convolutional-Coded Digital Communications Chap. 13
ty to tg tg ts
Path metric = 4
1 ia
State A=00
the 2“—! states at time ¢; and the survivors are retained; then the decoder moves to time tj+1
and repeats the process of updating the accumulated path metrics and creating the survivors.
To further demonstrate the process, we consider decoding the message for the demod-
ulator outputs and trellis diagram given in Fig. 13.11. Following Sklar [27], we assume that
the decoder is synchronized with the encoder and knows that the initial state at t = f), is
A = 00. In practice, knowledge of the initial state is not necessary; however, for purposes of
discussion there is no loss in generality. From state A = 00 the only possible transitions are
to state A = 00 or B = 10; see Fig. 13.13(a). At time ¢ = fy state “A to A” has branch metric
Aa = D2(A) = 2 while state “A to B” has branch metric 4p = D2(B) = O. At t = fo there
are two possible branches leaving states A = 00 and B = 10; see Fig. 13.13(b); the cumu-
lative path metrics of the codeword branches are labeled A4 = D3(A), AR = D3(B), Ac =
D3(C), Ap = D3(D) terminating at t = £3. At time t = & in Fig. 13.13(b) there will be two
possible trellis branches diverging from each state. As a result there are two paths entering
each state at t = f4; see Fig. 13.13(c). For each state, the path giving rise to the largest cu-
mulative can be eliminated with regard to further computation and the resulting decoding
trellis is illustrated in Fig. 13.13(d). Here we observe that 44 = D4(A) = D4(B) = Ag (ac-
cumulated path metrics) equal at t = t4. When this occurs in the decoding process, one path
can be chosen at random for elimination without affecting the average error probability. The
surviving path for each state is illustrated in Fig. 13.13(d).
Thus far in the decoding process, the cumulated branch distance metrics have been
the major concern and no mention has been made regarding the decoded bit stream. With
respect to this we note from Fig. 13.13(d) that there is only a single surviving path at time fy
and this path (shown dashed) has been produced by the input bit m = 1; the decoder outputs
a | as its first decoded bit. Here we can see how the decoding of the surviving branch is
facilitated by having drawn the lattice branches with solid lines for input zeros and dashed
lines for input ones. Note that the first bit was not decoded until the path metric computation
Sec. 13.3 Decoding Convolutional Codes 847
had proceeded three bit times in practice into the trellis. This so-called coding delay can be
as much as five times the constraint lenght in bits, that is, 5K Ty sec.
Now that we have demonstrated the decoder bit decision process and the concept of
coding delay, let us continue with the decoding process. At each succeeding step (t > t4)
in the decoding process there will always be two possible paths entering each state of
the encoder and decoder; one of the two will be eliminated by comparing cumulated path
metrics leaving the survivor paths. At t = t4, Fig. 13.13(e) demonstrates the two departing
paths which arrive at the four states at time fs. At t = fs there are two paths entering each
state and one of each pair can be eliminated; Fig. 13.13(f) shows the survivors. At this point
we note that we cannot make a decision on the second message bit because there still are two
paths leaving the state 10 node at time fy. At time f@ in Fig. 13.13(g) we see the pattern of
remerging paths, and in Fig. 13.13(h) we see the survivors at time fg. Also, in Fig. 13.13(h)
the decoder outputs one as the second decoded bit, corresponding to the single surviving
path between f2 and f3. The decoder continues in this way to advance deeper into the trellis
and to make decisions on the message bits by eliminating all paths but one.
The storage requirements of the Viterbi decoder grow exponentially with constraint length
K. For a code with rate 1/N,, the decoder retains a set of 2X—! paths after each decoding
step. With high probability, these paths will not be mutually disjoint very far back from the
present decoding depth. All of the 2“ ~! paths tend to have a common stem which eventually
branches to the various states. Thus, if the decoder stores enough of the history of the 2 ~!
paths, the oldest bits on all paths will be the same. A simple decoder implementation, then,
contains a fixed amount of path history and outputs the oldest bit on an arbitrary path each
time it steps one level deeper into the trellis. The amount of path storage required, S, is [13]
5 =h2*—! where h is the length of the information bit path history per state. A refinement,
which minimizes the value of h, uses the oldest bit on the most likely path as the decoder
output, instead of the oldest bit on an arbitrary path. It has been demonstrated, [13], that a
value of h of 4 or 5 times the code constraint length is sufficient for near-optimum decoder
performance. The storage requirement, S, is the basic limitation on the implementation of
Viterbi decoders. The current state of the art in microelectronics technology limits decoders
to a constraint length of about K = 15. For K > 10 such decoders have been called big
Viterbi decoders.
Codeword branch synchronization is the process of determining the beginning of a
branch word in the received sequence. Such synchronization can take place without new
information being added to the transmitted symbol stream because the received data appear
to have an excessive error rate when the branch synchronization is incorrect. Therefore, a
simple way of accomplishing synchronization is to monitor some concomitant indication of
this large error rate, that is, the rate at which the path metrics are increasing or the rate at
848 Convolutional-Coded Digital Communications Chap. 13
ty to Path Metric
2
A=00 ———— Ma =2
\0
B=10 ‘ N. i
d,,=0
(a)
Path Metric
ty to tg
2 1 2
Be eee A, =3
XO N1
B=10 - A373
2
C01 \
rare Nese
“
:
D=11_ . : \: do - 0
A=00
B=10
C=01
D=11
dX, =3
A
i =3
B
Xr. =0
< Figure 13.13 Selection of survivor paths:
(a) Survivors at t = f, (b) Survivors at
d=? t = 13, (c) Metric comparisons at t4, (d)
Survivors at t = t4 (Reprinted from [27])
Sec. 13.3 Decoding Convolutional Codes 849
ty to
A=00 \ 0 .
; “N
=10 - ~*~
C=01
D=11 :
(e)
A=00 da =1
de =i
B=10
Ne =3
C=01
D=11 x D7
A=00 Nn
B=10
C=01
D=11
4 to
~0
B=10 - aN d,=2
C=01 he a
D=11 x 5 =i
(h)
Figure 13.13 continued Selection of survivor paths: (e) Metric comparisons at fs, (f)
Survivors at t = ts, (g) Metric comparisons at f, (h) Survivors at t = t¢ (Reprinted from
[27])
850 Convolutional-Coded Digital Communications Chap. 13
which the surviving paths in the trellis merge. The monitored parameters are compared to a
threshold, and synchronization is acquired accordingly. The important details of this process
lie outside the scope of this book and are usually best understood in practice where decoders
are built to work in practical digital communication systems. :
where
Ro; 0<re=
Ro —€ < Ro
E(rc) = [email protected]).. Ro eak = =v.0)* < Cy, (13.18)
for 0 < op < 1 and Ro = Eo(1, Q) is the quantized channel cutoff rate defined in (12.102).
Analogous with block codes, the probability of error in decoding convolutional codes
can be made arbitrarily small by allowing N, to approach infinity in the region where
E(r-) > 0.
Figure 12.8 illustrates the random coding exponent E(r,) achievable in very noisy
channel for convolutional and block codes. The relative improvement increases with code
rate; in particular for r. = Ro = Cy,./2, the exponent for convolutional codes is almost six
times that for block codes. Since exact expressions for probability of error are difficult to
come by, we consider alternate approaches in what follows.
Like block codes, the distance properties of convolutional codes play a key role in under-
standing the error-correcting capability of the code. This point can best be made using the
encoder shown in Fig. 13.2 to demonstrate the distance between all possible pairs of code-
word sequences that can be generated by a random bit stream. In block codes, it was the
minimum distance Dmin that determined the f error-correcting capability of the code. For
convolutional codes, Dmin is replaced by the so-called minimum free distance D f and re-
lated to t via
Des
t= 5 (13.19)
where [x] denotes the largest integer which does not exceed the bracketed ratio.
To demonstrate the meaning of Ds, we make use of the fact that a convolutional
code is a linear code; hence, there is no loss in generality in simply finding the minimum
distance between each codeword sequence (a cascade of code branch word) and the all-zeros
sequence. First, we redraw the trellis diagram in Fig. 13.5 and relabel the branch outputs
by the Hamming distance from the all-zeros codeword. The resulting trellis is shown in
Fig. 13.14. Assuming that the all-zeros input sequence was transmitted, the trellis paths
of interest are those that start and end in state A = 00 anywhere in between. According
to the VA discussed in the previous section, at least one bit error will occur whenever the
Sec. 13.4 Probability of Decoding Error for Convolutional Codes
851
State A=00
B=10
C=01
D=11
Legend
Input bitO
ee Input bit1
Figure 13.14 Trellis diagram labeled with distances from the all-zeros path
distance of any other path that merges with state A = 00 at time ¢ = f; is less than that
of the all-zeros path up to time f;, thereby causing the all-zeros path to be eliminated in
the decoding process. The minimum distance for making such an error can be found by
exhaustively evaluating the distance between all paths from the 00 state to the 00 state in
trellis relative to the all-zeros path. For those paths that remerge to the 00 state that are
shorter in time length than the longest path, the sequence will be made of equal length by
appending the necessary number of zeros so as to make all paths have equal length. Using
Fig. 13.14, consider then all paths that diverge from the all-zeros path and then merge for
the first time at some arbitrary trellis node ending at t,. There is one path that departs at
= ft, and remerges to the all-zeros path at t = f4; this path has distance 5 from the all-zero
path and corresponds to the input data sequence of 100. Similarly, there are two paths at
distance 6 from the all-zeros path; one departs at f; and merges at time ts, while the other
departs at t; and merges again at t = f¢, etc. The input message bits for these two paths are,
respectively, 110 and 10100. Each of the message vectors differs in two bits from the all-
zeros path. Thus, the minimum distance is seen to be 5 since the distances for all arbitrarily
long paths in greater than or equal 5. Setting Dr = 5 in (13.19) verifies that Viterbi decoder
corresponding to the encoder in Fig. 13.2 can correct any two channel symbol errors.
For an arbitrary convolutional coder of rather short constraint length, the distance
properties can be examined qualitatively using the so-called code generating function
or transfer function; however, for long constraint lengths the complexity of the transfer
852 Convolutional-Coded Digital Communications Chap. 13
00
iy
: ale
Figure 13.15 State diagram labeled with distances from the all-zeros path
function increases exponentially and other methods must be explored [8], [12]. Table 13.1
summarizes the free distance for several optimum short constraint length rate 1/2 and 1/3
convolutional codes. The details associated with using the transfer function to study the dis-
tance properties of convolutional codes lie outside the scope of this chapter; however, in the
next section we will present one example to demonstrate the approach.
00
/ ee LN Ss ™» ‘
ce D°LN
a eee
Ue 10
DL
seis 11
D*L
ate
401 10
DLN \_ yy DL
aan de
10. i
/ DLN
Figure 13.16 State diagram labeled according to distance, length, and number of input ones
number of branches traversed by a path. Letter N characterizes the number of ones in a path.
The branch label includes N only if the information bit associated with that path is a one. ~
To demonstrate how this notation accounts for a codeword sequence ones, information bit
ones, and path lengths consider the path passing through the state sequence 00, 10, 01, 10,
11, 01, 00. The product of the labels on this path is D’L°N* = D?LN-DL-LN-DLN-
DL - D*L and the codeword sequence for this path is 11, 10, 00, 01, 11 which contains»
seven ones corresponding to the power of D in the overall path product. The corresponding ~
transmitted bit sequence for this path is 1 0 1 1 0 which contains three ones corresponding
to the power of N in the path product. On the other hand, the total number of branches
traversed is seen in the power of L and the path product is six. Consequently, we conclude
that the path label products for the parameters D, L, and N for all paths through the state
diagram characterize the desired features from which we will now take advantage.
Let S;(00), S(10), S(01), S(11), and S.(00) denote a sum of the path label products
for all paths starting with 00 and proceding with 10, 01, 11 and ending with 00. Then the
generating function, which also characterizes the transfer function of a signal flow graph,
can be found by simultaneous solution of the state equations written from Fig. 13.16, viz.
S,(00) = 1
S(10) = S,(00) - D?LN + S(01)LN
S(11) = S(10) - DLN + S(11)DLN (13.20)
S(01) = S(11)DL + S(10)DL
Each term in the above sum corresponds to the branch label sequence for a path which di-
verges and then merges with the all-zero path. The first term corresponds to the shortest
path which goes through the state sequence 00, 10, 01, 00. This path has length three cor-
responding to L* with Hamming weight five corresponding to D° and a message sequence
Hamming weight of one corresponding to N. The second and third terms in (13.21) enumer-
ate the two paths with Hamming weight of six and a message Hamming weight of two. This
analysis methodology which has just been described by example is applicable in principal
to all convolutional codes; however, for large K it is difficult to write down.
For convolutional codes with Viterbi decoding, the appropriate performance measure is the
probability that the correct path through the trellis is discarded at depth j (time ft = fj).
That is, the performance measure is the probability of making an error at each opportunity
for decoding error. An opportunity to discard the correct path exists each time the decoder
moves one step deeper into the trellis. This measure is comparable to the measure used for
block codes where an opportunity for error exists each time a new block is decoded and
the appropriate performance measure is the block decoding error probability. Note that the
probability of selecting an incorrect path over the full extent of the trellis is an inappropriate
performance measure. That is, it would not be correct to evaluate the probability that the
total semi-infinite decoder output sequence is not continuously merged with the correct
sequence. Over all time, the probability that the decoded path diverges from and remerges
with the correct path at least once approaches unity. Since convolutional codes are linear,
overall average decoding error probability is then equal to the decoding error probability
calculated assuming that the all-zero sequence is transmitted. For the remainder of this
discussion assume, therefore, that the correct path through the trellis is the all-zero path.
The decoding error probability is then the probability that a nonzero path entering state
00 at depth j of the trellis is selected as the survivor. In order to evaluate decoding error
probability, all potential decoding error events at depth j are enumerated, and their error
probabilities are calculated and summed. The summation is the result of a union bounding
argument similar to that used for the error probability bound for block codes. The reader is
referred to [2, 12, 18] for details of the error proabability derivation. There it is shown that
the decoding error probability P(E) is overbound by
[e,2)
where a, is the number of nonzero paths with Hamming weight k that merge with the all-
zero path at depth j and diverge once from the all-zero path at some previous depth, and
P, is the probability of incorrectly selecting a path with Hamming weight k. The values a,
can be found by differentiation from the generating function T(D, L, N) for the code with
L=1 and N = 1. Specifically
ka TOs 1)
ak = ki BRR yk ae, |D=0
The probability of incorrectly selecting a path which is Hamming weight k from the all-
zero path is a function of the channel and the received signal-to-noise ratio. For the binary
symmetric channel, P; is equal to the probability that symbol errors in the k positions where
the correct and incorrect paths differ cause the received sequence to be closer in Hamming
distance to the incorrect path than to the correct path. In particular, it can be shown that [12]
: 1
P= DO (5) reap) toe Ga pe py] (13.24)
e=(k/2)+1
for k even and
k
m= DO (E) pa = pk (13.25)
e=(k+1)/2
for k odd. Furthermore, for infinite quantization and convolutional encoded BPSK modula-
tion it can be shown that [2, 12]
2kreE
n=0/ ees) (13.26)
No
For soft-decision decoding, the calculation of Py, is much more complex and the reader is
referred to [2, 12] for the details.
Bit error probability is calculated using essentially the same techniques as used for the
decoding error probability. To calculate bit error probability, each decoding error event is
weighted by the number of bit errors associated with that event. Since the all-zero path is
the correct path, the number of bit errors associated with a nonzero path is equal to the
number of information ones associated with that path. The generating function is used to
enumerate error events along with the number of attendant bit errors. Recall that the number
of information ones on a nonzero path through the trellis is equal to the power of N in the
generating function term corresponding to that path. In order to change the exponent of N
into a multiplier to weight the path by the number of bit errors, the generating function is
differentiated with respect to N and N is then set to unity.
Consider the generating function given in (13.21). This function enumerates all
nonzero paths through the trellis of Fig. 13.5. Differentiating 7(D, L, N) with respect to
856 Convolutional-Coded Digital Communications Chap. 13
since the path length does not affect the result. This summation accounts for all possible
nonzero paths through the trellis which merge with the all-zero path at time t;. The first term
indicates that there is a single bit error associated with all nonzero paths which are Hamming
distance five from the all-zero path. The second term indicates that there are four bit errors
associated with all nonzero paths which are Hamming distance six from the all-zero path.
These four bit errors come from two different paths, each of which is associated with two
bit errors. In general, there are a total of cg bit errors associated with all paths which are
Hamming distance k from the all-zero path. Using the generating function to weight error
events by the number of bit errors, an upper bound on the bit error probability is
[o.@)
where Py is calculated using (13.24) for the hard-decision channel and (13.26) for the con-
tinuous output additive white Gaussian noise channel. The coefficients cz, can be found from
(13.27) by differentiation. Unfortunately, the performance analysis of a decoder which uses
the VA does not lend itself to closed form analysis and computer simulation techniques have
to be used. We shall present several such results in the next section for a set of convolutional
codes of greatest practical interest to date.
Constraint
length 3
ae
SNASGAR EEE
Log
Error
[Bit
Probability]
9;
3 4 5 6 i 8 10 11 12
E,/No, dB
Figure 13.17 Bit error probability performance for various rate 1/2 convolutional codes
using hard-decision decoding (Reprinted from [18])
Information bits from the data source are grouped into blocks of K, to create the 2Ke_ary
symbol alphabet used by the Reed-Solomon code. The coding system which appears first in
the chain, in this case the RS code, is called the outer code of a concatenated code system.
The RS output symbols are converted to their K,.-bit binary representations and input to a
symbol interleaver. The purpose of this interleaver is to enable the spreading of bursts of
channel symbol errors over RS codewords as uniformly as possible. As discussed in Chap-
ter 11, the interleaver changes the order in which the RS symbols are transmitted over the
channel. The symbols are put back in their original order in the deinterleaving buffer in the
receiver. This process of changing the transmission order, transmitting, and reestablishing
the original order causes contiguous bursts of channel errors to be spread uniformly over a
number of RS codewords rather than occurring all in a single codeword. RS codes, as well
as all of the other codes which have been studied, are designed to perform best when chan-
nel errors occur independently from one symbol to the next and do not occur in bursts. For
our purposes, we assume that the interleaver/deinterleaver combination may be considered
processors which take bursts of channel errors and spread these errors over some number of
codewords. The outer interleaver output of Fig. 11.16 serves as the input to a convolutional
coder whose output is sent over the channel. The convolutional code and Viterbi decoder op-
erate exactly as discussed previously; the convolutional code is called the inner code of the
system. The output of the Viterbi decoder is an estimate of the binary sequence which was
858 Convolutional-Coded Digital Communications Chap. 13
aulRi
se
Ae
NN
NWN
WO
lon
NAN rd
9
,
Log
Probability]
Error
[Bit
ee
es
ee
aA7
e177
ies
Pe
oe St
se
see
ENN
COL
A NS
ASAE
HHH Ae
MABHRONS: 6
[\q
[TTTTTT iiratrersa'ae
MRS
4/Ny, 4B
Figure 13.18 Bit error probability performance for various rate 1/2 convolutional codes
using soft-decision decoding (Reprinted from [18])
input to the encoder. The Viterbi decoder output is input to the outer deinterleaver. Detailed
measurements and simulations of Viterbi decoder performance have shown that output er-
rors tend to occur in bursts. It is for this reason that the interleaving scheme is required.
Without the interleaver, the channel would not appear to be memoryless. The outer deinter-
leaver output binary symbols are grouped into blocks of K, to construct the 2*¢-ary symbols
for input to the RS decoder. The RS decoder performs the processing required to estimate
the data source sequence. The RS decoder is able to correct some of the decoding errors of
the Viterbi inner decoder, thus improving communication reliability. Figure 13.23 illustrates
the bit error probability performance of the concatenated coding system of Fig. 11.16. The
convolutional code used for Fig. 13.28 has re = 1/2 and constraint length K = 7. Results
are shown for a number of different RS codes with K = 6,7, 8, and 9. Observe the cod-
ing gain at the bit error probability of P,(E) = 10~> of aproximately 7.0 dB relative to an
uncoded BPSK system.
Concatenated codes consisting of an inner convolutional code and an outer Reed-
Solomon (RS) code are used in several current and planned deep-space missions. The Voy-
ager spacecraft, for example, employs a concatenated coding system based on a (7, 1/2)
convolutional code as the inner code and an 8-bit (255, 223) RS code as the outer code. The
Galileo mission will use an experimental (15, 1/4) convolutional code concatenated with the
same 8-bit RS code. Future missions may use the recently discovered (15, 1/6) convolutional
code together with a 10-bit (1023, 959) RS code [19].
Sec. 13.7 Concatenated Reed-Solomon and Convolutional Codes
859
~ SCCEET EEE
ee Ed Uncoded He
Ov
Constraint
length 3
Log
Error
[Bit
Probability]
,
FEET SAS
CCCP
3 8
ANE
is/No, dB
Figure 13.19 Bit error probability performance for various rate 1/3 convolutional codes
using hard-decision decoding (Reprinted from [18])
Figures 13.24 to 13.32 show the performance of concatenated codes with ideal and
nonideal interleaver, and point out the difference in performance between 8-bit and 10-bit
RS codes [19]. The E,/No (bit SNR) shown in these figures is the bit signal-to-noise ratio
of the concatenated system, which includes a penalty of 0.58 dB due to the rate of the 8-bit
RS code, or 0.28 dB for the 10-bit RS code.
Figures 13.24 and 13.25 show the bit error rate of the (7, 1/2) code concatenated with
the 8-bit and 10-bit RS codes, respectively. Interleaving depths 7 = 2 and J = 4 are shown
along with ideal interleaving. Higher values of 7 had nonmeasurable performance degrada-
tion relative to ideal interleaving. Larger constraint length codes suffer more from shallow
interleaving since the average length of bursts grows with the constraint length. It must be
observed that results for nonideal interleaving need very large amounts of data (decoded
bits) and are not very accurate even with runs of 10 million or more bits. Figure 13.26 shows
a comparison of the bit error rates of the 8-bit and 10-bit codes concatenated with the (7, 1/2)
code and ideal interleaving. The advantage of the 10-bit RS code becomes apparent only at
very low bit error rates. A different 10-bit RS code specifically optimized for concatena-
tion with the (7, 1/2) code could offer a larger improvement over the 8-bit RS code than the
(1023, 959) RS code, which was optimized for the (15, 1/6) code.
Figures 13.27 and 13.28 show the bit error rate of the (15, 1/4) code concatenated
with the 8-bit and 10-bit RS codes, respectively. Figure 13.29 shows a comparison of the bit
error rates of the 8-bit and 10-bit codes concatenated with the (15, 1/4) code and ideal inter-
860 Convolutional-Coded Digital Communications Chap. 13
2B
8AS o
nif
EPacbe=
oI
6 N
a
5 aa SAE
VCE
eS
a
& V
ea)
a
S
=
a I
X=
Bs
a
EIFale
Slee
Pehl
SA FatDaw \
Figure 13.20 Bit error probability performance for various rate 1/2 convolutional codes
using soft-decision decoding (Reprinted from [18])
leaving. Now the advantage of the 10-bit RS code over the 8-bit RS code is approximately
0.2 dB at a bit error probability of 10~°. This advantage grows to approximately 0.3 dB at
ieee
Finally, Figs. 13.30 and 13.31 show the bit error rate of the (15, 1/6) code concate-
nated with the 8-bit and 10-bit RS codes, respectively. Figure 13.32 shows a comparison of
the bit error probability of the 8-bit and 10-bit codes concatenated with the (15, 1/6) code
and ideal interleaving. The advantage of the 10-bit code is slightly less than 0.2 dB at a bit
error probability of 10~°, and approximately 0.3 dB at a bit error probability of 107!2.
10°
=O—oO- Simulation
aa De?
10%
=
a
Ss
te
e
=ro 10°
°
S
s5)
=
Q
10°
A sequential decoder works [15], [16] by generating hypotheses about the transmitted code-
word sequence; it computes a metric between these hypotheses and the received signal. It
goes forward as long as the metric indicates that its choices are likely; otherwise, it goes
beckward, changing hypotheses until, through a systematic trial-and-error search, it finds a
likely hypothesis. Sequential decoders can be implemented to work with hard or soft deci-
sions, but soft decisions are usually avoided because they greatly increase the amount of the
required storage and the complexity of the computations.
While the error probability delivered by the VA decreases exponentially with K, the
number of states of the decoder complexity grows exponentially with constraint length.
Yet the computational complexity of the VA is independent of characteristics in the sense
that, compared to hard-decision decoding, soft-decision decoding requires a small increase
in the number of computations. The error probability associated with sequential decoding
algorithm is asymptotically equivalent to that of maximum-likelihood decoding and does
not require searching all possible codeword paths; in fact, the number of paths searched
is essentially independent of the constraint length, thereby making it possible to use large
constraint lengths (K = 4). This important factor provides for the ability to operate at very
small values of error probability. The major disadvantage of sequential decoding is that the
number of state metrics is a random variable, while the average number of poor hypotheses
and backward searches is a function of channel symbol energy-to-noise ratio, Ey./No =
862 Convolutional-Coded Digital Communications Chap. 13
=-O—O- Simulation
Probability
Error
Bit
rcEp/No. When Ey./No is small, more hypotheses must be tried than with high values
of Ew./No. Because of this variability in computational load, buffers must be provided to
store the arriving sequences. For small Ey./No the received sequences must be buffered
while the decoder is laboring to find a likely hypothesis. If the average symbol arrival rate
exceeds the average symbol decode rate, the buffer will overflow, no matter how large it
is, causing a loss of data. The sequential decoder typically puts out error-free data until the
buffer overflows, at which time the decoder has to go through a recovery procedure. The
buffer overflow threshold is a very sensitive function of E,/No. Therefore, an important
part of a sequential decoder specification is the probability of buffer overflow, see [5], [15],
and [16].
Figure 13.33 compares the bit error probability achievable versus E;,/No when Viterbi
and sequential decoding are used to decode convolutional codes. Their comparative perfor-
mance is made with BPSK operating in the AWGN channel. The curves compare Viterbi
decoding (rates 5 and 4 soft decision, K = 41). One can see from Fig. 13.33 that coding
gains of approximately 8 dB at Pg = 10~° can be achieved with sequential decoders. Rel-
ative to the approximate 11 dB of coding gain in E,/No promised by the “Shannon limit”
relative to uncoded BPSK, one might conclude that the major portion of that has already
been achieved. Sequential decoding has been used in various early (circa 1970) important
deep space missions.
Sec. 13.8 Other Convolutional Decoding Techniques 863
1073
t) = Number of symbol
errors that the R-S
| decoder is capable
1044 \ of correcting.
10-5
Probability
Error
Bit
10-6
Figure 13.23 Summary of concatenated coding bit error probability performance with
a K =7,r = 1/2 convolutional inner code and various RS outer codes (Reprinted from
[18])
Certain convolutional codes can be decoded using circuitry similar in complexity to that
used to decode Hamming codes. In order to be decoded so simply, the codes must have
additional structure as defined in detail by Massey [17] and by Lin and Costello [8]. The
simplicity of threshold decoders makes their implementation at very high speeds possible.
Although the performance of these special convolutional codes is not outstanding for the
AWGN channel, the structure is particularly convenient for applications where interleaving
is required; see Chapter 11. For channels where the interference is of the burst type, large
values of coding gain can be achieved.
864 Convolutional-Coded Digital Communications Chap. 13
INTERLEAVING DEPTH, |
Oo 1=2
a1=4
@ |=o
Probability
Error
Bit
E,/No, dB
Figure 13.24 Concatenated code performance (7, 1/2) convolutional code and 8-bit
(255, 223) RS code (Reprinted from [19])
INTERLEAVING DEPTH, |
Oni
4 1=4
@ l=o0
Probability
Error
Bit
E,/No, dB
Figure 13.25 Concatenated code performance (7, 1/2) convolutional code and 10-bit
(1023, 959) RS code (Reprinted from [19])
866 Convolutional-Coded Digital Communications Chap. 13
4 8-BIT RS CODE
B 10-BIT RS CODE
Probability
Error
Bit
E,/N, 6B
Figure 13.26 Concatenated code performance comparison: 8-bit (255, 223) and 10-bit
(1023, 959) RS code concatenated with (7, 1/2) convolutional code (Reprinted from [19])
Sec. 13.9 Punctured Convolutional Codes 867
INTERLEAVING DEPTH, |
Ol=2
a4l1=4
4 1=6
o!1=8
@ |=0
Bit 10-4
Error
Probability
INTERLEAVING DEPTH, |
o1=2
1=4
1=6
1=8
| = 00
Probability
Error
Bit
E,/No, dB
Figure 13.28 Concatenated code performance (15, 1/4) convolutional code and 10-bit
(1023, 959) RS code (Reprinted from [19])
Sec. 13.9 Punctured Convolutional Codes 869
Probability
Error
Bit
10-15
10-16
0.2 0.4 0.6 0.8 1.0 epg lid 1.4
E,/No, dB
Figure 13.29 Concatenated code performance comparison: 8-bit (255, 223) and 10-bit
(1023, 959) RS codes concatenated with (15, 1/4) convolutional code (Reprinted from
[19])
870 Convolutional-Coded Digital Communications Chap. 13
INTERLEAVING DEPTH,|
1=2
1=4
1=6
{=8
| ="e0
Error
Bit
Probability
0.4 06 08 i025) 12
E,/No, dB
Figure 13.30 Concatenated code performance (15, 1/6) convolutional code and 8-bit
(255, 223) RS code (Reprinted from [19])
Sec. 13.9 Punctured Convolutional Codes 871
10°!
1072 INTERLEAVING DEPTH, |
1073 Oo 1=2
1=4
10-4 1=6
10-5 Pak
| = 00
10-6
1077
10-8
10-9
Probability
Error
Bit 10°19
10711
10712
10-13
10°14
OME
0.1 0.3 0.5 0.7 0.9
Figure 13.31 Concatenated code performance (15, 1/6) convolutional code and 10-bit
(1023, 959) RS code (Reprinted from [19])
872 Convolutional-Coded Digital Communications Chap. 13
O 8-BIT RS CODE
@ 10-BIT RS CODE
Fd Probability
Error
Bit
E,/No, dB
Figure 13.32 Concatenated code performance comparison: 8-bit (255, 223) and 10-bit
(1023, 959) RS codes concatenated with (15, 1/6) convolutional code (Reprinted from
[19])
Sec. 13.9 Punctured Convolutional Codes 873
10°!
RATE = 1/3,K=41,
SEQUENTIAL
RATE =1/2,K=41,
SEQUENTIAL
10°2 RATE 1/2, K=7,
VITERBI, SOFT DECISION
RATE 1/3,K=7,
VITERBI SOFT DECISION
107 RAT ESE 25 Ki= 7
VITERBI
HARD DECISION
RATE 1/3,K=7,
VITERBI,
HARD DECISION
Error
Bit 10°4
Probability
UNCODED BPSK
10°?
=6
1giG 2 4 6 8 10 12 14
E,/No. dB
Figure 13.33 Bit error probability performance for various Viterbi and sequential
decoding schemes using coherent BPSK over an AWGN channel (Reprinted from [28]
© IEEE)
874 Convolutional-Coded Digital Communications Chap. 13
PROBLEMS
13.1 Assuming one input bit per clock cycle:
(a) What is the constraint length K and code rate for the convolutional encoder illustrated in
Fig. P13.1?
(b) Draw the tree, trellis, and state diagram generated by this encoder.
13.2 Assuming one input bit per clock cycle:
(a) What is the constraint length K and code rate for the convolutional encoder illustrated in
Fig. P13.2?
(b) Draw the tree, trellis, and state diagram generated by this encoder.
13.3. Assuming one input bit per clock cycle:
(a) What is the constraint length K and code rate for the convolutional encoder illustrated in
ign Piss,
(b) Draw the state diagram generated by this encoder.
13.4 Given the encoder illustrated in Fig. P13.2 and suppose the decoder employs hard decisions
to create the binary symmetric channel. Suppose that the received sequence for the first eight
branches is [12]
00 01 10 00 00 00 10 10
Trace the decisions on a trellis diagram labeling the surviving path’s Hamming distance metric
at each node. When a tie occurs in the decision metric, always choose the upper path.
13.5 Given the K =3,r = 1/2 code in Fig. P13.2, suppose the code is used on a BSC and the
transmitted sequence for the first eight branches is [12]
Bits
Coded Message
Bits
Figure P13.1
Figure P13.2
Problems 875
11 01 01 00 10 11 00
Suppose this sequence is received error free, but the first bit is lost and the received sequence is
incorrectly synchronized giving the assumed received sequence
10 10 10 01 01 10
This is the same sequence except the first bit is missing and there is now incorrect bit syn-
chronization at the decoder. Trace the decisions on a trellis diagram, labeling the survivor’s
Hamming distance metric at each node.
Hint: Note that when there is incorrect synchronization, path metrics tend to remain’, = -<~
relatively close together. This fact is used to detect incorrect bit synchronization.
13.6 Draw the state diagram, tree diagram, and trellis diagram for the K = 3, rate 1/3 convolutional
code generated by
gi(x) =x+ x?
82(x) =1+x
g3(x) =1 te
13.7 Suppose that you were trying to find the quickest way to get from London to Vienna by boat
or train. The diagram in Fig. P13.7 was constructed from various schedules. The labels on each
path are travel times. Using the Viterbi algorithm, find the fastest route from London to Vienna.
In a general sense, explain how the algorithm works, what calculations must be made, and what
information must be retained in the memory used by the algorithm.
, . : Coded Message
Bits Bits
Figure P13.3
Figure P13.7
876 Convolutional-Coded Digital Communications Chap. 13
13.8 Assuming the coded system operates at capacity Cy,, use Fig. 12.11 and determine the max-
imum coding gain (relative to uncoded BPSK operated at a bit error rate of 10-5) achievable
for a rate 1/2 code. Obtain this gain for hard-decision decoding (J = 2), 3-bit quantization
(J = 4), and infinite quantization. Recall that the communication efficiency required for BPSK
is E,/No = 9.6 dB.
13.9 Assuming that the coded system operates at Ro, repeat Problem 13.8 using Fig. 12.12. Compare
your results with those found in Problem 13.8.
13.10 What is the free distance, D+, achievable with the rate 1/3 convolutional coder illustrated in
Fig. 13.7? The free distance Df is the minimum Hamming distance between two distinct trellis
paths through the code trellis.
13.11 Assuming rate 1/2 convolutional coded BPSK and operation at a bit error probability of 1072;
what is the coding gain relative to uncoded BPSK if:
(a) arate 1/2, K =7 convolutional code and hard-decision decoding is used.
(b) arate 1/2, K =7 convolutional code and 3-bit soft-decision decoding is used.
(c) Repeat (a) and (b) assuming a rate 1/3, K = 7 convolutional code is used.
(d) Compare the coding gains achievable for the two code rates and comment on why the
results are different.
13.12 A concatenated Reed-Solomon/convolutional code is used to improve system performance in
the AWGN channel at a bit error probability of 10~>.
(a) Assuming a K =7, rate 1/2 convolutional code and coded BPSK, use Fig. 13.23 and find
the coding gain achievable (relative to uncoded BPSK) when the 2”-ary RS symbols are
constructed from n-bit blocks of input data bits, n = m = 6, 8.
(b) Compare your results with those found in Problem 13.11.
REFERENCES
. Elias, P., “Error-Free Coding,” JRE Trans. Information Theory, IT-4, 1954, pp. 29-37.
. Viterbi, A. J., “Error Bounds for Convolutional Codes and an Asymptotically Optimum Decoding
Algorithm,” JEEE Trans. Information Theory, vol. IT-13, April 1967, pp. 260-69.
. Forney, G. D., Jr., Concatenated Codes, Cambridge, MA: MIT Press, 1966.
. Gallager, R. G., “Information Theory and Reliable Communication,” New York: John Wiley and
Sons, 1968.
. Michelson, A. M. and A. H. Levesque, Error-Control Techniques for Digital Communications,
New York: John-Wiley and Sons, 1985.
. Berlekamp, E. R., Algebriac Coding Theory, New York: McGraw-Hill, 1968.
.J. Anderson, J: B. and S. Mohan, Source and Channel Coding, Boston, MA: Klumer Academic
Publishers, 1991.
. Lin, S. and D. J. Costello, Jr., Error Control Coding: Fundamentals and Applications, Englewood
Cliffs, NJ: Prentice Hall, 1983.
. Forney, G. D., Jr., “Convolutional Codes I: Algebraic Structure,’ IEEE Trans. Information The-
ory, vol. IT-16, 1970, pp. 720-38.
. Forney, G. D., Jr., “The Viterbi Algorithm,” Proceedings of the IEEE, vol. 61, March 1973, pp.
268-78.
. Forney, G. D., Jr., “Convolutional Codes, II: Maximum Likelihood Decoding,” Information and
Control, vol. 25, July 1974, pp. 222-26.
. Viterbi, A. J. and J. K. Omura, Principles of Digital Communication and Coding, New York:
McGraw-Hill Book Company, 1979.
References 877
13: Heller, J. A. and I. M. Jacobs, “Viterbi Decoding for Satellite and Space Communication,” IEEE
Trans. Communication Technology, vol. COM-19, no. 5, pp. 835-48, October 1971.
14. Odenwalder, J. P., “Error Control” in Data Communications, Networks, and Systems, Thomas
Bartee, ed., Indianapolis, IN: Howard W. Sams, 1985. Also see Error Control Coding Handbook,
San Diego, CA: Linkabit Corp., July 1976.
LS: Wozencraft, J. M. “Sequential Decoding for Reliable Communication Record,” vol. 5, part 2,
1957, pp. 11-15.
. Fano, R. M., Transmission of Information, New York: MIT Press and Wiley, 1961.
. Massey, J. L., Threshold Decoding, Cambridge, MA: MIT Press, 1963.
. Ziemer, R. E. and R. L. Peterson, Jntroduction to Digital Communication, New York: Macmillan
Publishing Company, 1992.
. Pollara, F. and K. M. Cheung, “Performance of Concatenated Codes Using 8-Bit and 10-Bit
Reed-Solomon Codes,” Telecommunications and Data Acquisition Progress Report 42- ce Jan.—
Mar. 1989, pp. 194-201, Pasadena, CA: Jet Propulsion Laboratory.
20. Begin, G. and D. Haccoun, “High-Rate Punctured Convolutional Codes: Structure Properties arid
Construction Technique,” IEEE Trans. on Communications, vol. 37, no. 12, December 1989, pp: ~:
1381-85.
Dc. Cain, J. B., G. C. Clark and J. Geist, ““Punctured Convolutional Codes of Rate (n — 1)/n and
Simplified Maximum Likelihood Decoding,’ JEEE Trans. Inform. Theory, vol. 1T-25, January
1979, pp. 97-100.
fuphe Yasuda, Y., Y. Hirata, K. Nakamura and O. Otani, “Development of a Variable-Rate Viterbi
Decoder and Its Performance Characteristics,” 6” Int. Conf. Digital Satellite Commun., Phoenix, 2
AZ, September 1983.
23. Yasuda, Y., K. Kashiki and Y. Hirata, “High-Rate Punctured Convolutional Codes for Soft Deci-
sion Viterbi Decoding,” IEEE Trans. Commun., vol. COM-32, March 1984, pp. 315-19.
24. Begin, G. and D. Haccoun, “Decoding of Punctured Convolutional Codes by the Stack Algo-
rithm,” Abstracts of Papers, 1986 IEEE Int. Symp. Inform. Theory, Ann Arbor, MI, October
1986, p. 159.
De Haccoun, D. and G. Begin, “High-Rate Punctured Convolutional Codes for Viterbi and Sequen-
tial Decoding,” IEEE Trans. Commun., vol. COM-37, November 1989, pp. 1113-25.
26. Hagenover, J., “Rate Compatible Punctured Convolutional Codes and Their Applications,” JEEE. |
Trans. Commun., vol. COM-36, April 1988, pp. 389-400.
Pay Sklar, B., Digital Communications, Englewood Cliffs, NJ: Prentice Hall, 1988.
28. Omura, J. K. and B. K. Levitt, “Coded Error Probability Evaluation for Antijam Communication
Systems,” JEEE Trans. Commun., vol. COM-30, no. 5, May 1982, pp. 896-903.
: ii Hae 7 ie
&
—
Index
———<§$§| ———— eh Ss
for independent binary symbol input, 75-80 Generalized M-ary symbol error probability bound,
Encoder state, 827 395-97
Equalization, 556 Global Positioning System (GPS), 15-16
techniques, 600-611 GlobalStar, 18, 24, 29
decision-feedback equalization, 605-11 Golay codes, 768, 799
mean-squared error linear equalizer (LE-MSE), error probability performance, 805
603-5 extended, 805
zero-forcing linear equalizer (LE-ZF), 601, Good error control code, and minimum Hamming
602-3 distance, 802-3
Equal a priori probabilities, binary communication Gram-Schmidt orthogonalization procedure, 154-56
with, 126 Gray code mapping, 210, 211, 214, 349
Equivalent noise bandwidth, 29-32 Group delay, 558
Equivalent Nyquist channel characteristic, 562
Erasure-and-error decoders, 804 H
Error correction/detection codes:
classification of, 744—47 Hamming codes, error probability performance,
Venn diagram for, 747 804-5
Error probability performance: Hamming distance, 759-60, 763-64, 769, 799,
coherent receivers, 186—221 800-802, 843, 847
convolutional codes, 854-56 and good error control code, 802-3
maximum-likelihood sequence estimation Hard-decision bounded distance decoder, 803
(MLSE), 591-600 Hard-decision demodulators, 733, 838
scalar communications, 113-15 binary coded BPSK, design of, 790-91
vector communications, 128-31 Hard-decision/soft-decision Viterbi decoding,
Excess bandwidth factor, 563 843-47
Exponential error bound, 765 Helical interleaver, 738-39, 741
Extended Binary Coded Decimal Interchange Code Hybrid interleaver, 741
(EBCDIC), 13 Hypothesis testing, 98, 100-102
Extended Golay code, 805
Eye pattern, 567-68
FE
Imperfect bit synchronization, 268-71
Imperfect carrier synchronizaiton, 267-68
False alarm, 145 Infinite bandwidth Shannon capacity limit, 783
Fast Fourier Transform (FFT), 321 Infinite bandwidth Shannon information transport,
Feedback filter, 605 783
Finite memory, 280 Information bits, 716
First-order phase difference modulation (PDM-1), Information symbols, 800
509-10 Information transportation system:
Forward filter, 605 basic building blocks of, 2-3
Fredholm equation of the first kind, 257 defined, 2
Fredholm equation of the second kind, 257 Information Transportation System (ITS) Pyramid, 7,
Frequency bands, electromagnetic spectrum, 33 9-11
Full response CPFM, 652, 698 Inmarsat, 18, 19-20, 24, 25
Full response CPM, 89 Inner code, 743
concatenated code system, 857-58
G Inphase-quadrature (I-Q) modulation and
#
demodulation, 261-84, 621° ++: nw. ae
Gabor bandwidth, 29 data asymmetry, effect of, 271-77 "pes
Gallager’s upper bound on ensemble error imperfect bit synchronization, effect of, 268-71 a
probability, 786-89 imperfect carrier synchronization, effect of,
very noisy channels, block vs. convolutional 267-68
coding for, 788-89 real and complex signal models, 261-66
Gamma pdf, 385 suboptimum detection, 277-84
Gaussian integral Q(x), 147-53 See also I-Q demodulation; I-Q modulations
series approximation for, 147-53 Integrated Services Digital Network (ISDN), 16
Generalized M-ary Markov sources: Interface control documents (ICDs), 3
power spectral density of, 63-68 Interleaver, 7
biphase (Manchester) baseband signaling, 66 Interleaver span, 738
delay modulation or Miller coding, 66-68 Interleaving-deinterleaving, 736-41
NRZ baseband signaling, 65 concatenated interleaver-deinterleaver systems,
RZ baseband signaling, 65 741
884 Index
“at
¢
~*~
2.*
<"
os
io ’ .
+
Information Information
Channel
Source Sink
Figure 1.6 Information Transportation System (ITS) pyramid; coded digital communication system functional architecture
Information
Sink
Source Information
Decoder Sink
Source Information s xe
Signal. Channel
Sink
ee Rog
Decoder Decoder
Detector
he ae “>
0
yy Rey
Source Informatio n ae? ee
Demodulator Signal Channel
Decoder Decoder Sink
i
Receiver Detector
2)
oe
——-
DIGIT
COMMUNICATION
TECHNIQUES
Signal Design and Detection
Marvin K. Simon »Sami M. Hinedi = William C. Lindsey
The global transition from analog to digital communication infrastructures makes it necessary
for practicing engineers, professors, and students to understand the theory associated with
Digital Communication Techniques. This timely volume, written by professionals with over 60
years of combined teaching, research, and product development experience, provides an
in-depth state-of-the-art treatment of the important concepts of signal design and detection,
modulation and demodulation, coding and decoding, and synchronization written from the
systems architecting, engineering, analysis, and simulation perspectives.
Digital Communication Techniques presents the methods of coherent, noncoherent, partially coherent,
and differentially coherent detection for various binary and M-ary coded and uncoded modulation
techniques (BPSK, QPSK, MSK, MFSK, QASK, QAM, etc.) that can be applied to the design of
uncoded systems while the Ro-criterion (channel throughput in bits/sec/Hz) and bit error
probability are used as the methodology for designing efficient coded digital communications.
Key concepts include scalar, vector, and waveform communications over the additive white
and nonwhite (colored) Gaussian noise channels. In addition, communications via Rayleigh
and Rician fading (due to multipath) channels are also covered.
ISBN O-13-200610-3
* PRENTICE HALL | | | (|
9'780132/006101 ||
- Englewood Cliffs, NJ 07632