ECE 2321 Signals and Systems Lecture Notes v2
ECE 2321 Signals and Systems Lecture Notes v2
First Edition
Martin Wafula
Multimedia University of Kenya
2
PREFACE
These notes for ECE 2321: Signals and Systems are designed to bridge your existing
knowledge of differential equations (from MTE 2220) with the fundamental tools used to
analyze and understand both continuous and discrete signals. Here, you will learn how to
apply Fourier and Laplace transforms to represent and examine signals and how to use these
techniques to study and design systems that handle these signals.
We will begin by classifying signals and systems, exploring the time and frequency domains,
and identifying the characteristics that distinguish linear, time-invariant systems from others.
In moving from the basics to more advanced methods, you will see how mathematical concepts
translate into practical engineering approaches—such as understanding how a given input
affects a system or how to synthesize a system that meets certain performance requirements.
To reinforce theory, laboratory exercises and MATLAB and Python simulations will pro-
vide hands-on experience. These activities will let you visualize signal properties, practice
sampling procedures, and investigate the effects of system parameters, ultimately giving you
a well-rounded skill set that combines conceptual understanding with practical know-how.
3
4
CONTENTS
Preface 3
Contents 5
List of Figures 9
List of Tables 11
5
6 CONTENTS
3.5 Conjugation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
3.6 Time-Reversal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
3.7 Differentiation in Frequency . . . . . . . . . . . . . . . . . . . . . . . 94
3.8 Parseval’s Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
3.9 Convolution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
3.10 Multiplication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
7 Sampling 99
1 The Sampling Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
2 Reconstruction of a Signal From Its Samples . . . . . . . . . . . . . . . . . . 101
2.1 Zero-Order Hold . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
2.2 First-Order Hold . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
3 Undersampling and Aliasing . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
4 Discrete-Time Processing of Continuous-Time Signals . . . . . . . . . . . . . 103
8 The Laplace Transform 107
1 The Laplace Transform . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108
2 The Region of Convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . 110
3 The Inverse Laplace Transform . . . . . . . . . . . . . . . . . . . . . . . . . 112
3.1 Partial Fraction Expansion . . . . . . . . . . . . . . . . . . . . . . . . 113
4 Some Properties of the Laplace Transform . . . . . . . . . . . . . . . . . . . 115
4.1 Convolution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115
4.2 Differentiation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115
4.3 Integration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116
5 Finding the Output of an LTI System via Laplace Transforms . . . . . . . . 116
6 Finding the Impulse Response of a Differential Equation via Laplace Transforms117
7.1 The frequency spectrum of the signal x(t) and the signal xp (t). . . . . . . . . 101
7.2 The impulse response of a zero-order-hold (left) and a first-orderhold (right). 102
7.3 Sampling x(t) = cos(t) at a frequency of ωs = 4. . . . . . . . . . . . . . . . . 103
7.4 The frequency response of a bandlimited differentiator. . . . . . . . . . . . . 105
1.1 The graph convolution as a shift register. Highlighted are the nodes that reach
node 1 on each consecutive shift; that is, the nodes j whose signal value xj
contributes to S x i . The resulting summary of each communication Sk x is
k
9
10 List of Figures
LIST OF TABLES
11
12 List of Tables
Part I
13
CHAPTER 1
INTRODUCTION TO SIGNALS AND
SYSTEMS
• Chemical and biological systems: concentrations of cells and reactants, neuronal activ-
ity, cardiac signals, ...
• Economic systems: stock prices, unemployment rate, tax rate, interest rate, GDP, ...
15
16 Introduction to Signals and Systems
• A continuous-time signal is a function of the form f (t), where t ranges over all real
numbers (i.e., t ∈ R ).
• A discrete-time signal is a function of the form f [n], where n takes on only a discrete
set of values (e.g., n ∈ Z ).
Note that we use square brackets to denote discrete-time signals, and round brackets to
denote continuous-time signals. Examples of continuous-time signals often include physical
quantities, such as electrical currents, atmospheric concentrations and phenomena, vehicle
movements, etc. Examples of discrete time signals include the closing prices of stocks at the
end of each day, population demographics as measured by census studies, and the sequence of
frames in a digital video. One can obtain discrete-time signals by sampling continuous-time
signals (i.e., by selecting only the values of the continuous-time signal at certain intervals).
Just as with signals, we can consider continuous-time systems and discrete time systems.
Examples of the former include atmospheric, physical, electrical and biological systems, where
the quantities of interest change continuously over time. Examples of discrete-time systems
include communication and computing systems, where transmissions or operations are per-
formed in scheduled time slots. With the advent of ubiquitous sensors and computing tech-
nology, the last few decades have seen a move towards hybrid systems consisting of both
continuous-time and discrete-time subsystems - for example, digital controllers and actu-
ators interacting with physical processes and infrastructure. We will not delve into such
hybrid systems in this course, but will instead focus on systems that are entirely either in
the continuous-time or discrete-time domain.
The term dynamical system loosely refers to any system that has an internal state and
some dynamics (i.e., a rule specifying how the state evolves in time).
This description applies to a very large class of systems, including individual vehicles,
biological, economic and social systems, industrial manufacturing plants, electrical power
grid, the state of a computer system, etc. The presence of dynamics implies that the behavior
of the system cannot be entirely arbitrary; the temporal behavior of the system’s state and
outputs can be predicted to some extent by an appropriate model of the system.
Example : 1
Consider a simple model of a car in motion. Let the speed of the car at any time t be
Introduction to Signals and Systems 17
given by v(t). One of the inputs to the system is the acceleration a(t), applied by the
throttle. From basic physics, the evolution of the speed is given by
dv
= a(t)
dt
The quantity v(t) is the state of the system, and equation (1.1) specifies the dynamics.
There is a speedometer on the car, which is a sensor that measures the speed. The value
provided by the sensor is denoted by s(t) = v(t), and this is taken to be the output of the
system. ■
Much of scientific and engineering endeavor relies on gathering, manipulating and un-
derstanding signals and systems across various domains. For example, in communication
systems, the signal represents voice or data that must be transmitted from one location to
another. These information signals are often corrupted en route by other noise signals, and
thus the received signal must be processed in order to recover the original transmission. Sim-
ilarly, social, physical and economic signals are of great value in trying to predict the current
and future state of the underlying systems. The field of signal processing studies how to
take given signals and extract desirable features from them, often via the design of systems
known as filters. The field of control systems focuses on designing certain systems (known
as controllers) that measure the signals coming from a given system and apply other input
signals in order to make the given system behave in an desirable manner. Typically, this is
done via a feedback loop of the form
Example : Inverted Pendulum
Suppose we try to balance a stick vertically in the palm of our hand. The sensor, controller
and actuator in this example are our eyes, our brain, and our hand, respectively, which
communicate using signals of various forms. This is an example of a feedback control
system. ■
• Signal and systems classifications: develop terminology and identify useful properties
of signals and systems
• Time domain analysis of LTI systems: understand how the output of linear time-
invariant systems is related to the input
• Sampling and Quantization: study ways to convert continuous-time signals into discrete-
time signals, along with associated challenges
The material in this course will lay the foundations for future courses in signal processing
(ECE438, 445), Digital Communications (ECE 2414), Digital Filters Design (ECE 2413),
control theory (ECE 2415, EEE 2422) and Digital Image Processing (ECE 2528).
CHAPTER 2
PROPERTIES OF SIGNALS AND SYSTEMS
19
20 Properties of Signals and Systems
v 2 (t)
p(t) = v(t)i(t) = i2 (t)R =
R
Thus the power is a scaled multiple of the square of the voltage and current signals.
Since the energy expended over a time-interval [t1 , t2 ] is given by the integral of the power
dissipated per-unit-time over that interval, we have
Z t2 Z t2 Z t2
2 1
E= p(t)dt = R i (t)dt = v 2 (t)dt
t1 t1 R t1
We will find it useful to discuss the energy and average power of any continuoustime
or discrete-time signal. In particular, the energy of a general (potentially complex-valued)
continuous-time signal f (t) over a time-interval [t1 , t2 ] is defined as
Z t2
E[t1 ,t2 ] ≜ |f (t)|2 dt
t1
Note that we are defining the energy of an arbitrary signal in the above way; this will end
up being a convenient way to measure the "size" of a signal, and may not actually correspond
to any physical notion of energy.
We will also often be interested in measuring the energy of a given signal over all time. In
this case, we define
Z ∞
E∞ ≜ |f (t)|2 dt
−∞
for discrete-time signals. Note that the quantity E∞ may not be finite.
Similarly, we define the average power of a continuous-time signal as
Z T
1
P∞ ≜ lim |f (t)|2 dt
T →∞ 2T −T
Based on the above definitions, we have three classes of signals: finite energy (E∞ < ∞),
finite average power (P∞ < ∞), and those that have neither finite energy nor average power.
An example of the first class is the signal f (t) = e−t for t ≥ 0 and f (t) = 0 for t < 0. An
example of the second class is f (t) = 1 for all t ∈ R. An example of the third class is f (t) = t
for t ≥ 0. Note that any signal that has finite energy will also have finite average power,
since
E∞
P∞ = lim =0
T →∞ 2T
for continuous-time signals with finite energy, with an analogous characterization for
discrete-time signals.
2 Transformations of Signals
Throughout the course, we will be interested in manipulating and transforming signals into
other forms. Here, we start by considering some very simple transformations involving the
time variable. For the purposes of introducing these transformations, we will consider a
continuous-time signal f (t) and a discretetime signal f [n].
Time-shifting: Suppose we define another signal g(t) = f (t − t0 ), where t0 ∈ R. In
other words, for every t ∈ R, the value of the signal g(t) at time t is the value of the signal
f (t) at time t − t0 . If t0 > 0, then g(t) is a "forward-shifted" (or time-delayed) version of
f (t), and if t0 < 0, then g(t) is a time-advanced version of f (t) (i.e., the features in f (t)
appear earlier in time in g(t)). Similarly, for a discrete-time signal f [n], one can define the
time-shifted signal f [n − n0 ], where n0 is some integer.
Time-reversal: Consider the signal g(t) = f (−t). This represents a reversal of the
function f (t) in time. Similarly, f [−n] represents a time-reversed version of the signal f [n].
Time-scaling: Define the signal g(t) = f (αt), where α is some real number. When
0 < α < 1, this represents a stretching of f (t), and when α > 1, this represents a compression
of f (t). If α < 0, we get a time-reversed and stretched (or compressed) version of f (t).
Analogous definitions hold for the discrete-time signal f [n].
The operations above can be combined to define signals of the form g(t) = f (αt+β), where
α and β are real numbers. To draw the signal g(t), we should first apply the time-shift by β
to f (t) and then apply the scaling α. To see why, define h(t) = f (t + β), and g(t) = h(αt).
Thus, we have g(t) = f (αt+β) as required. If we applied the operations in the other order, we
22 Properties of Signals and Systems
would first get the signal h(t) = f (αt), and then g(t) = h(t + β) = f (α(t + β)) = f (αt + αβ).
In other words, the shift would be by αβ rather than β.
Examples of these operations can be found in the textbook (OW), such as example 1.1.
f (t) = Ceat
where C and a are complex numbers. If both C and a are real, there are three possible
behaviors for this signal. If a < 0, then the signal goes to zero as t → ∞, and if a > 0, the
signal goes to ∞ as t → ∞. For a = 0, the signal is constant.
Now suppose f (t) = ej(ω0 t+ϕ) for some positive real number ω0 and real number ϕ (this
corresponds to C = ejϕ and a = jω0 in the signal given above). We first note that
ϕk (t) = ejkω0 t , k ∈ Z
In other words, it is the set of complex exponentials whose frequencies are multiples of the
fundamental frequency ω0 . Note that if ejω0 t is periodic with period T0 (i.e, ω0 T0 = 2πm for
some integer m ), then ϕk (t) is also periodic with period T0 for any k ∈ Z, since
f [n] = Cean
where C and a are general complex numbers. As before, let us focus on the case where
C = 1 and a = jω0 for some ω0 ∈ R in order to gain some intuition, i.e., f [n] = ejω0 n .
24 Properties of Signals and Systems
To see the differences in discrete-time signals from continuous-time signals, recall that a
continuous-time complex exponential is always periodic for any ω0 . The first difference
between discrete-time complex exponentials and continuous time complex exponentials is
that discrete-time complex exponentials are not necessarily periodic. Specifically, consider
the signal f [n] = ejω0 n , and suppose it is periodic with some period N0 . Then by definition,
it must be the case that
Due to periodicity, we must have ejω0 N0 = 1, or equivalently, ω0 N0 = 2πk for some integer
k. However, N0 must be an integer, and thus we see that this can be satisfied if and only
of ω0 is a rational multiple of 2π. In other words, only discrete-time complex exponentials
whose frequencies are of the form
k
ω0 = 2π
N
for some integers k and N are periodic. The fundamental period N0 of a signal is the
smallest nonnegative integer for which the signal is periodic. Thus, for discrete-time complex
exponentials, we find the fundamental period by first writing
ω0 k
=
2π N
where k and N have no factors in common. The value of N in this representation is then
the fundamental period.
2π 3π
Example : Consider the signal f [n] = ej 3 n +ej 4 n . Since both of the exponentials
have frequencies that are rational multiples of 2π, they are both periodic. For
the first exponential, we have
2π
3 1
=
2π 3
which cannot be reduced any further. Thus the fundamental period of the first
exponential is 3 . Similarly, for the second exponential, we have
3π
4 3
=
2π 8
Thus the fundamental period of the second exponential is 8 . Thus f [n] is
periodic with period 24 (the least common multiple of the periods of the two
signals).
■
The same reasoning applies to sinusoids of the form f [n] = cos (ω0 n). A necessary condi-
tion for this function to be periodic is that there are two positive integers n1 , n2 with n2 > n1
such that f [n1 ] = f [n2 ]. This is equivalent to cos (ω0 n1 ) = cos (ω0 n2 ). Thus, we must either
have
Properties of Signals and Systems 25
ω0 n2 = ω0 n1 + 2πk
or
ω0 n2 = −ω0 n1 + 2πk
for some positive integer k. In either case, we see that ω0 has to be a rational multiple of
2π. In fact, when ω0 is not a rational multiple of 2π, the function cos (ω0 n) never takes the
same value twice for positive values of n.
The second difference from continuous-time complex exponentials pertains to the period of
oscillation. Specifically, even for periodic discrete-time complex
exponentials, increasing the frequency does not necessarily make the period smaller. Consider
the signal g[n] = ej(ω0 +2π)n , i.e., a complex exponential with frequency ω0 + 2π. We have
The following table shows the differences between continuous-time and discretetime signals.
01
As we will see later in the course, the signals cos (ω0 n) and sin (ω0 n) correspond to continuous-time
signals of the form cos (ω0 t) and sin (ω0 t) that are sampled at 1 Hz . When 0 ≤ ω0 < π, this sampling rate
is above the Nyquist frequency ωπ0 , and thus the sampled signals will be an accurate representation of the
26 Properties of Signals and Systems
ejω0 t ejω0 n
Distinct signals for different values of ω0 Identical signals for values of ω0 sepa-rated by 2π
Periodic only if ω0 = 2π Nk for some in-
Periodic for any ω0
tegers k and N > 0.
Fundamental period undefined for ω0 =
Fundamental period: undefined for
ω0 = 0 and ω2π0 otherwise
0 and k ω2π0 otherwise
Fundamental frequency ω0 Fundamental frequency ω0
k
As with continuous-time signals, for any period N , we define the harmonic family of
discrete-time complex exponentials as
2π
ϕk [n] = ejk N n , k∈Z
This is the set of all discrete-time complex exponentials that have a common period N ,
and frequencies whose multiples of 2π N
. This family will play a role in our analysis later in
the course.
0 if n ̸= 0
δ[n] = .
1 if n = 0
The discrete-time unit step signal is defined as
(
0 if n < 0
u[n] =
1 if n ≥ 0
Note that by the time-shifting property, we have
In other words, the unit step function can be viewed as a superposition of shifted impulse
functions.
Suppose we are given some arbitrary signal f [n]. If we multiply f [n] by the time-shifted
underlying continuous-time signal.
2
Note that cos (ω0 n) = cos ((2π − ω0 ) n ) for any 0 ≤ ω0 ≤ π. The same is not true for sin (ω0 n). In fact,
one can show that for any two different frequencies 0 ≤ ω0 < ω1 ≤ 2π, sin (ω0 n) and sin (ω1 n) are different
functions.
Properties of Signals and Systems 27
impulse function δ[n − k], we get a signal that is zero everywhere except at n = k, where
it takes the value f [k]. This is known as the sampling or sifting property of the impulse
function:
f [n]δ[n − k] = f [k]δ[n − k]
More generally, for any signal f [n], we have
∞
X
f [n] = f [k]δ[n − k]
k=−∞
i.e., any function f [n] can be written as a sum of scaled and shifted impulse functions.
5.2 Continuous-Time
The continuous-time unit step function is defined by
(
0 if t < 0
u(t) =
1 if t ≥ 0
Note that u(t) is discontinuous at t = 0 (we will take it to be continuous from the right).
To define the continuous-time analog of the discrete-time impulse function, we first define
the signal
0 if t < 0
δϵ (t) = 1ϵ if 0 ≤ t ≤ ϵ
0 if t > ϵ
R∞
where ϵ ∈ R>0 . Note that for any ϵ > 0, we have −∞ δϵ (t)dt = 1. As ϵ gets smaller, the
width of this function gets smaller and the height increases proportionally. The continuous-
time impulse function is defined as the limit of the above function as ϵ approaches zero from
the right:
This function is drawn with an arrow at the origin (since it has no width and infinite
height). We will often be interested in working with scaled and timeshifted versions of the
continuous-time impulse function. Just as we did with discrete-time functions, we can take
a continuous-time function f (t) and represent it as
Z ∞
f (t) = f (τ )δ(t − τ )dτ
−∞
In other words, if we take an infinite sequence of shifted impulse functions, scale each of
them by the value of the function f (t) at the value of the time-shift, and add them together
(represented by the integration), we get the function f (t). For instance, we have
28 Properties of Signals and Systems
Z ∞ Z ∞
u(t) = u(τ )δ(t − τ )dτ = δ(t − τ )dτ
−∞ 0
Just as the discrete-time impulse function could be viewed as a difference of the discrete-
time unit step and its time-shifted version, the continuous-time impulse function can be
viewed as the derivative of the continuous-time unit step function.
6 Properties of Systems
As we discussed during the first lecture, a system can be viewed as an abstract object
that takes input signals and produces output signals. A continuous-time system operates
on continuous-time signals, and discrete-time systems operate with discrete-time signals.
Examples of the former include many physical systems such as electrical circuits, vehicles,
etc. Examples of the latter include computer systems, a bank account where the amount of
money is incremented with interest, deposits and withdrawals at the end of each day, etc.
y[n] = cos(x[n])
is memoryless, as the output at each time-step n ∈ Z only depends on the input at that
time-step.
However, the system whose input and output are related by
Z t
y(t) = x(τ )dτ
−∞
is not memoryless, as the output depends on all of the input values from the past.
Systems with memory are often represented as having some sort of state and dynamics, which
maintains the necessary information from the past. For example, for the system given above,
we can use the fundamental theorem of calculus to obtain
dy(t)
= x(t)
dt
30 Properties of Signals and Systems
where the state of the system is y(t) (this is also the output), and the dynamics of the
state are given by the differential equation above. Similarly for the
discrete-time system
n
X
y[n] = x[k]
k=−∞
we have
n−1
X
y[n] = x[k] + x[n] = y[n − 1] + x[n]
k=−∞
which is a difference equation describing how the state y[n] evolves over time.
Invertibility
A system is said to be invertible if distinct inputs lead to distinct outputs. In other words,
by looking at the output of a system, one can uniquely identify what the input was.
An example of an invertible system is y(t) = αx(t), where α is any nonzero real number.
Given the output y(t), we can uniquely identify the input as x(t) = α1 y(t). However, if α = 0,
then we have y(t) = 0 regardless of the input, and there is no way to recover the input. In
that case, the system would not be invertible.
Another example of a noninvertible system is y(t) = x2 (t), as the sign of the input is lost
when converting toP the output.
The system y[n] = nk=−∞ x[k] is invertible; to see this, we use the equivalent representation
y[n] = y[n − 1] + x[n] to obtain x[n] = y[n] − y[n − 1] for all n ∈ Z.
Causality
A system is causal if the output of the system at any time depends only on the input at that
time and from the past. In other words, for all t ∈ R, y(t) depends only on x(τ ) for τ ≤ t.
Thus, a causal system does not react to inputs that will happen in the future. For a causal
system, if two different inputs have the same values up to a certain time, the output of the
system due to those two inputs will agree up to that time as well. All memoryless systems
are causal.
There are various instances where we may wish to use noncausal systems. For example, if
we have time-series data saved offline, we can use the saved values of the signal for k > n to
process the signal at a given time-step n (this can be used for music and video editing, for
example). Alternatively, the independent variable may represent space, rather than time. In
this case, one can use the values of the signal from points on either side of a given point in
order to process the signal.
Example :
The signal y[n] = x[−n] is noncausal; for example, y[−1] = x[1], and thus the output at
negative time-steps depends on the input from positive time-steps (i.e., in the future).
The signal y(t) = x(t) cos(t + 1) is causal; the t + 1 term does not appear in the input,
and thus the output at any time does not depend on values of the input at future times.
Properties of Signals and Systems 31
Stability
The notion of stability is a critical system property. There are many different notions of
stability that can be considered, but for the purposes of this course, we will say that a
system is stable if a bounded input always leads to a bounded output. In other words, for
a continuous-time system, if there exists a constant B1 ∈ R≥0 such that the input satisfies
|x(t)| ≤ B1 for all t ∈ R, then there should exist some other constant B2 ∈ R≥0 such that
|y(t)| ≤ B2 for all t ∈ R. An entirely analogous definition holds for discrete-time systems.
Loosely speaking, for a stable system, the output cannot grow indefinitely when the input is
bounded by a certain value.
Example :
The system y(t) = tx(t) is memoryless and causal, but not stable. For example, if x(t) = 1
for all t ∈ R, we have y(t) = t which is not bounded by any constant.
Similarly,
Pn the system y[n] = y[n − 1] + x[n] is not stable. This is seen by noting that
y[n] = k=−∞ x[k]. So, for example, if x[n] = u[n], we have y[n] = (n + 1) if n ≥ 0, which
is unbounded.
An example of a stable causal memoryless system is y(t) = cos(x(t)). Another example of
a stable and causal system is
(
0 if n < 0
y[n] =
αy[n − 1] + x[n] if n ≥ 0
where α ∈ R satisfies |α| < 1. Specifically, if |x[n]| ≤ B1 for all n ∈ Z, then we have
B1
|y[n]| ≤ 1−|α| for all n ∈ Z. To see this, we prove by induction. Clearly |y[n]| ≤ 1−|α|
B1
for
n ≤ 0. Suppose that |y[n]| ≤ 1−|α| for some n ≥ 0. Then we have
B1
Example :
The system y(t) = cos(x(t)) is time-invariant. Suppose we define the input signal w(t) =
x (t − t0 ) (i.e., a time-shifted version of x(t) ). Let yw (t) be the output of the system when
w(t) is applied. Then we have
Linearity
A system is linear if it satisfies the following two properties.
1. Additivity: Suppose the output is y1 (t) when the input is x1 (t), and the output is
y2 (t) when the input is x2 (t). Then the output to x1 (t) + x2 (t) is y1 (t) + y2 (t).
2. Scaling: Suppose the output is y(t) when the input is x(t). Then for any complex
number α, the output should be αy(t) when the input is αx(t).
Both properties together define the superposition property: if the input to the system is
α1 x1 (t) + α2 x2 (t), then the output should be α1 y1 (t) + α2 y2 (t). Note that this must hold for
any inputs and scaling parameters in order for the system to qualify as linear. An entirely
analogous definition holds for discrete-time systems.
For any linear system, the output must be zero for all time when the input is zero for all
time. To see this, consider any arbitrary input x(t), and let the corresponding output be
y(t). Then, using the scaling property, the output to
αx(t) must be αy(t) for any scalar complex number α. Simply choosing α = 0 yields the
desired result that the output will be the zero signal when the input is the zero signal.
Example :
The system y(t) = tx(t) is linear. To see this, consider two arbitrary input signals x1 (t)
and x2 (t), and two arbitrary scalars α1 , α2 . Then we have
y3 [n] = x23 [n] = (x1 [n] + x2 [n])2 ̸= x21 [n] + x22 [n]
in general. Thus the additivity property does not hold, and the system is nonlinear.
The system y[n] = Re{x[n]} is nonlinear, where Re{·} denotes the real part of the
argument. To see this, let x[n] = a[n] + jb[n], where a[n] and b[n] are real-valued signals.
Consider a scalar α = j. Then we have
Properties of Signals and Systems 33
35
36 Analysis of Linear Time-Invariant Systems
i.e., x[n] can be written as a superposition of scaled and shifted impulse functions.
Since the system is time-invariant, the response of the system to the input δ[t − k] is
h[t − k]. By linearity (and specifically the scaling property),
P the response to x[k]δ[n − k] is
x[k]δ[n − k]. By the additivity property, the response to ∞ k=−∞ x[k]δ[n − k] is then
∞
X
y[n] = x[k]h[n − k]
k=−∞
The above is called the convolution sum; the convolution of the signals x[n] and h[n] is
denoted by
∞
X
x[n] ∗ h[n] = x[k]h[n − k]
k=−∞
Thus we have the following very important property of discrete-time LTI systems: if x[n]
is the input signal to an LTI system, and h[n] is the impulse response of the system, then
the output of the system is y[n] = x[n] ∗ h[n].
Example :
Consider an LTI system with impulse response
(
1 if 0 ≤ n ≤ 3
h[n] =
0 otherwise
Suppose the input signal is
(
1 if 0 ≤ n ≤ 3
x[n] =
0 otherwise
Then we have
∞
X
y[n] = x[n] ∗ h[n] = x[k]h[n − k]
k=−∞
n
X
y[n] = x[k]h[n − k]
k=0
When n = 1 we have
1
X
y[1] = x[k]h[1 − k] = x[0]h[1] + x[1]h[0] = 2.
k=0
Since both x[k] = 0 for k < 0 and h[n − k] = 0 for k > n, we have
n n
X 1 − αn+1
X
k
y[n] = x[k]h[n − k] = α =
k=0 k=0
1−α
for n ≥ 0, and y[n] = 0 for n < 0. ■
The expression on the right hand side is a superposition of scaled and shifted impulse
functions. Thus, when this signal is applied to an LTI system, the output will be a superpo-
sition of scaled and shifted impulse responses. More specifically, if h(t) is the output of the
system when the input is x(t) = δ(t), then the output for a general input x(t) is given by
Z ∞
y(t) = x(τ )h(t − τ )dτ
−∞
Example :
Suppose x(t) = e−at u(t) with a ∈ R>0 and h(t) = u(t). Then the output of the LTI system
with impulse response h(t) is given by
Z ∞
y(t) = x(t) ∗ h(t) = x(τ )h(t − τ )dτ
−∞
Z ∞
= e−aτ u(τ )u(t − τ )dτ
−∞
Z t
= e−aτ dτ
0
To see this, start with the definition of convolution and perform a change of variable by
setting r = n − k. This gives
∞
X ∞
X
x[n] ∗ h[n] = x[k]h[n − k] = x[n − r]h[r] = h[n] ∗ x[n]
k=−∞ r=−∞
The same holds for the continuous-time convolution. Thus it does not matter which of
the signals we choose to flip and shift in the convolution operation.
Let h1 [n] be the impulse response of System 1 , and let h2 [n] be the impulse response for
System 2. Then we have y1 [n] = x[n] ∗ h1 [n] and y2 [n] = x[n] ∗ h2 [n]. Thus,
y[n] = y1 [n] + y2 [n] = x[ n ∗ h1 [n] + x[n] ∗ h2 [n] = x[n] ∗ (h1 [n] + h2 [n])
The above expression indicates that the parallel interconnection can equivalently be viewed
as x[n] passing through a single system whose impulse response is h1 [n] + h2 [n] :
40 Analysis of Linear Time-Invariant Systems
In other words, it does not matter which order we do the convolutions. The above rela-
tionships can be proved by manipulating the summations (or integrals); we won’t go into the
details here.
Just as the distributive property had implications for parallel interconnections of systems,
the associative property has implications for series interconnections of systems. Specifically,
consider the series interconnection shown in Fig. 3.1.
We have
y[n] = y1 [n] ∗ h2 [n] = (x[n] ∗ h1 [n]) ∗ h2 [n] = x[n] ∗ (h1 [n] ∗ h2 [n])
Thus, the series interconnection is equivalent to a single system with impulse response
h1 [n] ∗ h2 [n], as shown in Fig. 3.2.
Further note that since h1 [n] ∗ h2 [n] = h2 [n] ∗ h1 [n], we can also interchange the order of
the systems in the series interconnection as shown in Fig. 3.3, without changing the overall
input-output relationship between x[n] and y[n].
Analysis of Linear Time-Invariant Systems 41
Figure 3.3: An equivalent series representation of the interconnection shown in Fig. 3.1
we see that y[n] will depend on a value of the input signal other than at timestep n unless
h[k] = 0 for all k ̸= 0. In other words, for an LTI system to be memoryless, we require
h[n] = Kδ[n] for some constant K. Similarly, a continuous-time LTI system is memoryless
if and only if h(t) = Kδ(t) for some constant K. In both cases, all LTI memoryless systems
have the form
In other words, if we have an LTI system with impulse response h[n], and another LTI
system with impulse response hI [n] such that h[n] ∗ hI [n] = δ[n], then those systems are
inverses of each other. The analogous statement holds in continuous-time as well.
Example :
Consider the LTI system with impulse response h[n] = αn u[n]. One can verify that this
impulse response corresponds to the system
n
X
y[n] = x[k]αn−k = αy[n − 1] + x[n]
k=−∞
Now consider the system yI [n] = xI [n] − αxI [n − 1], with input signal xI [n] and output
signal yI [n]. The impulse response of this system is hI [n] = δ[n]−
αδ[n − 1]. We have
Recall that a system is causal if its output at time t depends only on the inputs up to (and
potentially including) t. To see what this means for LTI systems, consider the convolution
sum
∞
X ∞
X
y[n] = x[k]h[n − k] = x[n − k]h[k]
k=−∞ k=−∞
where the second expression follows from the commutative property of the convolution. In
order for y[n] to not depend on x[n + 1], x[n + 2], . . ., we see that h[k] must be zero for k < 0.
The same conclusion holds for continuous-time systems, and thus we have the following: A
continuous-time LTI system is causal if and only if its impulse response h(t) is zero for all
t < 0. A discrete-time LTI system is causal if and only if its impulse response h[n] is zero for
all n < 0.
Note that causality is a property of a system; however we will sometimes refer to a signal
as being causal, by which we simply mean that its value is zero for n or t less than zero.
Analysis of Linear Time-Invariant Systems 43
∞
X
y[n] = x[k]h[n − k]
k=−∞
Note that
∞
X ∞
X
|y[n]| = x[k]h[n − k] ≤ |x[k]h[n − k]|
k=−∞ k=−∞
X∞
= |x[k]||h[n − k]|
k=−∞
Now suppose that x[n] is bounded, i.e., there exists some B ∈ R≥0 such that |x[n]| ≤ B
for all n ∈ Z. Then the above expression becomes
∞
X
|y[n]| ≤ B |h[n − k]|
k=−∞
Thus, if ∞k=−∞ |h[n − k]| < ∞ (which means that h[n] is absolutely summable), then
P
|y[n]| will
P∞also be bounded for all n. It turns out that this is a necessary condition as
well: if k=−∞ |h[n − k]| = ∞, then there is a bounded input that causes the output to be
unbounded.
The same conclusion holds in continuous-time
R∞ as well. Thus, we have: A continuous-time
LTI system isP stable if and only if −∞ |h(τ )|dτ < ∞. A discrete-time LTI system is stable
if and only if ∞ k=−∞ |h[k]| < ∞.
Example :
Consider the LTI system with impulse response h[n] = αn u[n], where α ∈ R. We have
∞ ∞
(
X X 1
if |α| < 1
|h[k]| = |α|k = 1−|α|
k=−∞ k=0
∞ if |α| ≥ 1
Thus, the system is stable if and only if |α| < 1.
Similarly, consider the continuous-time LTI system with impulse response h(t) = eαt u(t),
where α ∈ R. We have
(
∞
− α1 if α < 0
Z ∞ Z ∞
ατ 1 ατ
|h(τ )|dτ = e dτ = (e ) =
−∞ 0 α 0 ∞ if α ≥ 0
Thus, the system is stable if and only if α < 0. ■
44 Analysis of Linear Time-Invariant Systems
yh (t) = Aemt
for some m ∈ C. Substituting this into the homogeneous equation gives
mAemt + 2Aemt = 0 ⇒ m + 2 = 0 ⇒ m = −2
Thus, the homogeneous solution is yh (t) = Ae−2t , for any constant A.
Next, we search for a particular solution to the equation
dyp
+ 2yp (t) = Ke3t u(t)
dt
It seems reasonable to try yp (t) = Be3t , for some constant B. Substituting and evalu-
ating for t > 0, we have
46 Analysis of Linear Time-Invariant Systems
K
3Be3t + 2Be3t = Ke3t ⇒ B =
5
Thus, the particular solution is given by yp (t) = 5 e for t > 0.
K 3t
Together, we have y(t) = yh (t) + yp (t) = Ae−2t + K5 e3t for t > 0. Note that the coefficient
A has not been determined yet; in order to do so, we need more information about the
solutions to the differential equation, typically in the form of initial conditions. For exam-
ple, if we know that the system is at rest until the input is applied (i.e., y(t) = 0 until x(t)
becomes nonzero), we have y(t) = 0 for t < 0. Suppose we are given the initial condition
y(0) = 0. Then,
K K
y(0) = A + =0⇒A=−
5 5
Thus, with the given initial condition, we have y(t) = 5 (e3t − e−2t ) u(t).
K
The above example illustrates the general approach to solving linear differential equa-
tions of the form
N M
X dk y X dk x
ak k = bk k
k=0
dt k=0
dt
First find the homogeneous solution to the equation
N
X dk y h
ak k = 0
k=0
dt
by hypothesizing that yh (t) = Aemt for some m ∈ C. If there are N different values of m,
denoted m1 , m2 , . . . , mN for which the proposed form holds, then we take the homogeneous
solution to be yh (t) = A1 em1 t + A2 em2 t + · · · + AN emN t , where the coefficients A1 , . . . , AN
are to be determined from initial conditions. If there are not N different values of m, then
further work is required; we will see a more general way to solve these cases later in the
course.
Next, find a particular solution to the equation
N M
X dk yp X dk x
ak = bk k
k=0
dtk k=0
dt
where x(t) is some given function. The idea will be to make yp (t) a linear combination
of terms that, when differentiated, yield terms that appear in x(t) and its derivatives.
Typically this only works when x(t) involves terms like et , sin(t), cos(t), polynomials in t,
etc. Let’s try another example. ■
Example :
Consider the differential equation
The overall solution is then of the form y(t) = yh (t) + yp (t) = A1 e−3t + A2 e2t + 14 5 4t
e
for t > 0. If we are told that the system is at rest until the input is applied, and that
y(0) = y ′ (0) = 0, we have
5
y(0) = A1 + A2 + =0
14
20
y ′ (0) = −3A1 + 2A2 + =0
14
Solving these equations, we obtain A1 = 17 and A2 = − 12 . Thus, the solution is
1 −3t 1 2t 5 4t
y(t) = e − e + e u(t)
7 2 14
The overall solution will be of the form y[n] = yh [n] + yp [n], where yh [n] is a homogeneous
solution to
N
X
ak yh [n − k] = 0
k=0
and yp [n] is a particular solution satisfying the difference equation (3.2) for the given
function x[n]. In this case, we seek homogeneous solutions of the form yh [n] = Aβ n for some
48 Analysis of Linear Time-Invariant Systems
A, β ∈ C, and seek particular solutions that have the same form as the quantities that appear
in x[n]. Let’s do an example.
Example :
Suppose we have the difference equation
1
y[n] − y[n − 1] = x[n]
2
n
with x[n] = 31 u[n].
The solution to this difference equation will be of the form y[n] = yh [n] + yp [n], where
yh [n] is the homogeneous solution satisfying
1
yh [n] − yh [n − 1] = 0
2
and yp [n] is a particular solution to the given difference equation with the given input
signal x[n].
To find the homogeneous solution, we try yh [n] = Aβ n for some constants A and β.
Substituting into the homogeneous difference equation, we obtain
1 A 1
yh [n] − yh [n − 1] = Aβ n − β n−1 = 0 ⇒ β = .
2 2 2
Thus, the homogeneous solution is yh [n] = Aβ for some A that we will identify based
n
the system is at rest for n < 0, i.e., y[n] = 0 for n < 0. Looking at equation (3.3), we have
1
y[0] − y[−1] = 1 ⇒ y[0] = 1
2
Substituting the expression for y[n], we have
1 = y[0] = A − 2 ⇒ A = 3
Thus, the solution is given by
n n
1 1
y[n] = 3 −2 u[n]
2 3
■
Analysis of Linear Time-Invariant Systems 49
dN y(t) dN −1 y(t)
+ a N −1 + · · · + a0 y(t) = b0 x(t)
dtN dtN −1
y[n + N ] + aN −1 y[n + N − 1] + · · · + a0 y[n] = b0 x[n]
Drawing block diagrams for more general differential and difference equations (involving
more than just x[n] on the right hand side) is easier using Laplace
01
One can also calculate this using the homogeneous and particular solutions; in this case, the particular
n
solution would have the form yp [n] = Bδ[n] and B would be found to be zero, so that y[n] = yh [n] = A 12 .
Under the condition of initial rest and y[0] = 1 (obtained from the difference equation), we obtain A = 1,
thus matching the impulse response calculated recursively above.
50 Analysis of Linear Time-Invariant Systems
and z-transform techniques, and so we will defer a study of such equations until then.
For the above equations, we start by writing the highest derivative of y (or the most advanced
version of y ) in terms of all of the other quantities:
$$
dN y(t) dN −1 y(t)
= −aN −1 − · · · − a0 y(t) + b0 x(t)
dtN dtN −1
y[n + N ] = −aN −1 y[n + N − 1] − · · · − a0 y[n] + b0 x[n]
$$
Next, we use a key building block: the integrator block (for continuous-time) or the delay
block (for discrete-time). Specifically, the integrator block is a system whose output is the
integral of the input, and the delay block is a system whose output is a delayed version of the
input. Thus, if we feed dydt
into the integrator block, we get y(t) out, and if we feed y[n + 1]
into the delay block, we get y[n] out, as shown in Fig. 3.5.
To use these blocks to represent differential and difference equations, we simply chain a
sequence of these blocks in series, and feed the highest derivative into the first block in the
chain, as shown in Fig. 3.6.
This series chain of integrator (or delay) blocks provides us with all of the signals needed
N
to represent (3.4) and (3.5). Specifically, from equation (3.4), we see that d dty(t)
N is a lin-
dN −1 y(t) N
ear combination of the signals dtN −1 , · · · , y(t), x(t). Thus, to generate the signal d dty(t)
N ,
we simply take the signals from the corresponding integrator blocks, multiply them by the
coefficients, and add them all together. The same holds true for the signal y[n + N ] in (3.5).
CHAPTER 4
FOURIER SERIES REPRESENTATION OF
PERIODIC SIGNALS
51
52 Fourier Series Representation of Periodic Signals
Let us define
Z ∞
H(s) = h(τ )e−sτ dτ
−∞
If, for the given complex number s, the above integral exists (i.e., is finite), then H(s) is
just some complex number. Thus, we see that for an LTI system, if we apply the complex
exponential x(t) = est as an input, we obtain the quantity
y(t) = H(s)est
as an output. In other words, we get the same complex exponential out of the system,
just scaled by the complex number H(s). Thus, the signal est is called an eigenfunction of
the system, with eigenvalue H(s).
The same reasoning applies for discrete-time LTI systems. Consider an LTI system with
impulse response h[n], and input x[n] = z n . Then,
∞
X ∞
X
y[n] = x[n] ∗ h[n] = h[k]x[n − k] = h[k]z n−k
k=−∞ k=−∞
X∞
=z n
h[k]z −k .
k=−∞
Let us define
∞
X
H(z) = h[k]z −k .
k=−∞
If this sum converges for the given choice of complex number z, then H(z) is just some
complex number. Thus, we see again that for a discrete-time LTI system with the complex
exponential x[n] = z n as an input, we obtain the quantity
y[n] = H(z)z n
as an output. In this case z n is an eigenfunction of the system, and H(z) is the eigenvalue.
So, to summarize, we have the following:
Fourier Series Representation of Periodic Signals 53
• If the signal x(t) = est is applied to anR LTI system with impulse response h(t), the
∞
output is y(t) = H(s)est , where H(s) = −∞ h(τ )e−sτ dτ (assuming the integral exists).
• If the signal x[n] = z n is applied to anPLTI system with impulse response h[n], the
output is y[n] = H(z)z n , where H(z) = ∞ k=−∞ h[k]z
−k
(assuming the sum converges).
As we will see later in the course, the quantities H(s) and H(z) are the Laplace Transform
and z-Transform of the impulse response of the system, respectively.
Note that the above translates to superpositions of complex exponentials in an natural
way. Specifically, if the input is x(t) = i=1 ai e for some complex numbers a1 , . . . , an and
Pn si t
s1 , . . . , sn , we have
n
X
y(t) = ai H (si ) esi t
i=1
Thus, for real-valued impulse responses, we have H(−jω) = H(jω)∗ . We can equiva-
lently write these in polar form as
Suppose we apply the signal x(t) = cos (ω0 t) to the system. We expect the output
to be cos (ω0 (t − t0 )), based on the definition of the system. Let’s verify this using the
identities we derived earlier. We have
x(t) = ejω0 t
Recall that this signal is periodic with fundamental period T = ω2π0 (assuming ω0 > 0
). Based on this complex exponential, we can define an entire harmonic family of complex
exponentials, given by
ϕk (t) = ejkω0 t , k ∈ Z
In other words, for each k ∈ Z, ϕk (t) is a complex exponential whose fundamental frequency
is kω0 (i.e., k times the fundamental frequency of x(t) ). Thus, each of the signals ϕk (t) is
periodic with period T , since
Note that T may not be the fundamental period of the signal ϕk (t), however.
Since each of the signals in the harmonic family is periodic with period T , a linear combination
of signals from that family is also periodic. Specifically, consider the signal
∞ ∞
2π
X X
jkω0 t
x(t) = ak e = ak ejk T t
k=−∞ k=−∞
The terms corresponding to k = 1 and k = −1 are known as the first harmonic of the
signal x(t). The terms corresponding to k = 2 and k = −2 are known as the second harmonic
and so forth.
Fourier Series Representation of Periodic Signals 55
for some sequence of coefficients ak , k ∈ Z. Then the above representation is called the
Fourier Series representation of x(t). The quantities ak , k ∈ Z are called the Fourier Series
coefficients.
Example :
Consider the signal x(t) = cos (ω0 t), where ω0 > 0. We have
1 1
x(t) = ejω0 t + e−jω0 t
2 2
This is the Fourier Series representation of x(t); it has only first harmonics, with coef-
ficients a1 = a−1 = 12 .
Similarly, consider the signal x(t) = sin (ω0 t). We have
1 jω0 t 1
x(t) = e − e−jω0 t .
2j 2j
Once again, the signal has only first harmonics, with coefficients a1 = 2j1 and a−1 = − 2j1 .
Suppose that we have a periodic signal that has a Fourier Series representation
∞
X
x(t) = ak ejkω0 t
k=−∞
Now suppose that x(t) is real, i.e., x∗ (t) = x(t). Taking the complex conjugate of both
sides of the above expression, we have
∞
X
x∗ (t) = a∗k e−jkω0 t
k=−∞
Comparing the terms, we see that for any k ∈ Z, the coefficient of ejkω0 t is ak on the
left hand side, and is a∗−k on the right hand side. Thus, for real signals x(t), the Fourier
Series coefficients satisfy
a−k = a∗k
56 Fourier Series Representation of Periodic Signals
for all k ∈ Z. Substituting this into the Fourier Series representation (4.1) we have
∞
X
ak ejkω0 t + a−k e−jkω0 t
x(t) = a0 +
k=1
∞
X
ak ejkω0 t + a∗k e−jkω0 t
= a0 +
k=1
X∞
2 Re ak ejkω0 t
= a0 +
k=1
where Re is the real part of the given complex number. If we write ak in polar form as
rk ejθk , the above expression becomes
∞
X ∞
X
j(kω0 t+θk )
x(t) = a0 + 2 Re rk e = a0 + 2 rk cos (kω0 t + θk )
k=1 k=1
This is an alternate representation of the Fourier Series for real-valued signals (known
as the trigonometric representation). ■
We will soon see conditions under which a signal will have such a representation, but for
now, suppose that we are just interested in finding the coefficients ak , k ∈ Z. To do this,
multiply x(t) by e−jnω0 t , where n is some integer. This gives
∞
X ∞
X
−jnω0 t jkω0 t −jnω0 t
x(t)e = ak e e = ak ej(k−n)ω0 t
k=−∞ k=−∞
Now suppose that we integrate both sides of the above equation from t0 to t0 + T for any
t0 :
Z t0 +T Z t0 +T ∞
X ∞
X Z t0 +T
−jnω0 t j(k−n)ω0 t
x(t)e dt = ak e dt = ak ej(k−n)ω0 t dt
t0 t0 k=−∞ k=−∞ t0
Otherwise, if n ̸= k, we have
Fourier Series Representation of Periodic Signals 57
Z t0 +T 0 t +T
1
ej(k−n)ω0 t dt = ej(k−n)ω0 t
t0 j(k − n)ω0 t0
1
ej(k−n)ω0 (t0 +T ) − ej(k−n)ω0 t0
=
j(k − n)ω0
=0
Thus, we have
Z t0 +T ∞
X Z t0 +T
−jnω0 t
x(t)e dt = ak ej(k−n)ω0 t dt = an T
t0 k=−∞ t0
or equivalently,
Z t0 +T
1
an = x(t)e−jnω0 t dt
T t0
where t0 is any arbitrary starting point. In other words, we obtain the Fourier coefficient
an by multiplying the signal x(t) by e−jnω0 t and then integrating the resulting product over
any period.
Example :
Consider the signal
T
0 − 2 ≤ t < T1
x(t) = 1 −T1 ≤ t ≤ T1
0 T1 < t < T2
∞
X ∞
X
ak ejkω0 t + e−jkω0 t = a0 + 2
x(t) = a0 + ak cos (kω0 t)
k=1 k=1
if k ̸= n, and is equal to T otherwise. The functions ejkω0 and ejnω0 (for k ̸= n ) are said
to be orthogonal. More generally, a set of functions ϕk (t), k ∈ Z, are said to be orthogonal
on an interval [a, b] if
Z b
ϕk (t)ϕ∗n (t) = 0
a
if k ̸= n, and nonzero otherwise. Note that ϕ∗n (t) is the complex conjugate of ϕn (t).
We then derived the expressions for the coefficients by using the orthogonality property.
However, that derivation assumed that the signal could be written as a linear combination of
the functions in the harmonic family, and then derived the coefficient expressions. Here will
justify this by first trying to approximate a given signal by a finite number of functions from
the harmonic family and then taking the number of approximating functions to infinity. We
will start by reviewing how to approximate a given vector by other vectors, and then explore
the analogy to the approximation of functions.
x̂ = av1 + bv2
is "close" to x. A typical metric of "closeness" is taken to be the square of the approxi-
mation error. Specifically, the approximation error is given by
e = x − x̂ = x − av1 − bv2
Fourier Series Representation of Periodic Signals 59
Note that e is a vector, where the i-th component is the approximation error for the i-th
component of x. We try to minimize e21 + e22 + e23 , which is given by
∂e′ e
= −x′ v1 − v1′ x + 2av1′ v1 = 0
∂a
v′ x
⇒ a = ′1
v1 v1
′
∂e e
= −x′ v2 − v2′ x + 2bv2′ v2 = 0
∂b
v2′ x
⇒b= ′
v2 v2
where we used the fact that x′ v1 = v1′ x and x′ v2 = v2′ x (since these quantities are all
scalars). In terms of the vectors given above, we obtain
1 5
a= = 1, b= .
1 2
is "close" to x(t). Mathematically, we will use the notion of squared error to measure
"closeness." Specifically, the approximation error at any given point in time t is given by
N
X
e(t) = x(t) − x̂(t) = x(t) − ak ϕk (t)
k=−N
and the squared error over the entire interval [a, b] is then defined as
60 Fourier Series Representation of Periodic Signals
Z b
|e(t)|2 dt
a
Here, we will allow e(t) to a general complex valued function, so the absolute value in the
integral is interpreted as the magnitude of the complex number e(t), i.e., the square error
over the interval [a, b] is given by
Z b
e∗ (t)e(t)dt
a
Consider the harmonic family ϕk (t) = ejkω0 t , and suppose that we wish to find the best
approximation of a given T -periodic signal x(t) as a linear combination of ϕk (t) for −N ≤
k ≤ N , i.e.,
N
X
x̂(t) = ak ejkω0 t
k=−N
with error
N
X
e(t) = x(t) − x̂(t) = x(t) − ak ejkω0 t
k=−N
We evaluate the squared error over any interval of length T (since the functions ϕk (t) are
orthogonal over such intervals):
Z t0 +T Z t0 +T
Squared Error = 2
|e(t)| dt = e∗ (t)e(t)dt
t0 t0
N
! N
!
Z t0 +T X X
= x∗ (t) − a∗k e−jkω0 t x(t) − ak ejkω0 t dt
t0 k=−N k=−N
N N
!
Z t0 +T X X
= x∗ (t)x(t) − x∗ (t) ak ejkω0 t − x(t) a∗k e−jkω0 t dt
t0 k=−N k=−N
N N
!
Z t0 +T X X
+ a∗k e−jkω0 t ak ejkω0 t dt
t0 k=−N k=−N
N N
!
Z t0 +T Z t0 +T X X
= |x(t)|2 dt + −x∗ (t) ak ejkω0 t − x(t) a∗k e−jkω0 t dt
t0 t0 k=−N k=−N
N N
!
Z t0 +T X X
+ a∗k an e−jkω0 t ejnω0 t dt
t0 k=−N k=−N
Fourier Series Representation of Periodic Signals 61
Z t0 +T N
X Z t0 +T N
X Z t0 +T
∗
= 2
|x(t)| dt − ak x (t)e jkω0 t
dt − a∗k x(t)e−jkω0 t dt
t0 k=−N t0 k=−N t0
N N
!
X X Z t0 +T
+ a∗k an e−jkω0 t ejnω0 t dt
k=−N n=−N t0
Z t0 +T N
X Z t0 +T N
X Z t0 +T
∗
= 2
|x(t)| dt − ak x (t)e jkω0 t
dt − a∗k x(t)e−jkω0 t dt
t0 k=−N t0 k=−N t0
N
X
+T |ak |2 ,
k=−N
R t +T
where we used the fact that t00 e−jkω0 t ejnω0 t dt = 0 if k ̸= n and T otherwise.
Our job is to find the best coefficients ak , −N ≤ k ≤ N to minimize the square error. Thus,
we first write ak = bk +jck , where bk , ck ∈ R, and then differentiate the above expression with
respect to bk and ck and set the result to zero. After some algebra, we obtain the optimal
coefficient as
1 t0 +T
Z
ak = bk + jck = x(t)e−jkω0 t dt
T t0
which is exactly the same expression we found for the Fourier series coefficients earlier.
Note again why we bothered to go through this exercise. Here, we did not assume that a signal
x(t) had a Fourier series representation; we simply asked how to best approximate a given
signal by a linear combination of complex exponentials, and found the resulting coefficients.
These coefficients match exactly the coefficients that we obtained by assuming that the signal
had a Fourier series representation, and lends some justification for the validity of the earlier
analysis.
As N gets larger, the approximation error will get smaller and smaller. The question is
R t +T
then: will t00 |e(t)|2 dt go to zero as N goes to infinity? If so, then the signal would, in
fact, have a Fourier series representation (in the sense of having asymptotically zero error
between the true signal and the approximation). It turns out that most periodic signals of
practical interest will satisfy this property.
• In any finite interval of time x(t) has bounded variation, meaning that it has only a
finite number of minima and maxima during any single period of the signal.
• In any finite interval of time, there are only a finite number of discontinuities, and each
of these discontinuities are finite.
We won’t go into the proof of why these conditions are sufficient here, but it suffices to
note that signals that violate the above conditions (the last two in particular) are somewhat
pathological. The first condition guarantees that the Fourier series coefficients are finite,
since R t +T R t +T R t +T
|ak | = T1 t00 x(t)e−jkω0 t dt ≤ T1 t00 x(t)e−jkω0 t dt = T1 t00 |x(t)|dt.
Thus if the signal is absolutely integrable over a period, then |ak | < ∞ for all k ∈ Z.
Gibbs Phenomenon
If x(t) has a Fourier series representation, then that representation will exactly equal x(t)
at all points t where x(t) is continuous. At points of discontinuity in x(t), the value of the
Fourier series representation will be equal to the average of the values of x(t) on either side of
the discontinuity. One particularly interesting phenomenon occurs at points of discontinuity:
the Fourier series typically overshoots the signal x(t). The height of the overshoot stays
constant as the number of terms N in the approximation increases, but the width shrinks.
Thus, asymptotically, the error goes to zero (although technically the two signals are not
exactly the same at the discontinuity).
5.1 Linearity
Suppose we have two signals x1 (t) and x2 (t), each of which is periodic with period T . Let
FS FS
x1 (t) ←→ ak , x2 (t) ←→ bk .
For any complex scalars α, β, let g(t) = αx1 (t) + βx2 (t). Then
FS
g(t) ←→ αak + βbk .
The above property follows immediately from the definition of the Fourier series coefficients
(since integration is linear).
Fourier Series Representation of Periodic Signals 63
Example :
Consider x1 (t) = cos (ω0 t) and x2 (t) = sin (ω0 t). We have
α jω0 t α −jω0 t β β
g(t) = α cos (ω0 t) + β sin (ω0 t) = e + e + ejω0 t − e−jω0 t
2 2 2j 2j
α β α β
= + ejω0 t + − e−jω0 t
2 2j 2 2j
Thus, we see that each Fourier series coefficient of g(t) is indeed given by a linear
combination of the corresponding Fourier series coefficients of x1 (t) and x2 (t). ■
Thus we have
FS
x(−t) ←→ a−k ,
i.e., the Fourier series coefficients for a time-reversed signal are just the timereversal of the
Fourier series coefficients for the original signal. Note that this is not true for the output of
LTI systems (a time reversal of the input to an LTI system does not necessarily mean that
the output is a time-reversal of the original output).
64 Fourier Series Representation of Periodic Signals
Example :
Consider x(t) = sin (ω0 t) and define g(t) = x(−t). First, note that
1 jω0 t 1
x(t) = sin (ω0 t) = e − e−jω0 t = a1 ejω0 t + a−1 e−jω0 t
2j 2j
We have
1 −jω0 t 1
g(t) = sin (−ω0 t) = e − ejω0 t = b1 ejω0 t + b−1 e−jω0 t
2j 2j
and thus we see that b1 = − 2j1 = a−1 and b−1 = 1
2j
= a1 . ■
Thus, g(t) has the same Fourier series coefficients as x(t), but the Fourier series repre-
sentation has changed: the frequency is now ω0 α rather than ω0 , to reflect the fact that the
harmonic family is in terms of the new period Tα rather than T .
Example :
Consider x(t) = cos (ω0 t), which has series representation
1 1
x(t) = ejω0 t + e−jω0 t .
2 2
Then we have
1 1
g(t) = x(αt) = cos (ω0 αt) = ejω0 αt + e−jω0 αt .
2 2
■
5.5 Multiplication
Let x1 (t) and x2 (t) be T -periodic signals with Fourier series ak and bk respectively. Consider
the signal g(t) = x1 (t)x2 (t), and note that g(t) is also T -periodic. We have
∞
X ∞
X ∞
X ∞
X
jlω0 t jnω0 t
g(t) = x1 (t)x2 (t) = al e bn e = al bn ej(l+n)ω0 t .
l=−∞ n=−∞ l=−∞ n=−∞
Define k = l + n, so that
∞ ∞ ∞ ∞
!
X X X X
g(t) = al bk−l ejlω0 t = al bk−l ejkω0 t
l=−∞ l=−∞ k=−∞ l=−∞
∞
FS
X
g(t) = x1 (t)x2 (t) ←→ al bk−l .
l=−∞
In other words, the Fourier series coefficients of a product of two signals are given by the
convolution of the corresponding Fourier series coefficients.
Example :
Consider the signals x1 (t) = cos (ω0 t) and x2 (t) = sin (ω0 t). Define g(t) = x1 (t)x2 (t). The
Fourier series representations of x1 (t) and x2 (t) are given by
1 1 1 jω0 t 1
x1 (t) = ejω0 t + e−jω0 t , x2 (t) = e − e−jω0 t .
2 2 2j 2j
Denote the Fourier series coefficients of x1 (t) by the sequence ak , with a−1 = a1 = 12 ,
and ak = 0 otherwise. Similarly, denote the Fourier series coefficients of x2 (t) by the
sequence bk , with b−1 = − 2j1 , b(1) = 2j1 , and bk = 0 otherwise. Denote the Fourier series
coefficients of g(t) by ck , k ∈ Z. Then we have
∞
X
c k = ak ∗ b k = al bk−l .
l=−∞
Convolving the two sequences given above, we see that c−2 = − 4j1 , c2 = 1
4j
, and ck = 0
otherwise. Thus
1 j2ω0 t 1
g(t) = x1 (t)x2 (t) = cos (ω0 t) sin (ω0 t) = e − e−j2ω0 t
4j 4j
Noting that cos (ω0 t) sin (ω0 t) = 1
2
sin (2ω0 t), we see that the above expression is, in
fact, correct. ■
Substituting |x(t)|2 = x∗ (t)x(t) and the Fourier series for x(t) we have
∞
! ∞
!
Z t0 +T Z t0 +T
1 1 X X
|x(t)|2 dt = a∗k e−jkω0 t a∗n ejnω0 t dt
T t0 T t0 k=−∞ n=−∞
∞ ∞ Z t0 +T
X X 1
= a∗k an e−jkω0 t ejnω0 t dt
k=−∞ n=−∞
T t0
Using the fact that ejkω0 t and ejnω0 t are orthogonal for k ̸= n, we have
66 Fourier Series Representation of Periodic Signals
Z t0 +T ∞ ∞
1 X X
2
|x(t)| dt = a∗k ak = |ak |2
T t0 k=−∞ k=−∞
This leads to Parseval’s Theorem: for a T -periodic signal x(t) with Fourier series coeffi-
cients ak , k ∈ Z, we have
Z t0 +T ∞
1 X
|x(t)|2 dt = |ak |2
T t0 k=−∞
Example :
Consider the signal x(t) = cos (ω0 t), with Fourier coefficients a1 = a−1 = 1
2
and ak = 0
otherwise. We have
Z t0 +T Z t0 +T Z t0 +T
1 2 1 2 1 1 1
|x(t)| dt = cos (ω0 t) dt = (1 + cos (2ω0 t)) dt =
T t0 T t0 T t0 2 2
We also have
∞
X 1 1 1
|ak |2 = a21 + a2−1 = + =
k=−∞
4 4 2
which agrees with the direct calculation of average power, as indicated by Parseval’s
Theorem. ■
At this point, we encounter the first main difference between the discrete-time and continuous-
time Fourier series. Recall that a discrete-time complex exponential with frequency ω0 is the
same as a discrete-time complex exponential with frequency ω0 + 2π. Specifically, for any
k ∈ Z, consider
On the other hand, if r − k is not a multiple of N , we use the finite sum formula to obtain
+N −1
n1X
ej(k−r)ω0 n1 − ej(k−r)ω0 (n1 +N )
ej(k−r)ω0 n = =0
n=n1
ej(k−r)ω0 − 1
Example 4.11. Consider the N -periodic signal x[n], which is equal to 1 for −N1 ≤ n ≤ N1
and zero otherwise (modulo the periodic constraints).
We have
N1
1 X
ak = e−jkω0 n
N n=−N
1
If k = 0, we have a0 = 2N1 +1
N
. For k ∈ {1, 2, . . . , N − 1}, we have (via the finite sum
formula)
1 1
1 e−jkω0 2 ejkω0 (N1 + 2 ) − ejkω0 (N1 + 2 )
1
1 ejkω0 N1 − ejkω0 (N1 +1)
ak = =
N 1 − e−jkω0 N e−jkω0 12 1
ejkω0
2 − e
−jkω0 21
1 sin kω0 N1 + 21
=
sin k ω20
N
It is of interest to note that we do not have to worry about convergence conditions for
discrete-time Fourier series, as we did in the continuous-time case. Specifically, for an N -
periodic discrete-time signal x[n], we only require N numbers to completely specify the
entire signal. The Fourier series coefficients ak , k ∈ {0, 1, . . . , N − 1} thus contain as much
information as the signal itself, and form a perfect representation of the signal. In other
words, for discrete-time signals, the Fourier series representation is just a transformation of
the signal into another form; we do not encounter discrepancies like the Gibbs phenomenon
in discrete-time.
Thus, the Fourier series coefficients of g[n] are given by c0 = a0 b0 +a1 b1 and c1 = a0 b1 +a1 b0 .
We can write these in a unform way as follows:
1
X
c0 = a0 b0 + a1 b1 = a0 b0 + a1 b−1 = al b−l
l=0
1
X
c 1 = a0 b 1 + a1 b 0 = al b1−l
l=0
where we used the fact that b1 = b−1 by the periodic nature of the discrete-time Fourier
series coefficients. The above expressions show that the Fourier series coefficients of the
product of the two signals are given by a form of convolution of the coefficients of those
signals; however, the convolution is over a finite number of terms, as opposed to over all
time-indices.
Let us generalize this to functions with a larger period. Let x1 [n] and x2 [n] be two N -periodic
discrete-time signals, with discrete-time Fourier series coefficients ak and bk , respectively. We
have
N
X −1 N
X −1
jkω0 n
x1 [n] = ak e , x2 [n] = bk ejkω0 n
k=0 k=0
−1
N
! N −1
!
X X
g[n] = x1 [n]x2 [n] = al ejlω0 n br ejrω0 n
l=0 r=0
N
XN−1 X −1
= al br ej(l+r)ω0 n
l=0 r=0
where we have used l and r as the indices in the Fourier series in order to keep the terms
in the two sums distinct.
Define the new variable k = l + r. Substituting into the above expression, this gives
70 Fourier Series Representation of Periodic Signals
N
X −1 l+N
X−1
g[n] = al bk−l ejkω0 n
l=0 k=l
−1 −1 X−1
N N l+N
!
X X
jkω0 n jkω0 n
= al bk−l e + al bk−l e
l=0 k=l k=N
−1 −1
N N l−1
!
X X X
= al bk−l ejkω0 n + al bk+N −l ej(k+N )ω0 n
l=0 k=l k=0
−1 −1
N N l−1
!
X X X
= al bk−l ejkω0 n + al bk−l ejkω0 n
l=0 k=l k=0
N
X −1 N
X −1
= al bk−l ejkω0 n
l=0 k=0
−1 N −1
N
!
X X
= al bk−l ejkω0 n
k=0 l=0
N
X −1
ck = al bk−l
l=0
+N −1
n0X
1
|x[n]|2
N n=n0
where n0 is any integer. Parseval’s theorem for discrete-time signals states the following.
+N −1
n0X N −1
1 X
|x[n]|2 = |ak |2
N n=n0 k=0
Fourier Series Representation of Periodic Signals 71
The following example illustrates the application of the various facts that we have seen
about the discrete-time Fourier series.
Example 4.12. Suppose we are told the following facts about a discrete-time signal x[n].
•
P5
n=0 x[n] = 2.
•
P7 n
n=2 (−1) x[n] = 1.
• x[n] has the minimum power per period of all signals satisfying the preceding three
conditions.
The above facts are sufficient for us to uniquely determine x[n]. First, note that the
Fourier series representation of x[n] is
5 5
π
X X
jkω0 n
x[n] = ak e = ak ejk 3 n ,
k=0 k=0
73
74 The Continuous-Time Fourier Transform
However, we note that x(t) = x̃(t) in the interval of integration, and thus
Z T
1 2
ak = x(t)e−jkω0 t dt
T − T2
Furthermore, since x(t) is zero for all t outside the interval of integration, we can expand
the limits of the integral to obtain
1 ∞
Z
ak = x(t)e−jkω0 t dt
T −∞
Let us define
Z ∞
X(jω) = x(t)e−jωt dt
−∞
This is called the Fourier transform of the signal x(t), and the Fourier series coefficients
can be viewed as samples of the Fourier transform, scaled by T1 , i.e.,
1
ak = X (jkω0 ) , k ∈ Z
T
Now consider the fact that
∞ ∞
X
jkω0 t 1 X
x̃(t) = ak e = X (jkω0 ) ejkω0 t
k=−∞
T k=−∞
Since ω0 = 2π
T
, this becomes
∞
1 X
x̃(t) = X (jkω0 ) ejkω0 t ω0
2π k=−∞
The Continuous-Time Fourier Transform 75
Now consider what happens as the period T gets bigger. In this case, x̃(t) approaches
x(t), and so the above expression becomes a representation of x(t). As T → ∞, we have
ω0 → 0. Since each term in the summand can be viewed as the area of the rectangle whose
height is X (jkω0 ) ejkω0 t and whose base goes from kω0 to (k + 1)ω0 , we see that as ω0 → 0,
the sum on the right hand side approaches the area underneath the curve X(jω)ejωt (where
t is held fixed). Thus, as T → ∞ we have
Z ∞
1
x(t) = X(jω)ejωt dω
2π −∞
Thus we have the following.
Given a continuous-time signal x(t), the Fourier Transform of the signal is given by
Z ∞
X(jω) = x(t)e−jωt dt
−∞
The Inverse Fourier Transform of the signal is given by
Z ∞
1
x(t) = X(jω)ejωt dω
2π −∞
The Fourier transform X(jω) is also called the spectrum of the signal, as it represents the
contribution of the complex exponential of frequency ω to the signal x(t).
Example :
Consider the signal x(t) = e−at u(t), a ∈ R>0 . The Fourier transform of this signal is
Z ∞ Z ∞
−jωt
X(jω) = x(t)e dt = e−at e−jωt dt
−∞ 0
∞
1
=− e−(a+jω)t
a + jω 0
1
= .
a + jω
To visualize X(jω), we plot its magnitude and phase on separate plots (since X(jω) is
complex-valued in general). We have
1 −1 ω
|X(jω)| = √ , ∠X(jω) = − tan .
a2 + ω 2 a
The plots of these quantities are show in Fig. 4.5 of the text.
■
Example :
Consider the signal x(t) = δ(t). We have
Z ∞
X(jω) = δ(t)e−jωt dt = 1
−∞
In other words, the spectrum of the impulse function has an equal contribution at all
frequencies. ■
76 The Continuous-Time Fourier Transform
Example :
Consider the signal x(t) which is equal to 1 for −T1 ≤ t ≤ T1 and zero elsewhere. We have
Z T1
1 jωT1 2 sin (ωT1 )
X(jω) = e−jωt dt = e − e−jωT1 = .
−T1 jω ω
■
Example :
Consider the signal whose Fourier transform is
(
1, |ω| ≤ W
X(jω) =
0 |ω| > W
We have
Z ∞ Z W
1 jωt 1
x(t) = X(jω)e dω = ejωt dω
2π −∞ 2π −W
W
1 1 jωt
e =
2π jt −W
sin(W t)
=
πt
The previous two examples showed the following. When x(t) is a square pulse, then
X(jω) = 2 sin(ωT
ω
1)
and when X(jω) is a square pulse, x(t) = sin(Wπt
t)
. This is an example of
the duality property of Fourier transforms, which we will see later.
Functions of the form sin(W
πt
t)
will show up frequently, and are called sinc functions. Specif-
ically
sin(πθ)
sinc(θ) =
πθ
Thus 2 sin(ωT1 ) ωT1
and sin(W t) W Wt
.
ω
= 2T1 sinc π πt
= π
sinc π
■
2. x(t) has a finite number of maxima and minima in any finite interval.
3. x(t) has a finite number of discontinuities in any finite interval, and each of these
discontinuities is finite.
The Continuous-Time Fourier Transform 77
If all of the above conditions are satisfied, x(t) is guaranteed to have a Fourier transform.
Note that this only a sufficient set of conditions, and not necessary.
An alternate sufficient condition is that the signal have finite energy (i.e., that it be square
integrable):
Z ∞
|x(t)|2 dt < ∞
−∞
X(jω) = 2πδ (ω − ω0 )
i.e., the frequency domain signal is a single impulse at ω = ω0 , with area 2π. Using the
inverse Fourier transform, we obtain
Z ∞ Z ∞
1 jωt
x(t) = X(jω)e dω = δ (ω − ω0 ) ejωt dω
2π −∞ −∞
jω0 t
=e
Thus, the Fourier transform of x(t) = ejω0 t is X(jω) = 2πδ (ω − ω0 ). Similarly, if
∞
X
X(jω) = ak 2πδ (ω − kω0 )
k=−∞
In other words, if x(t) is a periodic signal with Fourier series coefficients ak , then the Fourier
transform of x(t) consists of a sequence of impulse functions, each spaced at multiples of ω0 ;
the area of the impulse at kω0 will be 2πak .
Example :
Consider the signal x(t) = cos (ω0 t). The Fourier series coefficients are a1 = a−1 = 12 .
Thus, the Fourier transform of this signal is given by
78 The Continuous-Time Fourier Transform
X(jω) = F{x(t)}
x(t) = F −1 {X(jω)}
We will also use the notation
F
x(t) ←→ X(jω)
to indicate that x(t) and X(jω) are Fourier transform pairs.
3.1 Linearity
The first property of Fourier transforms is easy to show:
The Continuous-Time Fourier Transform 79
3.2 Time-Shifting
Suppose x(t) is a signal with Fourier transform X(jω). Define g(t) = x(t − τ ) where τ ∈ R.
Then we have
Z ∞ Z ∞
−jωt
G(jω) = g(t)e dt = x(t − τ )e−jωt dt = e−jωτ X(jω)
−∞ −∞
Thus
3.3 Conjugation
Consider a signal x(t). We have
Z ∞ ∗ Z ∞
∗ −jωt
X (jω) = x(t)e dt = x∗ (t)ejωt dt
−∞ −∞
Thus,
Z ∞
∗
X (−jω) = x∗ (t)e−jωt dt = F {x∗ (t)}
−∞
The above is true for any signal x(t) that has a Fourier transform. Now suppose addition-
ally that x(t) is a real-valued signal. Then we have x∗ (t) = x(t) for all t ∈ R. Thus
X(jω) = |X(jω)|ej∠X(jω)
Then we have
Example :
Consider again the signal x(t) = e−at u(t); we saw earlier that the Fourier transform of this
signal is
1
X(jω) =
a + jω
It is easy to verify that
1
X(−jω) = = X ∗ (jω)
a − jω
as predicted. Furthermore, we can see from the plots of the magnitude and phase of
X(jω) that the magnitude is indeed an even function, and the phase is an odd function.
Suppose further that x(t) is even (in addition to being real-valued). Then we have x(t) =
x(−t). Then we have
Z ∞ Z ∞ Z ∞
X(−jω) = jωt
x(t)e dt = jωt
x(−t)e dt = x(t)e−jωt dt
−∞ −∞ −∞
= X(jω)
This, together with the fact that X(−jω) = X ∗ (jω) for real-valued signals indicates
that X(jω) is real-valued and even.
Similarly, if x(t) is real-valued and odd, we have X(jω) is purely imaginary and odd.
■
Example :
Consider the signal x(t) = e−a|t| , where a is a positive real number. This signal is real-
valued and even. We have
Z ∞ Z 0 Z ∞
−jωt at −jωt
X(jω) = x(t)e dt = e e dt + e−at e−jωt dt
−∞ −∞ 0
1 1
= +
a − jω a + jω
2a
= 2 .
a + ω2
As predicted, X(jω) is real-valued and even. ■
3.4 Differentiation
Consider the inverse Fourier transform
Z ∞
1
x(t) = X(jω)ejωt dω
2π −∞
1 ω
F{x(at)} = X j .
|a| a
Thus we see again that shrinking a signal in the time-domain corresponds to expanding it
in the frequency domain, and vice versa.
3.6 Duality
We have already seen a few examples of the duality property: suppose x(t) has Fourier
transform X(jω). Then if we have a time-domain signal that has the same form as X(jω),
the Fourier transform of that signal will have the same form as x(t). For example, the square
pulse in the time-domain had a Fourier transform in the form of a sinc function, and a sinc
function in the time-domain had a Fourier transform in the form of a square pulse.
We can consider another example. Suppose x(t) = e−|t| . Then one can verify that
2
F e−|t| =
1 + ω2
Specifically we have
Z ∞
−|t| 1 2
e = e−jωt dω
2π −∞ 1 + ω2
If we multiply both sides by 2π and interchange ω and t, we obtain
Z ∞
−|ω| 2
2πe = 2
e−jωt dt
−∞ 1 + t
Thus, we have
82 The Continuous-Time Fourier Transform
2
F = 2πe−|ω|
1 + t2
Duality also applies to properties of the Fourier transform. For example, recall that differ-
entiation in the time-domain corresponds to multiplication by jω in the frequency domain.
We will now see that differentiation the frequency domain corresponds to multiplication by
a certain quantity of t in the time-domain. We have
Z ∞
dX(jω)
= x(t)(−jt)e−jωt dt
dω −∞
3.8 Convolution
Consider a signal x(t) with Fourier transform X(jω). From the inverse Fourier transform,
we have
Z ∞
1
x(t) = X(jω)ejωt dω
2π −∞
This has the interpretation that x(t) can be written as a superposition of complex expo-
nentials (with frequencies spanning the entire real axis). From earlier in the course, we know
that if the input to an LTI system is ejωt , then the output will be H(jω)ejωt , where
Z ∞
H(jω) = h(t)e−jωt dt
−∞
The Continuous-Time Fourier Transform 83
In other words, H(jω) is the Fourier transform of the impulse response h(t). This, together
with the LTI property of the system, implies that
Z ∞ Z ∞
1 jωt 1
x(t) = X(jω)e dω ⇒ X(jω)H(jω)ejωt dω = y(t).
2π −∞ 2π −∞
Thus, we see that the Fourier transform of the output y(t) is given by
Y (jω) = H(jω)X(jω)
In other words:
We have
Z ∞ Z ∞ Z ∞
−jωt
Y (jω) = y(t)e dt = x(τ )h(t − τ )dτ e−jωt dt
−∞
Z−∞
∞
−∞
Z ∞
−jωt
= x(τ ) h(t − τ )e dt dτ
−∞ −∞
Z ∞
= x(τ )e−jωτ H(jω)dτ
−∞
Z ∞
= H(jω) x(τ )e−jωτ dτ
−∞
= H(jω)X(jω)
In the third line, we used the time-shifting property of the Fourier transform. Thus we
see that convolution of two signals in the time-domain corresponds to multiplication of the
signals in the frequency domain, i.e.,
This is precisely the first condition in the Dirichlet conditions; thus, as long as the system
is stable and the impulse response satisfies the other two conditions (which almost all real
84 The Continuous-Time Fourier Transform
systems would), the Fourier transform is guaranteed to exist. If the system is unstable, we
will need the machinery of Laplace transforms to analyze the input-output behavior, which
we will defer to a later discussion.
The convolution - multiplication property is also very useful for analysis of interconnected
linear systems. For example, consider the series interconnection shown in Fig. 5.1.
We have
y(t) = y1 (t) ∗ h2 (t) = (x(t) ∗ h1 (t)) ∗ h2 (t) = x(t) ∗ (h1 (t) ∗ h2 (t)) .
Taking Fourier transforms, we obtain
sin(W t)
h(t) =
πt
However, there are various challenges with implementing an LTI system with this im-
pulse response. One is that this is noncausal, and thus one must potentially include a
sufficiently large delay (followed by a truncation of the signal) in order to apply it. An-
other problem is that it contains many oscillations, which may not be desirable for an
impulse response.
Instead of the above filter, suppose consider another filter whose impulse response is
3.9 Multiplication
We just saw that multiplication in the time domain corresponds to convolution in the fre-
quency domain. By duality, we obtain that multiplication in the frequency domain corre-
sponds to convolution in the time-domain. Specifically, consider two signals x1 (t) and x2 (t),
and define g(t) = x1 (t)x2 (t). Then we have
Z ∞ Z ∞
−jωt
G(jω) = g(t)e dt = x1 (t)x2 (t)e−jωt dt
−∞ −∞
Z ∞ Z ∞
1
= x2 (t) X1 (jθ)ejθt dθe−jωt dt
2π −∞ −∞
Z ∞ Z ∞
1
= X1 (jθ) x2 (t)e−j(ω−θ)t dtdθ
2π −∞ −∞
Z ∞
1
= X1 (jθ)X2 (j(ω − θ))dθ
2π −∞
Thus,
86 The Continuous-Time Fourier Transform
Z ∞
1 1
F {x1 (t)x2 (t)} = (X1 (jω) ∗ X2 (jω)) = X1 (jθ)X2 (j(ω − θ))dθ
2π 2π −∞
Multiplication of one signal x1 (t) by another signal x2 (t) can be viewed as modulating the
amplitude of one signal by the other. This plays a key role in communication systems.
Example :
Consider a signal s(t) whose frequency spectrum lies in some interval [−W, W ]. Consider
the signal p(t) = cos (ω0 t). The Fourier transform of p(t) is given by
P (jω) = πδ (ω − ω0 ) + πδ (ω + ω0 )
Now consider the signal x(t) = s(t)p(t), with Fourier transform given by
Z ∞
1
X(jω) = S(jθ)P (j(ω − θ))dθ
2π −∞
1 ∞
Z
= S(jθ)δ (ω − θ − ω0 ) dθ
2 −∞
1 ∞
Z
+ S(jθ)δ (ω − θ + ω0 ) dθ
2 −∞
1 1
= S (j (ω − ω0 )) + S (j (ω + ω0 ))
2 2
Thus, multiplying the signal s(t) by p(t) results in a signal x(t) whose frequency spec-
trum consists of two copies of the spectrum of s(t), centered at the frequencies ω0 and
−ω0 and scaled by 21 . ■
The above example illustrates the principle behind amplitude modulation (AM) in commu-
nication and radio systems. A low frequency signal (such as voice) is amplitude modulated to
a higher frequency that is reserved for that signal. It is then transmitted at that frequency to
the receiver. The following example illustrates how the receiver can recover the transmitted
signal.
Example :
Consider the signal x(t) = s(t)p(t) from the previous example. Its frequency spectrum
has two copies of the spectrum of s(t), located at ±ω0 . We want to recover the original
signal s(t) from x(t). To do this, suppose we multiply x(t) by cos (ω0 t) again, to obtain
87
88 The Discrete-Time Fourier Transform
where ω0 = 2π
N
. The Fourier series coefficients are given by
+N −1
n0X
1
ak = x̃[n]e−jω0 kn
N n=n0
where n0 is any integer. Suppose we choose n0 so that the interval [−N1 , N2 ] is contained in
[n0 , n0 + N − 1]. Then since x̃[n] = x[n] in this interval, we have
+N −1
n0X ∞
1 1 X
ak = x[n]e−jω0 kn = x[n]e−jω0 kn
N n=n0
N n=−∞
Let us now define the discrete-time Fourier transform as
∞
X
jω
x[n]e−jωn .
X e ≜
n=−∞
From this, we see that ak = , i.e., the discrete-time Fourier series coefficients are
1 jkω0
e N
X
obtained by sampling the discrete-time Fourier transform at periodic intervals of ω0 . Also
note that X (ejω ) is periodic in ω with period 2π (since e−jωn is 2π-periodic).
Using the Fourier series representation of x̃[n], we now have
N −1 N −1 N −1
X
jkω0 n 1 X jkω0
jkω0 n 1 X
X ejkω0 ejkω0 n ω0
x̃[n] = ak e = X e e =
k=0
N k=0 2π k=0
Once again, we see that each term in the summand represents the area of a rectangle of
width ω0 obtained from the curve X (ejω ) ejω . As N → ∞, we have ω0 → 0. In this case,
the sum of the areas of the rectangles approaches the integral of the curve X (ejω ) ejωn , and
since the sum was over only N samples of the function, the integral is only over one interval
of length 2π. Since x̃[n] approaches x[n] as N → ∞, we have
Z
1
X ejω ejωn dω
x[n] =
2π 2π
This is the inverse discrete-time Fourier transform, or the synthesis equation.
The main differences between the discrete-time and continuous-time Fourier transforms
are the following. (1) The discrete-time Fourier transform X (ejω ) is periodic in ω with
The Discrete-Time Fourier Transform 89
period 2π, whereas the continuous-time Fourier transform is not necessarily periodic. (2)
The synthesis equation for the discrete-time Fourier transform only involves an integral over
an interval of length 2π, whereas the one for the continuous-time Fourier transform is over
the entire ω axis. Both of these are due to the fact that ejωn is 2π-periodic in ω, whereas the
continuous-time complex exponential is not.
Since the frequency spectrum of X (ejω ) is only uniquely specified over an interval of length
2π, we have to be careful about what we mean by "high" and "low" frequencies. Recalling
the discussion of discrete-time complex exponentials, high-frequency signals in discrete-time
have frequencies close to odd multiples of π, whereas low-frequency signals have frequencies
close to even multiples of π.
Example :
Consider the signal
∞
X ∞
X
X ejω = x[n]e−jωn = a|n| e−jωn
n=−∞ n=−∞
−1
X ∞
X
−n −jωn
= a e + an e−jωn
n=−∞ n=0
X∞ ∞
X
= n jωn
a e + an e−jωn
n=1 n=0
jω
ae 1
= +
1 − aejω 1 − ae−jω
2
1−a
=
1 − 2a cos(ω) + a2
■
x[n] = ejω0 n .
We claim that the Fourier transform of this signal is
∞
X
X ejω =
2πδ (ω − ω0 − 2πl)
l=−∞
i.e., a set of impulse functions spaced 2π apart on the frequency axis. To verify this, note
that the inverse Fourier transform is given by
Z
1
X ejω ejωn dω
2π 2π
The integral is only over an interval of length 2π, and there is at most one impulse function
from X (ejω ) in any such interval. Let that impulse be located at ω0 + 2πr for some r ∈ Z.
Then we have
Z Z
1 jω jωn 1
2πδ (ω − ω0 − 2πr) ejωn dω = ej(ω0 +2πr)n = ejω0 n
X e e dω =
2π 2π 2π 2π
where ω0 = 2πN
. The Fourier transform of each term of the form ak ejkω0 n is a set of impulses
spaced 2π apart, with one located at ω = kω0 . Furthermore each of these impulses is scaled
by ak 2π. Since ak = ak+N l for any l (by the periodicity of the discrete-time Fourier series
coefficients), when we add up the Fourier transforms of all of the terms in the Fourier series
expansion of x[n], we obtain
∞
jω
X 2πk
X e = 2πak δ ω −
k=−∞
N
N −1
1 X 1
ak = x[n]e−jkω0 n =
N k=0 N
X ejω = X ej(ω+2π)
This comes out of the fact that discrete-time complex exponentials are periodic in fre-
quency with period 2π.
3.2 Linearity
It is easy to see that
F {x [n − n0 ]} = e−jωn0 X ejω
and
The first property is easily proved using the inverse Fourier transform equation, and the
second property is proved using the Fourier transform equation.
Example :
Consider a discrete-time low-pass filter, whose Fourier transform Hlp (ejω ) is a square pulse
centered at even multiples of π. Now consider the high pass filter Hhp (ejω ) which consists
of square pulses centered at odd multiples of π. We see that Hhp (ejω ) = Hlp ej(ω−π) .
Thus we have
3.5 Conjugation
For any discrete-time signal (that has a Fourier transform), we have
X ejω = X ∗ e−jω
3.6 Time-Reversal
Consider the time-reversed signal x[−n]. We have
Z 2π Z 2π
1 −jωn 1
jω
X e−jω ejωn dω
x[−n] = X e e dω =
2π 0 2π 0
The Discrete-Time Fourier Transform 93
which is obtained by performing a change of variable ω → −ω (note that the negative sign
introduced by this change is canceled out by the reversal of the bounds of integration that
arise because of the negation). Thus, we have
F{x[−n]} = X e−jω
Together with the conjugation property, we see that for real-valued even signals (where
x[n] = x[−n] ), we have
Thus, the Fourier transform of real, even signals is also real and even.
Thus, the signal x(k) [n] is obtained by spreading the points of x[n] apart by k samples and
placing zeros between the samples. We have
∞
X ∞
X
−jωn
x(k) [rk]e−jωrk
F x(k) [n] = x(k) [n]e =
n=−∞ r=−∞
since x(k) [n] is nonzero only at integer multiples of k. Since x(k) [rk] = x[r], we have
∞
X
x[r]e−jωrk = X ejkω
F x(k) [n] =
r=−∞
Example :
Consider the signal
94 The Discrete-Time Fourier Transform
1 n ∈ {0, 2, 4}
x[n] = 2 n ∈ {1, 3, 5}
0 otherwise
sin(3ω)
G ejω = H ej2ω = e−2jω
sin(ω)
Finally,
sin(3ω)
X ejω = G ejω + 2e−jω G ejω = e−2jω 1 + 2e−jω
sin(ω)
■
dX (ejω )
F{nx[n]} = j
dω
∞ Z
X 1 2 2
X ejω
|x[n]| = dω
n=−∞
2π 2π
3.9 Convolution
Just as in continuous time, the discrete time signal ejωn is an eigenfunction of discrete-time
LTI systems. Specifically, if ejωn is applied to a (stable) LTI system with impulse response
h[n], the output of the system will be H (ejω ) ejωn .
Thus consider a signal x[n] written in terms of its Fourier transform as
Z
1
X ejω ejωn dω
x[n] =
2π 2π
This is a linear combination of complex exponentials (where the scaling factor on the
complex exponential ejωn is 2π
1
X (ejω ). By the LTI property, we thus have
Z Z
1 jω jωn 1
X ejω H ejω ejωn dω = y[n]
x[n] = X e e dω →
2π 2π 2π 2π
The expression on the right hand side is the output y[n] of the system when the input is
x[n]. Thus we have
The signal w3 [n] is given by w3 [n] = (−1)n w2 [n], and thus W3 (ejω ) = W2 ej(ω−π) .
From the bottom path, we have W4 (ejω ) = Hlp (ejω ) X (ejω ). Thus, we have
Recall that Hlp ej(ω−π) is a high-pass filter centered at π. Thus, this system acts as a
bandstop filter, blocking all frequencies in a certain range and letting all of the low and high
frequency signals through.
96 The Discrete-Time Fourier Transform
3.10 Multiplication
Consider two signals x1 [n] and x2 [n], and define g[n] = x1 [n]x2 [n]. The discretetime Fourier
transform of g[n] is given by
∞
X
jω
x1 [n]x2 [n]e−jωn
G e =
n=−∞
sin π2 n sin π4 n
x1 [n] = , x2 [n] =
πn πn
The Fourier transforms of these signals are square pulses, where the pulse centered at
0 extend from − π2 to π2 (for X1 (ejω ) ) and from − π4 to π4 (for X2 (ejω ) ). The Fourier
transform of g[n] = x1 [n]x2 [n] is given by
Z
jω 1
X1 ejθ X2 ej(ω−θ) dθ
G e =
2π 2π
Since we can choose any interval of length 2π to integrate over, let’s choose the interval
[−π, π) for convenience. We also only need to determine the values of the Fourier transform
for values of ω between −π and π, since the transform is periodic. Depending on the value
of ω, there are different cases that occur:
• If − 3π4
≤ ω < − π4 , then there is partial overlap in the signals; the product is a
rectangle with support from − π2 to ω + π4 , and thus G(jω) evaluates to 2π
1
ω + 3π
4
.
• If 3π
4
≤ ω < π, there is no overlap and G(jω) is zero.
Note that since we are only integrating over θ between −π and π, the values of X ejθ
outside of that interval does not matter. Thus, we could also create a new signal jθ
X̂ 1 e
which is equal to X1 ejθ over the interval [−π, π) and zero everywhere else. The Fourier
Z Z ∞
jω 1 jθ j(ω−θ) 1
X̂1 ejθ X2 ej(ω−θ) dθ
G e = X1 e X2 e dθ =
2π 2π 2π −∞
i.e., it is the usual convolution of the signals X̂1 (ejω ) and X2 (ejω ). ■
98 The Discrete-Time Fourier Transform
CHAPTER 7
SAMPLING
99
100 Sampling
Note that the values of the signal x(t) are irrelevant outside of the points where the impulse
functions in p(t) occur (i.e., at the sampling instants). Let us consider the frequency spectra
of these signals. Specifically, by the multiplication property of Fourier transforms, we have
Z ∞
1
Xp (jω) = X(jθ)P (j(ω − θ))dθ
2π −∞
Furthermore, since p(t) is periodic, we saw that the Fourier transform of p(t) will be given
by
∞
2π X
P (jω) = δ (ω − nωs )
Ts n=−∞
Thus, the frequency spectrum of xp (t) consists of copies of the frequency spectrum of x(t),
where each copy is shifted (in frequency) by an integer multiple of the sampling frequency
ωs and scaled by T1s (see Fig. 7.1).
Sampling 101
Figure 7.1: The frequency spectrum of the signal x(t) and the signal xp (t).
If we want to be able to reconstruct x(t) from its sampled version xp (t), we would like to
make sure that there is an exact copy of X(jω) that can be extracted from Xp (jω). Based
on the above discussion, we see that this will be the case if no two copies of X(jω) overlap
in Xp (jω). Looking at Fig. 7.1, this will occur as long as
ωs − ωM > ωM
or equivalently,
ωs > 2ωM
where ωM is the largest frequency at which x(t) has nonzero content. This leads to the
sampling theorem.
If the sampling frequency ωs is larger than twice the largest frequency of the signal x(t),
then we can reconstruct the signal x(t) from its sampled version xp (t) by passing xp (t) through
an ideal low-pass filter, with cutoff ωc = ω2s .
The frequency 2ωM is called the Nyquist rate.
Figure 7.2: The impulse response of a zero-order-hold (left) and a first-orderhold (right).
point all samples are zero). This is shown in Fig. 7.2. It is easy to check that the transfer
function is given by
∞
sin ω T2s
Z
−jω T2s
H0 (jω) = h0 (t)e−jωt dt = e ω
−∞ 2
This has magnitude Ts at ω = 0 (like the ideal reconstructor), and the first frequency
at which it is equal to zero is at ωs (unlike the ideal reconstructor that cuts off at ω2s .
Furthermore, this frequency spectrum is not bandlimited, and thus the copies of X(jω) in
the spectrum of Xp (jω) will leak into the reconstructed signal under the ZOH .
!2
1 sin ω T2s 1
H1 (jω) = ω = |H0 (jω)|2
Ts 2
Ts
The magnitude of this filter is smaller than that of H0 (jω) outside of ω2s , although it is
still not bandlimited. Furthermore, the FOH is noncausal, but can be made causal with a
delay of Ts .
Higher order filters are also possible, and can be defined as a natural extension of zero
and first order holds.
Sampling 103
One the other hand, if we take the discrete-time Fourier transform of the sequence xd [n],
104 Sampling
we have
∞
X ∞
X
−jωn
jω
x (nTs ) e−jωn
Xd e = xd [n]e =
n=−∞ n=−∞
107
108 The Laplace Transform
Based Ron the above, we see that the output is the input signal est , multiplied by the
∞
quantity −∞ h(t)e−sτ dτ . We will call this the Laplace transform of the signal h(t).
The Laplace transform of a signal x(t) is given by
Z ∞
X(s) = x(t)e−st dt
−∞
x(t) by L{x(t)}.
Note that the limits of the integration go from −∞ to ∞, and thus this is called the
bilateral Laplace transform. When the limits only go from 0 to ∞, it is called the unilateral
Laplace transform. For the purposes of this course, if we leave out the qualifier, we mean
the bilateral transform. Note that when s = jω, then X(s) is just the Fourier transform of
x(t) (assuming the transform exists). However, the benefit of the Laplace transform is that
it also applies to signals that do not have a Fourier transform. Specifically, note that s can
be written as s = σ + jω, where σ and ω are real numbers. Then we have
Z ∞
X(s) = X(σ + jω) = x(t)e−σt e−jωt dt
−∞
Thus, for a given s = σ + jω, we can think of the Laplace transform as the Fourier
transform of the signal x(t)e−σt . Even if x(t) is not absolutely integrable, it may be possible
that x(t)e−σt is absolutely integrable if σ is large enough (i.e., the complex exponential can
be chosen to cancel out the growth of the signal in the Laplace transform).
Example :
Consider the signal x(t) = e−at u(t) where a is some real number. The Laplace transform
is given by
Z ∞ Z ∞
−st
X(s) = x(t)e dt = e−(s+a)t dt
−∞ 0
∞
1 −(s+a)t
=− e
s+a 0
1
=
s+a
as long as Re{s + a} > 0, or equivalently Re{s} > −a. Note that if a is positive, then
the integral converges for Re{s} = 0 as well, in which case we get the Fourier transform
The Laplace Transform 109
X(jω) = jω+a1
. However, if a is negative, then the signal does not have a Fourier transform
(but it does have a Laplace transform for s with a sufficiently large real part). ■
Example :
Consider the signal x(t) = −e−at u(−t) where a is a real number.
We have
Z ∞ Z 0
−st
X(s) = x(t)e dt = − e−(s+a)t dt
−∞ −∞
0
1 −(s+a)t
= e
s+a −∞
1
=
s+a
as long as Re{s + a} < 0, or equivalently, Re{s} < −a. ■
Comparing the above examples, we notice that both the signals e−at u(t) and −e−at u(−t)
had the same Laplace transform s+a 1
, but that the ranges of s for which each had a Laplace
transform was different.
Consider a signal x(t). The range of values of s for which the Laplace transform integral
converges is called the Region of Convergence (ROC) of the Laplace transform.
Thus, in order to specify the Laplace transform of a signal, we have to specify both the
algebraic expression (e.g., s+a
1
) and the region of convergence for which this expression is
valid. A convenient way to visualize the ROC is as a shaded region in the complex plane. For
example, the ROC Re{s} > −a can be represented by shading all of the complex plane to
the right of the line Re{s} = −a. Similarly, the ROC Re{s} < −a is represented by shading
the complex plane to the left of the line Re{s} = −a.
Example :
Consider the signal x(t) = 3e−2t u(t)−2e−t u(t). It is easy to see that the Laplace transform
is a linear operation, and thus we can find the Laplace transform of x(t) as a sum of the
Laplace transform of the two signals on the right hand side.
The Laplace transform of 3e−2t u(t) is s+2
3
, with ROC Re{s} > −2. The Laplace transform
of −2e u(t) is − s+1 , with ROC Re{s} > −1. Thus, for the Laplace transform of x(t)
−t 2
to exist, we need s to fall in the ROC of both of its constituent parts, which means
Re{s} > −1. Thus,
3 2 s−1
X(s) = − = 2
s+2 s+1 s + 3s + 2
with ROC Re{s} > −1. ■
In the above examples, we saw that the Laplace transform was of the form
N (s)
X(s) =
D(s)
where N (s) and D(s) are polynomials in s. The roots of the polynomial N (s) are called
the zeros of X(s) (since X(s) will be zero when s is equal to one of those roots), and the
110 The Laplace Transform
roots of D(s) are called the poles of X(s) (evaluating X(s) at a pole will yield ∞). We can
draw the poles and zeros in the s-plane using ◦ for zeros and × for poles.
Example :
Consider the signal
4 1
x(t) = δ(t) − e−t u(t) + e2t u(t)
3 3
The Laplace transform of δ(t) is
Z ∞
L{δ(t)} = δ(t)e−st dt = 1
−∞
for any value of s. Thus the ROC for δ(t) is the entire s-plane. Putting this with the
other two terms, we have
4 1 1 1 (s − 1)2
X(s) = 1 − + =
3s+1 3s−2 (s + 1)(s − 2)
with ROC Re{s} > 2. ■
Another way to think of the above property is as follows. No matter what σ we pick, the
signal x(t)e−σt will be absolutely integrable as long as x(t) is of finite duration and absolutely
integrable. The fact that x(t) is of finite duration allows us to overcome the fact that the
signal e−σt may be growing unboundedly outside of the interval [T1 , T2 ].
While the previous property considered the case where the signal is of finite duration, we
will also be interested in signals that are only zero either before or after some time. First,
a signal x(t) is right-sided if there exists some T1 ∈ R such that x(t) = 0 for all t < T1 . A
signal x(t) is left-sided if there exists some T2 ∈ R such that x(t) = 0 for all t > T2 . A signal
x(t) is two-sided if it extends infinitely far in both directions.
Property 4. If x(t) is right-sided and if the line Re{s} = σ0 is in the ROC, then
the ROC contains all values s such that Re{s} ≥ σ0 .
To see why this is true, first note that since x(t) is right-sided, there exists some T1 such
that x(t) = 0 for all t < T1 . If s with Re{s} = σ0 is in the ROC, then x(t)e−σ0 t is absolutely
integrable, i.e.,
Z ∞
|x(t)|e−σ0 t dt < ∞
T1
Now suppose that we consider some σ1 > σ0 . If T1 > 0, then e−σ1 t is always smaller than
e over the region of integration, and thus x(t)e−σ1 t will also be absolutely integrable. If
−σ0 t
T1 < 0, then
Z ∞ Z 0 Z ∞
−σ1 t −σ1 t
|x(t)|e dt = |x(t)|e dt + |x(t)|e−σ1 t dt
T1 T1 0
Z 0 Z ∞
−σ1 t
≤ |x(t)|e dt + |x(t)|e−σ0 t dt
T1 0
The first term is finite (since it is integrating some signal of finite duration), and the
second term is finite since x(t)e−σ0 t is absolutely integrable. Thus, once again, x(t)e−σ1 t is
absolutely integrable, and thus s with Re{s} ≥ σ0 also falls within the ROC of the signal.
The same reasoning applies to show the following property.
Property 5. If x(t) is left-sided and if the line Re{s} = σ0 is in the ROC, then
the ROC contains all values s such that
Re{s} ≤ σ0 .
If x(t) is two-sided, we can write x(t) as x(t) = xR (t) + xL (t), where xR (t) is a right-sided
signal and xL (t) is a left-sided signal. The former has an ROC that is the region to the right
of some line in the s-plane, and the latter has an ROC that is the region to the left of some
line in the s-plane. Thus, the ROC for x(t) contains the intersection of these two regions (if
there is no intersection, x(t) does not have a Laplace transform).
Property 6. If x(t) is two-sided and contains the line Re{s} = σ0 in its ROC,
then the ROC consists of a strip in the s-plane that contains the line Re{s} = σ0 .
Example :
Consider the signal x(t) = e−b|t| . We write this as
112 The Laplace Transform
Since this is just the Fourier transform of x(t)e−σt , we can use the inverse Fourier transform
formula to obtain
The Laplace Transform 113
Z ∞
−σt 1
x(t)e = X(σ + jω)ejωt dω.
2π −∞
This is the inverse Fourier transform formula. It involves an integration over the line in the
complex plane consisting of points satisfying Re{s} = σ. There are actually simpler ways to
calculate the inverse Fourier transform, using the notion of partial fraction expansion, which
we will consider here.
Example :
Consider X(s) = 1
s(s+1)
. First, we note that
1 1 1
= −
s(s + 1) s s+1
Now each of these terms is of a form that we know (they correspond to complex expo-
nentials). So, for example, if the ROC for X(s) is the region to the right of the imaginary
axis, since the ROC consists of the intersection of the ROCs of both of the terms, we know
that both terms must be right-sided signals. Thus,
(s + z1 ) (s + z2 ) · · · (s + zm )
X(s) = K
(s + p1 ) (s + p2 ) · · · (s + pn )
Recall that the zeros of X(s) are given by −z1 , −z2 , . . . , −zm , and the poles are −p1 , −p2 , . . . , −pn .
First, suppose each of the poles are distinct and that X(s) is strictly proper. We would like
to write
k1 k2 kn
X(s) = + + ··· +
s + p1 s + p 2 s + pn
for some constants k1 , k2 , . . . , kn , since the inverse Laplace Transform of X(s) is easy in
this form. How do we find k1 , k2 , . . . , kn ?
Heaviside’s Cover-up Method. To find the constant ki , multiply both sides of the expansion
of X(s) by (s + pi ) :
k1 (s + pi ) k2 (s + pi ) kn (s + pi )
(s + pi ) X(s) = + + · · · + ki + · · · + .
s + p1 s + p2 s + pn
Now if we let s = −pi , then all terms on the right hand side will be equal to zero, except
for the term ki . Thus, we obtain
ki = (s + pi ) X(s)|s=−pi
Example :
Consider X(s) = s+5
s3 +3s2 −6s−8
. The denominator is factored as
4 4
k1 = (s + 1)X(s)|s=−1 = =−
(−3)(3) 9
7 7
k2 = (s − 2)X(s)|s=2 = =
(3)(6) 18
1 1
k3 = (s + 4)X(s)|s=−4 = =
(−3)(−6) 18
■
The partial fraction expansion when some of the poles are repeated is obtained by following
a similar procedure, but it is a little more complicated. We will not worry too much about this
scenario here. One can also do a partial fraction expansion of nonstrictly proper functions by
first dividing the denominator into the numerator to obtain a constant and a strictly proper
function, and then applying the above partial fraction expansion.
4.1 Convolution
Consider two signals x1 (t) and x2 (t) with Laplace transforms X1 (s) and X2 (s) and ROCs R1
and R2 , respectively. Then
4.2 Differentiation
Consider a signal x(t), with Laplace transform X(s) and ROC R. We have
Z
1
x(t) = X(s)est ds
2π
Differentiating both sides with respect to t, we have
Z
dx(t) 1
= sX(s)est ds
dt 2π
116 The Laplace Transform
4.3 Integration
Rt
Given a signal x(t) whose Laplace transform has ROC R, consider the integral −∞
x(τ )dτ .
Note that
Z t
x(τ )dτ = u(t) ∗ x(t)
−∞
and thus using the convolution property, we have
Z t
1
L x(τ )dτ = X(s)
−∞ s
with ROC containing R ∩ {Re{s} > 0}.
Y (s) = H(s)X(s)
assuming all Laplace transforms exist. Using the expressions for H(s) and X(s), we can
thus calculate Y (s) (and its ROC), and then use an inverse Laplace transform to determine
y(t).
Example :
Consider an LTI system with impulse response h(t) = e−2t u(t). Suppose the input is
x(t) = e−3t u(t). The Laplace transforms of h(t) and x(t) are
1 1
H(s) = , X(s) =
s+2 s+3
with ROCs Re{s} > −2 and Re{s} > −3, respectively. Thus we have
1 1
Y (s) = H(s)X(s) =
s+2s+3
The Laplace Transform 117
with ROC Re{s} > −2. Using partial fraction expansion, we have
1 1
Y (s) = −
s+2 s+3
and thus y(t) = e−2t u(t) − e−3t u(t). ■
Example :
Consider an LTI system with impulse response h(t) = −e4t u(−t) and input x(t) = e2t u(t),
where we interpret u(−t) as being 1 for t < 0. The Laplace transforms are
1 1
H(s) = , X(s) =
s−4 s−2
with ROCs Re{s} < 4 and Re{s} > 2, respectively. Since there is a nonempty inter-
section, we have
1 1 1 1 1 1
Y (s) = H(s)X(s) = = −
s−4s−2 2s−4 2s−2
with ROC 2 < Re{s} < 4. Thus, y(t) is two-sided, and given by
1 1
y(t) = − e4t u(−t) + e2t u(t)
2 2
■
or equivalently
Pm
bk s k
Y (s) = Pnk=0 k
X(s).
k=0 ak s
| {z }
H(s)
Thus, the impulse response of the differential equation is just the inverse Laplace transform
of H(s) (corresponding to an appropriate region of convergence).
118 The Laplace Transform
Example :
Consider the differential equation
Y (s) 1 1
H(s) = = 3 2
=
X(s) s + 2s − s − 2 (s − 1)(s + 1)(s + 2)
In this case, we have the partial fraction expansion
1 1 1 1 1 1
H(s) = − +
6s−1 2s+1 6s+2
Suppose we are told the impulse response is causal (which implies it is rightsided).
Thus, the ROC would be to the right of the furthest pole and we have
1 t 1 −t 1 −2t
h(t) = e − e + e u(t)
6 2 6
■
Part II
119
CHAPTER 1
GRAPHS SIGNALS AND SYSTEMS
121
122 Graphs Signals and Systems
1 Graph Terminology
We denote a weighted graph by G = (V, E, W), where V = {1, . . . , N } is the set of nodes,
E ⊆ V × V is the set of edges such that (i, j) ∈ E if and only if there is an edge from node i
to node j, and W : E → R+ is a weight function. If all edge weights equal one, the graph is
said unweighted.
Graphs can be directed or undirected. In an undirected graph, there is no orientation
in the edges in E, W(i, j) = W(j, i), and the neighboring set for a node i is denoted by
Ni = {j ∈ V : (i, j) ∈ E}. In a directed graph (or digraph), an edge (i, j) ∈ E has an
orientation starting from node i and ending at node j. We say that node j is an out-neighbor
of i (and i an in-neighbor of j ). The out-neighboring set of node i is denoted by Niout =
{j ∈ V : (i, j) ∈ E} and, likewise, the in-neighboring set by Niin = {j ∈ V : (j, i) ∈ E}.
We represent graph G via the weighted adjacency matrix A, which is an N × N sparse
matrix with nonzero elements [A]ji = aji = W(i, j) > 0 representing the strength of edge
(i, j) ∈ E. Matrix A is symmetric aij = aji for an undirected graph and asymmetric aji ̸=
aij for a directed one. For undirected graphs, another widely used matrix is the graph
Laplacian L = D − A, where D = diag(A1) is a diagonal matrix whose P i th diagonal
element [D]ii = dii is the sum of all edge weights incident to node i, i.e., dii = j∈Ni aij [7].
Both matrices A and L are special cases for S [5], [21]. Other candidate GSOs include the
normalized adjacency matrix An = D−1/2 AD−1/2 , the normalized Laplacian matrix Ln =
D−1/2 LD−1/2 or the random walk Laplacian Lrw = D−1 L.
A graph and associated GSO can represent:
1. Physical networks: Here, nodes and edges physically exist. For example, in a sensor
network, nodes are sensors and edges are communication links [32]. A directed edge
indicates the communication direction and the edge weight captures the communication
channel properties. Other examples include:
(a) multi-agent robot systems where nodes are robots and edges are communication
channels [33];
(b) power networks where nodes are buses and edges are power lines [34];
(c) telecommunication networks where nodes are transceivers and edges are channels
[35], [36];
(d) water networks where nodes are junctions and edges are pipes [37], and;
(e) road networks where nodes are intersections and edges are roads [38].
Graphs Signals and Systems 123
2. Abstract networks: These graphs typically represent dependencies between the dat-
apoints. Consider N datapoints, each described by a feature vector fi ∈ RF , and let
dist (fi , fj ) be a distance measure (e.g., Euclidean) between datapoints i and j. Each
datapoint is considered as a node and two datapoints could be connected based on [39]:
The above approaches build undirected abstract networks based on similarities. Al-
ternatives for directed or causal dependencies are also possible; see [40]-[42]. In fact,
identifying the most suitable abstract network from the data is a wide research area by
itself.
1. recommender systems, where two items are connected, e.g., if their Pearson correlation
is greater than some value [43];
2. brain networks, where the nodes are brain regions and the edges are given, e.g., by the
cross-correlation or mutual information of electroencephalography (EEG) time series
in the different regions [44];
3. social networks, where nodes are users and edge weights may represent the number of
interactions between them; iv ) economic networks, where nodes are different economic
sectors and the edges maynrepresent the input and output production going from one
sector to another [45].
Since abstract networks represent dependencies between datapoints, they can be manipulated
by recomputing edge weights, clustering, or pruning to facilitate representation. However,
this is not typically the case for physical networks, as they often represent the medium
with respect to which processing is performed. Graph filters can leverage such structure for
distributed processing.
124 Graphs Signals and Systems
We denote the space of all graph signals defined on graphs with N nodes by XN = {x :
V → R, |V| = N }. An example of a graph signal is a recording in a brain network, i.e., each
brain area corresponds to a node, two nodes share a link based on structural connectivity,
and the brain EEG measurement is the signal of a particular node. We may want to process
such a signal to understand, e.g., how different individuals have mastered a specific task [46].
Processing and learning tasks with graph signals include:
2. Signal compression: When graph signals feature nice patterns such as having similar
values at neighboring vertices, it is possible to compress the signal (with or without
loss of information) by developing representations that require fewer coefficients, and
storing those coefficients rather than the original signal.
3. Signal classification: This task consists of classifying different graph signals observed
over a common underlying graph. One such example is classifying patients based on
their brain recordings, as discussed above [46].
4. Node classification: This task consists of classifying a subset of nodes in the graph
given the class labels on another subset. Depending on the available information, there
are two ways to approach this problem. First, when node features are available we
can treat them as a collection of graph signals and leverage their coupling with the
underlying connectivity to infer the missing labels. The state-of-the-art for this task is
achieved by GNNs, which, as we shall see in Section VIII rely heavily on graph filters
[23]. Second, when node features are unavailable, we treat the available labels as graph
signals and transform node classification into a label interpolation task that can be
solved with graph filters [22].
5. Graph classification / regression: These tasks start with a collection of different graphs
and (optionally) graph signals.
Graphs Signals and Systems 125
Figure 1.1: The graph convolution as a shift register. Highlighted are the nodes that reach
node 1 on each consecutive shift; that is, the nodes j whose signal value xj contributes to
S x i . The resulting summary of each communication S x is correspondingly weighted by
k k
a filter parameter hk . For each k, the parameter hk is the same for all nodes.
The classification task assigns a label to the whole graph; e.g., classifying molecules
into different categories such as soluble vs. non-soluble, whereas the regression task
assigns a continuous number to each graph (e.g., the degree of solubility) [48].
6. Link prediction: Here, the goal is to infer if two nodes are connected given the current
structure and the respective graph signals [49]. This is the case of friend recommenda-
tion in social networks, whereby leveraging the friendship connectivity between users
based on their feature signals (e.g., geo-position, workplace) we can infer missing links.
7. Graph identification: This task extends the link prediction to that of inferring the whole
graph structure given only the graph signals [41]. Here graph filters play a central role
in modeling the relationships between candidate graph structures and the observed
signals. We detail this problem in Sec. IX-C.
8. Distributed processing: Here the graph topology represents the structure of a sensor
network and we want to distributively solve a task related to graph signals. Graph filters
lend themselves naturally to this setup because they rely only on local information. We
shall discuss in Sec. IX-E their use for different tasks such as distributed denoising,
reconstruction, and consensus.
A convolutional filter is defined by a shift-and-sum operation of the input signal [51]. While a
shift in time implies a delay, a graph signal shift requires taking into account the underlying
topological structure.
If the GSO is the adjacency matrix A, the shifted signal represents a one-step propagation.
Instead, if the GSO is the graph Laplcaian
P L, the shifted signal is a weighted difference of
the signals at neighboring nodes [Lx]i = j∈Ni aij (xi − xj ).
Upon defining the signal shift, a graph convolutional filter is simply a weighted sum of
multiple shifted signals.
implies that there exists at least one path of length k between nodes i and j through which
the signals can diffuse. These signals are shifted repeatedly over the graph as per (1.3); see
also 1.1. The term convolution for (1.4) is rooted in the algebraic extension of the convolution
operation [20] and the discrete-time counterpart can be seen as a particular case over a cyclic
graph; see the box of 1.2 Because of this analogy, the filter in (1.4) is also referred to as a
finite impulse response graph filter of order K.
Graphs Signals and Systems 127
Discrete-time circular convolution. The graph signal shift (1.3), the graph convo-
lutional filter (1.4), and their spectral equivalents in Sec. IV generalize the respective
concepts developed for discrete-time periodic signals.
Figure 1.2: Discrete-time periodic signals as graph signals over a directed cycle graph.
Each node Vc = {1, . . . , 6} is a time instant with adjacencies captured in the matrix
Ac . The temporal signal forms the graph signal x = [x1 , . . . , x6 ]⊤ and the shift Ac x
acts as a delay operation that moves the signal to the next time instant node.
3.2 Properties
Graph convolutional filters satisfy the following properties.
Property 1 (Linear). The convolution is linear in the input signal. For two inputs x1 and
x2 and filter H(S) it holds that
where α, β are scalars; i.e., the linear combination of the outputs equals the output of the
inputs’ linear combination.
Property 2 (Shift invariance). The graph convolution is invariant to shifts, i.e.,
SH(S) = H(S)S. This implies that given two filters H1 (S) and H2 (S) with respective
parameters h1 and h2 operating over the same graph and input signal x, it holds that we
can switch the filters order
P = P ∈ {0, 1}N ×N : P1 = 1, P⊤ 1 = 1
Then for a graph with GSO S and P ∈ P, the permuted graph has the GSO Ŝ = P⊤ SP,
which describes the same topology but with a reordering of the nodes. Likewise, the permuted
signal corresponding to the ordering in Ŝ is x̂ = P⊤ x. Permutation equivariance for filter (1.4)
implies
H(Ŝ)x̂ = P⊤ H(S)x
i.e., the filter output operating on the permuted graph Ŝ with the permuted signal x̂ is
the permuted output of the same filter operating on the original graph S with the original
signal x.
Thus, graph convolutions are independent of the arbitrary ordering of the nodes. And
as temporal and spatial convolutions exploit symmetries in the Euclidean domain by being
translation equivariant, graph convolutions exploit symmetries in the graph domain by being
permutation equivariant. This is key to their success in learning input-output mappings from
a few training samples [4].
Property 4 (Parameter sharing). All the nodes share the parameters K among them. For
two nodes i, j, the respective outputs are yi = h0 xi +h1 [Sx]i +. . .+hK S x i and yj = h0 xj +
h1 [Sx]j + . . . + hK SK x j , which shows that the k-shifted signal Sk x is weighted by the same
parameter hk .
Props. 364 imply that graph convolutions are inductive processing architectures. That is,
they can be designed or trained over a graph G and transferred to another graph G (with
possibly a different number of nodes) without redesigning or retraining for this specific graph
[53]. This is particularly useful, e.g., when using graph filters for distributed SP tasks, as
the physical channel graph may change. In Sec. IV (Prop. 8), we characterize the degree of
transference for specific GCFs.
Property 5 (Locality). Graph convolutions are local architectures. To see this, set
z(0) = S0 x. The one shifted signal z(1) = Sx = Sz(0) is local by definition. The k > 1 shift
z(k) = Sk x can be computed recursively as z(k) = S S(k−1) x = Sz(k−1) , which implies that
the (k − 1) st shift z(k−1) needs to be shifted locally to the neighbors. Hence, to compute the
filter output, each node exchanges locally with neighbors all K shifts z(0) , . . . , z(K−1) .
Locality of computation makes the graph convolutional filters readily distributable, as we
discuss in Sec. IX-E
Graphs Signals and Systems 129
x̃ = V−1 x
and likewise, the inverse GFT is x = Vx̃.
In the definition of GFT, we are assuming the GSO S is diagonalizable. While definitions
of GFT for nondiagonalizable GSOs exist [22], [54], [55], we hold to the diagonalizability
assumption for a consistent and simple exposition.
The eigenvectors of S in the columns of V serve as the basis expansion for the GFT. In the
discrete-time case, the complex exponentials fulfill this role. The coefficients x̃ are the weights
each of these eigenvectors contribute to represent the signal. Following again this analogy,
the vector λ contains the so-called graph frequencies. Interpreting these graph frequencies λ
and the respective GFT coefficients x̃ requires understanding how the signal varies over the
graph. In turn, measuring variability requires defining a criterion that accounts for the graph
structure. Here, we review two basic criteria used for undirected [19] and directed graphs
[21], [22].
Undirected graphs. The local signal difference at two connected nodes i and j can be
√
computed as ∆xij = aij (xi − xj ), which is higher when the signal in two strongly connected
nodes is substantially different. The local signal variability at a node i accounts for the
variabilities in its neighboorhood
130 Graphs Signals and Systems
sX ! 21
X
∆xi = (∆xij )2 = aij (xi − xj )2
j∈Ni j∈Ni
which is higher if the signal at node i differs from that of its neighboors. Then, the
variability of a signal x over an undirected graph is the squared-sum of all local variabilities
1X
TV2 (x) = (∆xi )2 = x⊤ Lx
2 i∈V
The TV2 (x) is the Laplacian quadratic form for signal x and quantifies how much the
signal varies over the graph. In fact, the constant graph signal x = c1 has a zero variability.
We can use the quadratic form (6) to interpret the variability of the Laplacian eigenvectors
L = V diag(λ)V⊤ . Treating each eigenvector vi as a graph signal, we have
x = VK x̃K
Graphs Signals and Systems 131
with VK = [v1 , . . . , vK ] and x̃K ∈ RK . Without loss of generality, we assume the first K
⊤
eigenvectors express the signal variability, i.e., the GFT coefficients are x̃ = x̃⊤ ⊤
.
K , 0N −K
Fig. 3: The frequency response of the filter 12], given in the black solid line, is completely
characterized by the values of the filter parameters h. Given a graph, this frequency response
gets instantiated on the specific eigenvalues of that graph, determining the effect the filter
will have on an input 11.
5 Frequency Response
By subsituting the eigendecomposition S = V diag(λ)V−1 into (3), we can write the filter
output as
K
X K
X
k
hk V diag λ⊙k V−1 x
y= hk S x =
k=0 k=0
where λ⊙k ∈ CN : λ⊙k i := λki . Using then (4) and defining the GFT of the output
K
X
hk diag λ⊙k x̃
ỹ =
k=0
Convolution theorem for graph filters. It follows from (11) that any shift-and-sum convolu-
tional graph filter of the form (3) operates in the spectral domain as a pointwise multiplication
ỹi = h̃ (λi ) x̃i between the input signal GFT x̃ = V−1 x and the filter frequency response
K
X
h̃(λ) = hk λk
k=1
01
Normalizing the adjacency matrix is made for pure technical reasons to prevent excessive scaling of the
shifted signal [22].
132 Graphs Signals and Systems
Such a result is reminiscent of the convolution theorem [52], whereby the convolution in the
graph domain, translates into a multiplication in the frequency domain. The filter frequency
response is an analytic polynomial in λ and it is independent of the graph. That is, for fixed
filter parameters, the filter effect on all graphs is already set. The specific filter effect on a
given graph is on the positions where the frequency response is instantiated; see Fig. 3. This
allows anticipating how the frequency response will behave for a wide range of graphs.
In this context, graph convolutional filters satisfy the following properties in the spectral
domain.
Property 7 (GFT of the filter). Eq. 11) can be rewritten as
ỹ = diag(h̃)x̃ with h̃ = Λh
where Λ ∈ CN ×(K+1) is a Vandermonde matrix such that [Λ]ik = λk−1 i . The vector
h̃ ∈ CN is known as the GFT of the filter parameters. Unlike traditional DSP, the convolution
operation is not commutative, and thus the input signal is mathematically (and conceptually)
different from the filter parameters. This becomes evident in the fact that the GFT of the
signal depends on the eigenvectors of the GSO, while the GFT of the filter on the eigenvalues.
Property 8 (Lipschitz continuity to changes in S). Let S, Ŝ ∈ RN ×N be two GSOs, po-
tentially corresponding to different graphs with the same number of nodes N . Define the
relative difference of Ŝ with respect to S as
for ε > 0 such that d(Ŝ; S) ≤ ε. Thus, if the relative difference between two GSOs is
small, the output of the filters to the same input signal will also be small.
Filters whose frequency response satisfies λh̃′ (λ) ≤ C are known as integral Lipschitz
filters. These filters may exhibit high variability for low values of λ (because its derivative
can be high), but they have to be approximately constant for high values of λ (because its
derivative has to be small). An example is shown in Fig. 3. For finite graphs with finite edge
weights, all convolutional filters [cf. (3)] are integral Lipschitz within the spectrum interval of
interest but the constant C may be large. This constant depends only on the filter parameters
and, thus, filters can be designed or learned to have a small value of C guaranteeing a tighter
bound; see [56] for details.
Part III
133
135
136 Introduction to Multidimensional Signals and Systems
CHAPTER 1
INTRODUCTION TO
MULTIDIMENSIONAL SIGNALS AND
SYSTEMS
Isufi et al.: Graph filters for signal processing and machine learning on graphs
isufi2024graph
Elvin Isufi et al. “Graph filters for signal processing and machine learning on graphs”. In:
IEEE Transactions on Signal Processing (2024).
137