0% found this document useful (0 votes)
131 views312 pages

UG-Signals and Systems 17-05-2024

Uploaded by

btechece230731
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
131 views312 pages

UG-Signals and Systems 17-05-2024

Uploaded by

btechece230731
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 312

(i)

Signals and Systems

Author
Prof. Sanjay L. Nalbalwar
Dean (Academics - FoE&T),
Department of Electronics and Telecommunication Engineering (E&TC),
Dr. Babasaheb Ambedkar Technological University,
Lonere, Raigad, Maharashtra.

Reviewed by
Prof. Satyabrata Jit
Professor (HAG),
Department of Electronics Engineering, IIT(BHU), Varanasi

All India Council for Technical Education


Nelson Mandela Marg, Vasant Kunj,
New Delhi, 110070

(ii)
BOOK AUTHOR DETAILS

Prof. Sanjay L. Nalbalwar, Professor and Dean, Dept. of Electronics and Telecommunication Engineering
(E&TC), Dr. Babasaheb Ambedkar Technological University, Lonere, Raigad, Maharashtra.
Email ID: [email protected]

BOOK REVIEWER DETAIL

Prof. Satyabrata Jit, Professor (HAG), Department of Electronics Engineering IIT(BHU), Varanasi.
Email ID: [email protected]

BOOK COORDINATOR (S) – English Version

1. Dr. Ramesh Unnikrishnan, Advisor-II, Training and Learning Bureau, All India Council for Technical
Education (AICTE), New Delhi, India.
Email ID: [email protected]
Phone Number: 011-29581215

2. Dr. Sunil Luthra, Director, Training and Learning Bureau, All India Council for Technical Education
(AICTE), New Delhi, India.
Email ID: [email protected]
Phone Number: 011-29581210

May, 2024
© All India Council for Technical Education (AICTE)
ISBN : 978-93-6027-309-5

All rights reserved. No part of this work may be reproduced in any form, by mimeograph or any
other means, without permission in writing from the All India Council for Technical Education
(AICTE).
Further information about All India Council for Technical Education (AICTE) courses may be obtained
from the Council Office at Nelson Mandela Marg, Vasant Kunj, New Delhi-110070.
Printed and published by All India Council for Technical Education (AICTE), New Delhi.

Attribution-Non Commercial-Share Alike 4.0 International (CC BY-NC-SA 4.0)

Disclaimer: The website links provided by the author in this book are placed for informational, educational
& reference purpose only. The Publisher do not endorse these website links or the views of the speaker /
content of the said weblinks. In case of any dispute, all legal matters to be settled under Delhi Jurisdiction,
only.

(iii)
(iv)
ACKNOWLEDGEMENT
I am deeply grateful to the authorities of AICTE, with special acknowledgment to
Prof. T. G. Sitharam, Chairman; Dr. Abhay Jere, Vice-Chairman; Prof. Rajive Kumar,
Member-Secretary; and Dr. Ramesh Unnikrishnan, Advisor-II and Dr. Sunil Luthra, Director,
Training and Learning Bureau. Their collective vision and planning were instrumental in the
publication of the book on Signals and Systems. Sincere appreciation is extended to
Prof. Satyabrata Jit, Professor, Department of Electronics Engineering IIT(BHU), Varanasi.
As the reviewer of this book, Prof. Satyabrata Jit has made invaluable contributions, ensuring
the content is not only student-friendly but also aesthetically pleasing and well-organized.
I extend my heartfelt thanks to Hon’ble Vice Chancellor Prof. Dr. Karbhari Kale for his
motivational and continuous encouragement throughout the process. My gratitude also goes
to Prof. Dr. A W. Kiwelekar, Dr. B. F. Jogi, Dr. S. M. Pore, Dr. Brijesh Iyer and other senior
colleagues in the University for their motivation and unwavering institutional support which
was crucial in the successful completion of this book. I extend my deepest appreciation to Dr.
Snehal Gaikwad, Dr. Pallavi Ingale and Prof. Prashant Mahajan for their insightful feedback
and constructive criticism that has greatly enriched the quality of this work. Their
contributions have been indispensable in refining the ideas and arguments presented herein.
Special acknowledgment goes to my wife, daughter and son for their invaluable support for
showing patience, excusing my work during weekends and family. I wish to thank my parents,
whose nurturing and support have been my pillars through the significant events of my life.
This book is an outcome of various suggestions of AICTE members, experts and authors who
shared their opinion and thought to further develop the engineering education in our country.
Acknowledgements are due to the contributors and different workers in this field whose
published books, review articles, papers, photographs, footnotes, references and other
valuable information enriched me at the time of writing the book.

Prof. Sanjay L. Nalbalwar

(v)
PREFACE
Welcome to the world of signals and systems – a cornerstone of modern engineering that lies
at the heart of countless technological innovations shaping our world today. From
telecommunications to medical imaging, from audio processing to control systems, the
principles of signals and systems permeate virtually every facet of our lives.
This book serves as a comprehensive guide to understanding the fundamental concepts,
theories, and applications of signals and systems. Whether you are a student embarking on
your academic journey in engineering or a seasoned professional seeking to deepen your
understanding, this text aims to provide you with the knowledge and tools necessary to
navigate this fascinating field.
Throughout these pages, you will embark on a journey that explores the mathematics, physics,
and engineering principles that underpin signals and systems. From the basic properties of
signals to the intricacies of system analysis and design, each chapter is carefully crafted to
build upon the previous one, offering a structured approach to learning that facilitates
comprehension and retention.
Furthermore, this book emphasizes the practical relevance of signals and systems by
incorporating numerous real-world examples and applications. By grounding theoretical
concepts in practical scenarios, readers can gain a deeper appreciation for the significance of
signals and systems in solving real-world engineering challenges. Moreover, this text is
designed to be accessible to readers with a range of backgrounds and expertise levels. Whether
you are encountering signals and systems for the first time or seeking to deepen your
understanding of advanced topics, this book strives to provide clear explanations, illustrative
examples, and helpful insights to aid your learning journey.
As an author, my goal is to provide a valuable resource that inspires curiosity, fosters
understanding, and equips readers with the knowledge and skills needed to tackle the
complexities of signals and systems. We hope that this book serves as a trusted companion on
your exploration of this captivating subject and empowers you to make meaningful
contributions to the ever-evolving landscape of engineering.
Thank you for embarking on this journey with us.

Prof. Sanjay L. Nalbalwar

(vi)
OUTCOME BASED EDUCATION
For the implementation of an outcome based education the first requirement is to develop an
outcome based curriculum and incorporate an outcome based assessment in the education
system. By going through outcome based assessments evaluators will be able to evaluate
whether the students have achieved the outlined standard, specific and measurable outcomes.
With the proper incorporation of outcome based education there will be a definite
commitment to achieve a minimum standard for all learners without giving up at any level.
At the end of the programme running with the aid of outcome based education, a student will
be able to arrive at the following outcomes:

PO1. Engineering knowledge: Apply the knowledge of mathematics, science,


engineering fundamentals, and an engineering specialization to the solution of
complex engineering problems.
PO2. Problem analysis: Identify, formulate, review research literature, and analyze
complex engineering problems reaching substantiated conclusions using first
principles of mathematics, natural sciences, and engineering sciences.
PO3. Design / development of solutions: Design solutions for complex engineering
problems and design system components or processes that meet the specified needs
with appropriate consideration for the public health and safety, and the cultural,
societal, and environmental considerations.
PO4. Conduct investigations of complex problems: Use research-based knowledge and
research methods including design of experiments, analysis and interpretation of
data, and synthesis of the information to provide valid conclusions.
PO5. Modern tool usage: Create, select, and apply appropriate techniques, resources, and
modern engineering and IT tools including prediction and modeling to complex
engineering activities with an understanding of the limitations.
PO6. The engineer and society: Apply reasoning informed by the contextual knowledge
to assess societal, health, safety, legal and cultural issues and the consequent
responsibilities relevant to the professional engineering practice.
PO7. Environment and sustainability: Understand the impact of the professional
engineering solutions in societal and environmental contexts, and demonstrate the
knowledge of, and need for sustainable development.

(vii)
PO8. Ethics: Apply ethical principles and commit to professional ethics and responsibilities
and norms of the engineering practice.
PO9. Individual and team work: Function effectively as an individual, and as a member
or leader in diverse teams, and in multidisciplinary settings.
PO10. Communication: Communicate effectively on complex engineering activities with
the engineering community and with society at large, such as, being able to
comprehend and write effective reports and design documentation, make effective
presentations, and give and receive clear instructions.
PO11. Project management and finance: Demonstrate knowledge and understanding of
the engineering and management principles and apply these to one’s own work, as a
member and leader in a team, to manage projects and in multidisciplinary
environments.
PO12. Life-long learning: Recognize the need for, and have the preparation and ability to
engage in independent and life-long learning in the broadest context of
technological change.

(viii)
COURSE OUTCOMES
By the end of the course the students are expected to learn:
CO-1: Analyse the spectral characteristics of continuous-time periodic and Aperiodic
Signal.
CO-2: Analyse LTI systems in the time domain.
CO-3: Analyse signals using Fourier series and Fourier transform.
CO-4: Apply DFT to analyse discrete-time systems.
CO-5: Analyse LTI systems using Z-Transform.
CO-6: Understand sampling theorem and its implications.

Mapping of Course Outcomes with Programme Outcomes to be done according to the


matrix given below:
Expected Mapping with Programme Outcomes
Course (1- Weak Correlation; 2- Medium correlation; 3- Strong Correlation)
Outcomes
PO-1 PO-2 PO-3 PO-4 PO-5 PO-6 PO-7 PO-8 PO-9 PO-10 PO-11 PO-12

CO-1 3 3 2 1 1 1 1 2 1 1 1 3

CO-2 3 3 3 1 2 1 1 2 1 1 1 3

CO-3 3 3 3 3 2 3 3 2 1 3 1 3

CO-4 3 3 3 3 2 1 1 2 1 3 1 3

CO-5 3 3 3 3 2 1 1 2 1 3 1 3

CO-6 3 2 2 1 1 1 1 2 1 1 1 3

(ix)
GUIDELINES FOR TEACHERS
To implement Outcome Based Education (OBE) knowledge level and skill set of the students
should be enhanced. Teachers should take a major responsibility for the proper
implementation of OBE. Some of the responsibilities (not limited to) for the teachers in OBE
system may be as follows:
 Within reasonable constraint, they should manoeuvre time to the best advantage of all
students.
 They should assess the students only upon certain defined criterion without considering any
other potential ineligibility to discriminate them.
 They should try to grow the learning abilities of the students to a certain level before they
leave the institute.
 They should try to ensure that all the students are equipped with the quality knowledge as
well as competence after they finish their education.
 They should always encourage the students to develop their ultimate performance capabilities.
 They should facilitate and encourage group work and team work to consolidate newer
approach.
 They should follow Blooms taxonomy in every part of the assessment.
Bloom’s Taxonomy
Teacher should Student should Possible Mode of
Level
Check be able to Assessment

Students ability
Create Design or Create Mini project
to create

Students ability
Evaluate Argue or Defend Assignment
to justify

Students ability Differentiate or Project/Lab


Analyse
to distinguish Distinguish Methodology

Students ability
Operate or Technical Presentation/
Apply to use
Demonstrate Demonstration
information

Students ability
Explain or
Understand to explain the Presentation/Seminar
Classify
ideas

Students ability
Remember to recall (or Define or Recall Quiz
remember)

(x)
GUIDELINES FOR STUDENTS
Students should take equal responsibility for implementing the OBE. Some of the
responsibilities (not limited to) for the students in OBE system are as follows:
 Students should be well aware of each UO before the start of a unit in each and every
course.
 Students should be well aware of each CO before the start of the course.
 Students should be well aware of each PO before the start of the programme.
 Students should think critically and reasonably with proper reflection and action.
 Learning of the students should be connected and integrated with practical and real life
consequences.
 Students should be well aware of their competency at every level of OBE.

(xi)
LIST OF FIGURES
Unit 1: Introduction to Signals and Systems
Fig. 1.1: Sinusoidal Signal 5
Fig. 1.2: ECG Signal 5
Fig. 1.3: Representation of signal and system 6
Fig. 1.4: Continuous time signal 7
Fig. 1.5: Discrete time signal 8
Fig. 1.6: Pulse signal 8
Fig. 1.7: CT Unit impulse signal 9
Fig. 1.8: DT Unit impulse sequence 10
Fig. 1.9: Unit Step Function 11
Fig. 1.10: DT unit step sequence 12
Fig. 1.11: u[n] represented by shifted impulse sequences 13
Fig. 1.12: CT Signum function 14
Fig. 1.13: DT Signum sequence 15
Fig. 1.14: CT ramp function 15
Fig. 1.15: DT ramp sequence 16
Fig. 1.16: CT real exponential signal 17
Fig. 1.17: CT complex exponential signal 17
Fig. 1.18: DT complex exponential sequence 18
Fig. 1.19: CT sinusoid signal 19
Fig. 1.20: CT sinusoid signal (a) Increasing sinusoid signal (b) Decaying sinusoid signal 19
Fig. 1.21: CT deterministic signal 20
Fig. 1.22: CT random signal 21
Fig. 1.23: CT Even signal 21
Fig. 1.24: CT Odd signal 22
Fig. 1.25: CT periodic signal 24
Fig. 1.26: CT aperiodic signal 25
Fig. 1.27: DT periodic signal 25
Fig. 1.28: Examples of finite duration signals (a) CT finite-duration signal (b) DT finite duration 32
signal

(xii)
Fig. 1.29: Examples of infinite duration signals (a) CT infinite-duration signal (b) DT infinite 32
duration signal
Fig. 1.30: Examples of causal signals (a) CT causal signal (b) DT causal signal 33
Fig. 1.31: Examples of noncausal signals (a) CT noncausal signal (b) DT noncausal signal 33
Fig. 1.32 Example of application of cartesian coordinates 34
Fig. 1.33: Random (stochastic) Signal 34
Fig. 1.34: Continuous and Discrete Amplitude Signals 37
Fig. 1.35: Examples of bounded and unbounded signal 43

Unit 2: Behavior of Continuous and Discrete-time LTI Systems


Fig. 2.1: Linear scaling 55
Fig. 2.2: Superposition Principle 56
Fig. 2.3: Superposition Principle with Linear Scaling 56
Fig. 2.4: Time-Invariant Systems 57
Fig. 2.5: Linear Time-Invariant Systems 57
Fig. 2.6: Superposition in Linear Time-Invariant Systems 58
Fig. 2.7: CT impulse response 58
Fig. 2.8: DT impulse response 58
Fig. 2.9: Representation of Continuous-Time Signals in Terms of Impulses 59
Fig. 2.10: Convolution operation 61
Fig. 2.11: x(t) input to system 62
Fig. 2.12: δ(t) input to system 62
Fig. 2.13: Delayed input to system 62
Fig. 2.14: x[n] input to system 62
Fig. 2.15: δ[n] input to system 63
Fig. 2.16: δ [n-k] input to system 63
Fig. 2.17: x[k] δ [n-k] input to system 63
Fig. 2.18: Equivalences for the series interconnection of continuous-time LTI systems. The (a) first 64
equivalence (b) second equivalence
Fig. 2.19: Equivalence for the parallel interconnection of continuous-time LTI systems 65
Fig. 2.20: System interconnection example 65
Fig. 2.21: State Space representation of CT system 70

(xiii)
Fig. 2.22 Periodic input to LTI system 78

Unit 3: Fourier Series and Transform


Fig. 3.1: Types of signals and their Fourier representations 89
Fig. 3.2: Even, Half wave, Quarter wave symmetry waveform 98
Fig. 3.3: Even symmetry waveform 99
Fig. 3.4: Odd, Half wave, Quarter wave symmetry waveform 102
Fig. 3.5: Odd symmetry waveform 102
Fig. 3.6: Half wave symmetry waveform 103
Fig. 3.7: Approximation of square waveform using some harmonic frequency components 106

Unit 4: Laplace Transform


Fig. 4.1: s-plane 153
Fig. 4.2: RoC for right sided signal 155
Fig. 4.3: RoC for left sided signal 156
Fig. 4.4: RoC for two sided signal 157
Fig. 4.5: ROC of x(t)=A u (t) 158
Fig. 4.6: ROC of x(t)=t u (t) 159
-8t
Fig. 4.7: ROC of x(t)= e u (-t) 161
Fig. 4.8: ROC of x(t)= e-3|t| 162

Unit 5: Z - Transform
Fig. 5.1: Z-plane 191
Fig. 5.2: Poles and zeros in the Z-plane 192
Fig. 5.3: x(n) = anu(n) 196
Fig. 5.4: Pole-zero plot and ROC |z| > |a| 196
Fig. 5.5 x[n]= an u( n  1) 198
Fig. 5.6: Pole-zero plot and ROC |z| < |a| 198
Fig. 5.7: Pole-zero plot and ROC |a| < |z| < |b| 207

(xiv)
Fig. 5.8: Pole-zero plot and ROC : 13 < |z| 208
Fig. 5.9: Pole-zero plot and ROC |z| < 14 210

Unit 6: Sampling & Reconstruction


Fig. 6.1: Three continuous time signals 228
Fig. 6.2: Impulse-train sampling 230
Fig. 6.3: Frequency domain representation due to sampling in time domain: (a) Spectrum of 231
original signal; (b) Spectrum of sampling function; (c) Spectrum of sampled signal with
>2 ; (d) Spectrum of sampled signal <2
Fig. 6.4: Recovery of original signal by using ideal lowpass filter (a) System for sampling &
reconstruction (b) Representative spectrum for x(t) (c) Corresponding spectrum for 232
xp(t) (d) Ideal lowpass filter to recover X(jω) from (jω) (e) Spectrum of (t)
Fig. 6.5: Sampling using Zero-Order Hold 233
Fig. 6.6: Zero order hold as impulse- train sampling followed by an LTI system with rectangular 233
response
Fig. 6.7: Interpolation between samples, solid curve represents interpolation 234
Fig. 6.8: Band-limited interpolation using Sinc function: (a) Band-limited signal x(t) (b) Impulse 235
train sampling of x(t) (c) Ideal band-limited interpolation in which impulse train is
replaced by superposition of Sinc functions
Fig. 6.9: Transfer function for zero-order hold 236
Fig. 6.10: Linear Interpolation (First-order hold) as impulse-train sampling followed with 237
convolution a triangular impulse response: (a) system for sampling & reconstruction;
(b) Impulse train of samples; (c) Impulse response representing a first-order hold; (d)
First-order hold applied to sampled signal; (e) Comparison of transfer function of ideal
interpolating filter & first order hold.
Fig. 6.11: Effect of oversampling & under sampling: (a) Spectrum of original sinusoidal signal; 239
(b) (c) Spectrum of sampled signal with > 2 ; (d) (e) Spectrum of sampled signal
with <2 ;

Fig. 6.12: Effect of aliasing on sinusoidal signal. For each of the four values of , the original 240
sinusoidal signal (solid curve), its samples, and reconstructed signal (dashed curve) are
illustrated: (a) =(5 )/6, in (a) and (b) no aliasing occurs, whereas in (c) & (d) there
is aliasing
Sinusoidal signal for example 6.1 241
Fig. 6.13:
Strobe effect 242
Fig. 6.14:
Fig. 6.15: Discrete-time processing of Continuous Time signals 243

Fig. 6.16: Notation for A/D conversion and D/A conversion 244

(xv)
Fig. 6.17: Sampling then followed by conversion to Discrete-time sequence: (a) Overall System; 245
(b) (t) for two sampling rates. Dashed envelop represents (t); (c) The output
sequence for two different sampling rates.
Fig. 6.18: The relationship among (jω), (jω) and ( ) with different sampling rates 246
Fig. 6.19: Conversion of discrete time sequence to a continuous time signal 247
Fig. 6.20: Frequency Response of ideal band-limited differentiator 248
Fig. 6.21: Frequency Response of discrete-time filter used to implement a continuous time band- 248
limited differentiator
Fig. 6.22: Discrete time sampling 250
Fig. 6.23: Impulse-train sampling of discrete-time signal in frequency domain: (a) Original signal 251
spectrum; (b) Spectrum of sampling sequence; (c) Spectrum of sampled signal with >
; (d) Spectrum of sampled signal with < . No aliasing occurs.
Fig. 6.24: Exact recovery of discrete time signal from its samples using an ideal low pass filter: (a)
252
Block diagram for sampling & reconstruction; (b) spectrum of x[n];

Fig. 6.25: Relationship between [n] corresponding to sampling and [n] corresponding to
254
decimation
Fig 6.26: Frequency domain illustration of the relationship between sampling & decimation 255
Fig 6.27: Continuous time signal that was originally sampled at Nyquist rate. After discrete time 256
filtering, the resulting sequence can be further downsampled. Here (jω) is the
continuous time Fourier Transform of (t), ( ) and ( ) are the discrete
time Fourier transforms of [ ] & [ ] respectively. And ( ) is the frequency
response of the discrete time low pass filter depicted in the block diagram.
Upsampling: (a) Overall system; (b) associated sequences and spectra for upsampling
Fig 6.28: 257
by a factor of 2
Spectra associated with example 6.4: (a) Spectrum of x[n]; (b) Spectrum after
Fig 6.29: downsampling by 4; (c) Spectrum after upsampling of x(n) by factor of 2; (d) Spectrum 258
after upsampling x[n] by 2 then downsampling by 9

(xvi)
CONTENTS

Foreword iv
Acknowledgement v
Preface vi
Outcome Based Education vii
Course Outcomes ix
Guidelines for Teachers x
Guidelines for Students xi
List of Figures xii

Unit 1: Introduction to Signals and Systems 1-50


Unit specifics
Rationale
Pre-requisites
Unit outcomes
1.1 Signals and systems as seen in everyday life, and in various branches of engineering and science 4
1.1.1 Signal 4
1.1.2 System 6
1.2 Classification of signals 6
1.2.1 Continuous time (CT) signals 6
1.2.2 Discrete time (DT) signals 7
1.3 Basic signals 7
1.3.1 Continuous-Time Unit Impulse δ(t) 8
1.3.2 CT Unit Impulse Function Properties 9
1.3.3 Importance of Impulse Function 9
1.3.4 Discrete-Time Unit Impulse sequence δ[n] 9
1.3.5 DT Unit Impulse Sequence Properties 10
1.3.6 CT Unit Step Function u(t) 10
1.3.7 Importance of step function 11
1.3.8 DT Unit Step Sequence u[n] 11
1.3.9 CT Signum Function 13
1.3.10 DT Signum Sequence 14

(xvii)
1.3.11 CT Ramp Function 14
1.3.12 DT Ramp Sequence 15
1.3.13 CT Exponential signal 15
1.3.14 DT Exponential signal 17
1.3.15 CT Sinusoid Signal 18
1.3.16 DT Sinusoid Sequence 19
1.4 Classification of Continuous Time and Discrete Time signals 19
1.4.1 Deterministic and Random signals
19
1.4.2 Even and Odd signals
21
1.4.3 Periodic and Aperiodic signals
23
1.4.4 Energy and Power signals
27
1.4.5 Finite (Time-limited) and Infinite Duration signals
31
1.4.6 Causal and Noncausal signals 32
1.5 Some signal properties 33
1.5.1 Absolute integrability 33
1.5.2 Determinism 33
1.5.3 Stochastic character 33

1.6 Continuous/Discrete Amplitude Signals 35


1.6.1 Continuous Amplitude Signals 35
1.6.2 Discrete Amplitude Signals 36
1.7 Continuous –Time (CT) and Discrete-Time (DT) systems 37
1.8 System properties: Linearity: additivity and homogeneity, Shift- invariance, Causality, Stability, 37
Realizability
1.8.1 Linear and Nonlinear Systems 37
1.8.2 Time-Invariant and Time-Variant Systems
38
1.8.3 Causal and Noncausal Systems
40
1.8.4 Stable and Unstable Systems
41
1.8.5 Realizability
44
Unit summary
Exercises 44
Know more 45
References and suggested readings 49

(xviii)
Unit 2: Behavior of Continuous and Discrete-time LTI Systems 51-86
Unit specifics
Rationale
Pre-requisites
Unit outcomes
2.1 Introduction 55
2.1.1 Linear Time Invariant Systems 55
2.1.2 Time Invariant Systems 57
2.1.3 Linear Time Invariant Systems 57
2.2 Impulse response of LTI System 58
2.2.1 Discrete-Time Unit Impulse Response and the Convolution 58
2.2.2 Representation of Continuous-Time Signals in Terms of Impulses 59
2.3 The Unit Step Response of an LTI System 60
2.4 Convolution Integral 61
2.4.1 Properties of the Convolution Integral 61
2.5 Input-output behaviour with aperiodic convergent inputs 62
2.5.1 Response of a continuous time system
62
2.5.2 Response of a discrete time system 62
2.6 Cascade interconnections 64
2.7 Causality for LTI systems 66
2.8 Stability for LTI Systems 66
2.9 System Representation through Differential Equations and Difference Equations 67
2.9.1 Differential Equation Description of CT LTI systems
67
2.9.2 Difference Equation Description of DT LTI systems 68
2.10 State space representation of systems 69
2.11 State Space Analysis 70
2.11.1 State equations
70
2.11.2 Output equations
71
2.11.3 State Model 72
2.12 Transfer function of a Continuous Time System 72
2.13 State Transition Matrix 72
2.14 Multi-Input, Multi-Output Representation 75
2.15 Periodic Inputs to LTI system 78

(xix)
2.16 The Notion of frequency response and its relation to impulse response 79
Unit summary
81
Exercises
82
Know more
85
References and suggested readings 86

Unit 3: Fourier Series and Transform 87-150


Unit specifics
Rationale
Pre-requisites
Unit outcomes
3.1 Introduction 89
3.2 Trigonometric Form of Fourier Series 89
3.2.1 Definition 89
3.2.2 Conditions for existence of Fourier series 90
3.2.3 Equations for , and 91
3.3 Exponential form of Fourier Series 94
3.3.1 Definition 94
3.3.2 Derivation for equation for 95
3.3.3 Relationship between the trigonometric and exponential form of Fourier coefficient 95
3.4 Waveform Symmetries 96
3.4.1. Even Symmetry
96
3.4.2. Odd Symmetry
99
3.4.3 Half Wave Symmetry
103
3.4.4 Quarter Wave Symmetry 103
3.5 Properties of Fourier Series in terms of Exponential Form 104
3.5.1. Linearity
104
3.5.2. Time Shifting
104
3.5.3 Frequency Shifting
104
3.5.4 Time Reversal
104
3.5.5. Multiplication
104
3.5.6. Conjugation
105
3.5.7 Time Scaling
105
3.5.8 Differentiation
105
3.5.9. Integration
105

(xx)
3.5.10. Real and Even 105
3.5.11 Real and Odd 105
3.5.12 Parseval’s Relation 105
3.6 Gibb’s Phenomenon 106
3.7 Fourier Transform (FT) 111
3.7.1. Definition
111
3.7.2. Definition of Inverse Fourier Transform (IFT)
112
3.7.3 Magnitude and Phase Spectrum using Fourier Transform 112
3.8 Properties of Fourier Transform (FT) 112
3.8.1. Linearity
112
3.8.2. Time Shifting
113
3.8.3 Time Scaling
123
3.8.4 Time Reversal
114
3.8.5. Conjugation
114
3.8.6. Frequency Shifting
115
3.8.7 Time Differentiation
115
3.8.8 Time Integration
116
3.8.9 Differentiation in Frequency
116
3.8.10 Convolution
117
3.8.11 Parseval’s Theorem
118
3.8.12 Duality Property
119
3.9 Fourier Transform (FT) Representation of Continuous-Time (CT) LTI System in terms of Convolution 120
and Multiplication
3.9.1. Representation of Transfer Function of CT LTI System in Frequency Domain 120
3.9.2. Relation of Impulse Response and Transfer Function of CT LTI System 120
3.9.3 Response of CT LTI System in terms of Fourier Transform 121
3.9.4 Magnitude and Phase Response of CT LTI System 121

3.10 Discrete Time Fourier Series representation 128


3.10.1. Representation of Discrete Time Fourier Series 128
3.10.2. Properties of DTFS in terms of coefficients 128
3.11 Discrete Time Fourier Transform (DTFT) Representation 132
3.11.1. Definition of DTFT
132
3.11.2. Definition of Inverse DTFT 132

(xxi)
3.12 Properties of DTFT 132
3.12.1. Linearity 132
3.12.2. Time Shifting 133
3.12.3 Periodicity 134
3.12.4 Time Reversal 134
3.12.5 Conjugation 134
3.12.6 Frequency Shifting 135
3.12.7 Differentiation in Frequency 135
3.12.8 Convolution 136
3.12.9 Parseval’s Theorem 137
3.13 Discrete Fourier Transform (DFT) Representation 138
3.13.1 Definition of DFT
138
3.13.2 Definition of Inverse DFT
139
3.14
Properties of DFT
139
3.14.1. Linearity
139
3.14.2. Circular Time Shifting
140
3.14.3 Periodicity
140
3.14.4 Time Reversal 141
3.14.5 Conjugation 142
3.14.6 Circular Frequency Shifting
142
3.14.7 Multiplication
142
3.14.8 Convolution
143
3.14.9 Parseval’s Theorem
144
Unit summary
144
Exercises
145
Know more
149
References and suggested readings 150

Unit 4 Laplace Transform 151-187


Unit specifics
Rationale
Pre-requisites
Unit outcomes
4.1 Introduction 153
4.2 Definition of Complex Frequency 153

(xxii)
4.2.1 Complex Frequency Plane (s-Plane) 154
4.2.2 Definition of Laplace Transform (LT) 154
4.2.3 Definition of Inverse Laplace Transform (ILT) 155
4.3 Region of Convergence (RoC) 155
4.4 Properties of Laplace Transform 168
4.4.1 Scaling of amplitude 168
4.4.2 Linearity 168
4.4.3 Time differentiation 168
4.4.4 Integration in time domain 169
4.4.5 Shifting in frequency domain 170
4.4.6 Shifting in time domain 170
4.4.7 Differentiation in frequency
170
4.4.8 Time Scaling
171
4.4.9 Initial Value theorem
172
4.4.10 Final Value theorem
172
4.4.11 Convolution Property 172
4.5 Poles and Zeros of System Functions and Signals 173
4.5.1 Poles and Zeros 173
4.6 ROC Properties for Laplace Transform 174
4.7 Inverse Laplace Transform by Partial Fraction Expansion Method 175
4.8 Laplace domain analysis 178
4.8.1 Transfer Function of LTI Continuous Time System
178
4.8.2 Impulse Response and Transfer Function 178
4.9 Solving Differential Equations by Using Laplace Transform 179
Unit summary
179
Exercises
180
Know more
186
References and suggested readings 187

Unit 5 Z Transform 188-266


Unit specifics
Rationale
Pre-requisites
Unit outcomes

(xxiii)
5.1 Introduction 190
5.2 Need of Z-transform 190
5.3 Types of Z-transforms 191
5.3.1 Unilateral Z-Transform 191
5.3.2 Bilateral Z-Transform 191
5.4 The Z-plane 192
5.4.1 Poles 193
5.3.2 Zeros 193
5.5 Region of Convergence (ROC) for Z-Transform 194
5.6 Properties of ROC 199
5.7 Properties of the Z-transform 200
5.7.1. Linearity 200
5.7.2. Time Shifting 201
5.7.3 Time Reversal 202
5.7.4 Scaling in z-domain 202
5.7.5 Time scaling property 203
5.7.6 Convolution 204
5.7.7 Differentiation in z-domain 205
5.7.8 Conjugation 205
5.7.9 Initial Value Theorem 206
5.7.10 Final Value Theorem 206
5.7.11 Accumulation 207
5.8 Relationship between DTFT and z-transform 211
5.9 Inverse Z- transform 212
5.9.1 Power series expansion method 213
5.9.2 Partial fraction method 216
5.10 Z – Domain Causality and stability analysis 219
Unit summary 220
Exercises 221
Know more 225
References and suggested readings 226

Unit 6 Sampling & Reconstruction 267-283


Unit specifics

(xxiv)
Rationale
Pre-requisites
Unit outcomes
6.1 Introduction 230
6.2 Sampling Theorem 230
6.2.1 Impulse train sampling 231
6.2.2 Sampling with Zero-Order Hold 234
6.3 Reconstruction of a signal from its samples using interpolation 236
6.3.1 The effect of under sampling: Aliasing 238
6.4 Discrete Time Processing of Continuous Time Signals 245
6.4.1 Digital differentiator 249
6.5 Sampling of discrete time signals 251
6.5.1 Impulse train sampling 251
6.5.2 Discrete time decimation and interpolation 255
Unit summary 261
Exercises 275
Know more 282
References and suggested readings 283

CO and PO Attainment Table 284


Index 285

(xxv)
Introduction to
1
d

Signals and Systems

UNIT SPECIFICS
This unit presents information related to the following topics:
 Explore fundamentals of signals and systems; understand their significance in engineering
and science;
 Examine properties of signals: periodicity, integrability, determinism, stochastic character;
 Study special signals: unit step, unit impulse, sinusoid, complex exponential, time-limited
signals;
 Differentiate between continuous-time and discrete-time signals; analyze their
characteristics and representations;
 Compare continuous and discrete amplitude signals; understand their applications;
 Explore system properties: linearity, shift-invariance, causality, stability, realizability;
 Discuss practical applications of signals and systems in everyday life and various fields;
 Learning outcomes: clear understanding of concepts; ability to analyze signals and
comprehend system properties; recognition of signal processing importance;
 Unit provides foundational knowledge for further studies and applications in engineering
and science.
This unit provides an introduction to signals and systems, focusing on their fundamental
concepts and applications in engineering and science. Students will explore the properties of
signals, including periodicity, integrability, determinism, and stochastic character, and
understand how these properties relate to real-world scenarios. Special signals such as the unit
step, unit impulse, sinusoid, and complex exponential will be studied, along with their
characteristics and applications.
The unit will differentiate between continuous-time and discrete-time signals, as well as
continuous and discrete amplitude signals, and analyze their representations and
characteristics. Students will also explore system properties such as linearity, shift-invariance,
causality, stability, and realizability, and understand their impact on signal processing.
2 | Introduction to Signals and Systems
The practical applications of signals and systems in various fields will be discussed,
highlighting their relevance in everyday life. Students will be encouraged to apply their
knowledge to solve practical problems and develop critical thinking and problem-solving skills.
Effective communication of concepts using appropriate terminology and notation will be
emphasized.
By completing this unit, students will have a strong foundation in signals and systems, enabling
them to further their studies and apply their knowledge in engineering and scientific contexts.

RATIONALE

The unit on "Introduction to Signals and Systems" is to provide students with a solid foundation
in understanding the fundamental concepts of signals and systems. This knowledge is essential
as signals and systems are pervasive in various branches of engineering and science.

By exploring the properties of signals, such as periodicity, integrability, determinism, and


stochastic character, students gain insights into the behavior and characteristics of different
types of signals encountered in real-world applications. Special signals, including the unit step,
unit impulse, sinusoid, and complex exponential, are studied to understand their unique
properties and applications. Differentiating between continuous-time and discrete-time signals,
as well as continuous and discrete amplitude signals, helps students comprehend the
distinctions and representations in both domains. Understanding system properties, such as
linearity, shift-invariance, causality, stability, and realizability, provides insights into how
signals interact with systems and how system properties impact signal processing. The unit also
emphasizes the practical applications of signals and systems, highlighting their relevance in
everyday life and across various branches of engineering and science. By applying their
knowledge to solve practical problems, students develop critical thinking and problem-solving
skills. Effective communication is essential, and students are encouraged to communicate
concepts related to signals and systems using appropriate terminology and notation.

Overall, this unit aims to equip students with a strong foundational understanding of signals
and systems, enabling them to pursue more advanced topics and apply their knowledge in
engineering and scientific contexts.

PRE-REQUISITES

1. Mathematics: Solid understanding of calculus, algebra, trigonometry, and complex


numbers.

2. Physics: Basic knowledge of force, motion, energy, and waves.


Signals and Systems | 3
3. Programming: Familiarity with programming languages like MATLAB or Python.

4. Algebraic Manipulation: Proficiency in manipulating algebraic expressions and solving


equations.

5. Analytical Thinking: Ability to think analytically and solve problems systematically.

6. These prerequisites are essential for students to effectively engage with the unit on
"Introduction to Signals and Systems" and ensure a smooth transition into the study of the
subject matter.

UNIT OUTCOMES
After studying this unit students will be able to:
U1-O1: Understand fundamental concepts of signals and systems.
U1-O2: Identify and analyze different signal types: periodic, deterministic, stochastic.
U1-O3: Recognize the importance of signal properties in real-world applications.
U1-O4: Apply special signals (unit step, unit impulse, sinusoid, complex exponential) in
problem-solving and system analysis.
U1-O5: Differentiate and understand representations and characteristics of continuous-time
and discrete-time signals; analyze system properties and their impact on signal
processing.

EXPECTED MAPPING WITH COURSE OUTCOMES


Unit-1
(1- Weak Correlation; 2- Medium correlation; 3- Strong Correlation)
Outcomes
CO-1 CO-2 CO-3 CO-4 CO-5 CO-6
U1-O1 3 2 1 2 2 3
U1-O2 3 1 1 1 - 1
U1-O3 3 1 - - - 3
U1-O4 3 2 1 1 - 1
U1-O5 3 3 2 1 1 2
4 | Introduction to Signals and Systems

1.1 Signals and systems as seen in everyday life, and in various branches of
engineering and science
Signals and systems are ubiquitous in everyday life and have significant applications in engineering and
science. The term signal is generally applied to something that conveys information. Signals may, for
example, convey information about the state or behavior of a physical system. As another class of
examples, signals are synthesized for the purpose of communicating information between humans or
between humans and machines. Although signals can be represented in many ways, in all cases, the
information is contained in some pattern of variations. Signals are represented mathematically as functions
of one or more independent variables.

Signals can be any physical quantity that varies with time, space, or other independent variables. They can
be represented in either the time domain or frequency domain. Examples of signals include human speech,
electric current, and voltage. Signals can be dependent on one or more independent variables, such as time,
temperature, position, pressure, or distance. If a signal depends on only one independent variable, it is
called a one-dimensional signal, while a signal dependent on two independent variables is called a two-
dimensional signal. Examples include audio signals (speech, music), image and video signals,
communication signals, biomedical signals (ECG, EEG), control systems, digital signal processing,
electrical circuits, mechanical systems, and feedback systems. These concepts find practical use in areas
such as audio processing, image recognition, telecommunications, medical diagnostics, robotics, power
systems, and scientific research. Signals and systems provide the framework for analyzing, manipulating,
and understanding information in various fields.

1.1.1 Signal

A signal can be realized as a physical quantity which conveys the information related to some physical
phenomena like any voltage signal, any electromagnetic wave that is transmitted over the air from any base
station to the mobile station and this carries the information about the communication between two
individuals. A signal can also be predicted in terms of a video or an image carrying information. In this
way, a signal is a carrier of data or information.

The examples of the signal can be, when we power on the mobile handset, the electromagnetic field gets
associated with the antenna and we receive a signal from a base station. Another example can be, when we
listen a whistle of train passing nearby, that is a kind of signal which is identified in our brain. Basically,
signals arise in many forms, e.g. acoustic, light, pressure, flow, mechanical, thermal, electrical etc.

For us to study the signal theory, we consider signal as a variation or change of an entity with respect to
time or space or any other independent variable. When time signals are considered, they are represented
using x(t), y(t) etc., and such 1D signals which varies in time can also be extended for 2D, 3D signals.
Signals occurring in many different physical forms are often converted in electrical form by a transducer
Signals and Systems | 5

for ease of processing. For example, a microphone converts sound wave into electrical signal which is
convenient for further processing.

To study signals, we must see how the signals look! Below is one signal which is very commonly used
i.e., sinusoidal signal.

y-axis

Amplitude

Time x-axis
Fig. 1.1: Sinusoidal Signal

Graphically, the independent variable is represented by horizontal axis (x-axis) and the dependent
variable is represented by vertical axis (y-axis). In the Fig. 1.1, we must notice that as the time value is
changing, the instantaneous value of amplitude (height of the signal waveform) is also changing, so the
value plotted on y-axis are dependent on value on x-axis.

The general expression for sinusoidal signal can be written as sin( ), where is constant and is time
which is value on -axis. Thus, quantity on -axis is dependent variable and quantity on -axis is
independent variable, as we all know that time depends on nothing and it varies independently. So, from
the discussion the definition of Signal is any physical quantity that varies with time, space or any other
independent variable, which conveys some information.

Similarly, ECG signal shown in Fig. 1.2 is also a biological signal which depicts the function of heart to
find or study any abnormalities in the heart.

Fig. 1.2: ECG Signal


6 | Introduction to Signals and Systems

Other examples of the signals are speech signal, EEG signal, Image/Video signal, Radar signal, AM/FM
signal etc.

1.1.2 System
For the examples of the signals given in section 1.1.1, a system is always associated with their generation
and extraction of useful information. For example, any waveform like sinusoid can be generated by a
function generator and displayed on a cathode-ray oscilloscope (CRO). Similarly, other signals like ECG,
EEG are generated by our heart and brain, respectively. These signals are analyzed using the biological
equipment with suitable software systems.

From this discussion, a continuous time system accepts an input signal, ( ) and produces an output signal,
( ). A system is often represented as an operator on the input signal. Hence, a system can be defined as
any physical device that performs a certain operation or a set of operations on the input signal x(t) to result
in a new signal y(t) as its output. Fig. 1.3 shows the system to which an input signal ( ) is provided that
is further processed to get the output ( ).

Fig. 1.3: Representation of signal and system

1.2 Classification of signals


According to the nature of independent variable, time, signals can be classified into continuous-time (CT)
signals and discrete-time (DT) signals.

1.2.1 Continuous time (CT) signals


If the independent variable is continuous then signal is called as CT signal. Symbol is used to denote
continuous time. Also the independent variable is enclosed in (.). For example, a signal can be continuous
time signal for case,

x(t) = sin( ) for all (1.1)

Above Eq. (1.1) is of sinusoidal signal, which is a continuous time signal as it is defined continuously over
time from -∞ to +∞ or over any continuous time interval as shown in Fig. 1.4.
Signals and Systems | 7

Fig. 1.4: Continuous time signal

1.2.2 Discrete time (DT) signals


If the independent variable has discrete time instances, then signal is called as DT signal. These instances
can be either be positive or negative as shown by Fig. 1.5. Symbol is used to denote discrete time
instances. Also, the independent variable is enclosed in [.]. If a signal is plotted on such discrete time
instants the plot is called as a stem plot.

So, stem plot contains discrete time instants also called as sequence of numbers. From Fig. 1.5 it can be
seen that, at time instant 0, the instance is x (0) that shows the amplitude values at time 0, at time instant
1, the instance is x (1) and so on. Hence, these x (-2), x (-1), x (0), x (1), x (2) are called as time series or
sequence of numbers.

For example, a signal can be discrete time signal for case,

x[n] = sin for all (1.2)

Fig. 1.5: Discrete time signal

1.3 Basic signals


There are many elementary signals which must be studied to better understand the properties of both the
signals and systems. These include unit impulse, unit step, signum, ramp, exponential and sinusoid signals.
8 | Introduction to Signals and Systems

1.3.1 Continuous-Time Unit Impulse ( )


The unit impulse, denoted as ( ), is also known as the Dirac delta function. It is one of the most
fundamental signals to understand various properties of the system.

The mathematical representation of Impulse function is,

Fig. 1.6: Pulse signal

Consider the given signal which is a pulse from − , hence pulse has a width of and height equals
. So, area under the pulse equals × = 1. So, for each pulse ( ) for every , the area under the pulse
equal to 1.

Now, let us consider an impulse function as,

( )= → ( ) (1.3)

As, tends to 0, the width goes to 0 and height to infinity, but area still remains the unity (constant).

Hence, area under the given pulse function,

( ) 1 <0<
∫ = (1.4)
0 ℎ
The representation of CT unit impulse signal is represented by Fig. 1.7.

Fig. 1.7: CT Unit impulse signal


Signals and Systems | 9
In general, a CT impulse function can be described as,

( )= ∞ =0
and ∫ δ(t) dt = 1
0 ℎ
The arrow on the top of impulse indicates that it has infinite amplitude as shown in Fig 1.7.

1.3.2. CT Unit Impulse Function Properties

1. ∫ ( ) ( ) = (0)
i.e., if we multiply the impulse function by any signal x(t) and integrate it, it picks the value of the
function x(t) at t = 0
2. ∫ ( ) ( − ) = ( )
Here, ( − ) is basically the shifted impulse to =

3. Scaling property: ( ) = | |
( ), >0

4. Product property: ( ). ( ) = (0). ( )


Similarly, ( ). ( − ) = ( ). ( − )
5. Sifting property: ∫ ( ) ( − ) = ( )
6. CT Unit Impulse Function is even: (− ) = ( )

1.3.3 Importance of Impulse Function


By applying impulse signal to a system one can get the impulse response of the system. From impulse
response, it is possible to get the transfer function of the system. From the impulse response of the system,
one can easily get the step response and ramp response by integrating it once and twice respectively. For
a linear time invariant system, if the area under the impulse response curve is finite, then the system is said
to be stable.
Impulse signal is easy to generate and apply to any system.

1.3.4 Discrete-Time Unit Impulse sequence [ ]


A DT unit impulse sequence is represented by [ ] and is defined by
1, =0
[ ]= (1.5)
0, ℎ

Graphically [ ] is represented by Fig. 1.8,


10 | Introduction to Signals and Systems

Fig. 1.8: DT Unit impulse sequence

1.3.5 DT Unit Impulse Sequence Properties


1. Product/multiplication property:
Consider [ ] is a DT sequence multiplied with gives output [ ]

[ ] = [ ]. [ ] = [0] [ ] (1.6)
2. Sifting property:
The product property leads to sifting property,

∑ [ ] [ − ]= [ ] (1.7)

Eq. (1.7) is true when [ − ] is within the given summation limit otherwise the RHS of the
equation becomes zero.

1.3.6 CT Unit Step Function ( )


The unit step signal, denoted as ( ), is a fundamental signal that has a value of 0 for less than 0
and a value of 1 for greater than 0.

Mathematically, it can be represented as:


1 >0
( )= =0 (1.8)
0 <0

At = 0, we sometime define as . But, at = 0, there is a discontinuity, that means it jumps from 0


to 1.
Signals and Systems | 11

Fig 1.9: Unit Step Function

For example, consider a scenario where a light switch is turned on at time t = 0. The unit step signal
can be used to represent this event. Before t = 0, the signal is 0 (light off), and after t = 0, the signal
is 1 (light on).

Let us consider the relationship between the unit impulse and unit step function:

( )= ( ) (1.9)

Derivative of unit step function is unit impulse function. Since, ( ) is discontinuous at = 0, it is


not formally differentiable but Eq. (1.9) can be interpreted as approximation to unit step function.

1.3.7 Importance of step function


Step signal is easy to generate and apply to the system, is used in mathematics of control system and
signal processing as a signal which switches on at a specified time and stays switched on indefinitely.
It is also used in mechanics with impulse function to describe different types of structural loads.

By differentiating the step response impulse response can be obtained.by integrating the step
response, ramp response can be obtained.

Application of step signal is equivalent to the application of numerous sinusoidal signals with a wide
range of frequencies. Step response is considered as a white noise which is drastic if the system
response is satisfactory for a step signal; it is likely to give satisfactory response to other types of
signals.

1.3.8 DT Unit Step Sequence [ ]

A DT unit step Sequence can be represented by


1, ≥0
[ ]= (1.10)
0, <0
12 | Introduction to Signals and Systems
Fig. 1.10 shows the graphical representation of unit step sequence.

Fig 1.10: DT unit step sequence

The DT unit impulse sequence in forms of DT unit step sequence can be represented by,
[ ] = [ ] − [ − 1] (1.11)
Conversely, DT unit step sequence can be represented by DT unit impulse sequence as a running
sum of impulse. As,

[ ]= [ − ] (1.12)

[ ]= [ ] + [ − 1] + [ − 2] + ⋯ (1.13)

That mean, u[n] can be recognized as a linear combination of shifted impulse sequences as shown
by fig. 1.11

Fig 1.11: u[n] represented by shifted impulse sequences


Signals and Systems | 13
{2, −4,8,2,6,2}
Example 1.1: x[n] = . Represent the signal [ ] in terms of weighted shifted impulse

functions.

Solution:
x[n] = ∑∞
k=0 x[k]δ[n − k] (1.14)

[ ] has range -2 to 3 i.e. −2 ≤ ≤3

[ ]= x[k]δ[n − k] (1.15)

Expanding the above equation we get

x[n] = x[−2]δ[n + 2] + x[−1]δ[n + 1] + x[0]δ[n] + x[1]δ[n − 1] + x[2]δ[n − 2]


+ x[3]δ[n − 3] (1.16)

1.3.9 CT Signum Function


A Signum function has amplitude 1 for t> 0 and -1 for t<0. For t=0, the amplitude is
discontinuous. The signum function is written as ( ) and represented by Eq. (1.13)
As,
1, >0
( ) = (1.17)
−1, <0
The Signum Signal is mostly used in the communication theory.
The Signum function can be represented in terms of unit step function as,
14 | Introduction to Signals and Systems

( ) = 2 ( ) −1 (1.18)
Fig. 1.12 show the CT Signum function

Fig 1.12: CT Signum function


1.3.10 DT Signum Sequence
A DT Signum sequence can be written as
1, >0
[ ] = 0, =0 (1.19)
−1, <0

Graphically, DT Signum sequence can be represented by Fig. 1.13

Fig 1.13: DT Signum sequence

A DT Signum sequence can be represented in terms of unit step sequence as,


[ ] = [ ] − [− ] (1.20)

1.3.11 CT Ramp Function


A CT ramp function is represented by ( ) and given by,
, ≥0
( )= (1.21)
0, <0
The graphical representation of ramp signal is shown by Fig. 1.14.
Signals and Systems | 15

Fig 1.14: CT ramp function

r(t) can be expressed in terms of u(t) as,


r(t) = tu(t) (1.22)
Also, when unit step function is integrated it gives ramp function.

∫ u(τ)dτ = r(t) (1.23)

1.3.12 DT Ramp Sequence


A DT ramp sequence is represented by [ ] and given by
n, n≥0
( )= (1.24)
0, n<0
The graphical representation of ramp sequence is shown by Fig. 1.15.

Fig 1.15: DT ramp sequence

1.3.13 CT Exponential signal


It is another important type of signal which is represented by,


( )= (1.25)

Depending upon the values of parameters A and α, the exponential signal is categorized
16 | Introduction to Signals and Systems
into two types.
1. Real Exponential Signal
2. Complex Exponential Signal

1.3.13.1 Real Exponential Signal

For this signal, A and α are real. Depending upon the value of α, the signal increases or decreases
exponentially.

If α > 0, ( ) increases exponentially with time as shown in Fig. 1.16(a).

If α <0, ( ) decreases exponentially with time as shown in Fig. 1.16(b).

For the value of = 0, ( ) is DC signal as shown in Fig 1.16(c).

Fig 1.16: CT real exponential signal

1.3.13.2 Complex Exponential Signal


For exponential signal x(t) = Ae , A and are complex values.
Here, ( ) can be written as,
x(t) = |A|e |∅| e( )
(1.26)
= |A|e e ( ∅)
(1.27)
σt
= |A|e [cos( 0 t + ∅) + jsin( 0 t + ∅)] (1.28)
Signals and Systems | 17

The complex exponential signal can be represented graphically in terms of sine and cosine waves
depending upon value of σ in Eq. (1.28). If value of σ=0, real and imaginary parts of Eq. (1.28)
become sinusoidal signals as shown by Fig. 1.17.

Fig 1.17: CT complex exponential signal


1.3.14 DT Exponential signal
The discrete version of CT exponential signal is defined by DT exponential sequence as
x[n] = A (1.29)
Depending upon values of the parameters and , the DT exponential sequence is
categorized into two types.
1) Real Exponential Sequence
2) Complex Exponential Sequence

1.3.14.1 Real Exponential Sequence


For this sequence and are real. The exponential signal grows with the value of n for
| |>1 and decays with for | |<1. The nature of DT real exponential sequence is shown
in Fig. 1.18.

Fig 1.18 DT complex exponential sequence


18 | Introduction to Signals and Systems

1.3.14.2 Complex Exponential Sequence


As per the discussion in CT complex exponential signal, [ ] can be written as,
∅ ( ∅)
[ ]=| | | | (1.30)

( ∅)
[ ] = | || | (1.31)

[ ] = | || | [ ( + ∅) + ( + ∅)] (1.32)

1.3.15 CT Sinusoid Signal


From Eq. (1.28) of CT complex exponential signal, when = 0, the real and imaginary
parts of a complex exponential become sinusoids as shown by Fig. 1.20.

Fig 1.19 CT sinusoid signal

( ) = | |[ ( + ∅) + ( + ∅)] (1.33)

( )=| | ( + ∅) + | | ( + ∅)] (1.34)

|A| ( + ∅) and | | ( + ∅) are continuous time sinusoidal signals. Where, A


represents the amplitude value of the signal, ∅ is the phase angle.
When σ>0, Eq. (1.28) becomes the increasing sinusoidal signal and for σ<0, Eq. (1.28)
becomes the decaying sinusoidal signal as shown by Fig. 1.21.
Signals and Systems | 19

( ) σ> 0
( )
σ< 0

-t 0 t -t t

(a) (b)
Fig 1.20 CT sinusoid signal (a) Increasing sinusoid signal (b) Decaying sinusoid signal

For example, the sound produced by a tuning fork is a sinusoidal signal. As the prongs of the
fork vibrate back and forth, they create a pure tone that can be represented by a sinusoidal
waveform.

1.3.16 DT Sinusoid Sequence


From Eq. (5) of DT complex exponential sequence, when | |=1, the real and imaginary parts
of a complex exponential sequence becomes sinusoids.
i.e., [ ] = | | ( + ∅) + | | ( + ∅) (1.35)

|A| cos(ω n + ∅) and |A| sin(ω n + ∅) are discrete time sinusoidal sequence where
represents the amplitude, ω is frequency in radians/sample and ∅ is phase angle in radians.

1.4 Classification of Continuous Time and Discrete Time signals

1.4.1 Deterministic and Random signals


These types of signals class can be either continuous time or discrete time signals. A
deterministic signal is completely specified that can be deterministically represented at each
and every time instant or frequency instant. Deterministic is the fact where all the present,
past and future values of the signal amplitude at given time instant are known. For example,
sin(2 ), are the deterministic signals in the sense that at a given time instant t there is
no ambiguity, that means given a time instant, one can exactly determine what is the signal.
One such example is shown by Fig. 1.22.
20 | Introduction to Signals and Systems

Fig 1.21: CT deterministic signal

In Fig. 1.21, ( ) = ( for -∞<t<∞ which determines the values at different time instants.
)
Other examples of deterministic signals include voltage signals, current signals, etc.
Whereas, the random signals are random in nature that means it takes random values at
various time instants. Mathematically, the behaviour of random signals cannot be predicted.
For example, the outcome of a coin toss experiment can lead to a random signal. If its outcome
is a head, it is presented by +1 and if the outcome is tail, it is presented by -1. That means,
signal [ ] = +1, if outcome equals heads or [ ] = − 1, if outcome equals tails. Since
the outcome of the coin toss experiment is random, the signal itself is random in nature and
this is a discrete time random signal. On the other hand, a continuous time random signal
includes speech signal, record of temperature of the city in a particular month over a time or
it can be interpreted as noise with too many variations along the time axis t. This noise tends
to limit the performance of a system and hence it is important to understand the behavior of
the noise to characterize the performance and behavior of any system. Fig. 1.22 shows the
CT random signal.

Fig 1.22: CT random signal


Signals and Systems | 21

1.4.2 Even and Odd signals


Another classification of the signals is even and odd type of the signal. For continuous time
signal an even function is represented as,
( )= (− ) (1.36)

Or for a discrete time signal an even function is represented as,


[ ] = [− ] (1.37)
For example, cos(2 ) is a CT even signal as cos(2 ) = cos(−2 ) as shown in Fig.
1.23.

Fig 1.23: CT Even signal

For continuous time signal an odd function is represented as,


( ) = − (− ) (1.38)

Or for a discrete time signal an even function is represented as,


[ ] = − [− ] (1.39)

For example, sin(2 ) is an odd signal as sin(2 ) = −sin(2 ) as shown in Fig. 1.24.

Fig 1.24: CT Odd signal


22 | Introduction to Signals and Systems

Hence, it can be seen that the even signals are symmetric about 0 that means they have even
symmetry and the odd signals are anti-symmetric about the vertical axis that means they have
odd symmetry.
Any real valued CT or DT unsymmetric signal can be represented in terms of its even and
odd parts as,
( )= ( )+ ( ) (1.40)
and

[ ]= [ ]+ [ ] (1.41)

Where, base E represents even part of the signal and base O represents odd part of the signal.

( )= { ( )} = { ( ) + (− )} (1.42)

( )= { ( )} = { ( ) − (− )} (1.43)

[ ]= { [ ]} = { [ ] + [− ]} (1.44)

[ ]= { [ ]} = {[ ] − [− ]} (1.45)

1.4.2.1 Important properties of Even and Odd Functions


1. Product of two even or odd signals is an even signal.

Proof: Let ( ) and ( ) are even signals. If ( ) = ( ) ( ) then,

(− ) = (− ) (− ) = ( ) ( )= ( )

If ( ) and ( ) are odd signals. If ( ) = ( ) ( ) then,

(− ) = (− ) (− ) = (− ( )) − ( ) = ( ) ( )= ( )

2. Product of one even and one odd signal is an odd signal.

Proof: Let ( ) is an even signal and ( ) is an odd signal, then

− (− ) = −[ (− ) ] = −[ ( ) (− ( ))] = ( ) ( )= ( )

3. ∫ ( ) = 2∫ ( ) where, ( ) is an even signal.


Signals and Systems | 23

4. (0) = [0]=0 and ∫ ( ) = 0 where, [ ] is an odd signal.

1.4.3 Periodic and Aperiodic signals


A signal is called as periodic if there exist a time period T such that,
( + ) = ( ), −∞ ≤ ≤ ∞ (1.46)
In Fig 1.25, we can see a periodic continuous time triangular wave that is repeating itself after
every period T, 2T, 3T towards positive as well as negative direction. For any T which is the
smallest positive value for which Eq. (1.46) holds is called the fundamental period of the
signal. Eq. (1.47) shows the periodic CT signal where same structure is repeating and it
follows,
( )= ( + ) = ( +2 )= ( +3 )= ⋯= ( + ) (1.47)
for all and any integer . Therefore, a periodic signal is an everlasting signal that exists over
the entire interval −∞ ≤ ≤ ∞. However, no physical signals are actually periodic as they
all begin at some finite time in the past and stop at some finite time in the future.

Fig 1.25: CT periodic signal

A regular example of periodic signal is sinusoidal signal which is written as,

x(t) = Asin(Ω t + ∅) = (2 F + ∅) (1.48)


where, A represents amplitude, Ω is frequency in radians/second, ∅ is the phase in radians
and F is frequency in Hertz. The fundamental period T for this signal can be written as,
2 1
= =
Ω F

Now consider two periodic signals, ( ) and ( ), with fundamental periods and ,
respectively. The sum of two periodic signals may or may not be periodic.
( )= ( + ), ( )= ( + ) (1.49)
24 | Introduction to Signals and Systems

Where, and are integers such that


( )= ( + )+ ( + ) (1.50)

In order for ( ) to be periodic with period T,


( + )= ( + )+ ( + )= ( + )+ ( + ) (1.51)
Thus,

= = (1.52)

or

= (1.53)

The sum of two periodic signals is periodic only if the ratio of their respective periods can be
expressed as a rational number. We can say that the fundamental period of ( ) is the smallest
positive value of that is an integer multiple of both and and this value is called as least
common multiple (LCM) of and . If is an irrational number, then ( ) and ( ) do
not have common period and ( ) is aperiodic. The Fig. 1.26 shows the example of aperiodic
signal which does not follows the conditions satisfied by the periodic signal that means
aperiodic signals are not periodic with respect to defined fundamental period.

Fig 1.26: CT aperiodic signal

For discrete signal,


[ + ] = [ ] for all (1.54)
Where is positive integer and the period of discrete time signal. It means that signal [ ]
repeats itself every samples as shown in Fig. 1.28. The smallest for above condition
holds is called fundamental period .
Signals and Systems | 25

Fig 1.27: DT periodic signal

A discrete time periodic signal follows,

[ ]= [ + ]= [ +2 ]= [ +2 ]=⋯= [ + ] (1.55)
for all and any integer . If [ ] is periodic with period N, then it is also periodic with
period 2 , 3 , ….

A discrete time sinusoid can be written as,


[ ]= [ + ∅] (1.56)

Where A is amplitude, is frequency in radians/sample and ∅ is phase in radians.


When a DT sequence is not periodic with the fundamental period then it is called as DT
aperiodic sequence.

Example 1.2:
Determine whether the following continuous time signals are periodic or not. If periodic, find
the fundamental period T.
1. ( ) = (50 )
2. ( ) = 20 (10 + /6)
3. ( ) = 2 cos(10 + 1) – sin(4 − 1)
4. ( ) = 60 + 50
5. ( ) = 3 cos 4 + 2 sin

Solution:

1. Given, ( ) = ( )

( ) = (50 ), where = 50 π.

Time period T = 2 π / = 2 π /50 π = 1/25.


26 | Introduction to Signals and Systems

2. Given, ( ) = ( + / )

= 10 π

T=2 / = 2 π / 10 π = 1/5

3. , ( ) = ( + )– ( − )

let ( )=2cos(10t+1)

( )= sin(4t-1)

= 2 π / 10 = π /5 and = 2 π/4 = π /2

Since, ratio / = (π /5) / (π /2) = 2/5 is a rational number, signal x (t) is periodic with
fundamental period

T=LCM( , )=LCM(π /5, π /2)= π

4. Given, ( ) = +

T1 = 2 / 60 = 1 / 30 sec

T2 = 2 50 = 1 25 sec

T = 1 / 2 = (1 / 30) / (1 /25) = 5 / 6

T= 6T1 = 5T2

T = 1 / 5 sec.

5. , ( ) = +

T1 = 2 /4 = / 2 sec

T2 = 2 / = 2 sec

T= 1/ 2= /4

Since it is not a rational number the signal is not periodic.


Signals and Systems | 27

1.4.4 Energy and Power signals


Energy and power are the two physical characteristics of the signal which are considered in
many applications. A familiar example may include, due to current ( ), the power is
delivered to the resistor and voltage source is ( ) can be written as,
( )
( )= ( ) ( )= ( ) = (1.57)

Hence the total energy at equals,


1
= ( ) = ( ) = ( )

Assume, = 1Ω, then

= ( ) = ( )

Hence for any continuous signal ( ), the energy E can be written as,

= ∫ | ( )| (1.58)

Similarly, energy of a discrete time signal [ ] can be written as,

=∑ | [ ]| (1.59)
A signal is called as energy signal if the energy is finite, i.e., 0 < < ∞. For example,
( ) or [ ] are the energy signals.
The power of a periodic continuous signal ( ) which is periodic with period T can be written
as,

= lim ∫ | ( )| (1.60)

The power of a discrete time signal ( ) can be written as,

= lim ∑ | [ ]| (1.61)

A signal is called as power signal if the power is finite (0 < < ∞) and non-zero ( ≠ 0).
A signal with finite energy has zero power and a signal with finite power has infinite energy.
A signal neither is an energy signal nor a power signal if it has infinite energy and zero power.
For example, CT and DT ramp signal i.e., ( ) and [ ] are neither energy nor power signal.
28 | Introduction to Signals and Systems

Example 1.3: Determine the energy E and power P of following signals.


1. ( ) = 2 ( ) >0

2. ( )= ( − 2)

3. [ ] = 3(0.5) [ ]

Solution:

1. Given, ( ) = ( ) >
=∫ | ( )| =∫ 2

=4


=4
−2 0

4 ∞
= [ ]
−2 0

4
= [ − ]
−2

4
= [ − ]
−2

4 4 2
= [ 0 − 1] = =
−2 2

As E value is finite, ( ) is an energy signal.


1
= lim | ( )|

Signals and Systems | 29

1
= lim 2 .

4
= lim

4
= lim [ ]
→ −2

4
= lim [ − ]
→ −2

=0
Since, P=0, ( ) is not a power signal.

2. Given, ( ) = ( − )


= | ( )| = = =
1 2
2

=∞
E is infinite, ( ) is not an energy signal
1
= lim | ( )|

1
= lim

1
= lim
→ 1
2
2
= lim

30 | Introduction to Signals and Systems

2
= lim − (2)

2
= lim √ − √2

=0
Since, = 0, ( )is also not a power signal.

3. Given, [ ] = ( . ) [ ]

[ ] = 3(0.5) [ ]

= | ( )|

= |3(0.5) |

= 9(0.5)

=9 (0.25)

Using, ∑ = for | | < 1

1
∴ =9
1 − 0.25
4
=9
3
= 12
E has finite value. Hence, ( ) is an energy signal.
Now,
Signals and Systems | 31

1
= lim | ( )|

1
= lim 9(0.25)

9
= lim (0.25)

Using, ∑ = , ≠1

9 1 − (0.25)
= lim
→ 1 − 0.25
=0
Since power P=0, [ ] is not a power signal.

1.4.5 Finite (Time-limited) and Infinite Duration signals


Finite or Time-limited signals are signals that exist only over a specific finite duration or time
interval. These signals are typically non-zero only within a specific time range. Outside that
range, they are considered to be zero. Time-limited signals are often encountered in practical
applications, such as audio signals that have a definite start and end time. The examples of
CT and DT finite duration signals are shown by Fig 1.28.

Fig 1.28: Examples of finite duration signals (a) CT finite-duration signal (b) DT finite-duration signal

Infinite duration signals are signals that exists from -∞ to ∞ i.e., for finite interval of time.
The examples of CT and DT infinite duration signals are shown by Fig 1.29.
32 | Introduction to Signals and Systems

Fig 1.29: Examples of infinite duration signals (a) CT infinite-duration signal (b) DT infinite-duration
signal

1.4.6 Causal and Noncausal signals


The signals that are zero for < 0 (for CT signals) or < 0 (for DT signals) are known as
causal (right-sided) signals. Examples of such signals are shown by Fig. 1.30.

Fig 1.30: Examples of causal signals (a) CT causal signal (b) DT causal signal

The signals that are zero for ≥ 0 (for CT signals) or ≥ 0 (for DT signals) are known as
Noncausal (left-sided) signals. Examples of such signals are shown by Fig. 1.31.

Fig 1.31: Examples of noncausal signals (a) CT noncausal signal (b) DT noncausal signal
Signals and Systems | 33

1.5 Some signal properties


1.5.1 Absolute integrability:
Absolute integrability is a property that describes the integrability of a signal over its defined
domain. A signal is said to be absolutely integrable if the integral of its absolute value exists
and is finite. Mathematically, if the integral of | ( )| from −∞ to +∞ is finite, then the
signal is absolutely integrable. This property is important in signal processing and analysis as
it ensures that the energy or power of the signal is well-defined.

Suppose we have a strictly time limited signal that is a rectangular pulse, so obviously this
curve has a finite area under it. Therefore, we can say this signal is absolutely integrable.

1.5.2 Determinism:
Determinism refers to the property of a signal that follows a predictable and deterministic
pattern.

A signal is said to be deterministic if there is no uncertainty with respect to its value at any
instant of time or the signals which can be defined exactly by a mathematical formula are
known as deterministic signals. Examples of deterministic signals include most mathematical
functions, such as polynomial functions or exponential functions.

Fig 1.32: Deterministic Signal

1.5.3 Stochastic character:


Stochastic character, also known as randomness or probabilistic behavior, refers to the
property of a signal that exhibits randomness or uncertainty. Stochastic signals are
characterized by having a random or unpredictable nature, and their future values cannot be
precisely determined. Stochastic signals are often modelled using statistical techniques and
probability distributions. Examples of stochastic signals include noise signals, stock market
fluctuations, or weather patterns.
34 | Introduction to Signals and Systems

Fig 1.33: Random (stochastic) Signal

Example 1.4: Represent the following discrete time signals graphically. (Arrow indicates
position of = 0)
[ ] = {1, 2,3,4,5}

[ ] = {−6, −3,2, 5, 1,3,7,8}

[ ] = { 4,3,1,0,5, 3}

Solution:
1. The given sequence is [ ] = { , , , , }

As the arrow position indicates, = 0
i.e. [0] = 1, [1] = 2, [2] = 3, [3] = 4, [4] = 5
So, the plot of [ ] (Amplitude) versus time is,
Signals and Systems | 35
2. The given sequence is, [ ] = {− , − , , , , , , }

i. e. [0] = 5, [1] = 1 , [2] = 3, [3] = 7, [4] = 8, [−1] = 2, [−2] = −3, [−3] = −6

3. The given sequence is [ ] = { − , , , , , }



i.e. [0] = 3, [1] = 1, [2] = 0, [3] = 5, [4] = 3, [−1] = −4

1.6 Continuous/Discrete Amplitude Signals

1.6.1 Continuous Amplitude Signals:


Continuous amplitude signals can take on any value within a continuous range. In other
words, the amplitude of the signal can vary continuously over time. These signals are often
encountered in analog systems, where the signal can have an infinite number of possible
36 | Introduction to Signals and Systems

amplitude values. Analog audio signals and continuous waveform signals are examples of
continuous amplitude signals.

1.6.2 Discrete Amplitude Signals:


Discrete amplitude signals can only take on a finite set of values. The amplitude of the signal
is restricted to specific levels or values. Discrete amplitude signals are commonly
encountered in digital systems, where the amplitude is represented using a finite number of
bits. Digital audio signals and discrete waveform signals are examples of discrete amplitude
signals.

Continuous-time discrete amplitude signals are basically digital signals. Discrete amplitude
is one which we get through quantization process and it depends how many levels of
quantization one wants. In simple words, quantization means assigning the amplitude values
of any analog signal to certain discrete levels, equidistant of each other based on certain
criteria. A square wave is a continuous-time discrete amplitude signal. For binary signals,
there are only two quantization levels (0, 1). Some examples of continuous-time discrete
amplitude signals are shown by Fig. 1.34 (a) and (b).

(a)

(b)
Fig. 1.34 Continuous and Discrete Amplitude Signals
Signals and Systems | 37

1.7 Continuous –Time (CT) and Discrete-Time (DT) systems


A system can be recognized as an interconnection of the physical components or subsystems
or software to perform the activities or tasks. It operates on the provided input signal,
processes it accordingly and provides the output. A system that processes continuous–time
(CT) signal and generates continuous–time (CT) signal is called continuous –time (CT)
system. The input to output relation for CT system is given as,
( )→ ( ) (1.62)
Similarly, a system that processes discrete–time (DT) signal and generates discrete–time (DT)
signal is called discrete–time (DT) system. The input to output relation for DT system is given
as,
[ ]→ [ ] (1.63)

1.8 System properties: Linearity: additivity and homogeneity, Shift- invariance,


Causality, Stability, Realizability.

1.8.1 Linear and Nonlinear Systems

A system is linear if it follows the two principles that are additivity and homogeneity.

1. Additivity property: Additivity means that the response of the system to the sum of two
inputs is equal to the sum of the individual responses to each input.
Mathematically, if input ( ) produces output ( ) and input ( ) produces output ( ),
then ( ) + ( ) must produce ( ) + ( ).
2. Homogeneity/scaling property: Homogeneity means that scaling the input signal by a
constant scales the output response by the same constant.

Mathematically, if input ( ) is scaled by a constant i.e., ( ), then it must produce the


scaled output ( ). Hence, a system that satisfies both additivity and homogeneity property
(combinely called as superposition principle) is called as linear system. Mathematically, for
CT system, if input ( ) produces output ( ) and input ( ) produces output ( ), then
a linearly combined input ( )= ( )+ ( ) must produce ( ) = ( )+
( ).
38 | Introduction to Signals and Systems

i.e., ( )+ ( )→ ( )+ ( ) (1.64)

Mathematically, for DT system, if input [ ] produces output [ ] and input [ ]


produces output [ ], then a linearly combined input [ ]= [ ]+ [ ] must
produce [ ] = [ ]+ [ ].

i.e., [ ]+ [ ]→ [ ]+ [ ] (1.65)

1.8.2 Time-Invariant and Time-Variant Systems


A system is called as time invariant if its behaviour does not change with respect to time.
That means, if a system is provided an shifted input in time by , then the system will produce
an output shifted by same time . For example in case of CT system, if input ( ) is delayed
by i.e., ( − ), then system produces delayed output ( − ). Similarly, for DT
system, if input [ ] is delayed by i.e., [ − ], then system produces delayed output
[ − ].
On the other hand, a system is called as time variant if its behaviour changes with respect to
time.

Procedure to check for time invariance:

Delay the input signal by and check the response of the system ( ).
Delay the output of the system for unshifted input by . Let this delayed response is ( ).
Check whether ( ) = ( ). If they are equal, system is time invariant otherwise time
variant.

Example 1.5
Check whether the following systems are time invariant or not.
1. ( )= 5 ( )
2. ( )= ( )sin(10 )
3. ( ) = 3 ( )
( )
4. ( ) = 4
5. ( ) =

Solution:
1. Given, ( ) = ( )

Delay the input signal by


( )= 5 ( − )

Delay the output of the system by


Signals and Systems | 39
( ) = 5( − ) ( − )
Here,

( )≠ ( ), hence the system is time variant.

2. Given, ( ) = ( ) ( )

Delay the input signal by


( )= ( − )sin(10 )

Delay the output of the system by

( )= ( − ) sin 10 ( − )
Here,

( )≠ ( ), hence the system is time variant.

3. Given, ( ) =

Delay the input signal by


( ) = 3 ( 2− )

Delay the output of the system by

( ) = 3 (( − )2 )
Here,

( )≠ ( ), hence the system is time variant.

( )
4. Given, ( ) =

Delay the input signal by


( )= 4 ( )

Delay the output of the system by

( )=4 ( )

Here,

( )= ( ), hence the system is time invariant.


40 | Introduction to Signals and Systems
5. Given, ( ) =

Delay the input signal by


( )= 2

Delay the output of the system by

( ) = ( − )2
Here,

( )≠ ( ), hence the system is time variant.

1.8.3 Causal and Noncausal Systems

A causal system produces an output response that depends only on present and past values of
the input signal. In other words, Causality means that the output of the system does not depend
on future inputs, but only on past input. On the other hand, the output of a noncausal system
depends upon present, past as well as future values of the input signal. All the physical
systems in the real world are the noncausal systems.

Mathematically, the output response ( ) at time is determined solely by the input signal
( ) for τ ≤ .

Example 1.6: Determine whether the given CT and DT systems are causal or noncausal.

1. ( ) = ( + 1)
2. ( ) = ( ) + ( − 1)
3. [ ] = [ ] + [ − 3]

Solution:

1. Given, ( ) = ( + )

When t = 0, y(0) = x(1), which implies that, The response at t = 0, i.e. , y(0) depends
on the future value of input x(0).
Signals and Systems | 41

When = 1, (1) = (2), which implies that, The response at = 1, i.e. , (1) depends
on the future value of input (2).

From the above analysis we can say that for any value of , the system output depends on
future inputs. Hence the system is noncausal.

2. Given, ( ) = ( ) + ( − )

When = 0, (0) = (0) + (−1), which implies that, The response at = 0, i.e. , (0)
depends on the present input (0) and past input (−1).

When = 1, (1) = (1) + (0), which implies that, The response at = 1, i.e. , (1)
depends on the present input (1) and past input (0).

From the above analysis we can say that for any value of , the system output depends on
present and past value of inputs. Hence the system is causal.

3. Given, [ ] = [ ]+ [ − ]

When = 0, [0] = 0 [0] + [−3], which implies that, The response at = 0, i.e. , [0]
depends on the present input [0] and past input [−3].

When = 1, [1] = 1 [1] + [−2], which implies that, The response at = 1, i.e. , [0]
depends on the present input [1] and past input [−2].

From the above analysis we can say that for any value of , the DT system output depends
on present and past value of inputs. Hence the system is causal.

1.8.4 Stable and Unstable Systems

Stability refers to the boundedness of the system's response. A system is considered stable if,
for bounded input signals, the output response remains bounded i.e., small inputs lead to
output that do not diverge. For example, if we apply only little pressure to push the object, it
42 | Introduction to Signals and Systems

will move only a little bit. In other words, if the input signal is finite, the output signal should
also be finite.

On the other hand, if a small signal causes the output signal to be arbitrarily large then that
system is called as unstable system. The examples of bounded and unbounded signals are
shown by Fig 1.35.

Fig. 1.35 Examples of bounded and unbounded signal

Stability is also defined by the boundedness. The input signal ( ) is said to be bounded if
there exist a constant (0 < < ∞), such that

| ( )|≤ for all t


(1.66)

Similarly, the output signal is bounded if it satisfies the condition |y(t)| ≤ My < ∞

Some of the examples of bounded input signal are step signal, decaying exponential signal
and impulse signal. Examples of unbounded input signal are ramp signal and increasing
exponential signal.

Note: For a bounded signal, the amplitude is finite.


Signals and Systems | 43

Example 1.7 Check for the stability of given systems.

1. ( ) = cos( ( ))
2. [ ] = [ − 1]
3. ( ) = ( )
4. ( ) = ∫ ( )

Solution:

1. Given, ( ) = ( ( ))

The value of θ lies between 1 to +1 for any value of θ. Therefore, the output ( ) is
bounded for any value of input ( ). Hence the given system is stable.

2. Given, [ ] = [ − ]
For an arbitrary signal [ ], | [ ]| ≤ for all
Then delayed input [ − 1] is also bounded by ,
| [[ − 1]| ≤ for all
The output is
[ ]=
Hence, the DT system is stable.
3. Given, ( ) = ( )
The given system is a time variant system, and so the test for stability should be performed
for specific inputs. There can be two cases for the existence of ( ).
Case 1: Let ( ) tends to ∞ or constant, as tends to infinity.
In this case, ( ) = ( ) will be infinity as t tends to infinity and so the system is unstable.
Case 2: Let ( ) tends to 0, as tends to infinity. In this case ( ) = ( ) will be zero as
tends to infinity and so the system is stable

4. Given, ( ) = ∫ ( )

Let the input ( ) be ( ) then


44 | Introduction to Signals and Systems

( ) = ∫ ( ) but ∫ ( ) = ( ) = ramp signal

It is unbounded because the amplitude of ramp is not finite and tends to become infinite when
t →infinite

Hence, the system is unstable.

1.8.5 Realizability:

Realizability is a property that determines whether a given system can be implemented


physically. It considers practical considerations like available resources, physical
constraints, and feasibility. A realizable system can be physically built and operated.

Consider the first system, ( ) = ( − 1) is a causal system, because its output is a time-
delayed version of the original signal.

On the other hand, consider the second system, ( ) = ( + 1), is non-causal, because
its output is a time-advanced version of the input signal. This means, that for example at
the output time t=0, the system requires access to the value of the input signal at time t=1.
Clearly, this is impossible in a realizable system, as nobody can predict the future.

UNIT SUMMARY

This this chapter, we have seen the introduction to signals and systems that covers the
fundamental concepts and properties of CT and DT signals. It explores different signal types,
including periodicity, determinism, and stochastic character. Special signals such as the unit
step, impulse, and ramp are discussed. Signals can exist in continuous or discrete domains with
continuous or discrete amplitudes. System properties, including linearity, time-invariance,
causality, stability, and realizability are also covered. Understanding signals and systems is
crucial for various engineering and scientific applications.
Signals and Systems | 45
EXERCISES
Multiple Choice Questions and Answers
1. Signals and systems are relevant in which areas?
a) Everyday life
b) Engineering
c) Science
d) All of the above
Answer: d) All of the above

2. Which of the following are signal properties?


a) Periodicity
b) Absolute integrability
c) Determinism
d) Stochastic character
e) All of the above
Answer: e) All of the above

3. Which of the following are special signals?


a) Unit step
b) Unit impulse
c) Sinusoid
d) Complex exponential
e) All of the above
Answer: e) All of the above

4. Continuous-time signals exist in which domain?


a) Continuous
46 | Introduction to Signals and Systems

b) Discrete
c) Both
d) None
Answer: a) Continuous

5. Discrete-time signals exist in which domain?


a) Continuous
b) Discrete
c) Both
d) None
Answer: b) Discrete

6. Linearity of a system refers to:


a) Additivity and homogeneity
b) Shift-invariance
c) Causality
d) Stability
Answer: a) Additivity and homogeneity

7. Which property ensures that a system's output remains bounded for bounded input signals?
a) Linearity
b) Shift-invariance
c) Causality
d) Stability
Answer: d) Stability
Signals and Systems | 47

8. Realizability of a system refers to:


a) Additivity and homogeneity
b) Shift-invariance
c) Causality
d) Feasibility of physical implementation
Answer: d) Feasibility of physical implementation

9. A special time-limited signal is characterized by:


a) Periodicity
b) Absolute integrability
c) Determinism
d) Time limitations
Answer: d) Time limitations

10. Which of the following signal properties relates to randomness?


a) Periodicity
b) Absolute integrability
c) Determinism
d) Stochastic character
Answer: d) Stochastic character

Short and Long Answer Type Questions

1. Give an example of a signal exhibiting periodicity.

2. What does it mean for a signal to have absolute integrability?

3. Define determinism in the context of signals.

4. Provide an example of a special time-limited signal.


48 | Introduction to Signals and Systems
5. Explain the concept of shift-invariance in systems.

6. Discuss the applications of signals and systems in engineering and science.

7. Explain the properties and characteristics of the sinusoidal signal.

8. Discuss the importance of linearity in systems and provide examples.

9. Explain the concept of causality in systems and its significance.

10. Discuss the differences between continuous and discrete amplitude signals.

Numerical Problems

1. Consider a periodic signal with a period of T = 4 seconds. Find the frequency of the signal.
2. Determine if the following signal is absolutely integrable: x(t) = e^(-2t) for t ≥ 0.
3. Calculate the average value of the signal x(t) = 3sin(2πt) over the interval 0 ≤ t ≤ 2 seconds.
4. Given a system with the impulse response h(t) = 2e^(-t)u(t), where u(t) is the unit step
function, find the response of the system to the input signal x(t) = 3u(t).
5. Determine if the system described by the difference equation y[n] = 0.5y[n-1] + x[n] is
linear or nonlinear.
6. Consider a discrete-time system with the input signal x[n] = {1, 2, 3, 4} and the impulse
response h[n] = {1, -1, 2, -2}. Calculate the output signal y[n] using the convolution sum.
7. Determine if the system described by the following difference equation is time-invariant
or time-varying: y[n] = x[n] + x[n-1].
8. Test the stability of the continuous-time system with the transfer function H(s) = 1/(s + 2).
9. Determine if the system with the transfer function H(z) = (1 - z^(-1))/(1 + z^(-1)) is stable
in the discrete-time domain.
10. Given a system with the input signal x(t) = 4sin(3πt) and the output signal y(t) = 2sin(3πt
+ π/4), calculate the gain of the system.
Signals and Systems | 49
KNOW MORE
Signals and systems are fundamental concepts that permeate our daily lives and play a vital role
in various fields of engineering and science. Signals exhibit different properties such as
periodicity, which describes their repetitive nature, and absolute integrability, which quantifies
the energy or power content of a signal. Signals can be deterministic, meaning they have a
predictable behavior, or stochastic, displaying random characteristics. Special signals of
significance include the unit step, representing abrupt changes, the unit impulse, denoting
instantaneous events, sinusoids, fundamental periodic waveforms, and complex exponentials,
integral to signal processing. Time-limited signals have finite duration, and they can be
continuous or discrete in both the time and amplitude domains. Systems, on the other hand,
possess distinct properties that govern their behavior. Linearity signifies that the response of a
system to a sum of inputs is the sum of their individual responses, while additivity and
homogeneity describe their scaling behavior. Shift-invariance indicates that shifting the input
signal leads to a corresponding shift in the output response. Causality denotes that the output
depends only on past and present inputs. Stability ensures that the system produces bounded
output responses for bounded inputs, and realizability signifies the practical implementability
of the system using realizable components or algorithms. Understanding these properties and
concepts is vital in comprehending the behavior and characteristics of signals and systems,
enabling their analysis, manipulation, and design in numerous applications across engineering
and scientific disciplines.

REFERENCES AND SUGGESTED READINGS


1. Signals and Systems by Simon Haykin
2. Signals and Systems by Ganesh Rao
3. Signals and Systems - Course (nptel.ac.in)
50 | Introduction to Signals and Systems
Dynamic QR Code for Further Reading
Signals and Systems | 51

2 Behavior of Continuous
and Discrete-time LTI
Systems
UNIT SPECIFICS
Through this unit we have discussed the following aspects:
 Impulse response and step response provide information about a system's characteristics
and its response to specific inputs.
 Convolution is used to compute the output of a system by integrating the product of the input
and the shifted impulse response.
 LTI systems can process aperiodic and convergent inputs, and their output can be computed
through convolution or other techniques.
 Cascade interconnections involve connecting multiple LTI systems in series by convolving
their impulse responses.
 Causality refers to a system's output depending only on past or present input values, while
stability means the output remains bounded for any bounded input.
 Differential equations represent continuous-time LTI systems, while difference equations
represent discrete-time LTI systems.
 State-space representation describes a system using first-order differential or difference
equations, including state variables and input-output relationships.
 State-space analysis allows for studying the behavior of systems in terms of their state
variables and can handle multi-input, multi-output systems.
 The state transition matrix relates the initial state to the state at any given time and is
essential in solving state-space equations.
 Frequency response describes how an LTI system responds to different frequencies in the
input, and it is related to the impulse response through Fourier analysis.
 Periodic inputs, such as sinusoidal waves, can be analyzed using the notion of frequency
response to understand the system's behavior in the frequency domain.

This unit focuses on the behavior of Continuous and Discrete-time Linear Time-Invariant (LTI)
Systems. It covers key topics including impulse response, step response, and convolution. The
52 | Behavior of Continuous & Discrete-time LTI Systems

unit explores how LTI systems respond to aperiodic and convergent inputs, emphasizing
techniques such as convolution to determine the output. Cascade interconnections, where
multiple systems are connected in series, are examined by convolving their impulse responses.
Causality and stability in LTI systems are characterized, where causality refers to the past and
present dependency between input and output, and stability ensures bounded output for any
bounded input.
The unit addresses system representation through differential equations for continuous-time
systems and difference equations for discrete-time systems. It introduces state-space
representation, which describes systems using first-order differential or difference equations,
facilitating state-space analysis and the study of multi-input and multi-output systems. The role
of the state transition matrix is explored, connecting the initial state to the state at any given
time. The unit also covers periodic inputs applied to LTI systems, investigating the notion of
frequency response and its relationship with the impulse response. This understanding offers
insights into how LTI systems respond to different frequencies in the input signal.
Overall, this unit provides comprehensive coverage of impulse response, step response,
convolution, input-output behavior, causality, stability, system representation, state-space
analysis, state transition matrix, periodic inputs, and the relationship between frequency
response and impulse response. It equips learners with a solid foundation in analyzing and
understanding the behavior of Continuous and Discrete-time LTI Systems.

RATIONALE

The unit on “Behavior of Continuous and Discrete-time LTI Systems" is to provide students
Understanding the behavior of Continuous and Discrete-time Linear Time-Invariant (LTI)
Systems is crucial in various engineering and scientific disciplines. This 8-hour unit is designed
to provide students with a comprehensive understanding of LTI systems and their
characteristics.

Impulse response and step response are fundamental concepts in LTI systems. By studying these
responses, students gain insights into how a system reacts to specific inputs and determine its
dynamic behavior. Convolution is a key operation used to compute the output of a system, and
it plays a vital role in analyzing LTI systems.

The unit focuses on the input-output behavior of LTI systems with a particular emphasis on
aperiodic and convergent inputs. Students learn how to analyze and determine the system's
response using techniques such as convolution. Cascade interconnections of LTI systems are
explored, as they are commonly encountered in real-world applications. By understanding the
convolution of impulse responses, students can analyze and predict the behavior of
interconnected systems.
Signals and Systems | 53

Characterizing causality and stability is essential for assessing the reliability and predictability
of LTI systems. Students examine the concepts of causality, where the output depends on past
and present inputs, and stability, which ensures bounded output for any bounded input. These
characterizations provide valuable insights into system behavior and performance.

System representation is an important aspect covered in the unit. Students learn to represent
LTI systems using differential equations for continuous-time systems and difference equations
for discrete-time systems. State-space representation is introduced as a powerful method to
describe and analyze complex systems. It provides a framework for studying multi-input, multi-
output systems and understanding their interactions.

The role of the state transition matrix is explored, emphasizing its significance in relating the
initial state to the state at any given time. This matrix plays a crucial role in solving state-space
equations and analyzing system behavior over time.

Periodic inputs and the notion of frequency response are examined to understand how LTI
systems respond to different frequencies in the input. The frequency response is related to the
impulse response through Fourier analysis, enabling students to analyze system behavior in the
frequency domain.

Overall, this unit equips students with the necessary knowledge and skills to analyze and
understand the behavior of Continuous and Discrete-time LTI Systems. The concepts covered,
such as impulse response, step response, convolution, input-output behavior, causality,
stability, system representation, state-space analysis, state transition matrix, periodic inputs,
and frequency response, are essential for successful engineering and scientific applications.

PRE-REQUISITES

1. Strong understanding of mathematics, including algebra, calculus, and complex numbers.


2. Familiarity with basic concepts in signals and systems, such as time-domain and frequency-
domain representations, Fourier analysis, and convolution.
3. Proficiency in solving ordinary differential equations and understanding linear algebra
concepts.
4. Basic knowledge of electronics and circuit analysis for understanding continuous-time LTI
systems.
5. Knowledge of digital signal processing concepts for understanding discrete-time LTI
systems.
6. These pre-requisites provide the necessary foundations to effectively engage with the
content and concepts covered in the unit on the behavior of Continuous and Discrete-time
LTI Systems.
54 | Behavior of Continuous & Discrete-time LTI Systems

UNIT OUTCOMES
List of outcomes of this unit is as follows:
U2-O1: Understand the concept of impulse response and step response.
U2-O2: Apply convolution to analyze the behavior of LTI systems.
U2-O3: Analyze the input-output behavior of LTI systems with aperiodic convergent inputs.
U2-O4: Understand and apply cascade interconnections of LTI systems.
U2-O5: Characterize the causality and stability of LTI systems.
U2-O6: Represent LTI systems through differential equations and difference equations.
U2-O7: Understand and apply state-space representation of systems.
U2-O8: Perform state-space analysis of LTI systems.
U2-O9: Analyze multi-input, multi-output systems.
U2-O10: Understand the role and application of the State Transition Matrix.
U2-O11: Analyze the behavior of LTI systems with periodic inputs.
U2-O12: Understand the concept of frequency response and its relation to the impulse
response.

EXPECTED MAPPING WITH COURSE OUTCOMES


Unit-2
(1- Weak Correlation; 2- Medium correlation; 3- Strong Correlation)
Outcomes
CO-1 CO-2 CO-3 CO-4 CO-5 CO-6
U2-O1 3 3 1 2 2 3
U2-O2 3 3 2 1 - 1
U2-O3 3 2 2 - - 3
U2-O4 3 3 2 1 - 2
U2-O5 3 3 2 1 1 2

"Education is the most powerful weapon which you can use to change the world."

- Nelson Mandela
Signals and Systems | 55

2.1 Introduction
In chapter 1, we have discussed types of systems and their properties. Two properties namely
linearity and time-invariance play very important roles in the analysis of signals and systems
since most of the practical systems possess these two properties. We call such systems as linear
time-invariant (LTI) systems. In our study of signals and systems, we will be especially
interested in systems that demonstrate both properties, which together allow the use of some of
the most powerful tools of signal processing.

2.1.1 Linear Time Invariant Systems


a) Linear Systems
A system is said to be a linear system if the system follows the linear scaling and superposition
principle as discussed below.
Linear Scaling: When the input to a given system is scaled by a constant value, if the output
of the system is also scaled by the same amount, the system is said to follow the linear scaling
property. It is demonstrated in Figure 2.1.

Fig 2.1 Linear scaling

In above Figure 2.1(a), the input to the linear system gives the output . If is scaled by a
value α and passed through this same system, as in Figure 2.1(b), the output will also be scaled
by α.
Superposition Principle: The linear system also obeys the principle of superposition. This
means that if two inputs are added together and passed through a linear system, the output will
be the sum of the outputs corresponding to their individual inputs. It is demonstrated in Figure
2.2.
56 | Behavior of Continuous & Discrete-time LTI Systems

Fig 2.2 Superposition Principle

That is, if cases in Figure 2.2 (a) and (b) are true then Figure 2.2 (c) is also true for a linear
system. The scaling property mentioned above still holds in conjunction with the superposition
principle. Therefore, if the inputs and are scaled by factors α and β, respectively, then the
sum of these scaled inputs will give the sum of the individual scaled outputs:

Fig 2.3 Superposition Principle with Linear Scaling


Signals and Systems | 57

2.1.2 Time Invariant Systems


A time-invariant (TI) system has the property that a certain input will always give the same
output (up to timing), without regard to when the input was applied to the system. In other
words, if y(t) is the output of the system corresponding to its input x(t), then y(t − t ) will be
the output when a delayed input x(t − t ) is applied to the system for all values of t as
demonstrated in Figure 2.4.

Fig 2.4 Time-Invariant Systems

In this figure, ( ) and ( − ) are passed through the system TI. Because the system TI is
time-invariant, the inputs ( ) and ( − ) produce the same output. Whether a system is
time-invariant or time-variant can be seen in the differential equation (or difference equation)
describing it. Time-invariant systems are modelled with constant coefficient equations. A
constant coefficient differential (or difference) equation means that the parameters of the system
are not changing over time and an input now will give the same result as the same input later.

2.1.3 Linear Time Invariant Systems


Certain systems are both linear and time-invariant, and are thus referred to as LTI systems.

Fig 2.5 Linear Time-Invariant Systems


58 | Behavior of Continuous & Discrete-time LTI Systems

As LTI systems are a subset of linear systems, they obey the principle of superposition. In the
figure below, we see the effect of applying time-invariance to the superposition definition in
the linear systems section above.

Fig 2.6 Superposition in Linear Time-Invariant Systems

2.2 Impulse response of LTI System

Fig 2.7 CT impulse response

Fig 2.8 DT impulse response

2.2.1 Discrete-Time Unit Impulse Response and the Convolution


Let ℎ [ ] be the response of the LTI system to the shifted unit impulse [ − ], then from the
superposition property for a linear system, the response of the linear system to the input [ ] in
Eq. (2.1) is simply the weighted linear combination of these basic responses:
Signals and Systems | 59

[ ]= [ ]h[n] (2.1)

If the linear system is time invariant, then the responses to time-shifted unit impulses are all
time-shifted versions of the same impulse responses:
ℎ [ ]=ℎ [ − ] (2.2)

Therefore, the impulse response ℎ[ ] = ℎ [ ] of an LTI system characterizes the system


completely. This is not the case for a linear time-varying system: one has to specify all the
impulse responses ℎ [ ] (an infinite number) to characterize the system. For the LTI system,
equation becomes

[ ]= [ ]h[n − k] (2.3)

This result is referred to as the convolution sum or superposition sum and the operation on the
right-hand side of the equation is known as the convolution of the sequences of x[n] and h[n].
The convolution operation is usually represented symbolically as
[ ] = [ ] ∗ ℎ[ ] (2.4)

2.2.2 Representation of Continuous-Time Signals in Terms of Impulses


A continuous-time signal can be viewed as a linear combination of continuous impulses:
( )=∫ ( ) ( − ) (2.5)
The result is obtained by chopping up the signal ( ) in sections of width , and taking sum

Fig 2.9 Representation of Continuous-Time Signals in Terms of Impulses


60 | Behavior of Continuous & Discrete-time LTI Systems

Recall the definition of the unit pulse δ ; we can define a signal ( ) as a linear combination of
delayed pulses of height (k )

( )= ( (k ( − k ))Δ (2.6)

Taking the limit as = 0, we obtain the integral of Eq. (2.6), in which when Δ =0
(1) The summation approaches to an integral
(2) k = and x(k )= x( )
(3) Δ=d
(4) δ(t-k )= δ(t- )
Eq. (2.7) can also be obtained by using the sampling property of the impulse function. If we
consider t is fixed and t is time variable, then we have

x( )δ(t - )= x(( )δ (-(t - )) = x(τ)δ(t - ). (2.7)

∫ ( ) ( − ) =∫ ( ) ( − ) = ( )∫ ( − ) = ( ) (2.8)

As in discrete time, this is the sifting property of continuous-time impulse.

2.3 The Unit Step Response of an LTI System


The step response of an LTI system is simply the response of the system to a unit step. It conveys
a lot of information about the system. For a discrete-time system with impulse response ℎ[ ],
the step response is [ ] = [ ] ∗ ℎ[ ]. However, based on the commutative property of
convolution, [ ] = ℎ[ ] ∗ [ ], and therefore, s[n] can be viewed as the response to input
ℎ[ ] of a discrete time LTI system with unit impulse response. We know that [ ] is the unit
impulse response of the accumulator. Therefore,

[ ]= h[k] (2.9)

From this equation, ℎ[ ] can be recovered from [ ] using the relation


ℎ[ ] = [ ] − [ − 1] (2.10)
Signals and Systems | 61

It can be seen that the step response of a discrete-time LTI system is the running sum of its
impulse response. Conversely, the impulse response of a discrete-time LTI system is the first
difference of its step response.
Similarly, in continuous time, the step response of an LTI system is the running integral of its
impulse response,
( )=∫ ℎ( ) (2.11)
and the unit impulse response is the first derivative of the unit step response,
( )
ℎ( ) = = ( ) (2.12)
Therefore, in both continuous and discrete time, the unit step response can also be used to
characterize an LTI system.

2.4 Convolution Integral:


Convolution of two continuous-time signals ( ) and ℎ( )denoted by,

( ) = ( ) ∗ ℎ( ) = ∫ ( )ℎ( − ) (2.13)
Equation (2.13) is commonly called the convolution integral. Thus, we have the fundamental
result that the output of any continuous-time LTI system is the convolution of the input ( )
with the impulse response ℎ( ) of the system. Fig. 2.10 illustrates the definition of the impulse
response ℎ( ) and the relationship of Eq. (2.13).

Fig 2.10 Convolution operation

2.4.1 Properties of the Convolution Integral:


The convolution integral has the following properties.

1. Commutative:
( ) ∗ ℎ( ) = ℎ( ) ∗ ( )

2. Associative:
{ ( ) ∗ ℎ ( )} ∗ ℎ ( ) = ( ){ℎ ( ) ∗ ℎ ( )}
62 | Behavior of Continuous & Discrete-time LTI Systems

3. Distributive:
( ) ∗ {ℎ ( ) + ℎ ( )} = ( ) ∗ ℎ ( ) + ( ) ∗ ℎ ( )

2.5 Input-output behaviour with aperiodic convergent inputs


2.5.1 Response of a continuous time system
If an input ( ) is applied and then output of the system is ( )

Fig 2.11 ( ) input to system

So, instead of ( ) if we apply a standard elementary signal that is known as impulse signal. If
we apply impulse signal then the system's response will be impulse response if delta is replaced
by Tau Delta.

Fig 2.12 ( ) input to system


If input is ( − ), so when input is delayed output will be delayed, this is time invariant
system.

Fig 2.13 Delayed input to system

( )=∫ ( )ℎ( − ) (2.14)

So, this equation this convolution in integral equation is very useful for calculating the
responses of the systems

2.5.2 Response of a discrete time system


We have input [ ] and discrete time system and this discrete time system is giving output as
[ ].
Signals and Systems | 63

Fig 2.14 [ ] input to system

If this input is an impulse signal [ ] then the system will provide the output corresponding to
impulse signal the output corresponding to impulse signal is called impulse response ℎ[ ].

Fig 2.15 [ ] input to system

If the input impulse signal is delayed by , i.e., [ − ] and it is applied to the same discrete
time system DTS then the output will also be delayed by , i.e., ℎ[ − ].

Fig 2.16 δ [n-k] input to system

If another input [ ] is multiplied with [ − ] and provided to the system, because it is a


linear time invariant system so multiplying is a scaling of input which will also result in scaling
of the output and hence the output will become [ ]ℎ[ − ].

Fig 2.17 x[k] δ [n-k] input to system

Dividing summation of the input from −∞ to +∞ and then passing through the system, the
summation will also result in the right side in terms of which is called convolution sum.
Convolution sum is important for calculating output of the discrete time systems.

[ ]= [ ]ℎ[ − ] (2.15)

[ ]= [ ] ∗ ℎ[ ] (2.16)
64 | Behavior of Continuous & Discrete-time LTI Systems

2.6 Cascade interconnections


Suppose that we have a LTI system with input , output , and impulse response ℎ. We know
that and are related as = ∗ ℎ. In other words, the system can be viewed as performing
a convolution operation. From the properties of convolution introduced earlier, we can derive
several equivalences involving the impulse responses of series and parallel-interconnected
systems.

Figure 2.18: Equivalences for the series interconnection of continuous-time LTI systems.
The (a) first equivalence (b) second equivalence.

Consider two LTI systems with impulse responses ℎ and ℎ that are connected in a series
configuration, as shown on the left-side of Fig 2.18(a). From the block diagram on the left side
of Fig. 2.18(a), we have
= ( ∗ ℎ ) ∗ℎ (2.17)
Due to the associativity of convolution, however, this is equivalent to
= ∗ (ℎ ∗ ℎ ) (2.18)
Thus, the series interconnection of two LTI systems behaves as a single LTI system with
impulse response ℎ ∗ ℎ . In other words, we have the equivalence shown in Fig. 2.18(a).

Consider two LTI systems with impulse responses ℎ and ℎ that are connected in a series
configuration, as shown on the left-side of Figure 2.18(b). From the block diagram on the left
side of Figure 2.18(b), we have

= ( ∗ ℎ ) ∗ ℎ (2.19)

Due to the associativity and commutativity of convolution, this is equivalent to

= ∗ (ℎ ∗ ℎ ) = ∗ (ℎ ∗ ℎ )
Signals and Systems | 65

= ( ∗ ℎ ) ∗ℎ (2.20)

Thus, interchanging the two LTI systems does not change the behaviour of the overall system
with input and output . In other words, we have the equivalence shown in Figure 2.18(b).

Consider two LTI systems with impulse responses h1 and h2 that are connected in a parallel
configuration, as shown on the left-side of Figure 2.19. From the block diagram on the left side
of Figure 2.19, we have

= ∗ ℎ + ∗ ℎ (2.21)

Due to convolution being distributive, however, this equation can be rewritten as

= ∗ (ℎ + ℎ ) (2.22)

Thus, the parallel interconnection of two LTI systems behaves as a single LTI system with
impulse response ℎ + ℎ . In other words, we have the equivalence shown in Figure 2.19.

Figure 2.19: Equivalence for the parallel interconnection of continuous-time LTI


systems

Figure 2.20: System interconnection example


66 | Behavior of Continuous & Discrete-time LTI Systems

Similarly, from the right half of the block diagram from Figure 2.20, we can write
( ) = ∗ ℎ (2.23)
Substituting the expression for into the Eq. (2.23) we obtain,
= ( ∗ [ + ℎ + ℎ ]) ∗ ℎ
= ∗ [ℎ + ℎ ∗ ℎ + ℎ ∗ ℎ ] (2.24)
Thus, from Eq. (2.24) the impulse response ℎ of the overall system is
ℎ( ) = ℎ + ℎ ∗ ℎ + ℎ ∗ ℎ (2.25)

2.7 Causality for LTI systems


A system is causal if its output depends only on the past and present values of the input signal.
Specifically, for a discrete-time LTI system, this requirement is [ ] should not depend on
[ ]for > . Based on the convolution sum equation, all the coefficients ℎ[ − ] that
multiply values of [ ] for > must be zero, which means that the impulse response of a
causal discrete-time LTI system should satisfy the condition
ℎ[ ] = 0, < 0 (2.26)
A causal system is causal if its impulse response is zero for negative time; this makes sense as
the system should not have a response before impulse is applied.
A similar conclusion can be arrived for continuous-time LTI systems, namely
ℎ( ) = 0, < 0 (2.27)

Examples: The accumulator ℎ[ ] = [ ], and its inverse ℎ[ ] = [ ] − [ − 1] are causal.


The pure time shift with impulse response ( ) = ( – ) > 0 is causal, but is not
causal for < 0.

2.8 Stability for LTI Systems


Recall that a system is stable if every bounded input produces a bounded output. For LTI
system, the input [ ] is bounded in magnitude [ ] ≤ , for all . If this input signal is
applied to an LTI system with unit impulse response ℎ[ ], the magnitude of the output,

(2.28)
[ ] is bounded in magnitude, and hence is stable if
ℎ[ ]˂ ∞ (2.29)
Signals and Systems | 67

So discrete-time LTI system is stable is Eq. (2.29) is satisfied. The similar analysis applies to
continuous-time LTI systems, for which the stability is equivalent to

∫ ℎ( ) ˂∞ (2.30)

Example: consider a system that is pure time shift in either continuous time or discrete time.
In discrete time,

While, in continuous time,

Hence, we can conclude that both systems are stable.

Example: The accumulator h[n] = u[n] is unstable because

2.9 System Representation through Differential Equations and Difference Equations.


Linear, time-invariant (LTI) systems are represented by linear constant-coefficient differential
equations in continuous-time (CT) and linear constant-coefficient difference equations in
discrete-time (DT).
CT LTI system through differential equations can be of form electrical circuits, mechanical
systems etc. DT LTI system through difference equations can be described by a wide variety of
data filtering, time series analysis, and digital filtering systems.

2.9.1 Differential Equation Description of CT LTI systems


Differential equations are used to represent CT systems, where the system variables are
functions of continuous time, typically denoted as . These equations describe the relationship
68 | Behavior of Continuous & Discrete-time LTI Systems

between the system's input, output, and their derivatives with respect to time. The general
structure of such a representation is given by Eq. (2.31),

( ) ( )
∑ =∑ (2.31)

Where, ( ) is the input to the system and ( ) is the response.

Any general output ( ) can also be represented in terms of two signals and that are particular
solution and homogeneous solution.

The homogeneous solution can be given as,

( )
∑ =0 (2.32)

To solve the above equation, we required the M auxiliary (initial) conditions.

The system is said to be linear when described by the above differential equation if all the initial
conditions are equal to 0. Similarly, the system is said to be time invariant if it is at initial rest,
i.e., if ( ) = 0 for ≤ then assume, ( ) = 0 for ≤ . Hence, the initial condition
becomes,

( ) ( )
( )= =…..= =0 (2.33)

That means the value of the output at and its derivatives up to degree ( − 1) is 0. So, if the
system satisfies both the conditions of linearity and time invariance then it is linear time-
invariant (LTI) system.

2.9.2 Difference Equation Description of DT LTI systems


Similar to CT LTI systems we can be represented by differential equations, a DT LTI system is
represented by Difference equations. A general structure of difference equation is,
Signals and Systems | 69

∑ ( − ) =∑ ( − ) (2.34)

Where, ( ) is an input and ( ) is an output.

2.10 State space representation of systems


Earlier we have seen LTI systems based on input-output relationships known as external
description of a system. Now we will examine the state space representation of systems known
as internal description of systems.
State space representation of systems consists of two parts, state equations and output equations.
State equations represent set of equations relating state variables to inputs. Whereas, Output
equations represents set of equations relating outputs to state variables and inputs. Advantage
of state space representation is it provides new insight into the system behaviour with the use
of matrix linear algebra. It can also handle multiple-input multiple-output (MIMO) systems.
The state of a CT system at time is defined as the minimal information that is sufficient to
determine the state and output of a system for all times.
State variables are the variables that contain all state information related to the memory. For
LTI system with output signal ( ) with input signal ( ) and impulse response ℎ( ).
Therefore,
( ) = ℎ( ) ∗ ( ) −> ( ) = ( ) ( ) (2.35)
This representation of the system expresses the input output relation. It does not provide us the
internal specification of the system. State space representation not only provides information
on I/O but also gives good view on the internal specification of the system. The states of the
system at time to include min required information to express the system situation at time .
These are the first-degree equations. CT LTI state space representation of LTI system is given
by,

= ̇( ) = ( ) + ( ) (2.36)

( ) = ( ) + ( ) (2.37)
70 | Behavior of Continuous & Discrete-time LTI Systems

Eq. (2.36) and Eq. (2.37) represents the state equations and output equations respectively.

Where, = ̇ ( ) is the vector whose entries are the derivative .

2.11 State Space Analysis


The state of a CT system is the condition of a system at any time instant. The state variables are
the variables that completely describe the state of a system at any given time. These variables
at t = 0 with inputs for t > 0 completely describes the behavior of system for ≥ 0.

2.11.1 State equations


Let us consider a CT system has P inputs, Q state variables and Y outputs.
Hence, let ( ), ( ), … … ( ) be the Q state variables,
( ), ( ), … … ( ) be the P input.
( ), ( ), … … ( ) be the Y outputs

Now, the CT system can be represented by Fig. 2.21 as shown below,

Fig. 2.21 State Space representation of CT system


Take first derivative of state variables as a function of state variables and input to form the state
equations.
( )= ( ), ( ), … … ( ) ( ), ( ), … … ( )



( )= ( ), ( ), … … ( ) ( ), ( ), … … ( )

Above state equation can be represented in q matrix form given by Eq. (2.38),
Signals and Systems | 71

( ) ⋯ ( ) ⋯ ( )
( ) = ⋯ ( ) + ⋯ ( ) (2.38)
( ) ⋯ ( ) ⋯ ( )
⋮ ⋮ ⋮ ⋮ ⋮
⋮ ⋮ ⋮ ⋮ ⋮
( ) ( ) ( )

Where
= System matrix
= Input matrix
( ) = State vector
( ) = Input vector

Hence, State equation can be written as,


( )= ( )+ ( ) (2.39)

2.11.2 Output equations


The output equation in terms of y can be written as,
( )= ( ), ( ), … … ( ) ( ), ( ), … … ( )



( )= ( ), ( ), … … ( ) ( ), ( ), … … ( )

The output equations can be written in terms of matrix form as in Eq. (2.40)
( ) ⋯ ( ) ⋯ ( )
( ) = ⋯ ( ) + ⋯ ( ) (2.40)
( ) ⋯ ( ) ⋯ ( )
⋮ ⋮ ⋮ ⋮ ⋮
⋮ ⋮ ⋮ ⋮ ⋮
( ) ( ) ( )

Where
C = Output matrix
D = Transition matrix
72 | Behavior of Continuous & Discrete-time LTI Systems

Q(t) = State vector


P(t) = Input vector
Hence,
( )= ( )+ ( ) (2.41)

2.11.3 State Model


The state model of a CT system can be represented in terms of state equations and output
equations.
( )= ( )+ ( ) (2.42)
( )= ( )+ ( ) (2.43)

2.12 Transfer function of a Continuous Time System


Consider the equation
( )= ( )+ ( )
Take Laplace transform on both sides assuming zero initial condition.
( )= ( )+ ( )
( )− ( )= ( )
( )( − ) = . ( ) … . . is unit matrix
( ) = . ( )( − ) (2.44)
Now consider output equation
Y(t) = CQ(t) +D P (t)
Take Laplace transform on both Sides
Y(s) = CQ (s) + D P (s) (2.45)
Put Eq. (2.7) in Eq. (2.8)
Y (s) = C [B.P(s) (sI-A) -1 ] + D P (s)
Y (s) +[C.B (Is-A)-1 PC(s)
( )
= C.B. (Is-A)-1 + D (2.46)
( )

Hence, Eq (2.46) is called transfer function of CT system.

2.13 State Transition Matrix


The State Equation can be written as
(̇ )– ( ) = ( )
Signals and Systems | 73

Multiplying above equation by e –At we get,


(̇ ) − ( ) = ( )
(̇ ) − ( )= ( ) (2.47)

Using derivative formula of, ( )= + in the above equation we get,


( ) = (̇ ) + (− ) ( ) (2.48)
Using Eq. (2.48) in Eq. (2.47) we get,

( ) = ( ) (2.49)

On integrating Eq. (2.49) we get,


( )=∫ . ( ) + (0) (2.50)
We have, (0) is the intitial condition vector, is the dummy variable used instead of t.
On multiplying Eq. (2.50) by we get,

( )= . ( ) + (0)

( )= ( )
∫ . ( ) + (0) (2.51)

Eq. (2.51) is the time domain solution of state equations of the CT system.
Matrix is called state transition matrix of CT System.

Example 2.1: Compute the continuous convolution of the following functions:

x (t) = e^(-t) * u(t)

h (t) = 2 * e^(2t) * u(t)

Solution:

To compute the continuous convolution, we can use the integral formula:

y(t) = ∫(x(τ) * h(t-τ)) dτ


74 | Behavior of Continuous & Discrete-time LTI Systems

Let's calculate it step by step:

For t > 0:

y(t) = ∫(e^(-τ) * 2 * e^(2(t-τ))) dτ

= 2 * ∫(e^(-τ) * e^(2(t-τ))) dτ

= 2 * ∫(e^(2t - 3τ)) dτ

= 2 * e^(2t) * ∫(e^(-3τ)) dτ

= 2 * e^(2t) * (-1/3) * e^(-3τ) | from 0 to t

= -2/3 * e^(2t) * (e^(-3t) - 1)

For t <= 0:

y(t) = 0

Therefore, the continuous convolution of x(t) and h(t) is:

y(t) = -2/3 * e^(2t) * (e^(-3t) - 1) for t > 0

0 for t <= 0

Example 2.2: Compute the discrete convolution of the following sequences:

x[n] = [1, 2, 3]

h[n] = [2, -1, 3, 0]

Solution:

To compute the discrete convolution, we can use the formula:

y[n] = ∑(x[k] * h[n-k])

Let's calculate it step by step:

For n = 0:

y[0] = (1 * 2) = 2
Signals and Systems | 75

For n = 1:

y[1] = (1 * -1) + (2 * 2) = 3

For n = 2:

y[2] = (1 * 3) + (2 * -1) + (3 * 2) = 5

For n = 3:

y[3] = (2 * 3) + (3 * -1) = 3

For n = 4:

y[4] = (3 * 3) = 9

Therefore, the discrete convolution of x[n] and h[n] is:

y[n] = [2, 3, 5, 3, 9]

2.14 Multi-Input, Multi-Output Representation

MIMO refers to systems that have multiple inputs and multiple outputs. These systems are
characterized by their ability to handle multiple control inputs and generate multiple outputs
simultaneously. The state-space representation for a MIMO system is an extension of the single-
input, single-output (SISO) case. In the MIMO representation, the state equations and output
equations are modified to accommodate multiple inputs and outputs.

Let's consider an example of a multi-input, multi-output (MIMO) system with a numerical


representation. Suppose we have a MIMO system with two inputs (u1, u2) and two outputs (y1,
y2). The state-space representation of the system is given by the following equations:

State equations:

−2 0 1 0 1
= . + . (2.52)
1 −3 0 2 2
76 | Behavior of Continuous & Discrete-time LTI Systems

Output equations:

1 1 2 1 0 1
= . + . (2.53)
2 −1 0 0 −1 2

In this representation:

1
 x= is a 2-dimensional state vector representing the internal state of the system.
2
1
 u= is a 2-dimensional input vector representing the inputs to the system.
2
1
 y= is a 2-dimensional output vector representing the outputs from the system.
2
−2 0
 The state matrix A is .
1 −3
1 0
 The input matrix B is
0 2
1 2
 The output matrix C is
−1 0
1 0
 The feed forward matrix D is .
0 −1
To analyze this system, we can perform various analyses such as stability analysis,
controllability analysis, observability analysis and system response analysis.

For stability analysis, we can determine the eigenvalues of the state matrix A. In this case, the
eigenvalues are -2 and -3, indicating that the system is stable. For controllability analysis, we
can check if the controllability matrix Co has full rank. The controllability matrix Co is formed
by concatenating the input matrix B with powers of the state matrix A. For observability
analysis, we can examine if the observability matrix Ob has full rank. The observability matrix
Ob is formed by concatenating the output matrix C with powers of the state matrix A. To
analyze the system's response, we can simulate the state equations and output equations with
appropriate inputs. For example, if we apply a step input ( ) = 1 and ( ) = 0, we can
Signals and Systems | 77

numerically solve the state equations and output equations to obtain the time-domain response
of the system.

By analyzing the state variables ( ), ( ), and the outputs ( ), ( ), we can observe the
behavior of the system, including transient response, steady-state behavior, and the interaction
between inputs and outputs.

Example 2.4: Find Laplace domain and time domain state transition matrix if

0 1
=
−2 −3

Solution:

The inverse of a 2×2 matrix is

To find φ(t) we must take the inverse Laplace Transform of every term in the matrix
78 | Behavior of Continuous & Discrete-time LTI Systems

Using partial fraction expansion of each term we get

2.15 Periodic Inputs to LTI system

When a periodic input signal is applied to a linear time-invariant (LTI) system, the response of
the system also becomes periodic. In this scenario, the output of the LTI system exhibits the
same periodicity as the input signal, but with potentially different amplitudes and phases.

Figure 2.22: Periodic input to LTI system

When a periodic signal with period is passed through a linear, time-invariant system, the
output is also periodic and can be expressed as a sum of complex exponential signals.

( / ∗ )
Let's consider the input signal [ ] =

Where is the discrete time index and k range from 0 to -1. This signal represents a complex
exponential with a frequency of 2 / .

According to the properties of linear, time-invariant systems, the output signal can be obtained
by multiplying the frequency response of the system, denoted by ℎ[2 / ], with the input
signal's complex exponential component [ ].
Signals and Systems | 79

Hence, the output for each component signal can be written as:

[ ]= [ ]∗ ℎ[ ( / ∗ )
] ∗ (2.54)

Now, let's simplify the expression further.

Since ℎ[2 / ] is a constant value for each component signal, we can combine it with the
exponential term:

( ( / ∗ ))
[ ] = ℎ[ ] ∗ (2.55)

Here, θ represents the phase shift introduced by the frequency response of the system.

When a periodic input signal is passed through a linear, time-invariant system, the output signal
remains periodic and can be represented as a sum of complex exponential signals. Each
component signal's frequency response, ℎ[2 / ], is multiplied by the input signal's complex
exponential component, resulting in a modified exponential term in the output.

The linearity property of the filter allows us to express the output signal as

( ( / ∗ ))
[ ] = ℎ[ ] ∗ (2.56)

Where, ℎ represents the frequency response of the system.

2.16 The Notion of frequency response and its relation to impulse response

The notion of frequency response is an important concept in the field of signal processing. It
provides information about how a system or a filter responds to different frequencies present in
the input signal. It is obtained by taking the Fourier transform of the impulse response. The
80 | Behavior of Continuous & Discrete-time LTI Systems

frequency response reveals how the system responds to different frequencies, while the impulse
response shows its response to an impulse signal.

Let's start with the convolution sum equation:

( ) = ∑[ ( − )ℎ( )] (2.57)

Where x (n) is the input signal, h (n) is the impulse response of the system, and y (n) is the
output signal. The convolution sum represents the mathematical operation of convolving the
input signal with the impulse response to obtain the output signal.

Now, let's consider a specific input signal in the form of a complex exponential function:

( ) =

Where represents the imaginary unit, is the angular frequency, and belongs to the set of
integers.

By substituting this input signal into the convolution sum equation, we get:

[ ]
[ ] = ∑ ℎ[ (2.58)

Next, we can simplify this expression by factoring out the term :

[ ]
[ ] = ∑ ℎ[ ] (2.59)

Notice that the term inside the summation, , is the complex conjugate of . Therefore,
we can rewrite it as:

[ ] = ∑ ℎ[ ]
Signals and Systems | 81

Now, if we compare this expression to the form of the output signal when the input is a complex
exponential with frequency ω, given as ( ) = ( ) , we can conclude that:

( ) = ∑ [ℎ [ ] ] (2.60)

The term ( ) in this equation is referred to as the frequency response. It represents the
relationship between the input signal's frequency ω and the output signal's amplitude and phase
shift. The frequency response provides information about how the system or filter amplifies or
attenuates specific frequencies.

Therefore, the frequency response H (ω) is obtained by taking the discrete-time Fourier
transform (DTFT) of the impulse response h (n). The frequency response describes how the
system or filter responds to different frequencies present in the input signal, and it is a
fundamental concept in signal processing.

UNIT SUMMARY

In this unit on the behavior of continuous and discrete-time LTI systems, we covered various
important topics. These include understanding impulse and step responses, analysing input-
output behavior with aperiodic convergent inputs, cascade interconnections, causality and
stability characterization, system representation through differential equations and difference
equations, state-space representation, state-space analysis, multi-input multi-output systems,
the role of the state transition matrix, periodic inputs and the notion of frequency response.
Overall, this unit provided a comprehensive overview of the behavior and analysis of LTI
systems in both continuous and discrete-time domains.
82 | Behavior of Continuous & Discrete-time LTI Systems

EXERCISES

Multiple Choice Questions and Answers


1. The impulse response of an LTI system provides information about:
a) The system's stability
b) The system's causality
c) The system's input-output behavior
d) The system's frequency response
Answer: c) The system's input-output behavior

2. Convolution is a mathematical operation used to:


a) Compute the impulse response of an LTI system
b) Determine the step response of an LTI system
c) Analyze the behavior of LTI systems
d) Calculate the transfer function of an LTI system
Answer: c) Analyze the behavior of LTI systems

3. Causality in LTI systems implies that the output of the system depends on:
a) Future inputs
b) Past inputs
c) Present inputs only
d) Both past and future inputs
Answer: b) Past inputs

4. Stability of an LTI system ensures that:


a) The output is bounded for any bounded input
b) The output is zero for any input
c) The system has a linear transfer function
d) The system has a constant impulse response
Answer: a) The output is bounded for any bounded input
Signals and Systems | 83

5. System representation through differential equations is commonly used for:


a) Analyzing the frequency response of LTI systems
b) Describing the input-output behavior of LTI systems
c) Evaluating the stability of LTI systems
d) Modeling continuous-time LTI systems
Answer: d) Modeling continuous-time LTI systems

6. State-space representation of systems involves describing the system behavior using:


a) Impulse response
b) Transfer function
c) Differential equations
d) Convolution
Answer: c) Differential equations

7. The State Transition Matrix in state-space analysis represents the:


a) Impulse response of the system
b) Transfer function of the system
c) Evolution of state variables over time
d) Frequency response of the system
Answer: c) Evolution of state variables over time

8. Multi-input, multi-output (MIMO) representation of LTI systems deals with:


a) Systems with multiple state variables
b) Systems with multiple inputs and multiple outputs
c) Systems with nonlinear behavior
d) Systems with periodic inputs
Answer: b) Systems with multiple inputs and multiple outputs

9. The frequency response of an LTI system is obtained through:


a) Fourier transform
b) Laplace transform
84 | Behavior of Continuous & Discrete-time LTI Systems

c) Z-transform
d) Convolution
Answer: b) Laplace transform

10. The relationship between the frequency response and impulse response of an LTI system
given by:
a) Convolution theorem
b) Parseval's theorem
c) Nyquist criterion
d) Plancherel's theorem
Answer: a) Convolution theorem

Short and Long Answer Type Questions


1. What is the impulse response of an LTI system?

2. How is the step response of an LTI system defined?

3. What role does convolution play in analyzing LTI systems?

4. How would you characterize causality in LTI systems?

5. What does stability refer to in the context of LTI systems?

6. How are LTI systems represented using differential equations?

7. What are difference equations, and how are they used to represent LTI systems?

8. What is the state-space representation of a system?

9. What is the significance of the State Transition Matrix in state-space analysis?

10. How do periodic inputs affect LTI systems, and what is the relationship between the
frequency response and the impulse response?

Numerical Problems
1. Given the impulse response h(t) = 3e^(-2t), calculate the step response of the system.

2. Find the convolution of the two sequences x[n] = {1, 2, 3} and h[n] = {2, -1, 0}.
Signals and Systems | 85

3. Consider an LTI system with the input x(t) = 2cos(3t). If the impulse response is h(t) = e^(-
t), determine the output y(t).

4. Analyze the stability of the difference equation y[n] = 0.5y[n-1] + x[n].

5. Represent an LTI system through a differential equation: y''(t) + 3y'(t) + 2y(t) = x(t), where
y(0) = 0 and y'(0) = 1.

6. Calculate the state-space representation of a multi-input, multi-output LTI system with two
state variables and two inputs.

7. Find the State Transition Matrix for a second-order LTI system described by the state-space
equations: ẋ(t) = Ax(t) + Bu(t) and y(t) = Cx(t) + Du(t).

8. Determine the response of a continuous-time LTI system with the impulse response h(t) =
sin(2πt) to a periodic input signal with a frequency of 10 Hz and an amplitude of 3.

9. Analyze the frequency response of an LTI system with the impulse response h(t) = e^(-
t)cos(2πt).

10. Investigate the relationship between the impulse response and the frequency response of an
LTI system using a specific example.

KNOW MORE

The study of continuous and discrete-time LTI systems involves analyzing their behavior using
concepts like impulse response, step response, convolution, and input-output behavior with
periodic inputs. Understanding causality and stability is crucial for characterizing these
systems. System representation can be done through differential equations or difference
equations, with state-space analysis providing a comprehensive approach. Multi-input, multi-
output systems and the role of the state transition matrix are also studied. Additionally, the
notion of frequency response and its relation to the impulse response is explored. These
concepts form the foundation for analyzing and designing LTI systems in various applications.
In summary, the study of continuous and discrete-time LTI systems includes analyzing impulse
86 | Behavior of Continuous & Discrete-time LTI Systems

and step responses, convolution, input-output behavior with periodic inputs, causality,
stability, system representation through differential or difference equations, state-space
analysis, multi-input multi-output systems, state transition matrix, and the relation between
frequency response and impulse response. These concepts are essential for understanding and
designing LTI systems in diverse applications.

REFERENCES AND SUGGESTED READINGS

1.Signals and Systems by Simon Haykin


2.Signals and Systems by Ganesh Rao
3.Signals and Systems - Course (nptel.ac.in)

Dynamic QR Code for Further Reading


Signals and Systems | 87

d
3 Fourier Series and
Transform

UNIT SPECIFICS
Through this unit we have discussed the following aspects:
 What is Fourier series, why it was developed?
 Fourier series representation of periodic signals, Waveform Symmetries, Calculation of
Fourier
Coefficients.
 Fourier Transform, convolution/ multiplication and their effect in the frequency
domain,
 Magnitude and phase response, Fourier domain duality.
 The Discrete-Time Fourier Transform (DTFT) and the Discrete Fourier Transform (DFT).

RATIONALE

The unit on “Fourier Series and Transform" provide students to understand the behavior of
Continuous and Discrete-time signals. This 6-hour unit is designed to provide students with a
comprehensive understanding of periodic and non-periodic signals along with their frequency
domain representation.

The unit focuses on Fourier series representation of periodic signals along with the properties.
The students can analyze the behavior of signal by extracting the frequency components from
the signal. The behavior will be found for both CT and DT signals.

The concepts covered, such as Fourier series representation of periodic signals, waveform
symmetries, calculation of Fourier coefficients, Fourier transform, convolution/ multiplication
and their effect in the frequency domain, magnitude and phase response, Fourier domain
88 | Fourier Series and Transform

duality, the Discrete-Time Fourier Transform (DTFT) and the Discrete Fourier Transform
(DFT).

PRE-REQUISITES
1. Strong understanding of mathematics, including algebra, calculus, and complex numbers.

2. Familiarity with basic concepts in signals and systems, such as time-domain and periodic,
non-periodic signals

3. Proficiency in solving ordinary differential equations and understanding linear algebra


concepts.

UNIT OUTCOMES
List of outcomes of this unit is as follows:
U3-O1: Be able to compute the frequency components of the signal.
U3-O2: Be able to predict how the signal will interact with linear systems and circuits using
frequency response methods.
U3-O3: Be able to extract the Fourier coefficients from the signal.
U3-O4: Be able to analyze the signal directly from its properties.
U3-O5: Be able to analyze the magnitude and phase response of the given signal.

EXPECTED MAPPING WITH COURSE OUTCOMES


Unit-3
(1- Weak Correlation; 2- Medium correlation; 3- Strong Correlation)
Outcomes
CO-1 CO-2 CO-3 CO-4 CO-5 CO-6
U3-O1 3 - 3 3 - -
U3-O2 3 - 3 - - -
U3-O3 3 - 3 3 - -
U3-O4 3 - 3 3 - -
U3-O5 3 - 3 2 - -
Signals and Systems | 89

3.1 Introduction
Earlier we have discussed the time domain representation of signals and systems. We have
also seen the types of signals that are periodic and non-periodic. The French mathematician
Jean Baptiste Joseph Fourier showed that any periodic non-sinusoidal signal could be
represented in terms of linear weighted sum of harmonic sinusoidal signals. This
representation is called as Fourier series representation.
On the other hand, the Fourier representation of the aperiodic or non-periodic signals is
performed by treating them as periodic signals with an infinite fundamental period. In this
case, the non-periodic signals are represented as a function of frequency called Fourier
Transform. Fourier domain representation of the signal is another name for Fourier
Transform.
The Fourier representations of the signals are used for the frequency domain analysis of
signals. It enables us to extract the amplitude and phase of various frequency components
present in the signal.
Figure 3.1 shows the types of signals and their corresponding Fourier representations.

Fig. 3.1 Types of signals and their Fourier representations

3.2 Trigonometric Form of Fourier Series


3.2.1 Definition
In this form of Fourier representation, any periodic signal ( ) with a fundamental period T
is represented in terms of trigonometric functions as,
( )= +∑ cos + ∑ sin (3.1)

Where = and, and are constants.


90 | Fourier Series and Transform

It may be noted that cos and sin are orthogonal functions over one period for
all integer values of and . This implies that ∫ cos m Ω t sin n Ω t dt = 0 for any
arbitrary value of .
Equation (3.1) can be extended as,
( )= + Cos + 2 +⋯+ Sin + 2 + ⋯ (3.2)

Where, =2 = = Fundamental frequency in rad/sec


F = Fundamental frequency in H
n Ω = Harmonic frequencies
a , a , b = Fourier coefficients of trigonometric form
Here, is a constant representing the average value or dc component of ( ).
The Fourier Coefficients in Eq. (3.1) can be expressed as,

= ∫ ( ) or

or = ∫ ( ) (3.3)

= ∫ ( ) Cos or = ∫ ( ) Cos (3.4)

= ∫ ( ) Sin or = ∫ ( ) Sin (3.5)

The integration is over one period i.e. to or 0 to .

3.2.2 Conditions for existence of Fourier series


The following conditions called as Dirichlet’s conditions must be satisfied by the signals for
the existence or convergence of Fourier series:
1. The signal x(t) should be absolutely integrable over one period.
i.e. ∫ |x(t)| dt < ∞

2. The signal x(t) must have finite number of maxima and minima in one period.
3. The signal x(t) must have finite number of discontinuities in one period.

If any signal x(t) satisfies the above Dirichlet’s conditions, the Fourier series represented by
equation (3.1) converges i.e., the sum becomes equal to x(t) except at the point of
discontinuities.
Signals and Systems | 91

3.2.3 Equations for , and .


1) Derivation equation for

a = ∫ ⁄
x(t)dt
or
a = ∫ x(t)dt
Proof: Consider Trigonometric FS in Eq. (3.1)
a
x(t) = + a cos nΩ t + b sin nΩ t
2
Integrate the above equation for 0 to

a
x(t)dt = dt + a cos nΩ t dt + b sin nΩ t dt
2

= ∫ dt + ∑ a ∫ cos nΩ t dt + ∑ b ∫ sin nΩ t dt

= [ t] + ∑ a +∑ b

= [ T − 0] + ∑ a − +∑ b +
2π 2π
T sin n T . T −cos n T . T 1
= a + a + b +
2 2π 2π 2π
n T n T n T

As, Ω = , cos0 = 1, sin0 = 0.

T sin n2π − cos n2π 1


∴ x(t)dt = a + a T + b T +
2 n2π n2π n2π

T 1 1
= a + a T. 0 + b T − +
2 n2π n2π
T
x(t)dt = a
2
92 | Fourier Series and Transform

2
∴a x(t)dt
T

2) Derivation equation for

a = ∫ x(t) cos n Ω tdt or a = ∫ x(t)cosnΩ t dt


Proof: Consider trigonometric FS in Eq. (3.1)
x(t) = +∑ a cos nΩ t + ∑ b sin nΩ t
Let us multiply above equation by cosmΩ t
∴ x(t)cosmΩ t
a
= cos m Ω t + cos m Ω t a cos n Ω t
2

+ cos m Ω t b sin n Ω t

Integrate the above equation over 0 to and expand,


a
∴ x(t) cos mΩ t dt = cos mΩ t + a cosΩ t. cosmΩ t dt + ⋯
2

+ a cos mΩ t dt + ⋯ + b sinΩ t. cos mΩ tdt + ⋯

+ b sin m Ω t. cosmΩ t dt + ⋯

In above equation all the definite integrations becomes zero except ∫ a cos mΩ t dt
∴ ∫ x(t) cos mΩ t dt = ∫ a cos mΩ t dt
=∫ a dt

= ∫ (1 + cos2m Ω t)dt

= t+

= T+ −0−
Signals and Systems | 93

= T+

= a … … as 2 2 = 0 for integer . Hence,

a = ∫ x(t) cos mΩ t dt (3.6)


As is the coefficient, the coefficient will be,

2
a = x(t) cos n Ω t dt
T

3) Derivation Equation for


/
b = ∫ /
x(t)sin nΩ t or b = ∫ x(t) sin nΩ t dt
Proof: Consider trigonometric FS in Eq. (3.1)
a
x(t) = + a cos n Ω t + b sin n Ω t
2
Let us multiply above equation by sin m Ω t
a
x(t) sin m Ω t = Sin mΩ t + sin m Ω t a cos n Ω t
2

+ sin m Ω t b sin n Ω t

Integrate above equation over 0 to and expand.

x(t) sin m Ω t dt

a
= sin m Ω t + a cos Ω t sin mΩ t + ⋯
2

+ a cos Ω t sin mΩ t dt + ⋯ + b sin Ω t sin m Ω t dt

+ ⋯..+ b sin m Ω t dt + ⋯
94 | Fourier Series and Transform

In above equation all the definite integrations becomes zero except ∫ b Sin mΩ t dt

x(t) sin m Ω t dt = b sin mΩ t dt

=b ∫ dt

= ∫ [1 − cos2mΩ t] dt

= t−

= T− −0+

= T−

= b …….. as sin 2m 2π = 0 for integer m.

b = ∫ x(t) sin m Ω t dt (3.7)


as m is the m coefficient, the n coefficient will be,

2
b = x(t)sin n Ω t dt
T

3.3 Exponential form of Fourier Series


3.3.1 Definition
A periodic signal can also be represented in terms of exponential form of Fourier Series. Let
us consider ( ) be a periodic signal with period . Then we can write
( )=∑ (3.8)
where, =2 = =Fundamental frequency in rad/sec
F =Fundamental frequency in H
±n Ω =Harmonic frequencies
=Fourier coefficients of exponential form
The value of can be evaluated as

= ∫ ( ) or = ∫ ( ) (3.9)
Signals and Systems | 95

The integration is over one period i.e. to or 0 to .


This form consists of both positive and negative frequencies of exponential harmonic
components. When these exponential components are added, we get real sine and cosine
signals.

3.3.2 Derivation for equation for


As we know the Fourier coefficient of exponential form is written as,

= ∫ ( ) or = ∫ ( )

Proof: Consider the exponential form of Fourier series of ( ) given by Eq. (3.8),
( )=∑ = ⋯+ +…+ + + +⋯+
+⋯
Multiply the above equation by
( ) = ⋯+ +… + +
+ +⋯+ +⋯
Now integrate the above equation over a period 0 .
∫ ( ) = ⋯+ ∫ +… +∫ +
∫ +∫ +⋯+∫ +⋯
In the above equation all the definite integrations beocmes zero except ∫ . Hence,
∫ ( ) =∫ = [ ] = [ − 0] =
1
= ( )

Above equation is for coefficient. Hnece we can write for coefficient ,


1
= ( )

3.3.3 Relationship between the triginimetric and exponential form of Fourier coefficient
The relationship between the triginimetric and exponential form of Fourier coefficient can be
written as,
= (3.10)

±= ( ∓
) for = ±1, ±2, ±3, … (3.11)
96 | Fourier Series and Transform

| |= + for all values of , except when = 0 (3.12)

3.4 Waveform Symmetires


3.4.1 Even Symmetry
If ( ) is an even signal, then it should satisfy the condition (− ) = ( ). The waveform
of an even symmetry signal is symmetric about the vertical axis or it is symmetric at = 0.
A waveform can be represented in terms of even symmetry by folding the waveform with
respect to vertical axis. After folding, if the shape remains same then waveform adopts even
symmetry.
The value of Fourier coefficient is zero if the average value of one period is equal to zero.
For an even signal , exists but does not exist. Hnece the Fourier coefficients are given
by,

= ∫ ( ) (3.13)

= ∫ ( ) cos (3.14)

=0 (3.15)
Proof 1: Consider Eq. (3.3),

2 2 2
= ( ) = ( ) = ( )

Let = − ; ∴ =−

2 2
= (− )(− )+ ( )

2 2
= (− ) + ( )

= ∫ (− ) + ∫ ( )
Signals and Systems | 97

Since, ( ) is even, (− ) = ( ),

= ∫ ( ) + ∫ ( )

4
= ( )

Proof 2: Consider Eq. (3.4),

2 2 2
= ( ) cos = ( ) cos + ( ) cos

Let = − ; ∴ =−

2 2
= (− ) cos (− ) (− )+ ( ) cos

2 2
= (− ) cos + ( ) cos

2 2
= (− ) cos + ( ) cos

Since, ( ) is even, (− ) = ( ),

2 2
= ( ) cos + ( ) cos

4
= ( ) cos

Proof 3: Consider Eq. (3.5),


98 | Fourier Series and Transform

2 2 2
= ( ) sin = ( ) sin + ( ) sin

Let = − ; ∴ =−

2 2
= (− ) sin (− ) (− )+ ( ) sin

Since, sin(− ) = − sin( ),

2 2
=− (− ) sin + ( ) sin

2 2
=− (− ) sin + ( ) sin

Since, ( ) is even (− ) = ( ),

2 2
=− ( ) sin + ( ) sin

=0

3.4.1.1 Some even symmetry signals with their Fourier series expansion

Fig. 3.2 Even, Half wave, Quarter wave symmetry waveform


Signals and Systems | 99

Figure 3.2 shows the waveform with even symmetry as well as half wave and quarter wave
symmetry. Hence for this waveform,
=0
=0


( )= − + − +
⋯ (3.16)

Fig. 3.3 Even symmetry waveform


Figure 3.3 shows the waveform with even symmetry. Hence for this waveform,

=0
( )= + − + − +⋯ (3.17)
3.4.2 Odd Symmetry
If ( ) is an odd signal, then it should satisfy the condition (− ) = − ( ). The waveform
of an odd symmetry signal is anti-symmetric about the vertical axis or it is anti-symmetric at
= 0.
a waveform can be represented in terms of odd symmetry by folding the waveform with
respect to vertical axis. After folding, if the shape is exaclty opposite of the the another side
of vertical axis then waveform adopts odd symmetry.
The value of Fourier coefficient is zero, is zero but exists. Hnece the Fourier
coefficients are given by,
=0 (3.18)
=0 (3.19)
100 | Fourier Series and Transform

= ∫ ( ) sin (3.20)
Proof 1: Consider Eq. (3.3),

2 2 2
= ( ) = ( ) = ( )

Let = − ; ∴ =−

2 2
= (− )(− )+ ( )

2 2
= (− ) + ( )

= ∫ (− ) + ∫ ( )
Since, ( ) is odd, (− ) = − ( ),

=− ∫ ( ) + ∫ ( )
= 0
Proof 2: Consider Eq. (3.4),
2 2 2
= ( ) cos = ( ) cos + ( ) cos

Let = − ; ∴ =−

2 2
= (− ) cos (− ) (− )+ ( ) cos

2 2
= (− ) cos + ( ) cos
Signals and Systems | 101

2 2
= (− ) cos + ( ) cos

Since, ( ) is odd, (− ) = − ( ),

2 2
=− ( ) cos + ( ) cos

=0
Proof 3: Consider Eq. (3.5),
2 2 2
= ( ) sin = ( ) sin + ( ) sin

Let = − ; ∴ =−

2 2
= (− ) sin (− ) (− )+ ( ) sin

Since, sin(− ) = − sin( ),

2 2
=− (− ) sin + ( ) sin

2 2
=− (− ) sin + ( ) sin

Since, ( ) is odd (− ) = − ( ),

2 2
= ( ) sin + ( ) sin

4
= ( ) sin
102 | Fourier Series and Transform

3.4.2.1 Some odd symmetry signals with their Fourier series expansion

Fig. 3.4 Odd, Half wave, Quarter wave symmetry waveform

Figure 3.4 shows the waveform with odd symmetry as well as half wave and quarter wave
symmetry. Hence for this waveform,
=0
=0

exists for odd values of and equation of Fourier series consists of


odd harmonics of all cosine terms

( )= + + + +⋯ (3.21)

Fig. 3.5 Odd symmetry waveform

Figure 3.5 shows the waveform with odd symmetry. Hence for this waveform,
=0
=0
Signals and Systems | 103

exists for all values of and Fourier series consists both even and
odd harmonics of sine terms

( )= − + − +⋯ (3.22)

3.4.3 Half Wave Symmetry


A periodic waveform satisfying the condition of having two equal and opposite half
period/cycle in one period/cycle are said to have half wave symmetry.
The conditin of half wave symmetry should satisfy the below equation,
± =− ( ) (3.23)
If any waveform satisfies the half wave symmetry condition then, the Fourier series will
consist of odd harmonic terms alone.
The waveform shown in Fig. 3.6 shows the example of half wave symmetry.

Fig. 3.6 Half wave symmetry waveform


( )=− + + +⋯ + + + +
⋯ (3.24)
3.4.4 Quarter Wave Symmetry
A waveform that satisfies half wave symmetry condition along with even or odd symmetry
then that waveform is said to have quarter wave symmetry. In this type of symmetry, each
quarter period will be of same identical shape and may or may not have opposite sign. The
conditin of quarter wave symmetry should satisfy the below equation,
± =− ( ) (3.23)
Where ( ) has half wave symmetry and its Fourier series coefficients consists of all odd
hrmonic terms.
104 | Fourier Series and Transform

If ( ) has even and half wave symmetry, then (− ) = ( ) and Fourier series will have odd
harmonics of only cosine terms.
If ( ) has odd and half wave symmetry, then (− ) = − ( ) and Fourier series will have
odd harmonics of only sine terms.
The waveforms shown in Fig. 3.2, 3.4 are the examples of quarter wave symmetry.

3.5 Properties of Fourier Series in terms of Exponential Form


Suppose ( ) and ( ) have the Fourier series coefficients and in exponential form,
respectively.
3.5.1 Linearity:
For continuous time periodic signal,
( )+ ( )
In terms of Fourier coefficients,
+
3.5.2 Time Shifting:
For continuous time periodic signal,
( − )
In terms of Fourier coefficients,

3.5.3 Frequency Shifting


For continuous time periodic signal,

( )
In terms of Fourier coefficients,

3.5.4 Time Reversal


For continuous time periodic signal,
(− )
In terms of Fourier coefficients,

3.5.5 Multiplication
For continuous time periodic signal,
( ) ( )
In terms of Fourier coefficients,
Signals and Systems | 105

3.5.6 Conjugation
For continuous time periodic signal,

( )
In terms of Fourier coefficients,

3.5.7 Time Scaling


For continuous time periodic signal,
( ) ; where >0

In terms of Fourier coefficients,

3.5.8 Differentiation
For continuous time periodic signal,
( )
In terms of Fourier coefficients,

3.5.9 Integration
For continuous time periodic signal,
( )
In terms of Fourier coefficients,
1

3.5.10 Real and Even


For continuous time periodic signal,
( ) is real and even
In terms of Fourier coefficients,
are real and even
3.5.11 Real and Odd
For continuous time periodic signal,
( ) is real and odd
In terms of Fourier coefficients,
are imaginary and odd
3.5.12 Parseval’s Relation
For continuous time periodic signal,
Average power of signal ( ) is,
1
= | ( )|
106 | Fourier Series and Transform

In terms of Fourier coefficients, the average power is

= | |

3.6 Gibb’s Phenomenon


Consider a periodic signal ( ) and it’s exponential form of Fourier Series expansion

( )=

Above expansion consists the terms as a sum of infinite series of harmonic frequency
components. When the signal ( ) is to be reconstructed with some N terms from infinite
series of harmonic frequency components, the signal displays some oscillations also called as
ripples that have discontinuities with it.

Fig. 3.7 Approximation of square waveform using some harmonic frequency components
Signals and Systems | 107

Consider the Fig. 3.7(a) of periodic square waveform. Figure 3.7(b), (c), (d) shows the
reconstruction using some N harmonic frequency components. It can be easily seen that the
reconstructed signals have some ripples with it at the edge points. Also it can be observed
that, at the points of discontinuity, the Fourier series converges to average value of the signal
on either side of discontinuity. This phenomenon is called as Gibbs phenomenon after the
name of great scientist, Josiah Gibbs. He had shown that as the value of N increases, the peak
overshoot shifts towards the point of discontinuity.

Solved examples on Fourier series


Example 3.1 Determine the trigonometric form of Fourier series of the waveform shown in
fig 3.1.1.

Fig. 3.1.1.
Solution: The waveform has even symmetry, half wave symmetry and quarter wave
symmetry.

= 0, = 0, = ∫ ( )
Equation for square wave is written as,
( ) = ; for = 0

= − ; for =
Now,

4
= ( )
108 | Fourier Series and Transform

4 4
= + (− )

= -

= -

= − - −

4 4
= −0 − −
2 2 2 2 2
2 2 4
= + =
2 2 2
= 0;
for even values of
= ±1;
for odd values of
∴ = 0;
for even values of
= ;
for odd values of
( )=

4A 3 5 7
( )= − + − +⋯
1 3 5 7

Example 3.2 Determine the trigonometric form of Fourier series of the waveform shown in
fig 3.1.2.

Fig. 3.1.2.
Signals and Systems | 109

Solution: The waveform has even symmetry.

= ∫ ( ) ; = ∫ ( ) ; =0
( ) ( )
ℎ = = = =
( ) ( )

( , ( )) = [0, 0] and ( , ( )) = [ , ]
( ) ( )
∴ = => = => ( ) =

for = 0
Now,

= ∫ ( ) = ∫ = ∫ = = −0 =

4 4 2
= ( ) =

= +

= sin + cos −
= +1;
for even values of
= −1;
for odd values of
∴ = 0;
for even values of
= cos[ − 1] = −
; for odd values of

( )= +
2

4A 3 5 7
( )= − + + + +⋯
2 1 3 5 7
110 | Fourier Series and Transform

Example 3.3 Given the Fourier series coefficients as a determine the signalx(t):
1 1
= [ − 1] + [ + 1] + [ + 3] + [ − 3] + [ + 4] + [ − 4]
2 3 6
Solution:
1 1
= [ − 1] + [ + 1] + [ + 3] + [ − 3] + [ + 4] + [ − 4]
2 3 6
The Fourier series coefficients are identified as
1
= 0, = , = 0, = , =1
6
1
= , = 0, = , =1
2 3

( )=

Substituting the values of a , we get


1 1
( )= + + + + +
2 6 3
Expanding by using Euler’s relation, we get
1
( )= ( + )+ (cos − ) + (cos 3 + 3 )
2 6
1
+ (cos 3 − 3 )+2 4
3
Simplifying, we get
3 1 1
( )= − sin +2 4 − sin 3 + 3
2 2 6 2
Example 3.4 Given the Fourier series coefficients as determine the signal ( ):
( )= + + 2
Solution:
Given
1 1
( )= + + 2
2 4
Now cost is periodic with period T = = 2π and cos2t s periodic with period T = =
π.
Hence,
It is periodic with period T = LCM T , T = LCM(2π, π) = 2π
Signals and Systems | 111

The fundamental frequency Ω


, ℎ ( )
1 + 1 +
( )= + +
2 2 4 2
1 + +
= + +
2 2 8
1 1 1 1 1
= + + + +
2 2 2 8 8
Since = 1,
1 1 1 1 1
( )= + + + +
8 2 2 2 8
We can identify the Fourier series coefficients as,
= , = , = , = = , = 0, | | > 2

3.7 Fourier Transform (FT)


3.7.1 Definition
Let ( ) is a continuous time signal, ( ) is the Fourier transform of signal ( ). The
Fourier signal of signal ( ) is defined as,
( )=∫ ( ) =∫ ( ) =
Ƒ{ ( )}
(3.24)
Where, =Cyclic frequency

Condition for Fourier Transform Existence:


The Fourier transform exists for the signal ( ) if it follows Dirichlet condition.
1. Signal ( ) is absolutely integrable.

( ) <∞

2. The signal ( ) should have a finite number of maxima and minima within a finite
duration of interval.
3. The signal ( ) can have a finite number of discontinuities within a interval.
112 | Fourier Series and Transform

3.7.2 Definition of Inverse Fourier Transform (IFT)


Inverse Fourier transform of ( ) is given by,

( )= ∫ ( ) =∫ ( ) =Ƒ { ( )} (3.25)

Signal ( ) and its Fourier transform ( ) makes the Fourier trasnform pair and the relation
is given by,
( )⃖ , ⃗ ( )

3.7.3 Magnitude and Phase Spectrum using Fourier Transform

Let Fourier transform of ( ) is a complex function of frequency and it can be defined in


terms of real and imaginary parts as shown below,

( )= ( )+ ( )

Where, ( ) = eal part of ( )

( ) =Imaginary part of ( )

The Magnitude spectrum of ( ) can be written as,

| ( )| = ( )+ ( ) (3.26)

Magnitude spectrum has always even symmetry.


The Phase spectrum of ( ) can be written as,
( )
∠ ( ) = tan
( )
(3.27)

Phase spectrum has always odd symmetry.


Both magnitude spectrum and phase spectrum are called as frequency spectrum.

3.8 Properties of Fourier Transform (FT)


3.8.1 Linearity

If ( ) ⃗ ( ),
( ) ⃗ ( )
Then,
{ ( )+ ( )} = ( )+ ( ) (3.28)
Signals and Systems | 113

Proof: By definition of FT,


( )=∫ ( )
∴ ( )=∫ ( )
∴ ( )=∫ ( )
Consider linear combination, ( )+ ( )
∴ { ( )+ ( )} = ∫ [ ( )+ ( )]
=∫ ( ) +∫ ( )
= ( )+ ( )

3.8.2 Time Shifting:

If ( ) ( )
Then,
{ ( − )} = ( ) (3.29)
Proof: By definition of FT,
( )=∫ ( )
∴ { ( − )} = ∫ ( − )
Let − = , ∴ = + , =
( )
∴ { ( − )} = ∫ ( )
=∫ ( ) .
= ∫ ( )
= − 0 ( )

3.8.3 Time Scaling:

If ( ) ⃗ ( )
then,
{ ( )} = ∗ (3.30)
| |
Proof: By definition of FT,

( )= ( )
114 | Fourier Series and Transform

{ ( )} = ( )

Let = , = , dt =
,

1 −
∴ { ( )} = ( )
−∞
1
= ( )

‘a’ can be negative or positive. Hence, in general { ( )} = ∗


| |

3.8.4 Time Reversal:


If ( ) ⃗ ( )
then,
{ ( )} = ( ) (3.31)

Proof: By using the Time Scaling Property in Eq. (3.30)


1
{ ( )} =
| |
For = −1,
1
{ (− )} = = (− )
| − 1| −1

3.8.5 Conjugation:
If ( ) ⃗ ( )
then,
{ ∗ ( )} = ∗ (− ) (3.32)

Proof: By definition of FT,

( )= ( )

∴ { ∗ ( )} = ∗( )
Signals and Systems | 115

= ( )

( ) ( )
=∫
∗( )
= −

3.8.6 Frequency Shifting:


( ) ⃗ ( )
then,
( ) = ( − ) (3.33)
Proof: By definition of FT,

( )= ( )

∴ ( ) = ( )

( ) ( )
=∫
= ( − )
3.8.7 Time differentiation
If ( ) ⃗ ( )
then,
( ) = ( ) (3.34)

Proof: By definition of FT,

( )= ( )

∴ ( ) =∫ ( )

= ( )

, = −
116 | Fourier Series and Transform

∴ ( ) = ( )| − (− ) ( )

= 0+ ( )

= ( )
3.8.8 Time Integration

( ) ⃗ ( )
then,
∫ ( ) = ( ) (3.35)
Proof: Let,

( )= ( )

Take FT on both sides,

{ ( )} = ( )

( )= ( )

( )=( ) ( )

1
∴ ( ) = ( )

3.8.9 Differentiation in Frequency

If ( ) ⃗ ( )
then,
{ ( )} = ( ) (3.36)
Proof: By definition of FT
Signals and Systems | 117

( )= ( )

Take differentiation on both sides, w. r. t.

( )= ( )

=∫ ( )

( ) (− )

As, − =

1
( )= [ ( )]

1
= { ( )}

∴ { ( )} = ( )
3.8.10 Convolution

If ( ) ⃗ ( )

( ) ⃗ ( )
then,
{ ( )∗ ( )} = ( ) ( ) (3.37)
Proof: By definition of FT,

( )= ( )

( )= ( )
118 | Fourier Series and Transform

∴ { ( )∗ ( )} = [ ( )∗ ( )]

( )∗ ( − )

= − ,
= ,
= + ,
=

∴ { ( )∗ ( )} = ( ) ( )

∴ { ( )∗ ( )} = ( ) ( )

Replace ,
∴ { ( )∗ ( )} = ( ) ( )

3.8.11 Parseval’s Theorem:

If ( ) ⃗ ( )
then,
∫ | ( )| = ∫ | ( | (3.38)
Proof:
Let | ( )| = ( ) ∗ ( ) (3.39)
∴ ∫ | ( )| = ∫ ( ) ∗( ) (3.40)
Recall the inverse FT definition,
( )= ∫ ( )
∗( )= ∗( )
∴ ∫ (3.41)
R.H.S. of Eq. (3.40) becomes,
Signals and Systems | 119

1 ∗(
∴ ( ) )
2

1 ∗(
∴ | ( )| = ) ( )
2

1 ∗(
= ) ( )
2

1
| ( )| = | ( |
2

3.8.12 Duality Property:

If ( ) ⃗ ( )
then,
( ) ↔2 (− ) (3.42)

Proof: By definition of IFT,


1
( )= ( )
2

(− ) = ∫ ( ) (3.43)
Put t= j in Eq. (3.43)
(− )= ∫ ( ) d

∴2 (− )= ( )

Where R.H.S. is FT of ( ).
120 | Fourier Series and Transform

3.9 Fourier Transform (FT) Representation of Continuous-Time (CT) LTI System in


terms of Convolution and Multiplication

3.9.1 Representation of Transfer Function of CT LTI System in Frequency Domain


Transfer function if CT LTI system is defined as the ratio of output of Fourier transform to
the input of Fourier transform.
Let ( )=Input of CT system
( )=Output of CT system
( )=FT of ( )
( )=FT of ( )
Then, the transfer funtion (TF) can be written as,
( )
= ( )
(3.44)
Frequency domain representation of CT LTI system can be written in terms of differential
equation as,
( )+ ( )+ ( ) +⋯+ ( )= ( )+ ( )+

( ) +⋯+ ( ) (3.45)

Transfer function of CT LTI system can be obtained by taking the FT of above equation and
arranging as a ration of ( ) to ( ).

3.9.2 Relation of Impulse Response and Transfer Function of CT LTI System


Let ( ) and ( ) are the input and output of CT LTI system respectively. When input ( )
is replaced by impulse signal ( ), the system output is called as impulse response which is
denoted by ℎ( ).
The impulse response is represented as the convolution of input and impulse response itself
for any input to LTI system. Let the convolution opertaion,

( ) = ( ) ∗ ℎ( ) = ( )ℎ( − )

Let,
( )=FT of ℎ( )
( )=FT of ( )
( )=FT of ( )
Using the convolution property,
Signals and Systems | 121

Ƒ{ ( ) ∗ ℎ( )} = Ƒ{ ( )} = ( ) ( )
∴ ( )= ( ) ( )
( )
( )= (3.46)
( )

Equation (3.46) shows that the transfer function in frequency domain is given by Fourier
transform of impulse response which is the ratio of Fourier transform of output to input.

3.9.3 Response of CT LTI System in terms of Fourier Trasnform


Consider Eq. (3.46) of transfer function,
( )
( )=
( )
∴ ( )= ( ) ( )
( )( ) ⋯
∴ ( )=
( )( ) ⋯
(3.47)
Using partial fraction expansion,
( )= +( +( +⋯ (3.48)
( ) ) )

Where , , , .. are residues.


The Fourier transform of = (3.49)
Using Eq. (3.49), the inverse Fourier transform of Eq. (3.48) can be written as,
( )= ( )+ ( )+ ( )+⋯ (3.50)
The response obtained by using Eq. (3.50) is the time domain steady state response of the LTI
continuous time system as the transfer function is defined with zero initial conditions.

3.9.4 Magnitude and Phase Response of CT LTI System

The output of LTI system in terms of the convolution opertaion is written as,

( ) = ( ) ∗ ℎ( ) = ℎ( ) ∗ ( ) = ℎ( ) ( − )

Let,
( )= (3.51)
( − )= ( )
∴ (3.52)
( )=∫ ( )
ℎ( ) (3.53)
122 | Fourier Series and Transform

( )=∫ ℎ( ) (3.54)
( )= ∫ ℎ( ) (3.55)

Recall the definition of FT,


( ) = ∫ ℎ( ) = ∫ ℎ( ) (3.56)
Using Eq. (3.51), (3.56), the Eq. (3.55) can be written as,

( )= ( ) ( ) (3.57)

Equation (3.57) says that if a complex signal is given as input signal to CT LTI system, then
the output has the same frequency as that of input signal multiplied by ( ). Hence, ( )
is called the frequency response of the CT LTI system. The

The ( ) can be represented in terms of magnitude and phase as,


( ) = | ( )|∠ ( )
Where,
| ( )| =
∠ ( )= ℎ
( )= ( )+ ( )
The magnitude function is written as,
| ( )| = ( )+ ( ) (3.58)

The phase function is written as,


( )
∠ ( )=
( )
(3.59)

Solved examples on Fourier transform


Example 3.5 Determine the Fourier transform of given continuous time signals.
1 − ; for | | < 1
1. ( )=
0; for | | > 1
2. ( )= ( )

Solution:
− ; | |<
1. Given, ( ) =
; | |>

That means
( )=1− ; for = -1 to +1
Signals and Systems | 123

Using the definition of FT,

( )= ( )

( )= (1 − )

( )= −

, = −

= − − 2
− − −

− 2
= − +

, = −

− 2
= − + − 1
− − −

2
= − − + − −
− ( )

2
= − − − − − +
( ) −

2 2
= − − − + +

2 2 2 2
=− + − − + + + + −

Using, = and =
2 2
=− + + +
124 | Fourier Series and Transform

2 2
=− 2 + 2

4
= −
2. Given, ( ) = ( )

Given signal can be written as,


( )= for ≥ 0
Using the definition of FT,

( )= ( )

=∫ =∫

= ∫ + ∫

( ) ( )
= ∫ + ∫

( ) ( )
= +
( ) ( )

1
= −
2 −( − + ) −( − + )
1
+ −
2 −( + + ) −( + + )
1 1 1 1
= 0+ + 0+
2 ( − + ) 2 ( + + )
1 1 1
= +
2 ( + )− ( + )+
1 ( + )+ +( + )−
=
2 ( + ) +

1 2( + ) ( + )
= =
2 ( + ) + ( + ) +
Signals and Systems | 125

Example 3.6 Determine the Fourier transform of the rectangular pulse shown in fig 3.10.1.

Fig. 3.10.1
Solution:
From fig. 3.10.1, the equation can be written as,
x(t) = 1; for = − +
Using the definition of FT,

( )= ( )

= 1

= = − = ( − )= 2

=2 =2 =2 ……….as =

Example 3.7 Determine the inverse Fourier transform of the following functions.
( )
1. ( )=

2. ( )=
( )

Solution:
( )
1.Given, ( )=
3( ) + 14
( )=
( + 3)( + 4)
Using partial fraction,
( )= +
( + 3) ( + 4)
126 | Fourier Series and Transform

3( ) + 14
= ( + 3) =5
( + 3)( + 4)

3( ) + 14
= ( + 4) = −2
( + 3)( + 4)

( )= −( (3.60)
( ) )

Using Fourier transform of { ( )} =

The inverse Fourier transform of Eq. (3.60) can be written as,

( )=5 ( )−2 ( )

2. Given, ( )=
( )

Using partial fraction, above function can be written as,

X(jΩ) = +
(jΩ + 3) jΩ + 3
+7
= ( + 3) = −3 + 7 = 4
( + 3)
Using rule of repeating poles,
+7
= ( + 3) = ( + 7) =1
( ) ( + 3) ( )

( )= + (3.61)
( )

Using Fourier transform of { ( )} = ( )

The inverse Fourier transform of Eq. (3.61) cann be written as,

( )=4 ( )+ ( )
Signals and Systems | 127

Example 3.8 Given X(jΩ) = 2πδ(Ω − Ω )


Determine the time-domain signal x(t).
Solution: The inverse Fourier transform can be found as,
1
x(t) = X(jΩ)e dΩ

Substituting for X(jΩ), we get


1
x(t) = 2πδ(Ω − Ω ) e dΩ

x(t) = e δ(Ω − Ω )dΩ

Using the Sifting property of impulse,

x(t)δ(t − t ) dt = x(t )

∴ x ( t) = e
For this time–domain signal, the Fourier transform given as,
e ↔ 2πδ(Ω − Ω )

Example 3.9 Find the Fourier transform of the periodic signal


x(t) = sinΩ t
Solution: Using Euler’s identity, x(t) can be written as
e −e 1 1
x(t) = = e − e
2j 2j 2j
The Fourier series coefficients can be identified as,
1 −1
a = ,a = ,
2j 2j

Substituting the values, we get the Fourier transform ofsinΩ t as


π π
X(jΩ) = δ(Ω − Ω ) − δ(Ω + Ω )
j j
128 | Fourier Series and Transform

3.10 Discrete Time Fourier Series representation

A discrete time signal which is periodic with fundamental period can be decomposed into
harmonics of frequency components (frequency spectrum). When we combine these related
frequency components it must give the Fourier series representation of that particular periodic
discrete time signal which is a function of angular frequency denoted as ω, . This representation
of Fourier series of discrete time signal is called Discrete Time Fourier Series (DTFS).
3.10.1 Representation of Discrete Time Fourier Series

The Discrete Time Fourier Series, (DTFS) of discrete time periodic signal [ ] with period
is defined as,

[ ]=

, ω0 = Fundamental frequency of [ ]

[ ]= 0

, = harmonic frequency component of [ ]

[ ]=∑ (3.62)

Where, = Fourier coefficients


This coefficient represents the amplitude and phase of the kth frequency component that
provides the presentation of [ ] in the Fourier/frequency domain.
These Fourier coefficients can be evaluated using following formula,

= ∑ [ ] ; for = 0, 1, 2, . . −1 (3.63)

3.10.2 Properties of DTFS in terms of coefficients


The properties of DTFS coefficients are given in table 3.1 in terms of Fourier coefficients. Let
[ ] has Fourier series coefficients and [ ] has Fourier series coefficients .
Signals and Systems | 129

Table 3.1 : Properties of DTFS in terms of coefficients

Sr. No. Property name Discrete periodic FS coefficients


signal
1 Linearity [ ]+ [ ] +
2 Time shifting [ − ]

3 Frequency shifting
[ ]
∗ ∗
4 Conjugation [ ]
5 Time reversal [− ]
6 Time scaling where is 1
multiple of
7 Multiplication [ ] [ ]

8 Convolution
[ ] [

− ]
9 Symmetry If [ ] is real = ∗
| |=| |
∠ = −∠
If [ ] is real and are real and even
even
If [ ] is real and are imaginary and
odd odd
10 Parseval’s theorem
1 = | |
= | [ ]|

Solved examples on DTFS


Example 3.6 Determine the DTFS representation of the following signals.
1. [ ] = 4 √5
2. [ ] = 8
130 | Fourier Series and Transform

Solution:
1. Given, [ ] = √

Let us check for the periodicity,


[ + ]=4 √5 [ + ] = 4cos[√5 + √5 ]
For periodicity √5 should be equal to integral multiple of 2 .
Let √5 = × 2 ; where M and N are integers ⎯⎯⎯ =

Here N cannot be an integer for any integer value of M and so [ ] will not be periodic and
for nonperiodic signal FS does not exists.
2. Given, [ ] =

Let us check for the periodicity,


[ + ]
[ + ]=8 = 8cos[ + ]
4 4 4
For periodicity should be equal to integral multiple of 2 .

Let = × 2 ; where M and N are integers ⎯⎯⎯ =8


Here N is an integer for M = 1, 2, 3, .....
Let M = 1, \ N = 8
So [ ] will be periodic with fundamental period 8 and fundamental frequency
= = = .
Now go for the Fourier series expansion,
The Fourier coefficient is given by,

= ∑ [ ] ; for = 0, 1, 2, . . −1
Here, N=8 and [ ] = 8

= ∑ 8 ; for = 0, 1, 2,3, 4, 5, 6, 7

8
= = −
8 4 4 4 4
Signals and Systems | 131

= 0( 0− 0) + − + −
4 4 4 2 2 2
3 3 3
+ − + ( − )
4 4 4
5 5 5 3 3 3
+ − + −
4 4 4 2 2 2
7 7 7
+ −
4 4 4
√ √
As, 0 = 1, = −1, = = = , =− , = 0, =
√ √
1, = 0, = =− , = 0, = −1, = , =


√2 √2 √2 √2 √2 √2 √2 √2 √2
=1+ − − − − + − − +
2 2 2 2 2 2 2 2 2
√2 √2 √2
+ +
2 2 2
2 2 2 2
=1+ + + + +
4 4 4 4
When k=0, = =1
When k=1, = = 1+ + +1+ + = 4
When k=2, = = 1+1+1+2+1+1= 7
When k=3, = = 1 + + + 3 + + = 10
When k=4, = = 1 + 2 + 2 + 4 + 2 + 2 = 13
When k=5, = = 1 + + + 5 + + = 16
When k=6, = = 1 + 3 + 3 + 6 + 3 + 3 = 19
When k=7, = = 1 + + + 7 + + = 22
The Fourier series representation of [ ] is,

[ ]= = =

= + + + + + + +

[ ]=1+4 +7 + 10 + 13 + 16 + 19 + 22
132 | Fourier Series and Transform

3.11 Discrete Time Fourier Transform (DTFT) Representation


3.11.1 Definition of DTFT

The Fourier transform (FT) of discrete-time signals is called Discrete Time Fourier Transform
(DTFT).
Let, [ ]=Finite energy discrete time signal
=Discrete time Fourier transform of signal [ ]
= { [ ]} = ∑ [ ] (3.64)
The discrete time Fourier transform exists only for the absolutely summable signals. That means
the Fourier transform exists for the signal [ ] if,

| [ ]| < ∞

3.11.2 Definition of Inverse DTFT

Let, [ ]=Finite energy discrete time signal


=Discrete time Fourier transform of signal [ ]
The inverse Fourier transform (IFT) of discrete-time signal is written as,
[ ]= { }= ∫ (3.65)
As is periodic with period 2 , the limits of integral can be either from "− to + ", or
from "0 to 2 ", or any interval of 2 .

3.12 Properties of DTFT


3.12.1 Linearity
If [ ] ⃗ [ ],
[ ] ⃗ [ ]
Then,
{ [ ]+ [ ]} = [ ]+ [ ] (3.66)
Proof: By definition of FT,

= { [ ]} = [ ]
Signals and Systems | 133
+∞
∴ = [ ] −

=−∞

+∞
∴ = [ ] −

=−∞

Consider linear combination, [ ]+ [ ]


∴ { [ ]+ [ ]} = ∑+∞
=−∞( [ ]+ [ ]) −

= ∑+∞
=−∞[ [ ] − + [ ] − ]
= ∑+∞
=−∞ [ ] − + ∑+∞
=−∞ [ ] −

= [ ]+ [ ]

3.12.2 Time Shifting:


If [ ] ⃗ ,
Then,
{ [ − ]} = (3.67)
Proof: By definition of FT,

= { [ ]} = [ ]

∴ { [ − ]} = ∑+∞
=−∞ [ − ] −

Let − = , ∴ = +
+∞

∴ { [ − ]} = [ ] − ( )

=−∞
+∞

= [ ] − −

=−∞
+∞
= − [ ] −

=−∞
Replace m by n
=
134 | Fourier Series and Transform

3.12.3 Periodicity

If [ ] ⃗ [ ], then [ ] is periodic with period 2


∴ = [ ( +2 ] (3.68)

Where is an integer.

3.12.4 Time Reversal:


If [ ] ⃗ [ ], then,
{ [− ]} = [ ] (3.69)
If a signal is folded about the origin in discrete time, its magnitude spectrum does not change
and the phase spectrum changes in sign i.e., phase reversal happens.

Proof:

By definition of FT,

= { [ ]} = [ ]

∴ { [− ]} = ∑+∞
=−∞ [− ]

Let − = , ∴ ℎ → −∞, → +∞ ℎ → +∞, → −∞


+∞

∴ { [− ]} = [ ]
=−∞
+∞
{ [− ]} = [ ] (− )−

=−∞
{ [− ]} = [ ]

3.12.5 Conjugation:
If [ ] ⃗ [ ], then

{ ∗ [ ]} = ∗
[ ] (3.70)

Proof: By definition of FT,

= { [ ]} = [ ]
Signals and Systems | 135

{ ∗ [ ]} = ∗
[ ]

[ ] ( )
=


=

= [ ]

3.12.6 Frequency Shifting:

If [ ] ⃗ [ ], then

( )
[ ] = [ ] (3.71)

Proof: By definition of FT,

= { [ ]} = [ ]

[ ] = [ ]

( )
= [ ]

= [ ( − 0)]

3.12.7 Differentiation in Frequency


If [ ] ⃗ [ ], then
{ [ ]} = [ ] (3.72)
Proof: By definition of FT,

= { [ ]} = [ ]
136 | Fourier Series and Transform

{ [ ]} = [ ]

{ [ ]} = [ ] (− )

= [ ](− )

As, (− ) =

= [ ]

= [ ]

= [ ]

3.12.8 Convolution

If [ ] ⃗ [ ]
[ ] ⃗ [ ] then,
{ [ ]∗ [ ]} = [ ] [ ] (3.73)

Proof: By definition of FT,

= { [ ]} = [ ]

+∞
∴ = [ ] −

=−∞

+∞
∴ = [ ] −

=−∞
Signals and Systems | 137

∴ { [ ]∗ [ ]} = ( [ ]∗ [ ])

Using convolution formula,

{ [ ]∗ [ ]} = ( [ ]∗ [ − ])

= ( [ ]∗ [ − ])

[ ] [ − ] ( )
=

Let, n-k=m

= [ ] [ ]

Replace ,
∴ { [ ]∗ [ ]} = [ ] [ ]

3.12.9 Parseval’s Theorem:


If [ ] ⃗ [ ]

[ ] ⃗ [ ] then, Parseval’s relation says that,

∑ ( [ ]∗ [ ]) = ∫ [ ] (3.74)
Proof: By definition of FT,

= { [ ]} = [ ]

Using inverse FT
1
[ ]= [ ]
2

Using R.H.S of Eq. (3.74),


138 | Fourier Series and Transform

1 1 ∗
[ ] = 1[ ] 2 [ ]
2 2

1 ∗
= 1[ ] 2 [ ]
2

1 ∗
= 1[ ] [ ]
2

=∑ ( [ ]∗ [ ])

3.13 Discrete Fourier Transform (DFT) Representation

The above discrete time Fourier transform (DTFT) concept provides analysis for a discrete time
signal in frequency domain where it is a continuous function of and so it cannot be processed
by digital system. Hence, we have to represent this into a discrete function of , so that
frequency analysis of discrete time signals can be presented using digital system.
Basically, the DFT of a discrete time signal is obtained by sampling the DTFT of the signal at
uniform frequency intervals. These samples must be sufficient to avoid aliasing effect. DFT is
represented as a sequence of complex numbers represented as ( ) for k = 0, 1, 2, 3, …. The
magnitude and phase of each sample of ( ) can also be computed.
The plot of magnitude versus is called magnitude spectrum and the plot of phase versus is
called phase spectrum (frequency spectrum).

3.13.1 Definition of DFT

Let [ ] be DTFT of the discrete time signal [ ]. The discrete Fourier transform (DFT)
[ ]is calculated by sampling one period of the DTFT [ ] at a finite number of frequency
points.
Let one period consists of N equally spaced points, 0 ≤ ≤2 .
Each frequency point is represented by ratio,
= ; for = 0, 1, 2 … −1
Hence, the sampling of DTFT at frequency points is written as,
[ ]= [ ] ; for = 0, 1, 2 … −1 (3.75)

DFT is also called as N point DFT, where the number of samples N for a finite duration
sequence x[n] of length L should be such that, N ≥ L, in order to avoid aliasing.
Signals and Systems | 139

Let, [ ] = Discrete time signal having length as L


[ ]= DFT of [ ]
Hence, n-point DFT of [ ] is defined as,
[ ]=∑ [ ] ; for = 0, 1, 2 … −1 (3.76)

3.13.2 Definition of Inverse DFT

Let, [ ] = Discrete time signal having length as L


[ ]= N-point DFT of [ ]
Hence, inverse DFT of [ ] is defined as,
[ ]= ∑ [ ] ; for = 0, 1, 2 … −1 (3.77)
The relation between DTF and inverse DFT is expressed as,
[ ] ⃖ , ⃗ [ ]
3.14 Properties of DFT
3.14.1 Linearity
If [ ] ⃗ [ ],
[ ] ⃗ [ ]
Then,
{ [ ]+ [ ]} = [ ]+ [ ] (3.78)
Proof: By definition of FT,

[ ]= { [ ]} = [ ]

−1
−2
∴ [ ]= [ ]
=0

−1
−2
∴ [ ]= [ ]
=0
Consider linear combination, [ ]+ [ ]
−2
−1
∴ { [ ]+ [ ]} = ∑ =0 ( [ ]+ [ ])
−2 −2
−1
=∑ =0 [ [ ] + [ ] ]
140 | Fourier Series and Transform
−2 −2
−1 −1
= ∑ =0 [ ] + ∑ =0 [ ]
= [ ]+ [ ]

3.14.2 Circular Time Shifting:


If [ ] ⃗ [ ],
Then,
−2
{ [( − ) ]} = [ ] (3.79)

Proof: By definition of DFT,

[ ]= { [ ]} = [ ]

−2
−1
∴ { [( − ) ]} = ∑ =0 [( − ) ]
Let − = , ∴ = +
−1
−2 ( )
∴ { [ − ]} = [( − ) ]
=0
−1
−2 ( )
= [( ) ]
=0
−1
−2 −2
= [ ]
=0

Replace p by n
−2
= [ ]
3.14.3 Periodicity

If [ ] ⃗
[ ] , i. e., If a sequence [ ] is periodic with periodicity of N samples then
N-point DFT, X(k) is also periodic with periodicity of N samples.
Hence, if [ ] and [ ] are N point DFT pair then,
[ + ] = [ ] ; for all (3.80)
[ + ] = [ ] ; for all (3.81)

Proof: By definition of DFT,


Signals and Systems | 141

[ ]= { [ ]} = [ ]

−2 ( + )
−1
∴ [ + ]=∑ =0 [ ]
−1
−2 −2
= [ ]
=0
−1
−2
= [ ] −2

=0
−1
−2
= [ ]
=0
∴ [ + ]= [ ]

3.14.4 Time Reversal:


If [ ] ⃗ [ ], then,

{ [ − ]} = [ − ] (3.82)
If a signal is folded about the origin in discrete time, its magnitude spectrum does not changes
and the phase spectrum changes in sign i.e., phase reversal happens.

Proof:
By definition of DFT,

[ ] = { [ − ]} = [ − ]

Let, − = ,∴ = −
( − )
= [ ]

( − )
= [ ]

= [ − ]
142 | Fourier Series and Transform

3.14.5 Conjugation:

If [ ] ⃗ [ ], then

{ ∗ [ ]} = ∗
[ − ] (3.83)

Proof: By definition of DFT,

{ ∗ [ ]} = ∗
[ ]

= [ ]


= ∑ [ ] As, =1
( ) ∗
= ∑ [ ]

= [ − ]

3.14.6 Circular Frequency Shifting:

If [ ] ⃗ [ ], then

[ ] = [( − ) ] (3.84)

Proof: By definition of FT,

[ ]= [ ] = [ ]

( )
= [ ]

= [( − ) ]

3.14.7 Multiplication

If [ ] ⃗ [ ],
[ ] ⃗ [ ]
Signals and Systems | 143

Then,
{ [ ] [ ]} = [ [ ]⊛ [ ]] (3.85)
Proof: By definition of FT,

[ ]= { [ ]} = [ ]

{ [ ] [ ]} = [ ] [ ]

1
= [ ] [ ]

1
= [ ] [ ]

1 ( ( ))
= [ ] [ ]

1
= [ ] [( − )]

1
= [ [ ]⊛ [ ]]

3.14.8 Convolution

If [ ] ⃗ [ ],
[ ] ⃗ [ ]
Then,
{ [ ]⊛ [ ]} = [ ] [ ] (3.86)

Proof: By definition of FT,

[ ]= { [ ]} = [ ]
144 | Fourier Series and Transform
−1
−2
∴ [ ]= [ ]
=0

−1
−2
∴ [ ]= [ ]
=0
Considering the product of X [k] and X [k] and taking inverse DFT of the product the convolution
property ccan be proved.
Hence,
[ ] [ ]= { [ ]⊛ [ ]}

3.14.9 Parseval’s Theorem:

If [ ] ⃗ [ ],
[ ] ⃗ [ ]
then, Parseval’s relation says that,
∑ [ ] ∗ [ ]= ∑ [ ] ∗ [ ] (3.87)

Unit Summary
Fourier series and Fourier transform are fundamental tools in signal processing, mathematics,
physics, and engineering. They are used to analyze and represent periodic and non-periodic
functions in terms of sinusoidal or complex exponential functions. Fourier series decomposes
a periodic function into a sum of sinusoidal functions (sine and cosine). It's applicable to
functions with periodicity, allowing representation in terms of a discrete set of harmonics. The
series comprises a constant term (DC component) and an infinite sum of harmonic terms, each
with its own amplitude and phase. Fourier transform extends the concept of Fourier series to
non-periodic functions or signals. It transforms a function from the time or spatial domain into
the frequency domain. It decomposes a function into its constituent frequencies, represented
by a continuous spectrum. Fourier series and Fourier transform are indispensable tools in
various fields for analyzing, synthesizing, and processing signals and functions. Understanding
these concepts facilitates advanced analysis and manipulation of signals in diverse
applications.
Signals and Systems | 145

Exercises
1. Find the Fourier series coefficients for the signals shown in figure.

2. Find the inverse CTFT of the following:



a. ( Ω) =
Ω Ω

b. (jΩ) =


c. X(jΩ) = Ω Ω
3. Find the output y(t) of an LTI system for an input x(t) = e u(t) + e u(−t) and
impulse response h(t) = e u(t) by using CTFT.
4. Given the FT of x(t) as,
2 + jΩ
X(jΩ) =
(jΩ) + 6jΩ + 31
Find the transform for the following signals by using properties of CTFT:
a. x + x(6t)
b. x(−3t + 1)
c. x(t)
d. tx(t)

5. Find the convolution of following signals using CTFT:


x(t) = te u(t), h(t) = te u(t)
6. If the function y = x in the range 0 to π is expanded as a sine series, show that it is equal
to

sin x sin 2x sin 3x


2 − + + ⋯……
1 2 3
7. Expand (π − x) in a sine series valid when
146 | Fourier Series and Transform

0≤x≤π
8. Find a sine series for
π
f(x) = x; 0<x<
2
= 0; <x<π
9. Show that in the range (0, π), the sine series for πx − x is sin x + sin3x +
sin5x + ⋯

10. Find a Fourier cosine series corresponding to the function f(x) = x, defined in the
interval (0,π).

11. Find the Fourier sine series and the Fourier cosine series corresponding to the function ,

f(x) = π − x when 0 < x < π


Defined in the interval 0 to π.
12. A function f(t) can be expressed as a sum of an odd part and even part:
f(t) = f (t) + f (t)

Show that Re [F(ω)] is the transform of f (t)


jIm[F(ω)] is the transform of f (t)
13. Find the N-point Discrete Fourier Transform (DFT) of

h(n) = , for n = 0,1,2, and zero otherwise.


14. Find the DFT of x(n) if

x(n) = 1, for n = 2 to 6
x(n) = 0, for n = 0,1,7,8,9
Assume x(n) is periodic beyond this interval 0 − 9
15. Find the response of the following system to the input:

nπ 2πn π
x(n) = 2 + 2cos + cos +
4 3 2
System: H(ω) = e cos( )
16. Show that the Hilbert Transform of exp (jωt)is(−sgn f)exp (jωt)
Signals and Systems | 147

17. Without using a calculator or computer find the dot products of (a) w1 and w-1, (b) w1
and w-2 (c) w11 and w37, where

w
⎡ ⎤
⎢w ⎥
w =⎢ ⎥ and w /
⎢w ⎥
⎣w ⎦
to show that they are orthogonal.
18. Find the DTFS harmonic function of a signal x[n]with period 4 for which x[0]= 3,
x[1]=1, x[2] = -5, and x[3]=0 using the matrix multiplication X =
19. One period of a periodic function with period 4 is described by x[n] = δ[n] −
δ[n − 2], 0 ≤ n < 4. Using the summation formula for the DTFS harmonic function and
not using the tables or properties, find the harmonic function X[k].
20. Find the DTFS harmonic function of

x [ n] = δ [ m ] − δ [ m − 1]

with N = N = 3.
21. A periodic signal x[n] is exactly described for all discrete time by its DTFS
/
X[k] = (δ [k − 1] + δ [k + 1] + j2δ [k + 2] − j2δ [k − 2])e
Using one fundamental period as the representation time.
a) Write a correct analytical expression for x[n] in which √1 (j) does not appear

b) What is the value of x[n] at n = −10?

22. Based on a representation time N = 4, the DTFS harmonic function X[k] of a signal
x[n] has the following values.

X[−1] = 2 − j2, X[0] = 4, X[1] = 2 + j2, X [ 2] = 3


a) What is the Value of X[3]?

b) What is the Value of X[22]?

c) What is the average value of x[n]?


148 | Fourier Series and Transform

Multiple-Choice Questions
1. A CT periodic signal x(t) is represented in Fourier series representation as
a. x(t) = ∑ a e Ω
b. x(t) = ∑ a e Ω
c. x(t) = ∑ a e Ω
d. x(t) = ∑ a e Ω /k

2. If the function f(x) is even, then which of the following is zero?


a. a
b. b
c. a
d. nothing is zero

3. The spectrum of a continuous periodic rectangular signal is a


a. Discrete periodic sinc function
b. Continuous periodic sinc function
c. Discrete aperiodic sinc function
d. Continuous aperiodic sinc function

4. If the function f(x) is odd, then which of the only coefficient is present?
a. a
b. b
c. a
d. Everything is present

5. Find a if the function f(x) = x–x .


a. finite value
b. infinite value
c. zero
d. can’t be found

6. The Fourier series coefficients of a continuous periodic signal x(t) are a . Fourier series
coefficients of x(−t) are
a. a
b. −a
Signals and Systems | 149

c. a
d.

7. If the Fourier series coefficients of a signal x(t) are a , then


a. a =a ∗
b. a =a ∗
c. −a = a ∗
d. a = −a ∗

8. x(t) is a continuous periodic signal with period T, fundamental frequency Ω and


Fourier series coefficients a . The Fourier series coefficients of x(2t) are
a. a
b. a
c. 2a
d. a

9. The CTFT of a signum function is


a. 2jΩ
b. δ(Ω)
c. Ω

d. Ω

10. If the Fourier transform of ( ) is ( Ω), then Fourier transform of (− ) is


a. (− Ω)
b. − ( Ω)
c. − (− Ω)
d. ( Ω)

KNOW MORE
To delve deeper into the realm of Fourier series and Fourier transform is to embark on a
journey of profound mathematical elegance and practical utility. One can uncover the
secrets of convergence theorems, unravel the mysteries of orthogonality in function
150 | Fourier Series and Transform

spaces, and master the art of manipulating signals in both time and frequency domains.
Advanced analytical techniques open doors to solving complex differential equations,
paving the way for applications in fields as diverse as physics, engineering, and finance.
Moreover, the realm of Fourier transform beckons with promises of understanding the
very essence of signals, be they audio waves, images, or quantum phenomena. From fast
algorithms powering digital signal processors to cutting-edge applications in medical
imaging and quantum computing, Fourier analysis continues to shape our modern world.
As we journey forward, the exploration of Fourier series and transform promises not only
a deeper understanding of mathematics and science but also an endless stream of
possibilities for innovation and discovery.

REFERENCES AND SUGGESTED READINGS

1. Signals and Systems by Simon Haykin


2. Signals and Systems - Course (nptel.ac.in)

Dynamic QR Code for Further Reading


Signals and Systems | 151

4 Laplace Transform

UNIT SPECIFICS

Through this unit we have discussed the following aspects:


 What is Laplace transform, why it was developed?
 Review of the Laplace Transform for continuous-time signals and systems
 Poles and zeros of system functions and signals
 Laplace domain analysis of the signals
 The solution to differential equations, and system behavior.

RATIONALE

The unit on “Laplace Transform" provides students to understand the behavior of Continuous
and Discrete-time signals and systems in s-domain or Laplace domain. The Laplace transform
is one of the most important tools used for solving ODEs and specifically, PDEs as it converts
partial differentials to regular differentials.

Laplace transform can convert complex differential equations that describe the dynamic
behavior of a system into simpler algebraic equations that describe the frequency response of
a system

PRE-REQUISITES

1. Strong understanding of mathematics, including algebra, calculus, and complex numbers.


2. Familiarity with basic concepts in signals and systems, such as time-domain and periodic,
non-periodic signals.
3. Proficiency in solving ordinary differential equations and understanding linear algebra
concepts.
152 | Laplace Transform

UNIT OUTCOMES
List of outcomes of this unit is as follows:
U4-O1: Understand the need for bilateral and unilateral Laplace transform for continuous-
time signals and systems.
U4-O2: Understand the relationship between continuous-time Fourier transform and Laplace
transform.
U4-O3: Learn the properties of Laplace transform.
U3-O4: Learn the applications of Laplace transform in the analysis of CT LTI systems.
U3-O5: Learn to solve the differential equations using Laplace transform.

EXPECTED MAPPING WITH COURSE OUTCOMES


Unit-4
(1- Weak Correlation; 2- Medium correlation; 3- Strong Correlation)
Outcomes
CO-1 CO-2 CO-3 CO-4 CO-5 CO-6
U4-O1 3 - - - - -
U4-O2 - - 3 - - -
U4-O3 2 - - - - -
U4-O4 - 3 - - - -
U4-O5 3
Signals and Systems | 153

4.1 Introduction
In the 3rd chapter we have discussed the Fourier series and Fourier transform along with their
magnitude and frequency spectrum. In this chapter, the Laplace transform is discussed which
is used to transform a time signal to complex frequency domain and this complex frequency
domain is called as Laplace domain or s-domain. Laplace transformation was proposed by
Laplace in the year 1780; hence this transformation is called as Laplace transform. In time
domain the equations to represent a system are written in terms of differential equations
whereas in s-domain, the differential equations are transformed to algebraic equations for
easier analysis. In this chapter a brief discussion about Laplace transform, its properties and
applications for analysis of signals and systems are presented.

4.2 Definition of Complex Frequency


The complex frequency is represented as,
= + Ω
where, σ = Neper frequency in neper per second

Ω = Radian (or Real) frequency in radian per second


( Ω)
Let, ( ) = = (4.1)

Let us analyze the signal of Eq. (3.1) for various choice of and Ω.

When, = ,

x(t) = Ae = Ae( Ω)

= A(cosΩt + jsinΩt)= AcosΩt + jAsinΩt (4.2)

The real part of Eq. (3.2) represents a cosine signal and the imaginary part represents a
sinusoidal signal.

Real part= AcosΩt

Imaginary part= AsinΩt

When, Ω = ,
154 | Laplace Transform

( ) = = (4.3)

In Eq. (3.3), if is positive, signal will be an exponentially increasing.

In Eq. (3.3), if is negative, signal will be an exponentially decreasing.

4.2.1 Complex Frequency Plane (s-Plane)


The complex frequency is defined as,
Complex frequency, s = σ + jΩ
where, σ = Real part of s
Ω= Imaginary part of s
Real part σ and imaginary part Ω can take values from -∞ to +∞. They are represented on a
two dimensional complex plane along with horizontal axis and vertical axis as shown in fig
4.1 is called complex frequency plane or s-plane.

Fig 4.1 s-plane

4.2.2 Definition of Laplace Transform (LT)


A time domain signal ( ) can be transformed into s-domain by multiplying the signal by
and integrate from -∞ to +∞. The transformed signal is represented by ( ) and
the transformation is denoted by the letter L.
Laplace transform is represented as,
( ) =L{ ( )}
Let ( ) be a continuous time signal defined for all values of . Let ( ) be Laplace transform
of ( ) then the Laplace transform of ( ) is defined as,
L{ ( )}= ( ) = ∫ ( ) (4.4)
For causal input signal,
Signals and Systems | 155

L{ ( )}= ( ) = ∫ ( ) (4.5)

4.2.3 Definition of Inverse Laplace Transform (ILT)


The s-domain signal ( ) can be represented to time domain signal ( ) by using inverse
Laplace transform (ILT).
Laplace transform is represented as,

L-1= ( ) = ∫ Ω
( ) (4.6)
Input signal ( ) and transformed signal ( ) are called Laplace transform pair and are
expressed as,
( )⟺ ( )

4.3 Region of Convergence (RoC)


The Laplace transform of a signal is given by, ∫ ( ) . The values of for which the
given LT equation converges is called Region of Convergence (RoC). The RoC is
expressed for three types of cases given below:

Case I: Right Sided Signal


Let x(t) = e u(t), where > 0
=e where ≥ 0
Take Laplace transform on both sides,
Using definition of LT,

x(s) = x(t) e dt

= e u ( t) e dt

( )
= e e dt = e dt

e ( )
=
−(b + s)
156 | Laplace Transform

1
= e ( )
(
− b+s )
1
= [e − e ]
−(s + b)
1 1
= [ 0 − 1] =
−(s + b) s+b

X(s) Converges for σ = −b where s = σ +

Therefore, for causal signal ROC is the right side of pole at σ = −b as shown in following fig
4.2.

Fig 4.2 RoC for right sided signal


Case II: Left sided signal
Let x(t) = e u(t), where > 0, for ≤0
Take Laplace, transform on both sides,
Using definition of LT,

x(s) = x(t)e dt

= e u ( t) e dt

( )
e e dt = e dt
Signals and Systems | 157

( )
=
−(s + b)
1 ( )×
= e − e( )×
−(s + b)
Let = + ,
1 ( )
∴ X(S) = − −
( + )
( ) ×
1
=− +
+ +
Let = +
If + > 0 , > − , . ., = ∞
If + < 0 , < − . ., = 0
Hence, ( ) converges when σ < −b

Fig 4.3 RoC for left sided signal


1 0 1
∴ ( )= − + = −
+ + +
Therefore, for an anticausal signal, RoC is on the left of pole = − shown in figure.

Case III: Two sided signal


Let x(t) = e u(t) + e u(−t) where , , > 0 and > .
Using definition of Laplace Transform,

X(s) = x(t)e dt
158 | Laplace Transform

= [ ( )+ (− )]

= +

( ) ( )
= +

( ) ( )
= +
−( + ) −( + )
Here also,
When, s + a > 0, s > −a → e =0
s + a < 0, s < −a → e = ∞
s + b > 0, s > −b → e = ∞
s + b < 0, s < −b → e =0
1 1
∴ X(s) = −
s+a s+b
Therefore, for two sided signal, ROC includes all points on s-plane lies between poles –a to –b
as shown in figure.

Fig 4.4 RoC for two sided signal

Example 4.1
Determine the Laplace transform of the following continuous time signals & find their ROC
1) x (t) = A u (t)
Signals and Systems | 159

2) x (t) = t u (t)

3) x (t) = e u ( t)

4) x (t) = e u (−t)
||
5) x (t) = e

Solution:
1.Given, ( ) = ()

Using definition of LT,

X(s) = x(t)e dt

A u ( t) e dt

x(s) = Ae dt = A e dt

e A
=A = [e ]
−s −s

A A
∴ X(s) = [e − e ] = − [ 0 − 1]
s s
∴ X(s) = where for s > 0 the X(s) converges.

Fig 4.5 ROC of ( )= ( )


160 | Laplace Transform

∴ ROC is the Right half of S-plane

2. Given, ( ) = ( )

Using definition of LT,

X(s) = x ( t) e dt

= t u ( t) e dt

= t e dt

Using ∫ uv = u ∫ v − ∫[du ∫ v] rule,


d
X(s) = t e dt − (t) e dt
dt
=t − ∫ 1 dt

= t −

= [e − 0] − [e −e ]

=
When, > 0, ( ) converges and ROC lies to the right of line passing through σ = 0.

Fig 4.6 ROC of ( )= ( )


Signals and Systems | 161

3. Given , ( ) = ()

Using definition of LT,


x(s) = ∫ x ( t) e dt

= e u ( t) e dt

= e e dt

( )
( )
=∫ e dt = ( )

1 1
[e − e ] = = −
s+8 s+8
For, + 8 > 0, X (s) converges for > −8 and ROC lies to the right of line passing through
σ = −8

4. Given, ( ) = (− )

Using definition of LT,

X (s) = x(t)e dt

= e u (−t) e dt

( )
= e e dt = e dt

( )
( )
= = − e
( )

Here, for S < - ∞, X (s) converges


1
∴ x (s) = −
s+8
ROC contains all points in s − place on the left side of line passing through σ = −8
162 | Laplace Transform

Fig 4.7 ROC of ( ) = (− )

||
5. Given ( ) =

Using definition of LT,

x(s) = x(t) e dt

||
= e e dt

e e dt + e e dt

( ) ( )
e dt + e dt

( ) ( )
= + −
( ) ( )

=− [e − e ] − [e −e ]
1 1 8
X(s) = − + = −
s−4 s+4 s − 16

ROC lies between the points σ = −4 to σ = 4


Signals and Systems | 163

||
Fig 4.8 ROC of ( ) =

Example 4.2: Determine the Laplace Transform of the following signals.

1. x(t) = sin Ω t u(t)


2. x(t) = cos Ω t u(t)
3. x(t) = e sin Ω t u(t)
4. x(t) = e cos Ω t u(t)

Solution:
1. Given ( )= ( )= ; ≥

Using definition of Laplace Transform,


X(s) = ∫ x(t)e dt

=∫ sin Ω t u(t)e dt

= ∫ sin Ω t e dt

∅ ∅
Using formula, sin∅ =

X(s) = ∫ e dt
164 | Laplace Transform

= ∫ e −e e dt

= ∫ e e −e e dt

( ) ( )
= ∫ e −e dt

( ) ( )
= ( )
− ( )

= ( )
− ( )
− ( )
+ ( )

= 0−0+ −

= ( )( )

as, j = −1

∴L{sin Ω tu(t)} =

2. Given that ( )= ( )= ; ≥

Using definition of Laplace Transform,

X(s) = ∫ x(t)e dt
Signals and Systems | 165

=∫ cos Ω t u(t)e dt

= ∫ cos Ω t e dt

∅ ∅
Using formula, cos∅ =

X(s) = ∫ e dt

= ∫ e e +e e dt

( ) ( )
= ∫ e +e dt

( ) ( )
= ( )
+ ( )

= 0+0+ +

= ( )( )

3. Given that ( )= ( )

= e sin Ω t ; t ≥ 0
Using definition of Laplace Transform,
X(s) = ∫ x(t)e dt
=∫ e sin Ω t e dt
166 | Laplace Transform

1 ( ) ( )
= e dt
2j

1 e ( )t e ( )
= −
2j −(s + a − jΩ −(s + a + jΩ
Ω
=
(s + a) + Ω
Ω
∴ L{e sin Ω t u(t)} =
(s + a) + Ω
Given that x(t) = e cos Ω t u (t)
=e cos Ω t; t≥0
Using definitiion of laplace Trasnform.

x(s) = x(t) e dt

= e cosΩ t e dt

1 ( ) ( )
= e dt
2

1 e ( ) e ( )
= −
2 −(s + a − jΩ −(s + a + jΩ

= 0+0+ +
s+a
=
(s + a) + Ω
∴L{e cos Ω t u(t)} = ( )
Signals and Systems | 167

Table 4.1 Some Standard Pairs of Laplace Transform with ROC

( ) ( )

( ) 1 −

( ) 1 >0

( ) 1 >0

1 >0
( )
( − 1)!
( ) 1 >−
+
− (− ) 1 <−
+

( ) ! >0

( ) 1 >−
( + )

( ) ! >−
( + )

( ) >0
+

( ) >0
+

( ) >−
( + ) +

( ) + >−
( + ) +
168 | Laplace Transform

4.4 Properties of Laplace Transform


4.4.1 Scaling of amplitude
If amplitude is scaled in time domain by constant K then it’s Laplace transform is multiplied by
same constant
i.e. if L{x(t)} = X(s)
then,
L{Kx(t)} = KX(s)
Proof: Using definition of LT
X(s) = ∫ x(t)e d (4.7)

=∫ Kx(t)e dt

= K∫ x(t)e dt
= ( ) Using Eq. (4.7)

4.4.2 Linearity

This property states that weighted sum of two or more signals is equal to similar weighted sum
of individual’s Laplace transform.
i.e. if L{x (t)} = X (s)
L{x (t)} = X (s) , then
L{ax (t) + bx (t)} = aX (s) + bX (s)

Proof: Using definition of LT


X (s) = ∫ x (t)e dt (4.8)
X (s) = ∫ x (t)e dt (4.9)
L{ax (t) + bx (t)} = ∫ [ax (t) + bx (t)]e dt
= a∫ x ( t) e dt + b ∫ x ( t) e dt
= aX (s) + bX (s)

4.4.3 Time differentiation

If derivative is taken in time domain then it’s Laplace transform is SX(s) − x(0)
Signals and Systems | 169

i.e. if L{x(t)} = X(s), then


L x(t) = SX(s) − x(0) , Where x(0) is value of x(t) at t = 0

Proof: Using definition of LT


X(s) = ∫ x(t)e dt (4.10)

∴L x(t) = ∫ x(t)e dt
( )
=∫ e dt … for causal x(t)
Using ∫ uv = u ∫ v − ∫[du ∫ v]
L x(t) = [e x(t)] − ∫ −se x(t)dt
=e x(∞) − e x(0) + s ∫ x(t)e dt
= S ∫ x(t)e dt − x(0)
= SX(s) − x(0) Using Eq. (4.10)

4.4.4 Integration in time domain

If L{x(t)} = X(s), then


( ) [∫ ( ) ]|
L{∫ x(t)dt} = +
Proof: Using definition of LT
X(S) = ∫ x(t)e dt …for causal signal

L{∫ x(t)dt} = ∫ [Sx(t)dt] e dt


Using, ∫ uv = u ∫ v − ∫[du ∫ v]

L{∫ x(t)dt} = [∫ x(t)dt] − ∫ x(t) dt

= [∫ x(t)dt]| + ∫ x(t)e dt
( ) [∫ ( ) ]|
= +
170 | Laplace Transform

4.4.5 Shifting in frequency domain

If L{x(t)} = X(s), then


L{e± x(t)} = X(S ∓ a)
Proof : Using definition of LT,
X (S) = ∫ x(t)e dt
L{e± x(t)} = ∫ e± x(t)e dt
( ∓ )
=∫ x(t)e dt
= X(S ∓ a)

4.4.6 Shifting in time domain

If L {x + (t)} = X(s), then


L{x(t ± a)} = e± x(s)
Proof: Using definition of LT,
X(s) = ∫ x(t)e dt (4.11)

L{x(t ± a)} = x(t ± a)e dt

let t ∓ a = τ
∴ t = τ ∓ a, dt = dτ

( ∓ )
∴ L{x(t ± a)} = x(τ)e dτ

= x(t)e e± dτ

= e± x ( τ) e dτ

= e± X(s) Using Eq (4.11)

4.4.7 Differentiation in frequency

If L{x(t)} = X(s), then,


Signals and Systems | 171

d
L {tx(t)} = − X (s)
ds
Proof: Using definition of LT,
d
X(s) = x(t)e dt
ds

d
= x(t) e dt
ds

= x(t)(−t e ) dt

= [−t x(t)]e dt

= L{−t x (t)}

= −L{t x (t)}
d
∴ L{t x(t)} = − X(s)
ds
4.4.8 Time Scaling

If L{x (t)} = X(s), then


1 s
L{ x (at)} = X
|a| a
Proof: Using definition of LT,

X(s) = x(t) e dt

∴ L { x (at)} = x(at) e dt

let at = τ
τ dτ
t= , dt =
a a
172 | Laplace Transform


X(s) = x ( τ) e
a

1
= x(τ)e dτ
a
1 s
L{x(at)} = X (4.12)
a a
Above equation is true when a is positive for negative a,
L{x(at)} = − X (4.13)
∴ Combining Eq. (4.12)& Eq. (4.13)
1 s
L{x(at)} = X
|a| a
4.4.9 Initial Value theorem

If L{x(t)} = X(s), then


Initial Value of signal x (t) at t=0 is,
x(0) = lim ( ) = lim ( )
→ →
4.4.10 Final Value theorem

If L {x(t)} = X(s). then


Final value of signal x(t)at t = ∞ is,
x(∞) = lim ( ) = lim ( )
→ →
4.4.11 Convolution Property

The convolution theorem of LT says that the convolution of two signals in time domain is
equivalent to multiplication of their Laplace transforms in S domain.
i.e. If L {x, (t)} = X (s)
L {x (t)} = X (s) then,
L{x (t) ∗ x (t)} = X (s)X (s)
Signals and Systems | 173

4.5 Poles and Zeros of System Functions and Signals


The Laplace transform is represented in terms of rational, i.e. it is a ratio of polynomials in the
complex variable s.
N (s)
X (s) 
D (s)

Where N and D are the numerator and denominator polynomials respectively.

In fact, X(s) will be rational whenever x(t) is a linear combination of real or complex
exponentials. Rational transforms also arise when we consider LTI systems specified in
terms of linear, constant coefficient differential equations.

We can mark the roots of N and D in the s-plane along with the ROC

4.5.1 Poles and Zeros:


The roots of N(s) are known as the zeros. For these values of s, X(s) is zero.
The roots of D(s) are known as the poles. For these values of s, X(s) is infinite, the Region of
Convergence for the Laplace transform cannot contain any poles, because the corresponding
integral is infinite
The set of poles and zeros completely characterise X(s) to within a scale factor (+ ROC for
Laplace transform)
i ( s  zi )
X (s) 
 j (s  p j )
The graphical representation of X(s) through its poles and zeros in the s-plane is referred to as
the pole-zero plot of X(s).

Example 4.3: Poles and Zeros


Consider the signal below signal and arrange in pole zero form showing the ROC.
4 t 1
x (t )   (t )  e u (t )  e 2 t u (t )
3 3
Solution: By linearity we can evaluate the second and third terms
The Laplace transform of the impulse function is:

L { ( t )}    ( t ) e  st dt  1

174 | Laplace Transform

which is valid for any s. Therefore,


4 1 1 1
X (s)  1  
3 s 1 3 s  2
( s  1) 2
 , Re{s}  2
( s  1)(s  2)
That means ROC lies on the RHS of right most pole which is 2.

4.6 ROC Properties for Laplace Transform:


Depending upon the location of poles and the criteria of convergence the following properties
are given.

Property-1: The ROC of X(s) consists of strips parallel to the jW - axis in the s-plane.

Property-2: If x(t) is of finite duration and is absolutely integrable, then the ROC is the entire
s- plane.

Property-3: If x(t) is right sided, and if the line passing through Re(s) = s0 is in ROC, then all
values of s for which Re(s) > s0 will also be in ROC.

Property-4: If x(t) is left sided, and if the line passing through Re(s) = s0 is in ROC, then all
values of s for which Re(s) < s0 will also be in ROC.

Property-5: If x(t) is two sided, and if the line passing through Re(s) = s0 is in ROC, then the
ROC will consists of a strip in the s-plane that includes the line passing through Re(s) =
s0 .

Property-6: If X(s) is rational, (where X(s) is Laplace transform of x(t)), then its ROC is
bounded by poles or extends to infinity.

Property-7: If X(s) is rational, (where X(s) is Laplace transform of x(t)), then ROC does not
include any poles of X(s).

Property-8: If X(s) is rational, (where X(s) is Laplace transform of x(t)), and if x(t) is right
sided, then ROC is the region in s-plane to the right of the rightmost pole.
Signals and Systems | 175

Property-9: If X(s) is rational, (where X(s) is Laplace transform of x(t)), and if x(t) is left sided,
then ROC is the region in s-plane to the left of the leftmost pole.

4.7 Inverse Laplace Transform by Partial Fraction Expansion Method

Let Laplace transform of x(t) be X(s). The s-domain signal X(s) will be a ratio of two
polynomials in s (i.e., rational function of s). The roots of the denominator polynomial are called
poles. The roots of numerator polynomials are called zeros. In signals and systems, three
different types of s-domain signals are encountered. They are, with separate poles, with multiple
poles, with complex conjugate poles.

The inverse Laplace transform (ILT) by partial fraction expansion method of all the three cases
are explained with an example.

Type 1: When s-Domain Signal X(s) has Distinct Poles

Let ( ) = ( )( )

Using partial fraction expansion, above equation can be written as,


A1 A2 A3
( )= + +
s (s + p1) (s + p2)

The residues A1, A2, A3 will be found for s=0, s=-p1, s=-p2.

Type 2: When s-Domain Signal X(s) has Multiple Poles

Let ( ) = ( )( )

Using partial fraction expansion, above equation can be written as,


A1 A2 A3 A4
( )= + + +
s (s + p1) (s + p2) (s + p2)
176 | Laplace Transform

The residues A1, A2, A4, will be found for s=0, s=-p1, s=-p2.

For residue A3,

d
3= [ ( )(s + p2) ]
ds

Type 3: When s-Domain Signal X(s) has Complex Conjugate Poles

Let ( ) = ( )( )

Using partial fraction expansion, above equation can be written as,


A1 A2s + A3
( )= +
(s + p1) (s + + )

The residues A2 and A3 are solved by cross multiplying the above equation and then equating
the
coefficients of like power of s.

Examples 4.4: On Inverse Laplace Transform


1. Find the ILT of given signal

1
X (s)  Re{ s}   2
( s  1)( s  2 )

Solution: Like the inverse Fourier transform, expand as partial fractions


1 A B 1 1
X ( s)     
(s  1)(s  2) (s  1) (s  2) (s  1) (s  2)
Pole-zero plots and ROC for combined & individual terms

L 1
 e  t u ( t )  , Re{s}  1
s 1
L 1
 e  2t u (t )  , Re{s}  2
s2
L 1
x(t )  (e t  e  2t )u (t )  , Re{s}  2
( s  1)( s  2)
Signals and Systems | 177

2. Find the ILT of ( ) = ( if the ROC is


)( )
i. −2 > { } > −4
ii. { } < −4
iii. { } > −2

Solution: Given ( ) = ( = +
)( )

Residue A1 is,
4
1 = ( )(s + 2)| = × (s + 2) =2
(s + 2)(s + 4)
Residue A2 is,
4
2 = ( )(s + 4)| = × (s + 4) = −2
(s + 2)(s + 4)
2 2
( )= −
s+2 s+4
i. For −2 > { } > −4
Given that ROC lies between lines passing through s = 2 to s = 4. Hence, x(t) will be two sided
signal.
( )=− (− ) − ( )

ii. { } < −4
Given that ROC is left of the line passing through s = 4. Hence x(t) will be anticausal signal
( )=− (− ) + (− )
178 | Laplace Transform

iii. { } > −2
Given that ROC is right of the line passing through s = – 2. Hence x(t) will be causal signal.
( )= ( )− ( )

4.8 Laplace domain analysis

4.8.1 Transfer Function of LTI Continuous Time System


In general, the input-output relation of a LTI (Linear Time Invariant) continuous time system
is represented by the constant coefficient differential equation shown below,

( )+ ( )+ ( )+⋯+ ( )+ ( )= ( )+

( )+ ( ) +⋯+ ( )+ ( ) (4.14)

Where n is order of system and ≤

On taking LT of above equation and taking zero initial conditions,


( ) ⋯
= (4.15)
( ) ⋯

The transfer function of a continuous time system is defined as the ratio of Laplace transform
of output and Laplace transform of input. Hence the equation (4.15) is the transfer function of
an LTI continuous time system.

4.8.2 Impulse Response and Transfer Function


Let, ( )=input to LTI system
( )=output for input ( )
ℎ( )=Impulse response
Now, the response y(t) of the continuous time system is given by convolution of input and
impulse
response. Recall the formula of convolution,
( ) = ( ) ∗ ℎ( ) = ∫ ( )ℎ( − ) (4.16)
Taking LT,
( )= ( ) ( ) (4.17)
Signals and Systems | 179

( ) ( − )( − )( − ) … … ( − )
( )= =
( ) ( − )( − )( − ) … … ( − )
The transfer function of LTI continuous time system is also given by Laplace transform of the
impulse response.

4.9 Solving Differential Equations by Using Laplace Transform


Given a differential equation of LTI system as shown by Eq. (4.17), the system function can be
written as,
( )
( )= (4.18)
( )
Which is obtained by taking the Laplace transform of the differential equation. The inverse
Laplace transform of system function ( ) can be obtained to get the impulse response of
the system ℎ( ). For the stable system, i.e., if ROC includes the Ω axis, then by
substituting = Ω , the frequency response of the system ( Ω) can be obtained.

Unit Summary
The Laplace transform, a powerful mathematical tool, transcends mere computation to offer
profound insights into the behavior of dynamic systems across various disciplines. Beginning
with its historical roots and fundamental properties, the journey into Laplace transform theory
elucidates its superiority over conventional methods in solving differential equations, thanks to
its ability to convert complex time-domain problems into simpler algebraic ones. Techniques
such as partial fraction decomposition and theorems for initial and final values amplify its
practicality in diverse applications, from control systems analysis and electrical circuit design
to signal processing and probability theory. Advanced topics unveil its versatility in handling
generalized functions and partial differential equations, while its integration into modern
engineering workflows underscores its indispensable role in modeling and analysis. As we
reflect on its enduring impact, it becomes evident that the Laplace transform not only
revolutionizes problem-solving but also fosters a deeper understanding of dynamic systems,
paving the way for innovation and discovery across scientific and engineering frontiers.
180 | Laplace Transform

Exercise
1. Using final value theorem, find the final value of the signal x(t) given
a. X(s) = ( )

b. X(s) =

c. X(s) =

d. X(s) =

e. X(s) =

2. Find the inverse Laplace transform of the following:

a. X (s) = ( )( )
; ROC: −2 < Re{s} < 1

b. X (s) = ; −7 < Re{s} < 4

c. X (s) = ; ROC: −15 < Re{s} < 15

3. Using convolution property, find the following:

a. y (t) = e u(t) ∗ u(t)

b.y (t) = u(t) ∗ u(t − 5)

c. y (t) = δ(t) ∗ e u(t)

d.y (t) = e u ( t) ∗ e u(t)


Signals and Systems | 181

e. y (t) = e u(t) ∗ e u(−t)

4. Sketch the pole-zero plot and ROC (if exists) for the following signals:

a. x (t) = e u(t)

b. x (t) = e u(−t)

c. x (t) = te u(t)

│ │
d. x (t) = 3e

e. x (t) = e( ⁄ )
u(−t)

5. Using Laplace transform, solve the following differential equations:

a. y(t) + 5y(t) = 5; y(0 ) = 1

b. y(t) + 7 y(t) + 12y(t) = 3

c. x(t) + 3 x(t) + 2x(t) = 4; x(0 ) = 1 = x΄(0 )

d. y(t) + 11 y(t) + 30y(t) = e u(t); y(0 ) = y΄(0 ) = y΄΄(0 ) = 0

e. y(t) + 4 y(t) + 4y(t) = 1; y(0 ) = 1, y΄(0 ) = −1, y΄΄(0 ) = 0

6. Find the Laplace transform of the signal


( )=− (− )
7. Find the Laplace transform of the signal
( )= │ │
182 | Laplace Transform

8. Find the Laplace transform of the signal and its corresponding ROC.
( )= [ ( ) − ( − 4)]
9. Find the Laplace transform of ( ) = ( )+ ( )
10. Find the Laplace transform of the signal ( ) = ( )+ (− )

11. Find the Laplace transform of ( ) = ( ) − ( )+ ( )

12. Find the Laplace transform of ( ) = ( )+ (− − 1)


13. Find the Laplace transform of ( ) = ( )
14. Find the Laplace transform of ( ) = ( )
15. Find the Laplace transform of ( ) = ( ).
16. Find the Laplace transform of = (2 ) by using the time-scaling property

17. Find the Laplace transform of ( ) = ( )∗ ( ), where ( )= ( ), ( ) = ( ).

18. Find the Laplace transform of ( ) = − ( ). Using the differentiation in time domain
property.
19. Find the Laplace transform of the output of an LTI system ( ) = 3 ( ) by using the
differentiation in s-domain property. Given
+2
( )=
+4 +4

20. Find the initial value and final value of ( ) with Laplace transform
2
( )=
( + 3 + 5)

21. Find the inverse Laplace transform of


( )= , Re { } > −1

22. Find the inverse Laplace transform of


2 +1
( )=
+3
For (a) ROC: Re{ } > −3.
Signals and Systems | 183

23. Find the inverse Laplace transform of

1
( )=
( + 5)( − 3)
For the following ROCs:
( )−5 < { } < 3
(b) { } > 3
(c) { } < −5
24. Find the inverse Laplace transform of
( )
( )= ,
( )
ROC: { } > −3

25. Find the inverse Laplace transform of


(a) ( ) = ( )
; ROC: { } > −2

(b) ( ) = ; ROC: { } > −2

(c) ( ) = ( )
; ROC: { } > −1

Multiple-Choice Questions
1. For a causal signal x(t), the ROC of X(s) is
a) Right-half of s-plane
b) Left-half of s-plane
c) Entire s-plane
d) jΩ-axis

2. The Laplace transform of x(t) = 2δ(t) is


a) 2
b) b) 2⁄s
c) c) 2s
d) d) s

3. The Laplace transform of x(t) = e u(t) is


a)
184 | Laplace Transform

b)
c) ( )

d) ( )

4. ROC of X(s) of e u(−t) is

a) Re{s} > 4
b) Re{s} < −4
c) Re{s} < 4
d) Re{s} > −4

5. If X(s) = then x(0 ) is


a) 0
b) ∞
c) 1
d) 2
6. Given (t) ↔ X(s) , then −4 x(t) ↔ is
a) 4sX(s)
b) −4sX(s)
c) X(s)
d) − X(s)

7. The Laplace transform of δ(t) ∗ u(t) is


a)
b) s
c)
d)
8. The Laplace transform of x(2t) is
a) X
b) 2X(2s)
c) X(2s)
Signals and Systems | 185

d) 2 ⁄

9. The inverse Laplace transform of X(s) = ( )( )


is
a) 5(3e − e )u(t)
b) 5(3e − e )u(t)
c) 5(3e − e )u(t)
d) 5(3e − e )u(−t)

10. The Laplace transform of x(t) = e u(t) is


a) ; Re{s} > 1
b) ; Re{s} > 1
c) ; Re{s} < 1
d) ; Re{s} < 1

11. The Laplace transform of δ(4t) is


a) 1⁄4s
b) 1⁄4
c) s⁄4
d) 4s
12. The output of an LTI system y(t) + y(T) = x(t) initially at rest, for given x(t) = δ(t) is
a) e u(t)
b) u(t)
c) δ(t)
d) δ(t − 1)

13. If x(t) is a left-sided signal, then its ROC is


a) entire s-plane
b) left-half of s-plane
c) right-half of s-plane
d) entire jΩ- axis

14. A causal LTI system has a transfer fraction ( ), whose ROC will be
a) Right-sided in the s-plane
186 | Laplace Transform

b) Left-sided in the s-plane


c) Entire s-plane
d) s = jΩ- axis

15. The ROC of a stable LTI system will


a) include the jΩ- axis
b) be Re{s} > 1
c) be Re{s} + 1
d) be Re{s} > 0

KNOW MORE
The Laplace transform is a mathematical technique that is used to simplify solving differential
equations, particularly those with initial conditions. It transforms functions of time into
functions of complex frequency. This transformation makes it easier to solve linear differential
equations by turning them into algebraic equations. Understanding the Laplace transform's
definition and its properties is crucial. This includes linearity, time-shifting, scaling, and
differentiation properties. Additionally, understanding the region of convergence is important
for ensuring convergence of the transformed function. Knowing how to convert a Laplace-
transformed function back into the time domain is essential. Techniques such as partial fraction
decomposition, contour integration, and the use of tables of Laplace transforms are commonly
employed. Laplace transform is extensively used in solving ordinary and partial differential
equations, including those arising in engineering, physics, and other fields. It simplifies solving
differential equations with initial conditions, boundary conditions, and forcing functions.
Laplace transform finds applications in signal processing for analyzing and manipulating
continuous-time signals. It aids in filtering, modulation, demodulation, and system
identification tasks, contributing to various fields such as telecommunications, audio
processing, and medical imaging. We have to stay updated on recent advancements and ongoing
research in Laplace transform theory, applications, and computational methods. Investigating
emerging trends, challenges, and potential interdisciplinary collaborations in areas such as data
science, machine learning, and quantum computing along with Laplace transform is needed.
Signals and Systems | 187

REFERENCES AND SUGGESTED READINGS


1. Signals and Systems by Simon Haykin
2. Signals and Systems by Ganesh Rao
3. Signals and Systems - Course (nptel.ac.in)

Dynamic QR Code for Further Reading


188 | Z - Transform

5 Z - Transform

UNIT SPECIFICS
Through this unit we have discussed the following aspects:
 What is z- Transform, why it was developed?
 z-Transform and ROC of finite and infinite duration sequences
 Relation between Discrete Time Fourier Transform (DTFT) and z-Transform
 Properties of z-Transform.
 Inverse z-Transform and methods of analysis

RATIONALE

The unit on “z-Transform" provides students to understand the relationship between DTFT and
z-Transform. The students will understand the conversion of a discrete-time signal, which is a
sequence of real or complex numbers, into a complex valued frequency-domain (the z-domain
or z-plane) representation.

The unit focuses on z-Transform and ROC of finite and infinite duration sequences along with
the properties. The students can analyse the behaviour of the linear time-invariant (LTI) system
using the Z transform.

PRE-REQUISITES

1. Strong understanding of mathematics, including algebra, calculus, and complex numbers.


Signals and Systems| 189

2. Familiarity with basic concepts in signals and systems, such as periodic, non-periodic
signals, unilateral and bilateral sequences.
3. Proficiency in solving ordinary differential equations and understanding linear algebra
concepts.

UNIT OUTCOMES
List of outcomes of this unit is as follows:
U5-O1: Be able to understand the need for bilateral and unilateral z-transforms to analyze
discrete-time (DT) signals and systems.
U5-O2: Be able to understand the relationship between DT Fourier Transform (DTFT) and z-
transform.
U5-O3: Be able to learn the properties of z-transform.
U5-O4: Be able to learn the applications of bilateral and unilateral z-transform.

EXPECTED MAPPING WITH COURSE OUTCOMES


Unit-5
(1- Weak Correlation; 2- Medium correlation; 3- Strong Correlation)
Outcomes
CO-1 CO-2 CO-3 CO-4 CO-5 CO-6
U5-O1 - - - 2 3 -
U5-O2 - - - 2 3 -
U5-O3 - - - - 3 -
U5-O4 - - - - 3 -
190 | Z - Transform

5.1 Introduction
The Z-transform is a mathematical technique used in signal processing and control theory to
analyze and process discrete-time signals and systems. It is the discrete-time counterpart of the
Laplace transform, which is used for continuous-time signals and systems.

The Z-transform is beneficial for several reasons:


Frequency Analysis: It allows for the analysis of signals and systems in the frequency domain.
By taking the Z-transform of a discrete signal or system, you can analyze its frequency
components and behavior.
System Representation: The Z-transform can be used to represent and analyze discrete-time
linear time-invariant (LTI) systems, which are essential in fields like control theory and digital
signal processing.
Transfer Functions: It helps in finding the transfer function of a system, which relates the input
and output in the Z-domain. This is useful for designing and analyzing control systems and
filters.
Stability Analysis: The Z-transform can be used to analyze the stability of discrete-time
systems. It's crucial in control theory to ensure that a system doesn't become unstable over time.
Differential Equations: It can be used to solve linear difference equations, which often arise in
discrete-time systems modeling.
The Z-transform is a powerful tool for the analysis and design of discrete-time systems and is
widely used in various engineering and scientific fields. It helps in simplifying the analysis of
complex systems and understanding their behavior in both the time and frequency domains.

5.2 Need of Z-transform

There are some signals which are not absolutely summable and there Fourier transform does
not exist. Instead of taking transform of x[n], we can do a little change in the signal so that the
signal becomes absolutely summable and then apply the transform. Let us multiply the signal
x[n] with r-n. r-n is a slow decaying exponential signal. This transform is named as Z-transform.
. . { ( )} = ( ) = ∑ { [ ] }
=∑ [ ]

= [ ]
Signals and Systems| 191

Substituting into z = r ejw the above equation, the difficulty is resolved by generalizing the DTFT
of the signal x(n) can be expressed as sum of complex exponential, zn.
( )=∑ [ ] (5.1)
The Z-transform reduces to DTFT for the value of r = 1.
Inverse Z-transform is given by,
[ ]= ∮ ( )

It is denoted as,
.
( ) ( )
Z. T. { ( )} = ( )
I. Z. T. { ( )} = [ ]

5.3 Types of Z-transforms


There are mainly two types of Z-transforms: the unilateral Z-transform and the bilateral Z-
transform. Both are mathematical techniques used in the analysis and design of discrete-time
systems, but they differ in their definitions and applications.

5.3.1 Unilateral Z-Transform:


Definition: The unilateral Z-transform is defined for signals that are causal, meaning they start
at a finite time and continue indefinitely.
Formula: The unilateral Z-transform of a discrete-time signal x[n] is given by:
( )=∑ [ ] (5.2)
Application: It is commonly used when dealing with signals that are causal and have a starting
point in time.

5.3.2 Bilateral Z-Transform:


Definition: The bilateral Z-transform is defined for signals that may be non-causal, extending
over both past and future time indices.
Formula: The bilateral Z-transform of a discrete-time signal x[n] is given by:
( )=∑ [ ] (5.3)
Application: It is more general and can be applied to a wider range of signals, including non-
causal ones. However, it may not converge for all signals, and caution is needed in its
application.
192 | Z - Transform

In practical applications, the unilateral Z-transform is often more commonly used, especially in
the context of causal systems and signals. The choice between unilateral and bilateral Z-
transform depends on the nature of the problem and the characteristics of the signals involved.

5.4 The Z-plane


The Z-plane is a graphical representation used in the analysis and design of discrete-time
systems, particularly in the context of Z-transforms as shown in the figure below. The Z-plane
is a complex plane where the Z-transform is visualized and analyzed. In the Z-plane, complex
numbers are represented as points, and each point corresponds to a specific frequency response
or characteristic of a discrete-time system.

Fig 5.1: Z-plane


The Z-transform is a mathematical transformation that converts a discrete-time signal,
represented by a sequence x[n], into a complex function of a complex variable z. The Z-plane
provides a way to understand and interpret the properties of this complex function.

The point at the origin (z = 0) represents the DC (zero-frequency) component of the signal. The
unit circle in the Z-plane corresponds to the unit circle in the complex plane, where the
magnitude of z is equal to 1. Points on the unit circle are associated with frequencies equal to
the sampling frequency. The unit circle is particularly important for analyzing frequency
response characteristics.

Poles and zeros of the Z-transform are represented as points in the Z-plane. Zeros are locations
where the Z-transform is zero, and poles are locations where the Z-transform becomes infinite.
The distribution of poles and zeros in the Z-plane provides insights into the stability and
frequency response of a discrete-time system.
Signals and Systems| 193

Fig 5.2: Poles and zeros in the Z-plane

5.4.1 Poles:
Poles are the values of Z for which the transfer function becomes infinite (the denominator of
the transfer function becomes zero). The poles are denoted with the cross sign in the above
figure. They represent the natural frequencies of the system and provide information about the
system's stability and response to input signals.
The locations of the poles in the Z-plane determine how the system responds to different
frequencies. Poles closer to the origin (Z = 0) correspond to faster decaying modes, while poles
farther from the origin may represent dominant resonant frequencies.
The system is considered stable if all poles are inside the unit circle in the Z-plane. If any poles
are outside the unit circle, the system is unstable.

5.4.2 Zeros:
Zeros are the values of Z for which the transfer function becomes zero (the numerator of the
transfer function becomes zero). The zeros are denoted with a small circle sign in the above
figure. They represent the frequencies at which the system's response is zero, indicating points
in the Z-plane where the system does not respond to certain input frequencies.
Zeros can affect the system's frequency response, leading to resonant peaks or notches in the
frequency domain. The location of zeros in the Z-plane indicates the frequencies at which the
system has no response.
In summary, poles and zeros in the Z-transform provide valuable information about the
frequency response and stability of a discrete-time system. By analyzing the distribution of
poles and zeros in the Z-plane, one can understand how a system responds to different
frequencies and make design decisions to achieve desired system performance.
The Z-plane is also associated with the concept of the Region of Convergence (ROC), which is
the set of values of z for which the Z-transform converges. The ROC is often specified to ensure
the convergence of the Z-transform. We will discuss about this later in this chapter. The
194 | Z - Transform

location of poles in the Z-plane is crucial for stability analysis. For a discrete-time system to be
stable, all poles must lie inside the unit circle.

The Z-plane is a valuable tool for visualizing and analyzing the properties of discrete-time
systems in the context of Z-transforms. It provides insights into the frequency response,
stability, and convergence properties of these systems, making it an essential concept in the
field of digital signal processing and control system analysis.

5.5 Region of Convergence (ROC) for Z-Transform


We have already seen that the Z-Transform of x[n] is the Fourier Transform of x[n] r-n, that is,
( )=∑ { [ ] } (5.4)
Hence, is guaranteed to converge if x[n] r-n is absolutely summable. So, if x[n] r-n is absolutely
assumable then | ( )| < ∞.
For a given Z-transform X(z), the Region of Convergence (ROC) is the set of values of the
complex variable z for which the Z-transform converges. The ROC is crucial because it
determines the set of values for which the Z-transform is well-defined.

Let us see few signals and their Z-Transform.

Example 5.1 Determine the Z.T. and ROC of the following finite duration signals:
(a) [ ] = {1,2,3, −1,0,1}, −2 ≤ ≤ 3
(b) [ ] = {0, 0 ,1,2,1}, 0 ≤ ≤ 4
(c) [ ] = {1, 2, 3, −1,0, }, −4 ≤ ≤ 0

Solution:
(a) [ ] = {1, 2, , −1,0,1}, −2 ≤ ≤3

By definition,

( )= [ ]

The given signal is of finite duration 2 < n < 3.


Signals and Systems| 195

= [ ]

= (−2) + (−1) + (0) + (1) + (2) + (3)


Substituting values of x[n] we get,
( )=1 +2 +3− +
ROC for X(z) is entire z-plane except z = 0 and z = .

(b) [ ] = { , 0,1,2,1}, 0 ≤ ≤4
By definition,

( )= [ ]

The given signal is of finite duration 0  n  4.


Change the limits of summation,

( )= [ ]

= (0) + (1) + (2) + (3) + (4)


( ) =0+0 +1 +2 +1
= +2 +
∴ ( ) − = 0.

(c) [ ] = {1,2,3, −1, }, −4 ≤ ≤ 0.


By definition,

( )= [ ]

The signal x[n] has non-zero values of n = 4, 3, 2,  1, 0. i.e. 4  n  0.

( )= [ ]

( )
= (−4) + (−3) ( ) + (−2) ( )
+
(−1) ( ) + (0)
=1 + 2 + 3 + (−1) + 0
=1 + 2 + 3 + (−1) + 0
196 | Z - Transform

( )= +3 +2 +
The ROC for X(z) is entire z-plane except z = .

Example 5.2
Consider the Signal x[n] = anu(n). Find Z-Transform of x[n] Where |a| < 1.
Given : x[n] = an u(n), |a| < 1
Solution: The signal x[n] is right sided i.e. causal and infinite duration.
Z.T. of x[n] can be found by using equation below,

( )= [ ]

= ( )

We know,
1, ≥0
( )=
0, . .

( )= .1.

= ( . )

We know,
1
. ( )= , | |<1
1−
Applying this to X(z) we get,
1
( )= ( ) = =
1− −

| − | > 0, | | > | | It will be the ROC.


The zeros can be obtained by equating numerator to zero and poles by equation denominator to
zero.
z  a =0  z = a
One pole is located at z = a.
Signals and Systems| 197

Fig. 5.3 x(n) = anu(n)

Fig. 5.4 Pole-zero plot and ROC |z| > |a|

Observe here that the signal x[n] is causal, therefore the ROC of X(z) is outside the circle having
radius |a|.

Example 5.3: Determine the Z.T. of the another signal given below
[ ]=− [ − − 1] , | | < 1
Solution: It is an infinite duration left sided, anticausal signal.
The Z.T. of the x[n] can be evaluated as,

( )= [ ]

Substituting value of x[n] in X(z) we get,


198 | Z - Transform

( )= − (− − 1)

Here, u(n  1) is 1 only when n  [ ,  1].

∴ ( )=− ( ( − 1)

=− . 1.

=− ( )

Replacing n by n we get,

=− ( )

As the lower limit starts from 1, we cannot directly find the answer. We will do some
adjustments so that it will start from 0.

=− ( ) −1

=1−

1
=1− =
1− −

This series converges,


| |<1
| |<⌈ ⌉

∴ ( )= ; | |<| |

[ ]=− (− − 1) is a non-causal infinite duration signal.
The ROC is inside the circle of radius |a| as shown below.
Signals and Systems| 199

Fig. 5.5 x[n]= an u( n  1)

Fig. 5.6 Pole-zero plot and ROC |z| < |a|

It is an interesting thing to observe that the signals: [ ] = ( ), | | 1 [ ]=


− ( − − 1)are having same value of ( ) = Here the ROC is playing an important
role of differentiating the two signals. For causal signal the ROC of X(z) is outside the circle
having radial |a| and anti-causal signal the ROC of X(z) is inside the circle having radial |a|.
These two examples explain the importance of specifying ROC.

5.6 Properties of ROC


The ROC determines the validity and convergence of the Z-transform expression. If a specific
Z value is within the ROC, the Z-transform is well-defined and finite for that value. Outside the
ROC, the Z-transform might not exist or could be infinite, leading to undefined results. In this
section, let us go through the properties of ROC.
1. The ROC of X(z) consists of a ring in the z-plane centered about the origin.
2. If X(z) is rational, then The ROC must not contain any pole because at pole location the
Z.T. becomes infinity.
200 | Z - Transform

3. If x[n] is of finite duration, then the ROC is the entire z-plane, except possibility z = 0 and/or
z = .
4. If x[n] is right sided and of infinite duration sequence, then ROC is the region of the z-plane
outside the outermost pole.
5. If x[n] is a left sided and of infinite duration sequence, then ROC is the region of the z-plane
inside the innermost pole.
6. If x[n] is two sided and if the circle |z| = r0 is the ROC, then the ROC will consist of a ring
in the z-plane that includes the circle |z| = r0.
7. If the z-transform X(z) of x[n], is rational, then its ROC is bounded by poles or extends to
infinity.
8. If the Z.T. of X(z) of x[n] is rational and if x[n] is right sided, then ROC is the region in the
z-plane outside the outermost pole. i.e. outside the circle of radius equal to the largest magnitude
of the pole of X(z).
If signal x[n] is causal then ROC includes z = .
9. If the Z.T. X(z) of x[n] is rational and if x[n] is left sided, then ROC is the region in the z-
plane inside the innermost nonzero pole.
i.e. Inside the circle of radius equal to the smallest magnitude including z = 0 in particular, if
x[n] is anti-causal then the ROC also includes z = 0.

5.7 Properties of the Z-transform


The Z-transform is a mathematical tool used in the analysis and processing of discrete-time
signals and systems. It has several important properties that make it a valuable tool in various
engineering and scientific fields. Here are some of the key properties of the Z-transform:

5.7.1 Linearity:
The Z-transform is a linear operation, which means it satisfies the superposition principle. If
you have a linear combination of signals, you can compute the Z-transform of each signal
separately and then sum them to find the Z-transform of the combined signal.
Linearity property states that,
.
( ) ( ) ℎ ∶
.
( ) ( ) ℎ ∶

.
Then, y(n) = ( )+ ( ) ( )=a ( )+ ( ) with ROC  (5.5)
Proof : The Z.T. of Y[n] is given by,
Signals and Systems| 201

( )= ( ).

= { ( )+ ( )}

= ( ). + ( )

( ) ( )
( )= ( )+ ( ) ℎ 
Hence proved.
The linearity property can be generalized for any number of arbitrary signals.
It implies that the Z.T. of a linear combination of signals is the same as linear combination of
their Z.T.

5.7.2 Time Shifting:


If you delay (advance) a discrete signal in the time domain, the Z-transform of the shifted signal
is related to the original Z-transform by multiplying it by a power of Z.
.
If [ ] ( ) ℎ ∶
.
then, ( ) = ( − ) ( )= ( ) ∶ ⋂{0 < | | < ∞} (5.6)
Proof : Let, y[n] = x[n  n0]
By the definition of the Z.T. we have,

[ ]= [ ]

= ( − )

Let, − = , + =
[ ]= ∑ [ ] ( )
(5.7)
=∑ [ ] (5.8)
= ∑ ( ) (5.9)
( )= ( ) ∶ ∩ {0 < | | < ∞}
Hence proved.
202 | Z - Transform

5.7.3 Time Reversal:


The Z-transform of a time-reversed signal is related to the complex conjugate of the Z-transform
of the original signal.
Time reversal property states that,
.
if [ ] ( ) with ROC : R and ( )= (− )
then, ( )= ( ) ℎ (5.10)
Proof : We have, ( ) = (− )
By the definition of Z.T.

( )= ( )

=∑ (− ) (5.11)
Substituting − =

( )= [ ]

=∑ [ ][ ] (5.12)
( )= ( )

then, ( ) = ( ) ℎ .
which is same as R.H.S. Hence proved.

5.7.4 Scaling in z-domain:


If you multiply a discrete signal by a constant in the time domain, the Z-transform of the scaled
signal is obtained by multiplying the original Z-transform by the same constant.
.
Let, [ ] ↔ ( ) ℎ .
x[n] (z) with ROC R.
.
then, [ ] = . [ ] ( )= ℎ ∶ | | . (5.13)
Proof : Let . . [ ]= . [ ] [ ] ,
Z.T. of [ ] = . [ ] is,

[ ]= ( ) = . [ ]
Signals and Systems| 203

= [ ].

( )= ℎ ∶ | | .

then, ( ) = ℎ ∶ | | .
|| | R is scaled version of R.
z
If X(z) has a pole or zero at z = b then X a has a pole or zero at z = a  b.
 
If b is +ve number, the scaling can be interpreted as shrinking or expanding of the z-plane.

5.7.5 Time scaling:


if n is multiple of m;
, [ ]=
0 if n is not multiple of m ;

xm[n] is obtained from x[n] by inserting (m  1) zeros between successive sample values of
x(m).
if, ( ) = X (z)
z
then, y(n) = Xm(n)  X(zm) = Y(z) with ROC : R1/m. (5.14)
Proof : We know that,

( )= [ ]

( )= [ ]( )

= [ ]

k = mn then n=k/m,
( )=∑ (5.15)

Replacing k by n,

( )=
204 | Z - Transform

Y(z) = X(zm) with ROC : R1/m.

5.7.6 Convolution:
Convolution in the time domain corresponds to multiplication in the Z-domain. This property
is particularly useful for analyzing the behavior of linear time-invariant systems. Convolution
property states that,
( )↔ ( ) ℎ
( )↔ ( ) ℎ 2

then, ( ) = ( ) ∗ ( )↔ ( )= ( ). ( ) ∶  (5.16)
Proof : By definition of Z.T.
We have,

( )= ( )

[ ( )∗ ( )]

But, we know,
( )= ( )∗ ( )

=∑ ( ) ( − ) (5.17)
Substituting y (n) in y (z) as,
( )=∑ ∑ ( ). ( − ) (5.18)
=∑ ( )∑ ( − ) (5.19)
Using time shifting property,
=∑ ( ). ( ) (5.20)
Putting ∑ ( ). = ( ) into the above equation,
( )= ( ). ( ) ∶ 
Hence proved.
Signals and Systems| 205

5.7.7 Differentiation in z-domain:


Taking the Z-transform of a signal's derivative in the time domain is related to the Z-transform
of the original signal through multiplication by Z (i.e., differentiation in the time domain
corresponds to multiplication by Z in the Z-domain).
[ ]↔ ( ) ℎ .
Then, ( ) = . [ ] ↔ ( ) = − ( ) ℎ . (5.21)
Proof : We know,

( )= [ ]

Differentiate X(z) w.r.t. z we get,


( )=∑ [ ]. ( ) (5.22)
( )=∑ [ ] . (− ) . (5.23)
( )=− ∑ . [ ]. (5.24)
− . ( )=∑ . [ ] (5.25)
It is the . . { . [ ]}
Hence, . [ ] ↔ − ( )

5.7.8 Conjugation:
Convolution property states that
It states that, if [ ] ↔ ( ) ℎ
then, ( ) = ∗ ( ) ↔ ( ) = ∗ ( ∗ ) ℎ . (5.26)
Proof: The . . [ ] ,
( )=∑ [ ]
Taking conjugate of the above equation we get,
∗( ) )∗
= (∑ [ ] (5.27)
∗( ) ∗ ∗ ( )( ∗ )
=∑ ( ). = ∑ (5.28)

Putting z = z* we get,
206 | Z - Transform

∗ ( ∗) ∗( ).
=

∗( ∗
Hence, )↔ ( ∗)

5.7.9 Initial Value Theorem:


The initial value of a signal can be found from its Z-transform. This property is useful for
analyzing the behavior of signals as they start at time zero.
It states that, if x[n] is causal signal i.e. x[n] = 0, n < 0.
then, (0) = lim ( ) (5.29)

5.7.10 Final Value Theorem:


The final value of a signal can also be determined from its Z-transform. This property is useful
for understanding the long-term behavior of signals as time approaches infinity.
It states that, (∞) = lim(1 − −1. ) ( ) (5.30)

where, x() is the final value of causal signal.
This expression is possible only when the poles of
(1  z1) X(z) are inside the unit circle.
Proof : We have, [ ] − ( − 1) ↔ (1 − ) ( )
L.H.S. ( ) = [ ] − ( − 1)
Taking the Z.T. of above,

( )= { [ ] − ( − 1)}

= →∞ { [ ]− ( − 1)} −

=0

lim (1 − ) ( ) = lim { [ ] − ( − 1)} = (∞)


→ →
(∞) = lim (1 − ). ( )

Hence proved.
Signals and Systems| 207

5.7.11 Accumulation:

, [ ]= [ ]

Then
( )
( )= (5.31)
( )

Proof :

( ) − ( − 1) = [ ]− [ ]

( ) − ( − 1) = [ ]
Taking Z.T. on both sides, we get
( )− ( )= ( )
( )(1 − )= ( )

( )
( )=
(1 − )
This is the Z.T. of accumulator. It adds poles at z = 1 and ROC 1 < |z| < R.

These properties make the Z-transform a powerful tool for analyzing and solving problems in
discrete-time signal processing, control theory, and other related fields. They simplify the
analysis of discrete-time systems and help in understanding their behavior in the Z-domain.

Example 5.4
Determine the Z.T. of the following signals [ ] = . [ ] − [− − ].
for | | < | |, | | > ,| | < .
Solution: ∶ [ ]= [ ]− [ − 1]
The given signal is two sided; the Z.T. is given by,

( )= [ ]

=∑ ( ( )− (− − 1))
208 | Z - Transform

= ( ) − (− − 1)

= ( ) + −( )

1 1
= + −1
1− 1−

( )= +
− −
There are two different causes for ROC.

For | | < | |, ℎ ,| | > ,| | <


These two ROC's overlap as shown and X(z) exists and given by,

( )= + | |<| |<| |
− −

Fig. 5.7 Pole-zero plot and ROC |a| < |z| < |b|
For an infinite duration two sided sequence. The ROC is a ring in the z-plane.
From this example, it is clear that we cannot have two ROC's.
Signals and Systems| 209

Example 5.5
Determine the Z.T. of the signal,
1 1
[ ]= ( )+ ( ).
3 4
Solution: We know that Z.T. of,

. .
( ) , | | > | |.

Hence,
1 . z 1
u ( n) , |z| >
3 1 3
z−3
1 . z 1
u(n) ∶ |z| >
4 1 4
z−4
Therefore, z-transform of x[n] is given by,

( )= +
1 1
−3 −4

4 3 3 4
= + = +
4 −1 3 −1 3 −1 4 −1
1 1
The first series converges for |z| > . Second series converges for |z| > . The common
3 4
1
values of z for which both the series converges is |z| > . This is shown in the figure below.
3

1
Fig. 5.8 Pole-zero plot and ROC : < |z|
3
210 | Z - Transform

Example 5.6
Determine the Z.T. of x[n],

[ ]= (− − ) + (− − )

1 1
[ ]= (− − ) + (− − 1)
4 3
Solution: The Z.T. of x[n] is given by,

1 1
( )= [ (− − 1) + (− − 1)]
4 3

1 1
= +
4 3

= (4 ) + (3 )

4 3
( )= +
1−4 1−3
1
First series converges for |z| < because signal is non-causal and second series converges for
4
1
|z| < . Therefore, x(z) will be converge for common ROC of R1 and R2.
3
1
i.e. |z| <
4
1
The ROC of X(z) is inside the circle of radius |z| <
4
 Pole zero plot and ROC is shown below.
Signals and Systems| 211

1
Fig. 5.9: Pole-zero plot and ROC |z| <
4
5.8 Relationship between DTFT and z-transform:
The relationship between CTFT and L.T. we have seen.
DTFT and Z.T.exhibits similar relation as CTFT and L.T.
Let, z= r  ejω
Here, r is the magnitude and  z = ω is the phase.
We know, z = rejω

( )= [ ].

=∑ [ ] (5.32)

=∑ { . [ ]} (5.33)
and we have, DTFT given by
∑ [ ] (5.34)
On comparing (1) and (2), it is clear that DTFT of rn  x[n] is Z.T.

. .
[ ] ( )
When r = 1.
. .
, (1) [ ]= [ ] ( )

Hence, | | =| | = 1 , DTFT is Z.T. obtained on the unit circle.

( )| = = . . { [ ]}
212 | Z - Transform

5.9 Inverse Z- transform


We know that Z.T. of x[n] which is defined as

( )= [ ]

= .

= . .{ [ ] } (5.35)
Applying I.Z.T.to above equation we get,

[ ] = . . { }
[ ]= . . .{ }
Using I.F.T. expression, we have
[ ]= ∫ ( ) (5.36)

1
( )( . )
2
If we put,
=

= .  dω = /(jr . )
Substituting in above equation is obtained.

[ ] = 1/(2 ) ∮ ( ). . . 1/( . ) (5.37)


The symbol ∮ represents contour integration.
[ ] = 1/2( ) ( ). ( ) ( )

[ ] = 1/(2 ) ( ).

This is the expression for I.Z.T. indicates the integration around a counter clockwise closed
circular contour centered at the origin with radius r.
This is a formal definition of I.Z.T. This is a direct method of computing I.Z.T. I.Z.T.
There are other methods to find a time domain sequence when its z-transform is known.
These are,
Signals and Systems| 213

a) Power series expansion or Long division method


b) Partial fraction method

5.9.1. Power series expansion method


The power series expansion method for the inverse Z-transform is a useful technique for finding
the time-domain sequence when an explicit inverse Z-transform expression is not readily
available. The power series expansion method can be employed to find the inverse Z-transform
by representing X(z) as a power series and then finding the inverse Z-transform of each term in
the series.
The power series expansion is typically expressed as:

( )= .

where cn = x[n] are the coefficients in the power series. When X(z) is rational, the expansion can
be performed by long division. Thus, it is also called as long division method.

We will learn this method through solved examples.

Example 5.7
Using long division method find the I.Z.T. of
+
( )= assuming ROC to be | | > .
+
Solution:
The ROC is outside the circle of radius z = 1/3. So, its corresponding time domain signal is
causal.
Causal signals have negative power series expansion of z.

2 2 2
1 + z1  z2 + z3 
3 9 27

1
1 +   z ) 1 + z1
3
1
1 + z1
3
214 | Z - Transform

2
z1
3
2 2
 z1 + z2
3 9
2
 z2
9
2 2
 z2 + z3
9 27
2
z3
27
2 2 2
X(z) =1 + z1  z2 + z3
3 9 27
Taking I.Z.T.
 2 2 2 
x[n] =1     
 3 9 27 

Example 5.8
z
Using long division method find the I.Z.T. of X(z) = ; ROC |z| < |a|
za
Solution:
The ROC is inside the circle of radius z = a. So, its corresponding time domain signal is
anti-causal.
Anti-causal signals have positive power series expansion of z.
z
So, the I.Z.T. of can be evaluated by using following method (the terms are written in
za
opposite sequence so as to get positive power series expansion of z) as,
1 12 13
 z    z2    z3
a a a 
a+z ) z
1
z  z2 z ) z
a
1
z2
a
Signals and Systems| 215

1  12
z2    z3
a  a
12
a  z3
 
12 13
a  z3  a  z4
   
13
a  z4
 
13 14
a  z4  a  z5
   

We know that quotient is the power series of X(z).


Therefore,

1 1 1
( )=− − − …

1 1 1
=− + + …….

So, x[n] is given by,


1 12 13 
[ ] = −  z +   z2 +   z3 
 a  a  a 
={ a3, a2, a1, 0}
= {… … … , , , 0}

We can write, [ ]

[ ] = −− (− − 1)
 We can write relation X(z) and x[n] as,

− (− − 1) ↔ ; | |<| |
( − )
216 | Z - Transform

5.9.2 Partial fraction method


The inverse Z.T. can be evaluated by partial fraction method. If the Z.T.is represented as a
rational function given by,
( ) ⋯…
( )= = (5.38)
( ) ⋯…
With its corresponding ROC, then partial expansion can be used to find the inverse Z.T.
1. If the denominator D(z) can be factorized and if it has distinct real roots as given by,

( )
= + (5.39)

=

( ) .( )( )…..(( )
( )= =( (5.40)
( ) )(( )……(( )

Case I) P1, P2, P3, , Pn are different poles.

( )
= + + + ⋯..
− − −

= +∑ (5.41)

The C0, C1, C2  Cn can be evaluated by following equations,

= ( ) =0
=( − ) ( )

( )= +( +( + ⋯..( (5.42)
) ) )

We can compute I.Z.T. of , ….


Case II) If X(z) has poles with multiplicity with greater than one or if X(z) has more than one
poles at same location.
Signals and Systems| 217

Let Pi is pole location with multiplicity r then X(z)/z will have term of the following form

( )
= +( +⋯( )
(5.43)
)

( )
= ( − ) (5.44)
!

From the above equation we can find the coefficients of poles which are located at same
location.

Example 5.9
Determine the inverse Z.T. of the following X(z) by the partial fraction method.
+
( )=
− +
With ROC ( )| | > , ( )| | < , ( ) > | | < .
Solution: Let,
( ) ( + 2)
=
(2 − 7 + 3)

( ) +2
=
1
2 − 2 ( − 3)

( )
= + +
1 ( − 3)
−2

0+2 2
= ( )| = =
1
2+ ∗3 3
2

1 ( + 2)
= −
2 1
2 − ( − 3)
2

( + 2)
=
2 ( − 3)
218 | Z - Transform
1 5
+2
= 2 = 2 = −1
1 5
2∗ 1 −
2
2 ∗3
2

( + 2)
= ( − 3)
1
2 − ( − 3)
2

3+2 5 1
= + =
1 5 3
2∗3 −2 6∗2

 Substituting values of C0, C1 and C2 we get,

2 1
( )= − +
3 1 3 ( − 3)
−2

(a) ROC : |z| > 3


1
X(z) has 2 poles at z = , z = 3.
2
 The I.Z.T. of X(z) is given by,

2 . . . 2
⎯ [ ]
3 3

. . . 1
⎯ ( )
1 2

2

. . .
⎯ (3) ( )
( − 3)
2 1
[ ]= [ ]− ( ) + (3) ( )
3 2
1
(b) ROC |z| <
2
1
The ROC is inside the circle of radius z = .
2
2 1 1
[ ]= [ ]+ (− − 1) − (3) (− − 1)
3 2 3
(c) ROC : > | | <
Signals and Systems| 219
1
( ) ∶ | |<3
2
 The ROC is ring between the poles at = = 3.

z z
The corresponding time domain signal of is anti-causal and of is causal.
(z  3) 1
z
2
2 1 1
[ ]= [ ]+ ( ) − (3) (− − 1)
3 2 3

5.10 Z – Domain Causality and stability analysis


A causal linear time invariant system is one whose unit sample response h(n) satisfies the
condition,
ℎ[ ] = 0 <0 (5.45)
We have also seen that the ROC of the Z-transform of a causal sequence is the exterior of a
circle outside the outermost pole.
When we talk about stability we want to refer to the so called BIBO (bounded-input, bounded-
output) stability. As the name already implies, a BIBO stable system should respond with a
bounded output to a bounded input.
The output y[n] is commonly known as the convolution of x[n] and the impulse response, h[n].
So, with a bounded input, output will be bounded if and only if the impulse response is finite.
For the impulse response to be finite, we need to ensure that h[n] is absolutely summable. This
would ensure that the system will be stable. Hence, the bottom line is that we need an absolutely
summable impulse response, i.e.
∑ |ℎ[ ]| < ∞ (5.46)
By definition, ( ) = ∑ ℎ( ) , it follows
| ( ) |= ∑ |ℎ( ) |= ∑ | ℎ )||
( | (5.47)
For |z| = 1 i.e. on the unit circle, it will become,

( )= ∑ |ℎ[ ]| (5.48)

In turn this condition implies that H(z) must contain the unit circle within its ROC.

Example 5.10
A linear time-invariant system is characterized by the system function
3−4
( )=
1 − 3.5 + 1.5
220 | Z - Transform

Specify the ROC of H(z) and determine h(n) for the following conditions:
(a) The system is stable.
(b) The system is causal.
(c) The system is anticausal.
Solution:
3−4
( )=
1 − 3.5 + 1.5
By applying partial fractions,
1 2
( )= +
1 − 0.5 1−3
The system has poles at z= 0.5 and z= 3.
(a) Since the system is stable, its ROC must include the unit circle and hence
it is 0.5 < |z| < 3. Consequently, h(n) is non-causal and is given as,
ℎ( ) = (0.5) ( ) − 2(3) (− − 1)
(b) Since the system is causal, it ROC is |z| > 3. Hence,
ℎ( ) = (0.5) ( ) + 2(3) )
and the system is unstable in this case.
(c) If the system is anti-causal, it ROC is |z| < 0.5. Hence,
ℎ( ) = −(0.5) ( − 1) − 2(3) (− − 1)
and the system is unstable in this case.

Unit Summary
The Z-transform, a powerful mathematical tool in discrete-time signal processing, offers a
comprehensive framework for analyzing discrete signals and systems. Understanding its
definition and properties, including linearity, time-shifting, scaling, and convolution properties,
forms the foundation of its application. The inverse Z-transform facilitates the conversion of
transformed signals back into the time domain, employing techniques such as partial fraction
decomposition and contour integration. In practical terms, the Z-transform finds extensive use
in digital filter design, system analysis, and control theory, allowing engineers to analyze and
design discrete-time systems with precision. Its applications extend to fields such as
telecommunications, audio processing, and image processing, where it enables efficient
manipulation of discrete signals. Advanced topics, including the relationship between Z-
transform and Laplace transform, as well as multirate signal processing, deepen the
understanding of its theoretical underpinnings and broaden its scope of application. As
technology continues to evolve, the Z-transform remains a vital tool for digital signal
processing, offering insights into the behavior of discrete systems and paving the way for
innovative solutions in a digital world.
Signals and Systems| 221

Exercise
1) Determine the Z-transform of the following signals, also mention the ROC.

a) ( ) = {1, 2, 3, 3, 2, 1} 0≤ ≤5

b) ( ) = {−1, 0, 1, 0, −1, 0, +1} −2 ≤ ≤4

c) ( ) = {3, 2, 1, 1, 2, 3} −5≤ ≤0

2) Determine the z-transform of the following signals, also plot the ROC.

a) ( )= sin( ). ( )

b) ( ) = (0.3) [ ( ) − ( − 2)]

3) Determine the causal signal ( ) if Z-transform is given by

2 − 1.5
( )=
1 − 1.5 + 0.5
4) Determine the Z-transform of the signal ( ) = − (− − 1) and plot the ROC.

5) Evaluate the inverse z-transform of

a) ( )= |z|<0.5
.

b) ( )=( . )( . )

6) For the given sequences ( ) = 5 ( ) − 2 ( − 2) and ( ) = 3 ( − 3), find


( ) = ( ) ∗ ( ) using the convolution property of z-transform.

7) A linear time invariant system is characterized by the system function ( )=

. .
Specify the ROC of H(z) and determine h(n) for the following
convolutions

a) The system is stable

b) The system is causal

c) The system is anticausal.


222 | Z - Transform

Multiple choice questions


1. The Ƶ – transform of (− − 1) is,

a) ⁄
b) ⁄

c)

d)

2. The ROC of the sequence ( ) = (− ) is,

a) |z|>1

b) b) |z|<1

c) c) no ROC

d) d) -1<|z|<1

3. The inverse Ƶ − ,| | > 4 ,


a) 3(4) ( − 1)
b) 3(4) ( )
c) 3(4) ( + 1)
d) 3(4) ( − 1)

4. ROC of x(n) contains


a) Poles
b) zeros
c) no poles
d) no zeros

5. The Ƶ − ( ) = [ ( ) − ( − 3))], | |>1 ,


a) ( ) =
b) ( ) =
( )

c) ( ) =
( )
d) none of the above
Signals and Systems| 223

6. If all the poles of the system function H(z) have magnitude smaller than 1, then the system will
be,

a) Stable
b) unstable
c) BIBO stable
d) a and c

7. ( ) = {0.5, −0.25, 1}, ℎ Ƶ− ℎ ,


a)
. .
b)
. .
. .
c)

d)

8. ℎ ℎ ( )= −5 < <5 ,
a) Entire z-plane
b) entire z-plane except z=0 and z=∞
c) Entire z-plane except z=0
d) entire z-plane except z=∞

9. Ƶ− ( ) ( )ℎ Ƶ− (− ) ,
a) − ( )
b) (− )
c) − ( )
d) ( )

10. ℎ Ƶ− ( ) ,
a) ( ) = ∮ ( )
b) ( ) = ∮ ( )

c) ( ) = ∮ ( )
d) ( ) = ∮ ( )
224 | Z - Transform

11. ℎ Ƶ − ,
a) finite series
b) infinite power series
c) geometric series
d) both a and c

12. ℎ Ƶ − ℎ ( ) ( ) ,
∗( ) ∗
a) ( ) b) ( ) ( ) c) ( ) ∗ ( ) d) ( ) ( )

13. For a stable LTI discrete time system poles should lie_______ and unit circle should be_______
a) Outside unit circle, included in ROC
b) inside unit circle, outside of ROC
c) inside unit circle, included in ROC
d) outside unit circle, outside of ROC

14. ℎ , ℎ ( ) = (− ) ( ) − < −1 ,
a) stable system
b) unstable system
c) anticausal system
d) neither stable nor causal

15. ( )ℎ ℎ , ℎ , ( ) ,
a) signed constant sequence
b) signed decaying sequence
c) signed growing sequence
d) constant sequence

Ƶ
16. ( )ℎ − ( ) ℎ → ℎ ( )↔ ,

a) b) a c) d)

17. ℎ Ƶ − ( )= ( ) ,
a) ( ) = ; | |>1 b) ( ) = ; | |>1
( ) ( )
Signals and Systems| 225

c) ( ) = ( )
; | |<1

d) ( ) = ; | |<1
( )

Answers to the multiple choice questions:


1. a 2. b 3. d 4. c 5. d 6. a 7. c 8. b 9. d
10. c 11. b 12. b 13. c 14. a 15. a 16. a 17. a

KNOW MORE
Z-transform reveals a rich tapestry of mathematical intricacies and practical applications.
Beyond its fundamental properties lie advanced techniques and insights that amplify its utility
in discrete signal analysis and system design. Understanding the intricacies of Z-transform
inversion methods, such as residue calculus and contour integration, empowers engineers and
researchers to unravel complex system behaviors with precision and accuracy. The Z-
transform's role extends far beyond mere signal processing; it serves as a cornerstone in areas
ranging from digital control theory to communication systems design, enabling the
development of robust algorithms and efficient data processing techniques. Exploring the
connections between the Z-transform and other mathematical tools, such as Fourier analysis
and Laplace transform, unveils deeper insights into the interplay between time and frequency
domains in discrete systems. Moreover, ongoing research in areas such as multirate signal
processing and adaptive filtering continues to push the boundaries of Z-transform theory,
paving the way for innovative applications in emerging technologies. As we delve deeper into
the intricacies of the Z-transform, we unlock a world of possibilities, where mathematical
abstraction converges with real-world engineering challenges, driving forward progress and
innovation in the digital age.
226 | Z - Transform

REFERENCES AND SUGGESTED READINGS


1. Signals and Systems by Simon Haykin
2. Signals and Systems by Ganesh Rao
3. Signals and Systems - Course (nptel.ac.in)

Dynamic QR Code for Further Reading


Signals & Systems | 227

Sampling &
d
6 Reconstruction

UNIT SPECIFICS
Through this unit we have discussed following aspects:
• The necessity of sampling theorem
• Sampling theorem for Continuous Time & Discrete time signal
• Understanding of discrete time processing of continuous time signals
• Frequency domain spectra of the sampled signals
• Interpolation methods for reconstruction of sampled signals
• Zero-order & first order hold interpolation methods
• Effects of under sampling and oversampling on the signals
• Using spectra to understand aliasing and its effects
• Understanding continuous and discrete time systems

RATIONALE
The unit “Sampling and Reconstruction” is not only important to understand signals but also
will be helpful in Communications. We can call on simple intuition to motivate and describe the
processes of sampling and reconstruction from samples, because in communication systems
that are closely related to sampling or rely fundamentally on using a sampled version of the
signal to be transmitted.
228 | Sampling & Reconstruction

This unit focuses on sampling of continuous and discrete time signals. The effects of under
sampling and oversampling are discussed in detail with various examples. Frequency domain
analysis i.e. Fourier transform of signals is extensively used for the better understanding of the
topic. The effects of ‘aliasing’ are explained in lucid manner. Various methods to avoid aliasing
are also discussed in this topic. Both continuous time and discrete time signals are considered
for discussions. For reconstruction of the sampled signal, various interpolation methods are
discussed with detailed mathematical analysis and distinct examples. Different filtering
techniques are studied for proper reconstruction of the signal.

The discrete and continuous time systems are also discussed to give overview of the working of
how systems work.

PRE-REQUISITES
1. Strong understanding of mathematics, including algebra, calculus, and complex
numbers.
2. Familiarity with basic concepts in signals and systems, such as time-domain and
frequency-domain representations, Fourier analysis, and convolution.
3. Proficiency in solving ordinary differential equations and understanding linear
algebra concepts.
4. Basic knowledge of electronics and circuit analysis for understanding continuous-
time LTI systems.
5. Knowledge of digital signal processing concepts for understanding discrete-time
LTI systems.

UNIT OUTCOMES
List of outcomes of this unit is as follows:
U6-O1: Understand need of sampling theorem
U6-O2: Apply sampling theorem to continuous and discrete time signals
U6-O3: Study the effects of under sampling and oversampling
U6-O4: Perform Zero-order and first order hold interpolation
U6-O5: Study aliasing and its effects
U6-O6: Understand aliasing and its effects through Fourier analysis
U6-O7: Study relationship between continuous time & discrete time systems
Signals & Systems | 229

EXPECTED MAPPING WITH COURSE OUTCOMES


Unit-6
(1- Weak Correlation; 2- Medium correlation; 3- Strong Correlation)
Outcomes
CO-1 CO-2 CO-3 CO-4 CO-5 CO-6
U6-O1 - - - - - 3
U6-O2 2 - - 2 - 3
U6-O3 - - - - - 3
U6-O4 - - - - - 3
U6-O5 - - - - - 3
U6-O6 - - 3 - - 3
U6-O7 - 2 - 2 - 2
230 | Sampling & Reconstruction

6.1 Introduction
A continuous time signal can entirely be represented by its samples which are equally spaced
in time. The sampling theorem is associated with these samples. This theorem is widely used in
applications where digital data is preferred over analog. Sampling theorem is one of the most
important theorems in signals & systems as it acts bridge between continuous time signals and
discrete time signals.
Nowadays, technically advanced digital systems are developed to effectively process
continuous time signals. Hence, there is need to convert these continuous time signals into
discrete time signals. Sampling process provides some insight, to deal with the problem of
conversion mentioned above. Sampling is the process which involves conversion of continuous
time signal to discrete time signal. The sampling is performed by taking samples of continuous
time signal at definite intervals of time. This sampled (discrete time) signal is easily processed
by discrete time systems. The original continuous time signal is reconstructed from this discrete
time signal.
In the following discussion, we introduce and develop concept of sampling and process of
reconstructing CT signals from its samples. In the following discussion, we will analyze the
conditions under which sampling rate is sufficient to be able to exactly reconstruct the original
continuous time signal. And we will also observe when sampling rate is low what will happen
to original continuous time signal while trying to reconstruct from its samples. Finally, we
examine the sampling of discrete time signals & related concepts of decimation & interpolation.
6.2 Sampling Theorem
First, we need to clearly see some examples of continuous time signals which can be uniquely
specified by a sequence of equally spaced samples. For example, figure 6.1 illustrate three
different continuous time signals, all of which have identical values at integer multiples of T (T
is sampling time/sampling period/sampling interval) that is,
x (kT) = x (kT) = x (kT)

Fig 6.1: Three continuous time signals


Signals & Systems | 231

6.2.1 Impulse train sampling


To sample a continuous time signal at regular intervals we need a convenient way to develop
sampling theorem. One useful way is to use periodic impulse train multiplied by continuous
time signal ( ) which is the signal we wish to sample. This mechanism is known as impulse-
train sampling and is shown in figure 6.2. The periodic impulse train ( ) is referred to as
sampling function, the period T as sampling period and the fundamental frequency of ( ),
= as the sampling frequency. In time domain, we can write,
( )= ( ) ( ) (6.1)
Where,

( )= ( − ) (6.2)

Now, by using sampling property of unit impulse function, we know that multiplying x(t) by a
unit impulse sample the value of signal at the point at which the impulse is located i.e.
( ) ( − ) = ( ) ( − ). Applying this to Eq. (6.1) we see the result which is illustrated
in figure 6.2, that ( ) is an impulse train with amplitudes of impulses equal to the samples of
x(t) at intervals spaced at time interval T; that is,

( )= ( ) ( − ) (6.3)

From multiplication property, we know that


1
( )= ( ) ( − ) (6.4)
2
Using earlier properties,
2
( )= ( − ) (6.5)

Since, convolution with the impulse function simply shifts a signal, it follows that,

1
( )= (( − )) (6.6)
232 | Sampling & Reconstruction

Fig 6.2: Impulse-train sampling

So, ( ) is periodic function of consisting of superposition of shifted replicas of ( ),


scaled by 1/ as illustrated in figure 6.3. In figure 6.3(c), <( − ), or equivalently,
>2 and hence there is no overlap between shifted replicas of ( ), whereas in figure
6.3(d), with > 2 , x(t) can be recovered exactly from xp(t) by means of a low pass filter
with gain T and cut off frequency greater than and less than − and is indicated in
figure 6.4.
Signals & Systems | 233

Fig. 6.3: Frequency domain representation due to sampling in time domain: (a) Spectrum of
original signal; (b) Spectrum of sampling function; (c) Spectrum of sampled signal with
> ; (d) Spectrum of sampled signal <

Sampling Theorem:
Let x(t) be a band limited signal with ( ) = 0 | |> . Then x(t) is uniquely
determined by its samples x(nT), = 0, ±1, ±2, … …, if
>2
Where,
2
=
Given these samples, we can reconstruct x(t) by generating a periodic impulse train in which
successive impulses have amplitudes that are successive sample values. This impulse train is
then processed through an ideal lowpass filter with gain T and cutoff frequency greater than
and less than − . The resulting output will be exactly equal to x(t).
We have seen sampling theorem, where impulse train sampling method was discussed.
However, in practice the frequency of the original continuous time signal which comes under
sampling theorem must be excess than the sampling frequency is referred to as ‘Nyquist Rate’.
In real life applications a non-ideal lowpass filter is used instead of ideal lowpass filter as
shown in figure 6.4. The non-ideal filter has filter characterictics as | ( )| where | ( )| ≅
1 < and | ( )| ≅ 0 − . For understanding basic principles of
sampling theorem for convenience we will regularly use ideal filters throughout this chapter.
234 | Sampling & Reconstruction

Fig 6.4: Recovery of original signal by using ideal lowpass filter


(a) System for sampling & reconstruction (b) Representative spectrum for x(t) (c)
Corresponding spectrum for xp(t) (d) Ideal lowpass filter to recover ( ) from ( ) (e)
Spectrum of xr(t)

6.2.2 Sampling with Zero-Order Hold


In sampling theorem a band limited signal is uniquely represented by its samples using the
impulse-train sampling. However, in practice, large narrow amplitude pulses are difficult
Signals & Systems | 235

generate and transmit. So, it is often more convenient to generate sampled signal in a form
referred to as ‘Zero-Order hold. Such system samples the given continuous time signal at given
instant and holds that value until the next instant at which sample is taken as shown in figure
6.5.

Fig 6.5: Sampling using Zero-Order Hold


Now, the output ( ) of zero order hold can be generated by impulse train sampling followed
by LTI system. The impulse response is shown in figure 6.6 as shown below,

Fig 6.6: Zero order hold as impulse- train sampling followed by an LTI system with ractangular
response
236 | Sampling & Reconstruction

6.3 Reconstruction of a signal from its samples using interpolation


Interpolation, to fit a continuous signal to a set of sample values, is a widely used for
reconstructing a function from its samples. One simple interpolation procedure is the zero-order
hold discussed in Section 6.1.2. Another useful form of interpolation is linear interpolation,
whereby adjacent sample points are connected by a straight line, as illustrated in Figure 6.7

Fig 6.7: Interpolation between samples, solid curve represents interpolation


For a band-limited signal if the sampling instants are sufficiently close then signal can be
reconstructed exactly. By using a lowpass filter, exact interpolation can be carried out between
sample points. For the reconstruction of ( ) using the interpolation process we will consider
the effect of lowpass filter in time domain as shown in figure 6.4. The output of reconstruction
is written as,
( ) = ( ) ∗ ℎ( )
Substituting for ( ) using Eq. (6.3) in above equation, we can write

( )= ( )ℎ( − ) (6.7)

Eq. (6.7) describes how to fit a continuous curve between the sample points ( ) and
consequently represent an interpolation formula. For the ideal lowpass filter ( ) in Figure
6.4, the impulse response ℎ( ) is,

sin( )
ℎ( ) = (6.8)

we will get,

( )= ( )ℎ( − ) (6.9)

( ( − ))
( )= ( ) (6.10)
( − )
Signals & Systems | 237

The reconstruction according to Eq. (6.10) with = /2 is illustrated in Figure 6.8 Figure
6.8(a) represents the original band-limited signal ( ), and Figure 6.8(b) represents ( ), the
impulse train of samples. In figure 6.8(c), the superposition of the individual terms in Eq. (6.10)
is illustrated.

Interpolation using the impulse response of an ideal lowpass filter as in Eq. (6.10) is commonly
referred to as band-limited interpolation, since it implements exact reconstruction if ( ) is
band limited and the sampling frequency satisfies the conditions of the sampling theorem. As
we have indicated, in many cases it is preferable to use a less accurate, but simpler, filter or,
equivalently, a simpler interpolating function than the function in Eq. (6.8). For example, the
zero-order hold can be viewed as a form of interpolation between sample values in which the
interpolating function ℎ( ) is the impulse response ℎ ( ) depicted in figure 6.6. In that sense,
with ( ) in the figure corresponding to the approximation to ( ), the system ℎ ( ) represents
an approximation to the ideal lowpass filter required for the exact interpolation. Figure 6.9
shows the magnitude of the transfer function of the zero-order-hold interpolating filter,
superimposed on the desired transfer function of the exact interpolating filter.

Fig 6.8: Band-limited interpolation using Sinc function: (a) Band-limited signal x(t) (b) Impulse
train sampling of x(t) (c) Ideal band-limited interpolation in which impulse train is
replaced by superposition of Sinc functions.
238 | Sampling & Reconstruction

Fig 6.9: Transfer function for zero-order hold

If the interpolation provided by zero-order hold is insufficient then we can opt for interpolation
strategies which are of higher order holds. We know from figure 6.5 that the zero-order hold
produces an output signal that is not continuous in time. On the other hand, the linear
interpolation, as shown in figure 6.7 gives us reconstructions which are continuous. The linear
interpolations are sometimes also known as first-order hold and can also be viewed as in figure
6.6. The associated transfer function is also shown in figure and is given by

1 sin 2
( )= (6.11)
2
The transfer function of the first-order hold is shown superimposed on the transfer function for
the ideal interpolating filter. Now, we can define second- and higher order holds that produce
reconstructions with a higher degree of smoothness. For example, the output of a second-order
hold provides an interpolation of the sample values that is continuous and has a continuous first
derivative and discontinuous second derivative.

6.3.1 The effect of under sampling: Aliasing

Till now we assumed that the sampling frequency was sufficiently high that the conditions of
the sampling theorem were satisfied. As illustrated in figure 6.3, with >2 the spectrum
of the sampled signal consists of scaled replications of the spectrum of ( ), and this forms the
basis for the sampling theorem. When < 2 , ( ) the spectrum of ( ), is no longer
replicated in ( ) and thus it is not possible to recover original continuous time signal by
lowpass filtering. This effect is known as aliasing, and in this section, we explore its effect and
consequences.

Clearly, if the system of figure 6.4 is applied to a signal with <2 the reconstructed
signal ( ) will no longer be equal to x(t). However, as explored in earlier section, the original
Signals & Systems | 239

signal, and the signal ( ) that is reconstructed using band­ limited interpolation will always
be equal at the sampling instants; that is, for any choice ,

( )= ( ), = 0, ±1, ±2 … … (6.12)

Fig 6.10: Linear Interpolation (First-order hold) as impulse-train sampling followed


with convolution a triangular impulse response: (a) system for sampling &
reconstruction; (b) Impulse train of samples; (c) Impulse response representing
a first-order hold; (d) First-order hold applied to sampled signal; (e)
Comparison of transfer function of ideal interpolating filter & first order hold.

We will try to understand the relationship between ( ) and ( ) when <2 for the
simple case of sinusoidal case thus let,
( ) = cos( ) (6.13)

with Fourier transform ( ) as indicated in Figure 6.11(a). In this figure, we have graphically
distinguished the impulse at w0 from that at -w0 for convenience. Let us consider ( ), the
spectrum of the sampled signal, and focus on the effect of a change in the frequency with
the sampling frequency fixed. In figure 6.11(b)-(e), we illustrate ( ) for several values
of . Also indicated by a dashed line is the passband of the lowpass filter of figure 6.4 with
= /2. Note that no aliasing occurs in (b) and (c), since < /2, whereas aliasing does
occur in (d) and (e). For each of the four cases, the lowpass filtered output ( ) is given as
follows:
240 | Sampling & Reconstruction

( ) = ; ( ) = cos = ( )
6
2
( ) = ; ( ) = cos = ( )
6
4
( ) = ; ( ) = cos( − ) ≠ ( )
6
5
( ) = ; ( ) = cos( − ) ≠ ( )
6
When aliasing occurs, the original frequency takes on the identity of a lower frequency,
− . For /2 < < , as increases relative to , the output frequency −
decreases. When = , for example, the reconstructed signal is a constant. This is consistent
with the fact that, when sampling once per cycle, the samples are all equal and would be
identical to those obtained by sampling a constant signal ( = 0). In figure 6.12, we have
depicted, for each of the four cases in Figure 6.11, the signal ( ), its samples, and the
reconstructed signal ( ). From the figure, we can see how the lowpass filter interpolates
between samples. Consider another sinusoidal signal given by Eq. (6.14) as,

( ) = cos( + ) (6.14)

In this case, the Fourier transform of ( ) is essentially the same as Figure 6.11(a), except that
the impulse indicated with a solid line now has amplitude , while the impulse indicated
with a dashed line has amplitude with the opposite phase, namely, If we now consider
the same set of choices for as in Figure 6.11, the resulting spectra for the sampled versions
of cos( + ) are exactly as in the figure, with all solid impulses having amplitude
and all dashed ones having amplitude Again, in cases (b) and (c) the condition of the
sampling theorem is met, so that ( ) = cos( + ) = ( ), while in (d) and (e) we again
have aliasing. but we now see that there has been a reversal in the solid and dashed impulses
appearing in the passband of the lowpass filter. As a result, we find that in these cases, ( ) =
cos[( − ) + )], where we have a change in the sign of the phase i.e., a phase reversal.

It is important to note that the sampling theorem explicitly requires that the sampling frequency
be greater than twice the highest frequency in the signal, rather than greater than or equal to
twice the highest frequency. The next example illustrates that sampling a sinusoidal signal at
exactly twice its frequency (i.e., exactly two samples per cycle) is not sufficient.
Signals & Systems | 241

Fig 6.11: Effect of oversampling & under sampling: (a) Spectrum of original sinusoidal signal;
(b) (c) Spectrum of sampled signal with > ; (d) (e) Spectrum of sampled signal
with < ;
242 | Sampling & Reconstruction

Fig 6.12: Effect of aliasing on sinusoidal signal. For each of the four values of , the original
sinusoidal signal (solid curve), its samples, and reconstructed signal (dashed curve) are
illustrated: (a) = , in (a) and (b) no aliasing occurs, whereas in (c) & (d) there is
aliasing
Signals & Systems | 243

Example 6.1

Consider a sinusoidal signal

( ) = cos + ,
2
And suppose that the above signal is sampled using impulse sampling at exactly the twice the
frequency of the sinusoid that is with sampling frequency . Now if this impulse sampled
signal is applied as input to an ideal low pass filter with cut-off frequency /2 the resulting
output is

( ) = (cos ) cos
2
It is observed that the perfect reconstruction of ( ) is only possible when = 0 or when is
integer multiple of 2 . Otherwise ( ) ≠ ( ). So, the perfect reconstruction of original
continuous time signal becomes conditional, which is not desirable.

Now, let us consider the case in which = − /2 so that,

( ) = sin
2
The signal corresponding to above signal is sketched in figure 6.12. The values of the signal at
integer multiples of the sampling period 2 / are zero. So, sampling at this rate will produce
a signal which is zero. Now these zero inputs will be given to the ideal low pass filter, the
resulting output ( ) will also be zero.

Fig 6.13: Sinusoidal signal for example 6.1

Due to under sampling, stroboscopic effect is observed where higher frequencies are reflected
into lower frequencies, is the principle on which the it is based. Consider, for example, the
situation depicted in Figure 6.13, in which we have a disc rotating at a constant rate with a
single radial line marked on the disc.
244 | Sampling & Reconstruction

Fig 6.14 Strobe effect

The flashing strobe can be considered to act as a sampling system, since it illuminates the disc
for extremely brief time intervals at a periodic rate. It is observed that when the rotational speed
of the disc is less than the strobe frequency then the speed of the rotation of the disc is perceived
correctly. Now, when the strobe frequency is less than twice the rotational frequency of the
disc, then rotation appears to be at lower frequency than the actual. Sometimes because of phase
reversal disc will appear to be rotating in the wrong direction. Now, if we track the position of
line on the disc at successive samples then when < < 2 such that sampling rate per
revolution is increased so that so that we sample somewhat more frequently than once per
revolution, samples of the disc will show the fixed line in positions that are successively
displaced in a counterclockwise direction, opposite to the clockwise rotation of the disc itself.
At one flash per revolution, corresponding to = , the radial line appears stationary (i.e.,
the rotational frequency of the disc and its harmonics have been aliased to zero frequency). A
similar effect is commonly observed in western movies, where the wheels of a stagecoach
appear to be rotating more slowly than would be consistent with the coach's forward motion,
and sometimes in the wrong direction. In this case, the sampling process corresponds to the fact
that moving pictures are a sequence of individual frames with a rate (usually between 18 and
24 frames per second) corresponding to the sampling frequency.

The preceding discussion suggests interpreting the stroboscopic effect as an example of a useful
application of aliasing due to under sampling. Another practical application of aliasing arises
in a measuring instrument referred to as a sampling oscilloscope. This instrument is intended
for observing very high-frequency waveforms and exploits the principles of sampling to alias
these frequencies into ones that are more easily displayed.
Signals & Systems | 245

6.4 Discrete Time Processing of Continuous Time Signals


In many applications, there is a significant advantage offered in processing a continuous time
signal by first converting it to a discrete-time signal and, after discrete-time processing,
converting back to a continuous-time signal. The discrete-time signal processing can be
implemented by a special purpose computer, with various DSP processors, or with any of the
variety of devices that are specifically designed for discrete-time signal processing.

In broad terms, this approach to continuous-time signal processing can be viewed as the cascade
of three operations, as indicated in Figure 7.13, where ( ) and ( ) are continuous-time
signals and ( ) and ( ) are the discrete-time signals corresponding ( ) to ( ) and.
The overall system is, of course, a continuous-time system in the sense that its input and output
are both continuous-time signals. The theoretical basis for converting a continuous-time signal
to a discrete-time signal and reconstructing a continuous-time signal from its discrete-time
representation lies in the sampling theorem. By satisfying simple conditions of sampling
theorem and through the process of periodic sampling with the sampling frequency consistent
with the conditions of the sampling theorem, the continuous-time signal ( ) is exactly
represented by a sequence of instantaneous sample values ( ); that is, the discrete-time
sequence ( ) is related to ( ) by

Fig 6.15: Discrete-time processing of Continuous Time signals

( )= ( ) 6.15

The continuous time signal is applied to the first block of the figure 6.13 and ( ) is converted
to the discrete time signal ( ). It will be abbreviated as C/D conversion. The third block in
figure 6.13 converts the discrete time signal ( ) to continuous signal. This conversion is
246 | Sampling & Reconstruction

known is abbreviated as D/C. The D/C operation uses the interpolation technique between
sample values provided to it as input. The continuous time signal produced is expressed as

( )= ( )

This notation is made explicit in Figure 6.14. In systems such as digital computers and digital
systems for which the discrete-time signal is represented in digital form, the device commonly
used to implement the C/D conversion is referred to as an analog-to-digital (A­ to-D) converter,
and the device used to implement the D/C conversion is referred to as a digital-to-analog (D-
to-A) converter.

Fig 6.16: Notation for A/D conversion and D/A conversion

To understand further the relationship between the continuous-time signal ( ) and its
discrete-time representation ( ) , it is helpful to represent C/D as a process of periodic
sampling followed by a mapping of the impulse train to a sequence. These two steps are
illustrated in Figure 6.15. First step is to represent using the sampling process, the impulse train
( ) corresponds to a sequence of impulses with amplitudes corresponding to the samples of
( ) and with a time spacing equal to the sampling period T. In the process of conversion from
the impulse train to the discrete-time sequence, we obtain ( ), corresponding to the same
sequence of samples of ( ), but with unity spacing in terms of the new in­ dependent variable
n. Thus, in effect, the conversion from the impulse train sequence of samples to the discrete-
time sequence of samples can be thought of as a normalization in time. This normalization is
evident in Figures 6.15 (b) and (c).

It is also instructive to examine the processing stages in Figure 6.13 in the frequency domain.
Since we will be dealing with Fourier transforms in both continuous and discrete time, in this
section only we distinguish the continuous-time and discrete-time frequency variables by using
in continuous time and Ω in discrete time. For example, the continuous-time Fourier
Signals & Systems | 247

transforms of ( ) and ( ) are ( ) and ( ), respectively, while the discrete-time


Fourier transforms of ( ) and ( ) are ( Ω) and ( Ω), respectively.

Fig 6.17: Sampling then followed by conversion to Discrete-time sequence: (a) Overall System;
(b) ( ) for two sampling rates. Dashed envelop represents ( ); (c) The output
sequence for two different sampling rates.

Let us, apply Fourier Transform to ( ) to get ( ), then we will get,

( )= ( ) ( − ) (6.16 )

Now, we know the transform of ( − ) is it follows that

( )= ( ) (6.17 )

On similar lines the discrete-time Fourier Transform of ( ) will be


248 | Sampling & Reconstruction

( Ω) = ( ) (6.18 )

Using the equation 6.15

= ( ) (6.19)

Comparing equation 6.17 and 6.19 we see that & ( ) are related as

= ( Ω/ ) (6.20)

With the help of equation 6.6, we can further write

1
( )= ( (ω − )) (6.22)

Similarly,

1
= ( (Ω − 2πk)/ ) (6.23)

The relationship among ( ), ( ) and ( ) is illustrated in figure 6.16 for two


different sampling rates. Here, ( ) is a frequency scaled version of ( ) and in
particular is periodic in Ω with period 2 . This is the characteristic of any discrete-time Fourier
transform. The spectrum of ( ) is related to that of ( ) through periodic replication,
represented by eq. (6.22), followed by linear frequency scaling, represented by eq. (6.20). The
periodic replication is a consequence of the first step in the conversion process in Figure 6.15,
namely, the impulse-train sampling. The linear frequency scaling in eq. (6.20) can be thought
of informally because of the normalization in time introduced by converting from the impulse
train ( ) to the discrete-time sequence ( ).

In the overall system of Figure 6.13, after processing with a discrete-time system, the resulting
sequence is converted back to a continuous-time signal. This process is the reverse of the steps
in Figure 6.15. Specifically, from the sequence ( ), a continuous time impulse train ( )
can be generated. Recovery of the continuous-time signal ( ) from this impulse train is then
accomplished by means of lowpass filtering, as illustrated in Figure 6.17.
Signals & Systems | 249

Fig 6.18: The relationship among ( ), ( ) and ( ) with different sampling rates

Fig. 6.19: Conversion of discrete time sequence to a continuous time signal

6.4.1 Digital differentiator


Consider the discrete-time implementation of a continuous-time band-limited differentiating
filter. As discussed in earlier section, the frequency response of a continuous-time
differentiating filter is,
( )= (6.24)
And band-limited differentiator with cutoff frequency is
, | |<
( )= (6.25)
0, | |>
250 | Sampling & Reconstruction

With sampling frequency =2 , we see that corresponding discrete-time transfer function


is,
Ω
= , | Ω| < (6.26)

For the above filters, the magnitude and phase response are shown in figure 6.19 and figure
6.20

Fig 6.20: Frequency Response of ideal band-limited differentiator

Fig 6.21: Frequency Response of discrete-time filter used to implement a continuous time band-
limited differentiator
Example 6.2
By considering the output of the digital differentiator for a continuous time sine input, we may
conveniently determine the impulse response ℎ ( ) of the discrete-time filter in the
implementation of the digital differentiator.
Signals & Systems | 251

sin( / )
( )= (6.27)

Where T is sampling period, then


1, | |< /
( )= ( )=
0, ℎ
which is sufficiently band limited to ensure that sampling ( ) at frequency =2 /
does not give rise to any aliasing. It follows that the output of the digital differentiator is,

cos( / ) sin
( )= ( )= − (6.28)

The ( ) is given by equation (6.27) the corresponding signal ( ) is


1
( )= ( )= ( ) (6.29)

Which can be verified from l’Hospital’s rule,


−1
( )= ( )= , ≠0 (6.30)
0, =0
So when input to the filter is given by equation 6.26 is the scaled unit impulse in equation (6.29)
The resulting output is given by eq. 6.30. So, impulse response of this filter is given by,
−1
ℎ ( )= , ≠0 (6.30)
0, =0

6.5 Sampling of discrete time signals


Thus far in this chapter, we have considered the sampling of continuous-time signals, and in
addition to developing the analysis necessary to understand continuous-time sampling, we have
introduced several of its applications. As we will see in this section, a very similar set of
properties and results with several important applications can be developed for sampling of
discrete-time signals.
6.5.1 Impulse train sampling
In analogy with continuous-time sampling as carried out using the system of Figure 6.14,
sampling of a discrete-time signal can be represented as shown in Figure 6.20. Here, the new
sequence ( ) resulting from the sampling process is equal to the original sequence ( ) at
integer multiples of the sampling period N and is zero at the intermediate samples; that is,
252 | Sampling & Reconstruction

( ), =
( )= (6.31)
, ℎ
Using the multiplication property developed early in this section, discrete time sampling in
frequency domain can be written as,

( )= ( ) ( )= ( ) ( − ) (6.32)

We have, in frequency domain,

1 ( )
= (6.33)
2

Fig 6.22: Discrete time sampling


Fourier Transform of sampling space ( ) is,

2
= ( − ) (6.34)
Signals & Systems | 253

Equation (6.42) is the counterpart for discrete-time sampling of eq. (6.6) for continuous-time
sampling and is illustrated in figure 6.21. In Figure 6.21(c), with − > , or
equivalently, > there is no aliasing (i.e., the nonzero portions of the replicas of ( )
do not overlap), whereas with < , as in figure 6.21(d), frequency domain aliasing results.
In the absence of aliasing, ( ) is faithfully reproduced around = 0 and integer multiples
of 2 . Consequently, ( ) can be recovered from ( ) by means of a lowpass filter with gain
cutoff frequency greater than and less than − , where we have specified cutoff
frequency of the low pass filter as /2. If overall system from figure 6.22 (a) is applied to a
sequence for which < , so that there are aliasing results ( ) will no longer will equal
to ( ). But with continuous time sampling, the two sequences will be equal at multiple time
of sampling period. Now, we have,
( ) = ( ), = 0, ±1, ±2 … … …, (6.35)
independently of whether aliasing occurs

Fig 6.23: Impulse-train sampling of discrete-time signal in frequency domain: (a) Original signal
spectrum; (b) Spectrum of sampling sequence; (c) Spectrum of sampled signal with >
; (d) Spectrum of sampled signal with < . No aliasing occurs.
254 | Sampling & Reconstruction

Fig 6.24: Exact recovery of discrete time signal from its samples using an ideal low pass filter:
(a) Block diagram for sampling & reconstruction; (b) spectrum of ( );
Example 6.3
Consider a sequence of ( ) whose Fourier transform ( ) has property that
=0 2 /9 ≤ | | ≤
To determine the lowest rate at which ( ) may be sampled without the possibility of aliasing,
we must find the largest N such that
2 2 9
≥2 ⟹ ≤
9 2
We conclude that = 4 and corresponding sampling frequency is =
Signals & Systems | 255

The reconstruction of ( ) using a lowpass filter applied to ( ) can be interpreted in the


time domain as an interpolation formula like eq. (6.10). With ℎ( ) denoting the impulse
response of the lowpass filter, we have
sin( )
ℎ( ) = (6.36)

The reconstructed sequence is then


( )= ( ) ∗ ℎ( ) (6.37)
Or equivalently,

( )= ( )ℎ ( − ) (6.38)

Where ℎ ( ) is impulse response of interpolating filter

6.5.2 Discrete time decimation and interpolation


There are a variety of important applications of the principles of discrete-time sampling, such
as in filter design and implementation or in communication applications. In many of these
applications it is inefficient to represent, transmit, or store the sampled sequence ( ) directly
in the form depicted in Figure 6.20, since, in between the sampling instants, ( ) is known to
be zero. Thus, the sampled sequence is typically replaced by a new sequence ( ), which is
simply every Nth value of ( ); that is,
( )= ( ) (6.39)
Similarly,
( )= ( ) (6.40)
since ( ) and ( ) are equal at integer multiples of N. The operation of extracting every
sample is commonly referred to as decimation. The relationship between ( ), ( ), ( )
is illustrated in figure 6.23.
To determine the effect of decimation in the frequency domain, we wish to determine the
relationship between ( ) the Fourier transform of ( ) and ( ). To this end, we note
that,

= [ ] (6.41)

Using the Eq. (6.39)


256 | Sampling & Reconstruction

= [ ] (6.42)

If we let = , or equivalently = / , we can write

/
= [ ] (6.43)

And since [ ] = 0 when n is not integer multiple of N, we can also write,

[ ] /
= (6.44)

Fig 6.25: Relationship between [ ] corresponding to sampling and [ ] corresponding to


decimation

Also, right hand side of equation (6.44) as the Fourier transform of [ ]; so,

[ ] / /
= ( ) (6.45)

Thus, from equation (6.44) & (6.45) we can conclude that


/
= ( ) (6.46)
Signals & Systems | 257

This relationship is illustrated in Figure 6.24, and from it, we observe that the spectra for the
sampled sequence and the decimated sequence differ only in a frequency scaling or
normalization. If the original spectrum ( ) is appropriately band limited, so that there is no
aliasing present in ( ), then, as shown in the figure, the effect of decimation is to spread
the spectrum of the original sequence over a larger portion of the frequency band.

Fig 6.26: Frequency domain illustration of the relationship between sampling & decimation

If the original sequence ( ) is obtained by sampling a continuous-time signal, the process of


decimation can be viewed as reducing the sampling rate on the signal by a factor of N. To avoid
aliasing, ( ) cannot occupy the full frequency band. In other words, if the signal can be
decimated without introducing aliasing, then the original continuous­ time signal was
oversampled, and thus, the sampling rate can be reduced without aliasing. With the
interpretation of the sequence ( ) as samples of a continuous-time signal, the process of
decimation is often referred to as downsampling.
258 | Sampling & Reconstruction

Fig 6.27: Continuous time signal that was originally sampled at Nyquist rate. After discrete time
filtering, the resulting sequence can be further downsampled. Here ( ) is the
continuous time Fourier Transform of ( ), ( ) are the discrete time
Fourier transforms of ( ) & ( ) respectively. And is the frequency response
of the discrete time low pass filter depicted in the block diagram.

In some applications in which a sequence is obtained by sampling a continuous­ time signal,


the original sampling rate may be as low as possible without introducing aliasing, but after
additional processing and filtering, the bandwidth of the sequence may be reduced. An example
of such a situation is shown in Figure 6.25. Since the output of the discrete-time filter is band
limited, downsampling or decimation can be applied.

Just as in some applications it is useful to downsample, there are situations in which it is useful
to convert a sequence to a higher equivalent sampling rate, a process referred to as upsampling
or interpolation. Upsampling is basically the reverse of decimation or downsampling. As
Signals & Systems | 259

illustrated in Figures 6.23 and 6.24, in decimation we first sample and then retain only the
sequence values at the sampling instants. To upsample, we reverse the process. For example,
referring to Figure 6.23, we consider upsampling the sequence [ ] to obtain [ ]. From
[ ], we form the sequence [ ] by inserting N - 1 points with zero amplitude between each
of the values in [ ]. The interpolated sequence [ ] is then obtained from [ ] by lowpass
filtering. The overall procedure is summarized in Figure 6.26.

Fig 6.28: Upsampling: (a) Overall system; (b) associated sequences and spectra for
upsampling by a factor of 2.

Example 6.4

In this example, we illustrate how a combination of interpolation and decimation may be used
to further downsample a sequence without incurring aliasing. It should be noted that maximum
possible downsampling is achieved once the non-zero portion of one period of the discrete-time
spectrum has expanded to fill the entire band from − + .
260 | Sampling & Reconstruction

Consider the sequence [ ] whose Fourier transform ( ) is illustrated in Figure 6.27(a).


The lowest rate at which impulse-train sampling may be used on this sequence without
incurring aliasing is 2 /4. This corresponds to,

Fig 6.29: Spectra associated with example 6.4: (a) Spectrum of [ ]; (b) Spectrum
after downsampling by 4; (c) Spectrum after upsampling of ( ) by factor of 2;
(d) Spectrum after upsampling [ ] by 2 then downsampling by 9
Signals & Systems | 261

sampling every 4th value of ( ). If the result of such sampling is decimated by a factor of 4,
we obtain a sequence [ ] whose spectrum is shown in Figure 6.27(b). Clearly, there is still
no aliasing of the original spectrum. However, this spectrum is zero for 8 /9 ≤ | | ≤ , which
suggests there is room for further downsampling.

Specifically, examining Figure 6.27(a) we see that if we could scale frequency by a factor of
9/2, the resulting spectrum would have nonzero values over the entire frequency interval from
− + . However, since 9/2 is not an integer, we cannot achieve this purely by
downsampling. Rather we must first upsample [ ] by a factor of 2 and then downsample by a
factor of 9. In particular, the spectrum of the signal [ ] obtained when [ ] is upsampled by
a factor of 2, is displayed in Figure 6.27(c). When [ ] is then downsampled by a factor of 9,
the spectrum of the resulting sequence [ ] is as shown in Figure 6.27(d). This combined
result effectively corresponds to downsampling [ ] by a no integer amount, 9/2. Assuming
that [ ] represents unaliased samples of a continuous-time signal ( ), our interpolated and
decimated sequence represents the maximum possible (aliasing-free) downsampling of ( ).

Unit Summary
In this chapter we have developed the concept of sampling, whereby a continuous-time or
discrete-time signal is represented by a sequence of equally spaced samples. The conditions
under which the signal is exactly recoverable from the samples is embodied in the sampling
theorem. For exact reconstruction, this theorem requires that the signal to be sampled be band
limited and that the sampling frequency be greater than twice the highest frequency in the signal
to be sampled. Under these conditions, exact reconstruction of the original signal is carried out
by means of ideal lowpass filtering. The time-domain interpretation of this ideal reconstruction
procedure is often referred to as ideal band­ limited interpolation. In practical implementations,
the lowpass filter is approximated and the interpolation in the time domain is no longer exact.
In some instances, simple interpolation procedures such as a zero-order hold or linear
interpolation (a first-order hold) suffice.
If a signal is undersampled (i.e., if the sampling frequency is less than that required by the
sampling theorem), then the signal reconstructed by ideal band-limited interpolation will be
related to the original signal through a form of distortion referred to as aliasing. In many
instances, it is important to choose the sampling rate to avoid aliasing. However, there are a
variety of important examples, such as the stroboscope, in which aliasing is exploited.
Sampling has several important applications. One particularly significant set of applications
relates to using sampling to process continuous-time signals with discrete­ time systems, by
means of minicomputers, microprocessors, or any of a variety of devices specifically oriented
toward discrete-time signal processing.
The basic theory of sampling is similar for both continuous-time and discrete­ time signals. In
the discrete-time case there is the closely related concept of decimation, whereby the decimated
262 | Sampling & Reconstruction

sequence is obtained by extracting values of the original sequence at equally spaced intervals.
The difference between sampling and decimation lies in the fact that, for the sampled sequence,
values of zero lie in between the sample values, whereas in the decimated sequence these zero
values are discarded, thereby compressing the sequence in time. The inverse of decimation is
interpolation.

Solved examples on DTFS


Example 6.1:
Shown in figure below is a system in which sampling signal is an impulse train with alternating
sign

The sampling signal ( ), the Fourier transform of signal ( ) and frequency response of filter
are shown below

(a) For Δ < , Sketch the Fourier Transform of ( ) and ( )

(b) For Δ < , Determine a system that will recover ( ) from ( ) and another that will
recover ( ) from ( )
(c) What is the maximum value of ∆ in relation to for which ( ) can be recovered from
either ( ) or ( )
Solution:
We know that ( ) = ( ) ( ), by dual of convolution theorem, we have,
Signals & Systems | 263

( ) = ( ) ( ),
So we first find Fourier transform of ( ) as follows
The Fourier transform of periodic function is an impulse train at intervals of = =

Strength of impulse at being

= ( )
Δ
= 1− = 1−
Δ Δ
= (1 − (−1) )
Δ
Thus, we can sketch ( )

Thus, we can also sketch ( ) and hence ( ):

(b) Recover ( ) from ( )


Modulate ( ) with cos( )
264 | Sampling & Reconstruction

cos( ) has spectrum with impulses of equal strength at & − . Thus, new signal will have
copies of the original spectrum (modulated by constant of course) at all even multiples of

Now, an appropriate low pass filter can extract the original spectrum
To recover ( ) from ( )
Here too, notice from the figures that modulation with cos( ) will do the job. Here too the
modulated signal will have copies of the original spectrum at all even multiples of
(c) So long as adjacent copies of the original spectrum do not overlap in ( ), theoretically
one can reconstruct the original signal. Therefore, the condition is,
2
2 < ⟹Δ<
Δ
Example 6.2:
The signal ( ) is obtained by convolving signals ( ) and ( ) where:
| ( )| = 0 | | > 1000 &
| ( )| = 0 | | > 2000
Impulse train sampling is performed on ( ) to get

( )= ( ) ( − )

Specify the range of values of T so that ( ) may be recovered from ( )


Solution:
By Convolution theorem
( )= ( ) ( )
( )=0 | | > 1000
Thus, from the sampling theorem, the sampling rate must exceed 2 ∗ = 1000
So, T must be less than 10 i.e. 1mSec
Example 6.3:
In the figure below, we have a sampler, followed by an ideal low pass filter, for reconstruction
of ( ) from its samples ( ). From sampling theorem, we know that if = is
greater than twice the highest frequency present in ( ) = , then reconstructed signal
will exactly equal ( ). If this condition on bandwidth of ( ) is violated, then ( ) will
not equal ( ). We seek to show in this problem that if = then for any choice of T,
Signals & Systems | 265

( ) & ( ) will be always equal at the sampling instants; that is, ( )= ( ), =


0, ±1, ±2, … …

To obtain this result, consider the following equation which expresses ( ) in terms of samples
( ):
sin[ ( − )]
( )= ( )
( − )

With = this becomes,

sin ( − )
( )= ( )
( − )
[ ]
By considering value of for which = 0, show that without any restrictions on
( ), ( )= ( ) for any integer value of k.
Solution:
To show that ( ) and ( ) are equal at the sampling instants, consider
sin ( − )
lim ( ) = lim ( )
→ → ( − )

sin ( − )
= lim ( )
→ ( − )

sin ( − ) sin ( − )
= ( ) + lim ( )
( − ) → ( − )
,

sin[ ( − )] sin ( − )
= ( ) + ( ) lim
( − ) → ( − )
,
266 | Sampling & Reconstruction

=0+ ( )×1 ∵ ( − ) ∈ ⟺ sin[ ( − )] = 0 & lim =1



Thus,
lim ( )= ( )

Assuming continuity of ( ) at = , ( )= ( ), ∀ ∈
Example 6.4:
This problem deals with one procedure of bandpass sampling & reconstruction. This procedure,
used when ( ) is real, consists of multiplying ( ) by a complex exponential and then
sampling the product. The sampling system is shown below figure a with ( ) real with
( ) non-zero only for < | | < , the frequency chosen to be = ( + )
and low pass filter ( ) has cutoff frequency ( + )
(a) For ( ) shown in figure b, Sketch ( )
(b) Determine maximum sampling period T such that ( ) is recoverable from ( )
(c) Determine a system to recover ( ) from ( )

Solution:
(a) Multiplication by the complex exponential in time domain is equivalent to shifting
left the Fourier transform by an amount in frequency domain. Therefore, resultant
transform looks as shown below
Signals & Systems | 267

After passing through filter, the Fourier transform becomes,

Now sampling the signal amounts to making copies of the Fourier transform, the center of each
separated from the other by the sampling frequency in the frequency domain. Thus,
( ) has the following form
268 | Sampling & Reconstruction

(b) ( ) is recoverable from ( ) only if the copies of Fourier transform obtained by sampling
do not overlap with each other. For this to happen, the condition set down by Shannon-
Nyquist theorem for a band-limited signal must be satisfied i.e. the sampling frequency
should be greater than twice the bandwidth of original signal. Mathematically,
>2
>2
2
<

Hence the maximum sampling period for ( ) to be recoverable from ( ) is
(c) The system to recover ( ) from ( ) is outlined below:

Example 6.5:
Shown in the figures is a system in which the sampling signal is an impulse train with alternating
sign. The Fourier transform of the input signal is as indicated in figures below.
(i) For Δ < , Sketch the Fourier Transform of ( ) and ( )

(ii) For Δ < , determine a system that will recover ( ) from ( )

(iii) For Δ < , determine a system that will recover ( ) from ( )


(iv) What is the maximum value of Δ in relation to for which ( ) can be recovered from
either ( ) or ( )?
Signals & Systems | 269
270 | Sampling & Reconstruction

Solution:
(a) As ( ) = ( ) ( ), by dual of convolution theorem we have ( )= ( ) ( ) so
we first find Fourier transform of ( ) as follows
The Fourier transform of a periodic function is an impulse train at intervals of = =
Each impulse being of magnitude:

( ) = 2 /2Δ ( )

= /Δ(1 − cos( ))
Here, we see that the impulses on axis vanish at even values of k
Hence, Fourier transform of ( ) is as shown in figure (a). In the frequency domain, the
output signal Y can be found by multiplying the input with the frequency response. Hence
( ) is as shown below in figure (b)
Signals & Systems | 271

(b) To recover ( ) from ( ), we do the following two things:


1. Modulate the signal by
cos((2 /Δ) )
2. Apply a low pass filter of bandwidth /2Δ
(c) To recover ( ) from ( ) we do following two things
1. Modulate signal by 2 cos((2 /Δ) )
2. Apply a low pass filter of bandwidth /2Δ
(d) Maximum value for recoverability is / as can be seen from graphs
272 | Sampling & Reconstruction

Example 6.6:
A signal ( ) with Fourier transform ( ) undergoes impulse train sampling to generate

( )= ( ) ( − )

Where = 10 For each of the following set of constraints on ( ) and/or ( ), does the
sampling theorem guarantee that ( ) can be recovered exactly from ( ) ?
(a) ( ) = 0 | | > 5000
(b) ( ) = 0 | | > 15000
(c) ℜ{ ( )} = 0 | | > 5000
(d) ( ) real and ( ) = 0 | | > 5000
(e) ( ) real and ( ) = 0 | | < −15000
(f) ( ) ∗ ( ) = 0 | | > 15000
(g) | ( )| = 0 > 5000
Solution:
We have = 10
So, = 20000
(a) ( ) = 0 | | > 5000
2 = 10000
Here, obviously <2
Hence ( ) can be recovered exactly from ( )
(b) ( )=0 | | > 15000
2 = 30000
Here, obviously <2
Hence ( ) can be recovered exactly from ( )

(c) ℜ{ ( )} = 0 | | > 5000


Real part of ( ) = 0, but we cannot say anything particular about imaginary part of the
( ), thus, not necessary that ( ) = 0 for this range
Hence ( ) cannot be recovered exactly from ( )
(d) ( ) real and ( )=0 | | > 5000
As ( ) is real we have ( ) = (− )
Taking mod on both sides
Signals & Systems | 273

( ) = (− )=0 > 5000


⟹ | (− )| = | (− )| = 0 > 5000
⟹ (− )=0 > 5000
⟹ ( )=0 < −5000
So, we get
( )=0 | | > 5000
Here, obviously >2
Hence ( ) can be recovered exactly from ( )
(e) ( ) real and ( )=0 | | < −15000
Proceeding as above we get
( )=0 | | > 15000
Here, obviously <2
Hence ( ) can not be recovered exactly from ( )
(f) ( )∗ ( )=0 | | > 15000
When we convolve two functions with domain and then domain of their
convolution function varies from + to +
Here, = & =
2 = 15000
⟹ = 7500
Therefore,
( )=0 | | > 7500
Here, obviously >2
Hence ( ) can be recovered exactly from ( )
(g) | ( )| = 0 > 5000
We can not say anything about ( ) < −5000 ,
Hence ( ) can not be recovered exactly from ( )
274 | Sampling & Reconstruction

Example 6.7:
Figure I shows overall system for filtering a continuous time signal using a discrete time filter.
If ( ) and ( ) are as shown in figure II with = 20 sketch ( ), ( ),
( ), ( )& ( )

Solution:
Signals & Systems | 275

EXERCISES

1) The signal ( ) = sin(14000 ), where t is in seconds, is sampled at a rate of 9000 samples


per second. The sampled signal is the input to an ideal lowpass filter with frequency
response ( ) as following:
1, | | ≤ 12
( )=
0, | | > 12

What is number of sinusoids in the output and their frequency in KHz?


276 | Sampling & Reconstruction
( / )
2) Consider a continuous time signal defined as ( ) = ∗∑ ( − 10 ). Where
( )
* denotes the convolution operation and t is in seconds. Calculate the Nyquist sampling
rate for x(t) in (samples/sec).
( ) ( )
3) Calculate the Nyquist sampling rate for the signal ( ) = ×
4) Calculate minimum sampling frequency (in samples/sec) required to reconstruct the
following signal from its samples without distortion
sin 2 1000 sin 2 1000
( )=5 +7

5) A signal ( ) with bandwidth 500Hz is first multiplied by a signal g(t) where

( )= (−10) ( − 0.5 10 )

The resulting signal is then passed through an ideal low pass filter with bandwidth 1KHz.
Write the output of the low pass filter.
6) A 1KHz signal is ideally sampled at 1500 samples/sec and the sampled signal is passed
through an ideal low pass filter with cut off frequency of 800 Hz. Calculate frequency of
the output signal
7) A signal ( ) = 100 cos(24 × 10 ) is ideally sampled with sampling period of 50
and then passed through an ideal low pass filter with cutoff frequency of 15 KHz. What
will be the frequencies at the output?
8) A 4 GHz carrier is DSB-SC modulated by a low pass message signal with maximum
frequency of 2 MHz. The resultant signal is to be ideally sampled. Find the minimum
frequency of the sampling impulse train.
9) Find the Nyquist sampling interval for the signal (700 ) + (500 )
10) A low pass signal ( ) band-limited to B Hz is sampled by a periodic rectangular pulse
train ( ) of period = 1/3 sec. Assuming natural sampling and that the pulse
amplitude and pulse width are A volts 1/30 sec., respectively, obtain an expression for
the frequency spectrum of the sampled signal ( ).
11) A real valued signal ( ) is known to be uniquely determined by its samples, when the
sampling frequency is = 10000 . For what values of is ( ) guaranteed to be
zero.
12) A continuous time signal ( ) at the output of of ideal low pass filter with cutoff frequency
= 1000 . If impulse train sampling is performed on ( ), Which of the following
sampling periods would guarantee that ( ) can be recovered from its sampled version
using an appropriate low pass filter
(a) = 0.5 × 10
Signals & Systems | 277

(b) = 2 × 10
(c) = 10
13) For the following figure shows a system consisting of a continuous time LTI system
followed by a sampler, conversion to a sequence, an LTI discrete time system. The
continuous time LTI system is causal and satisfies the linear, constant coefficient
differential equation,
( )
+ ( )= ( )
The input ( ) is unit impulse ( )
(a) Determine ( )
(b) Determine frequency response ( ) and impulse response ℎ[ ] such that [ ]= [ ]

14) A signal ( ) is obtained through impulse train sampling of a sinusoidal signal ( ) whose
frequency is equal to half the sampling frequency
( ) = cos +
2
And

( )= ( ) ( − )

Where = 2 /
(a) Find ( ) such that
( ) = cos cos + ( )
2
(b) Show that
( )=0 = 0, ±1, ±2, … …
15) Suppose [ ] = cos + with 0 ≤ ≤ 2 and [ ] = [ ] ∑ [ − 4 ],
what additional constraints must be imposed on to ensure that
278 | Sampling & Reconstruction

sin 4
[ ]∗ = [ ]
4
16) With reference to the filtering approach, assume that sampling period used is and input
( ) is band limited so that ( ) = 0 for | | ≥ / . If the overall system has the
property ( ) = ( − 2 ) determine the impulse response ℎ[ ] of the discrete time
filter.
17) Repeat the previous problem except this time assume that
( )= − .
18) Impulse train sampling of ( ) is used to obtain

[ ]= [ ] [ − ]

If = 0 for 3 /7 ≤ | | ≤ , determine the largest value for the sampling interval


which ensures that no aliasing takes place while sampling [ ].
19) The following facts are given about the signal [ ] and its Fourier transform:
1. [ ] is real
2. ≠0 0< <
3. [ ] ∑ [ −2 ] = [ ]
Determine [ ].

20) Consider a signal ( ) = ,


Which we wish to sample with a sampling frequency of = 150 to obtain a signal ( ) with
Fourier transform ( ). Determine the maximum value of for which it is guaranteed that
( ) = 75 ( ) | |≤
Where ( ) is Fourier transform of ( ).

Multiple-Choice Questions
1. Sampling can be done by:
a) Impulse train sampling
b) Natural sampling
c) Flat-top sampling
d) All of above
Ans: d)
Signals & Systems | 279

2. Sampled data technique is appropriate as:


a) For long distance data transmission
b) Pulses are transferred by little loss of accuracy
c) More than one channel of information is sequentially sampled and transmitted.
d) All of the mentioned
Ans: d)
3. Signal sampling reduces the power demand made on the signal.
a) True
b) False
Ans: a)
4. The signal is reconstructed back with the help of
a) Zero order hold circuits
b) Extrapolations
c) Signal is reconstructed with zero order holds and extrapolations
d) Signal is not reconstructed
Ans: c)
5. Aliasing is caused when:
a) Sampling frequency must be equal to the message signal
b) Sampling frequency must be greater to the message signal
c) Sampling frequency must be less to the message signal
d) Sampling frequency must be greater than or equal to the message signal
Ans: c)
6. The first step required to convert Analog signal to digital is:
a) Sampling
b) Holding
c) Reconstruction
d) Quantization
Ans: a)
7. _______________ is a sampling pattern which is repeated periodically
a) Single order sampling
b) Multi order sampling
c) Zero order sampling
d) Unordered sampling
Ans: b)
8. Choose minimum sampling rate required to avoid aliasing when continuous time signal
( ) = 5 cos 400 is sampled:
280 | Sampling & Reconstruction

a) 200 Hz
b) 500 Hz
c) 400 Hz
d) 450Hz
Ans: c)
9. Find Nyquist rate and Nyquist interval of sinc[t]
a) 1 Hz, 1 sec
b) 2 Hz, 2 sec
c) ½ Hz, 2 sec
d) 2 Hz, ½ sec
Ans: a)
10. Determine the Nyquist rate of signal ( ) = 1 + cos 2000 + sin 4000
a) 2000 Hz
b) 4000 Hz
c) 500 Hz
d) 3000 Hz
Ans: b)
11. Which of the following requires interpolation filtering?
a) UP-Sampler
b) D to A Converter
c) Both (a) & (b)
d) None of these
Ans: c)
12. Which process requires Low Pass Filter.
a) UP-sampling
b) Down-sampling
c) Up-sampling & Down-sampling
d) None of the above mentioned
Ans: c)
13. Which device is needed for the reconstruction of signal?
a) Low pass filter
b) Equalizer
Signals & Systems | 281

c) Low pass filter & Equalizer


d) None of the above mentioned
Ans: c)
14. Decreasing the data rate is called as
a) Aliasing
b) Down-sampling
c) Up-sampling
d) None of the above mentioned
Ans: b)
15. Instantaneous sampling
a) Has a train of impulses
b) Has the pulse width approaching zero value
c) Has negligible power content
d) All of the above
Ans: d)
16. The spectrum of sampled signal may be obtained without overlapping only if
a) ≥ 2
b) <2
c) >
d) <
Ans: a)
17. Decimation is a process in which the sampling rate is __________
a) Enhanced
b) Stable
c) Reduced
d) Unpredictable
Ans: c)
18. To change the sampling rate for better efficiency in two or multiple stages, the decimation
& interpolation factors must be ______________ unity.
a) Greater than
b) Less than
c) Equal to
d) None of the above
282 | Sampling & Reconstruction

Ans: a)
19. How is the sampling rate conversion achieved by factor I/D
a) By increase in sampling rate with (I)
b) By filtering the sequence to remove unwanted images of spectra of original signal.
c) By decimation of filtered signal with factor D
d) All of the above
Ans: d)
20. The first step required to convert Analog signal to digital is:
a) Aliasing
b) Holding
c) Quantization
d) Sampling
Ans: d)

KNOW MORE
Our treatment of sampling is concerned primarily with the sampling theorem and its
implications. However, to place this subject in perspective we begin by discuss the general
concepts of representing a continuous-time signal in terms of its samples and the reconstruction
of signals using interpolation. After using frequency-domain methods to derive the sampling
theorem, we consider both the frequency and time domains to provide intuition concerning the
phenomenon of aliasing resulting from under sampling. One of the very important uses of
sampling is in the discrete-time processing of continuous time signals. Which is discussed
thoroughly. Following this, we turn to the sampling of discrete-time signals. The basic result
underlying discrete-time sampling is developed in a manner that parallels that used in
continuous time, and the applications of this result to problems of decimation and interpolation
are described. Again, a variety of other applications, in both continuous and discrete time, are
addressed in the problems.
Both the sampling and reconstruction are critical in maintaining the integrity and fidelity of
signals as they transition between the continuous and discrete domains. Careful consideration
of the sampling rate, anti-aliasing, and reconstruction techniques is essential to avoid signal
degradation and ensure accurate representation. These concepts are foundational in various
fields, including telecommunications, audio processing, image processing, and more.
Signals & Systems | 283

REFERENCES AND SUGGESTED READINGS


1. Signals and Systems by Simon Haykin
2. Signals and Systems by Ganesh Rao
3. Signals and Systems - Course (nptel.ac.in)

Dynamic QR Code for Further Reading


284 | Mapping

CO AND PO ATTAINMENT TABLE

Course outcomes (COs) for this course can be mapped with the programme outcomes (POs)
after the completion of the course and a correlation can be made for the attainment of POs to
analyze the gap. After proper analysis of the gap in the attainment of POs necessary measures
can be taken to overcome the gaps.

Table for CO and PO attainment

Expected Mapping with Programme Outcomes


Course (1- Weak Correlation; 2- Medium correlation; 3- Strong Correlation)
Outcomes
PO-1 PO-2 PO-3 PO-4 PO-5 PO-6 PO-7 PO-8 PO-9 PO-10 PO-11 PO-12

CO-1

CO-2

CO-3

CO-4

CO-5

CO-6
Index| 285
INDEX

Aliasing, 238 Laplace transform, 153, 154


Convolution, 61, 204 Left sided signal, 156
Cascade systems, 64 Magnitude, 112
Complex frequency, 153, 154 Periodic Signal, 23, 24
Conversion, 246, 247 Parseval’s Theorem, 118, 137
Continuous time signal, 6 Poles, 173
Continuous-Time Unit Impulse, 8 Power series, 213
CT Unit Impulse Function Properties, 9 Partial fraction, 216
CT Unit Step Function, 10 Ramp signal, 14, 15
CT Signum Function, 13 Region of convergence, 155, 194
CT Ramp Function, 14 Right sided signal, 155
CT Exponential signal, 15 Reconstruction, 236
Complex Exponential Signal, 16 Real Exponential Signal, 16
Complex Exponential Sequence, 17 Real Exponential Sequence,17
CT Sinusoid Signal, 19 Step signal, 11, 12
Discrete signal, 25 Signum signal, 14
Differential equation, 67, 68 Sinusoid signal, 18, 19
Exponential signal, 16, 17, 18 System, 37, 38, 39, 42
Energy signal, 27 State space representation, 69, 70
ECG Signal, 5 Sampling theorem, 230, 233
Fourier series, 89, 90, 94 Strobe effect, 244
Fourier transform, 111, 112, 120 Waveform symmetry, 96, 97
Final value theorem, 172 Z transform, 190, 191
Gibbs phenomenon, 106, 107 Z-plane, 192
Impulse signal, 8, 9, 10 Zeros, 193
Inverse Laplace transform, 155, 175
Initial value theorem, 206
Inverse Z transform, 212
LTI system, 57, 58

You might also like