Ad8701 DL Unit5 Notes
Ad8701 DL Unit5 Notes
This document is confidential and intended solely for the educational purpose of
RMK Group of Educational Institutions. If you have received this document through
email in error, please notify the system manager. This document contains proprietary
information and is intended only to the respective group / learning community as
intended. If you are not the addressee you should not disseminate, distribute or
copy through e-mail. Please notify the sender immediately by e-mail if you have
received this document by mistake and delete this document from your system. If
you are not the intended recipient you are notified that disclosing, copying,
distributing or taking any action in reliance on the contents of this information is
strictly prohibited.
AD8701
DEEP LEARNING
Department : AI & DS
Batch/Year : 2020- 2024 / IV
Created by : Ms.A.AKILA
Date : 29.09.2023
Signature :
1. TABLE OF CONTENTS
S NO CONTENTS SLIDE NO
1 Contents 5
2 Course Objectives 7
9
3 Pre Requisites (Course Names with Code)
4 Syllabus (With Subject Code, Name, LTPC details) 11
5 Course Outcomes 12
7 Lecture Plan 17
16 Assessment Schedule 61
Course Objectives
2. COURSE OBJECTIVES
PRE REQUISITES
3. PRE REQUISITES
PRE-REQUISITE CHART
MA8391 Probability
and Statistics
AD8402 Artificial
Intelligence I
AD8552 Machine
Learning
Syllabus
4. SYLLABUS
Course Outcomes
4. COURSE OUTCOME
Cognitive/
Affective
Course
Course Outcome Statement Level of the
Code Course
Outcome
CO1 Explain the basics in deep neural networks Understand
K2
CO – PO/PSO Mapping
5. CO-PO/PSO MAPPING
P P P P P P P P P P P P PS PS PS
Course O O O O O O O O O O O O O O O
Outcomes 1 2 3 4 5 6 7 8 9 10 11 12 1 2 3
(Cos)
K K K3 A
K4 K5 /K A2 A3 A3 A3 A3 A2 K3 K3 K3
3 5 3
5
CO1 K
2 2 1 2 3 3 3
CO2 K
3 2 1 2 3 3 3
CO3 K
2 3 2 1 1 3 3 3 3 3
K
5
CO4
3 3 3 2 2 3 3 3 3 3
CO5 K
3 3 2 1 1 3 3 2 2 2
12
Lecture Plan
Unit V
Detection in chest X-
1 ray images -object CO5 K3
1 PPT
detection and
classification 20.10.2023
CO5 K3
RGB and depth image 1 PPT
2 fusion - NLP tasks 20.10.2023
dimensionality CO5 K3
1
estimation - time series PPT
3 forecasting 21.10.2023
CO5 K3
dimensionality
estimation - time series PPT
4 forecasting 1 02.11.2023
building electric power CO5 K3
grid for controllable PPT
1
5 energy resources 03.11.2023
building electric power CO5 K3
6 grid for controllable PPT
1
energy resources 03.11.2023
guiding charities in CO5 K3
maximizing donations 1
17
ACROSS
1. step in the data science life cycle where hypotheses are formed and data is explored.
3. step in data modeling where relevant features are selected or created.
4. used for data visualization and exploration.
5. for defining the project scope, objectives, and key performance indicators (KPIs).
6. person responsible for extracting insights from data and making data-driven decisions.
7. phase where insights are communicated to stakeholders and used for decision-making.
8. energy-based modeling approach used in Boltzmann Machines.
DOWN
2. for designing experiments, collecting and analyzing data, and drawing conclusions.
3. for converting categorical data into a numerical format for modeling.
4. used for data storage, retrieval, and preprocessing.
6. final step in the data science life cycle where the model is deployed and monitored.
7. phase where data is collected, prepared, and made ready for analysis.
Lecture Notes
UNIT V
Chest X-ray image analysis holds a significant position within the field of
medical imaging. These images serve as invaluable diagnostic tools for
identifying a wide range of respiratory and cardiac conditions.
One of the primary reasons for the importance of chest X-ray image analysis
is its potential to facilitate early detection of abnormalities. This early
detection can lead to timely treatment interventions, ultimately improving
patient outcomes. Additionally, automated analysis of chest X-rays allows
radiologists to efficiently review a large volume of patient data, increasing
the efficiency of healthcare delivery.
• Chest X-ray images play a crucial role in medical diagnostics, particularly for
detecting respiratory abnormalities.
2. Labeling:
3. Model Architecture
Convolutional Neural Networks (CNNs) are a popular choice for
analyzing chest X-ray images due to their effectiveness in image-
related tasks.
It consists of multiple convolutional layers with rectified linear unit
(ReLU) activation functions, followed by max-pooling layers to
reduce the spatial dimensions of the feature maps.
The output of the convolutional layers is then flattened and fed
into a fully connected layer with a ReLU activation function.
4. Convolutional Layers
• The core building blocks of a CNN are convolutional layers.
• These layers use filters (kernels) to scan the input image for specific
patterns or features.
• In the context of chest X-rays, these patterns could be indicative of
abnormalities like nodules, opacities, or other relevant structures.
5. Pooling Layers
Pooling layers follow convolutional layers and are used to reduce the
spatial dimensions of the input volume.
Max pooling, for example, retains the most important information
from a group of pixels, helping to decrease the computational load
and focus on the most relevant features
6. Activation Functions
Non-linear activation functions like ReLU (Rectified Linear Unit) are
commonly used to introduce non-linearity into the model.
This allows the CNN to learn complex patterns and relationships
within the data.
8. Output Layer
The output layer typically consists of one or more neurons,
depending on the specific classification task.
9. Evaluation:
• Evaluate the trained model on the test set to assess its generalization
performance.
• Metrics such as accuracy, precision, recall, and F1 score are commonly
used to quantify performance.
10.Deployment:
• Once satisfied with the model's performance, deploy it for real-world
use.
• This could involve integrating it into a healthcare system to assist
radiologists in detecting abnormalities in chest X-rays.
Model Architecture
R-CNN
The feature extractor used by the model was the AlexNet deep CNN.
Training is a multi-stage pipeline. Involves the preparation and operation
of three separate models.
The output of the CNN is then interpreted by a fully connected layer then the
model bifurcates into two outputs, one for the class prediction via a softmax
layer, and another with a linear output for the bounding box. This process is
then repeated multiple times for each region of interest in a given image.
Object Detection
Introduction
RGB and depth image fusion is a technique that combines information from
conventional RGB images and depth maps. This fusion process enhances the
perception and understanding of the environment, enabling various applications in
robotics, augmented reality, and computer vision.
The fusion of RGB and depth information allows machines to perceive the world
more like humans. It enhances the ability to understand object shapes, distances,
and their interactions with the environment. This has significant implications for
fields such as robotics and augmented reality.
Applications
Robotics
RGB and depth fusion is vital for robotic perception. Robots equipped with depth
sensors can navigate and interact with their surroundings more effectively. They
can identify obstacles, manipulate objects, and plan paths with greater precision.
Augmented Reality
In augmented reality applications, combining RGB and depth data allows for realistic
object placement and interaction with the real world. This technology is used in
games, simulations, and various interactive experiences.
Fusion Techniques
Depth sensors like Microsoft Kinect and LiDAR systems capture depth information using
structured light or time-of-flight principles. These sensors provide accurate depth maps
that complement RGB images.
Use Cases
Gesture Recognition
RGB and depth fusion is pivotal in gesture recognition systems. By analyzing both RGB
and depth data, computers can accurately detect and interpret hand gestures. This is
used in gaming consoles, human-computer interaction, and sign language recognition.
In computer vision tasks, combining RGB and depth information enhances object
recognition and tracking. It enables algorithms to differentiate between objects with
similar appearances but different 3D shapes, improving the accuracy of tracking and
identification.
In conclusion, RGB and depth image fusion is a powerful technique with diverse
applications in robotics, augmented reality, and computer vision. Advancements in
depth sensing technology and deep learning approaches are poised to drive further
innovation in this field, enabling more accurate and robust fusion systems for a wide
range of applications.
4. NLP tasks
Overview of NLP
● Pre-processing
● Training
Researchers use the pre-processed data and machine learning to train NLP models to
perform specific applications based on the provided textual information. Training NLP
algorithms requires feeding the software with large data samples to increase the
algorithms' accuracy.
Machine learning experts then deploy the model or integrate it into an existing
production environment. The NLP model receives input and predicts an output for the
specific use case the model's designed for. You can run the NLP application on live
data and obtain the required output.
Supervised NLP : Supervised NLP methods train the software with a set of labeled
or known input and output. The program first processes large volumes of known data
and learns how to produce the correct output from any unknown input. For example,
companies train NLP tools to categorize documents according to specific labels.
NLP benefits
● The ability to analyze both structured and unstructured data, such as speech, text
messages, and social media posts.
● Improving customer satisfaction and experience by identifying insights using sentiment
analysis.
● Reducing costs by employing NLP-enabled AI to perform specific tasks, such as
chatting with customers via chatbots or analyzing large amounts of text data.
● Better understanding a target market or brand by conducting NLP analysis on relevant
data like social media posts, focus group surveys, and reviews.
Natural language processing (NLP) is critical to fully and efficiently analyze text and
speech data. It can work through the differences in dialects, slang, and grammatical
irregularities typical in day-to-day conversations.
Sentiment Analysis
Deep learning models like Convolutional Neural Networks (CNNs) and Recurrent
Neural Networks (RNNs) have demonstrated excellent performance in sentiment
analysis tasks. CNNs can capture local patterns in text, while RNNs can model the
sequential nature of language. This allows them to discern nuances in sentiment,
making them valuable tools for sentiment analysis applications.
Sentiment Lexicons
Text Generation
Text generation involves the creation of coherent and contextually relevant text by
machines. Generative models, such as GPT-3 (Generative Pre-trained Transformer 3),
have gained prominence for their ability to generate human-like text. They are
pretrained on vast amounts of text data and can generate text for applications like
chatbots, content creation, and automated storytelling.
Generative models like GPT-3 have been employed in various applications. Chatbots
powered by GPT-3 can engage in more natural and context-aware conversations. In
content creation, these models can assist in generating articles, reports, and creative
writing. They have also been used in automated content summarization and translation
tasks.
Machine Translation
Machine translation is the task of automatically translating text from one language
to another. Neural machine translation models, which are deep learning-based, have
significantly improved translation quality. They work by learning to map input
sentences to their corresponding translations in multiple languages.
5. Dimensionality Estimation
Introduction
Basics of PCA
PCA begins with mean-centering the data to have a zero mean. It then calculates
the covariance matrix of the features. The eigenvalues and eigenvectors of this matrix
determine the principal components. The components are ranked in descending order
of explained variance.
intricate patterns and dependencies in the data, making them suitable for complex
datasets.
Autoencoders
The architecture of an autoencoder depends on the specific problem and data. Deep
autoencoders have multiple hidden layers, allowing them to capture hierarchical
features. Training involves backpropagation and gradient descent techniques to
optimize the encoding and decoding functions.
Anomaly Detection
Time series forecasting is a critical area of data analysis with applications in various
domains. It involves predicting future values based on historical data patterns.
Time series data is characterized by its temporal nature, where observations are
recorded at successive time points. This type of data is commonly encountered in fields
such as finance (e.g., stock price prediction) and meteorology (e.g., weather
forecasting).
Recurrent Neural Networks (RNNs) are a class of deep learning models specifically
designed for sequential data analysis. In the context of time series forecasting, RNNs
are well-suited to capture temporal dependencies and patterns.
Introduction to RNNs for Time Series Analysis: RNNs are capable of maintaining a
hidden state that captures information from previous time steps, making them suitable
for modeling sequences.
LSTM and GRU Architectures: Long Short-Term Memory (LSTM) and Gated
Recurrent Unit (GRU) are specialized RNN variants that address the vanishing gradient
problem, allowing them to capture long-term dependencies in time series data
effectively.
Forecasting Models
Deep learning-based models have gained popularity for time series forecasting due
to their ability to handle complex patterns. These models leverage neural networks for
predictive tasks.
Deep Learning-Based Time Series Forecasting Models: Models like the recurrent neural
network (RNN) and its variants, as well as convolutional neural networks (CNNs), are
used to extract features and make predictions.
Evaluation Metrics
Evaluating the accuracy of time series forecasting models is essential to assess their
performance and make improvements.
Metrics for Measuring Forecasting Accuracy: Common metrics include Mean Absolute
Error (MAE), Mean Squared Error (MSE), and Root Mean Squared Error (RMSE). These
quantify the difference between predicted and actual values.
Real-World Applications
Time series forecasting has practical applications across various industries, enhancing
decision-making and planning.
In summary, time series forecasting is a powerful tool for making predictions based
on historical sequential data. Deep learning models, especially RNNs, have proven
effective in capturing temporal patterns and improving forecasting accuracy in diverse
domains. Evaluation metrics and cross-validation techniques play a vital role in
assessing and refining these models, ensuring their real-world applicability and
reliability.
Smart grids have emerged as a critical innovation in the field of electric power
distribution. Traditional power grids face significant challenges that hinder their
efficiency and reliability. These challenges include outdated infrastructure, limited
visibility into grid operations, and difficulties in accommodating the growing demand
for electricity. Consequently, there is a pressing need to transition towards smart grids.
Smart grids offer a range of benefits that address these shortcomings. They enable
enhanced monitoring and control of the entire electricity network, from power
generation to consumption. This leads to improved reliability, reduced power losses
during transmission, and quicker response to faults. Additionally, smart grids facilitate
the integration of renewable energy sources, enabling a cleaner and more sustainable
energy future.
Deep learning plays a pivotal role in optimizing the management of smart grids. Its
applications extend to various aspects of grid operations, offering data-driven solutions
to complex challenges.
Smart grids face the challenge of managing fluctuating energy sources, such as
wind and solar power, which are inherently variable. Deep learning algorithms aid in
addressing this challenge.
Deep learning models analyze data from renewable energy sources and predict their
output. These predictions help grid operators anticipate fluctuations and adjust energy
generation accordingly, ensuring grid stability and reliability even in the presence of
intermittent renewable sources.
Case Studies
To illustrate the practical application of smart grids and deep learning in the energy
sector, several real-world case studies serve as examples.
Numerous regions and utilities worldwide have implemented smart grid solutions.
For instance, cities have adopted advanced metering infrastructure (AMI) to monitor
electricity usage in real-time, and some regions have integrated distributed energy
resources (DERs) seamlessly into their grids. These implementations demonstrate the
potential of smart grids to enhance energy management and sustainability.
In conclusion, the integration of deep learning techniques into smart grids offers
substantial advantages in terms of energy efficiency, reliability, and sustainability. By
addressing the challenges faced by traditional power grids and harnessing the benefits
of smart grids, we can create a more resilient and responsive energy infrastructure for
the future.
Due to rising electricity demand and finite fossil fuel resources, a transition to
renewable energy is imperative. While renewables can be intermittent, combining
them with other sources can mitigate this issue. We present a hybrid renewable system
integrated with grids for efficient energy prediction using an energy management
system. Our study focuses on optimizing energy production, grid interactions, and
storage, all in tandem.
• PREDICTIVE MODELING
• LOADFORECASTING
• ENERGY MANAGEMENT
• GRID OPTIMATION
• Deep reinforcement learning (DRL) can be applied to optimize the operation of the
power grid.
• DRL agents can make real-time decisions about the distribution of electricity from
various sources to meet demand while minimizing costs and emissions.
This is especially useful when integrating controllable energy resources like battery
storage and flexible demand
society. However, fundraising for charitable causes presents several challenges. These
challenges include the need to reach potential donors effectively, maintain donor
engagement, and secure sustainable funding sources.
In the digital age, charities are leveraging data-driven approaches to gain insights
into donor behavior. Predictive modeling techniques analyze historical donation data,
demographic information, and engagement patterns to anticipate how donors might
respond to different appeals and campaigns. This data-driven understanding allows
organizations to tailor their fundraising strategies more effectively.
Armed with insights from predictive models, charities can implement targeted
marketing strategies. These strategies involve tailoring communication and outreach
efforts to specific donor segments. By delivering personalized messages and content
to donors who are likely to be interested in a particular cause, charities can significantly
enhance their fundraising efficiency.
Recommender Systems
Deep learning, particularly recommender systems, plays a pivotal role in the world
of charitable giving. These systems employ sophisticated algorithms to analyze donor
preferences and recommend causes or campaigns that align with an individual's values
and interests. By leveraging deep learning techniques, charities can provide donors
with highly personalized recommendations, fostering greater engagement and
participation.
In DRL for robot control, robots are trained using a reward-based system. They
explore different actions and learn which actions lead to favorable outcomes. Over
time, the robot's neural network improves its decision-making abilities, enabling it to
perform tasks with increasing precision and efficiency. DRL has been applied to a wide
range of industrial tasks, from robotic arm control in manufacturing to autonomous
navigation in warehouse environments.
Training robotic agents using DRL involves creating simulation environments where
robots can learn without the risk of physical damage. The agents receive rewards for
accomplishing tasks and penalties for making mistakes. Through a process of trial and
error, the agents gradually refine their policies and behaviors. Once trained in
simulations, the models can be deployed to physical robots in real-world industrial
settings.
Applications
offering assistance and augmenting human capabilities while maintaining safety. This
collaborative approach maximizes the benefits of automation while minimizing risks.
Assignments
ASSIGNMENT :
Choose a real-world business scenario (e.g., fraud detection in financial transactions) and
identify three data science tools or libraries that would be most useful for solving this
problem. Provide a brief explanation of each tool's functionality and how it can contribute to
the project's success. Discuss any advantages and limitations of these tools in your chosen
scenario.
Part A – Q & A
Unit - V
PART A
object and does not provide the location of the object within the visual.
4. What is image segmentation? CO5,K1
It is the task of grouping pixels with comparable properties together instead of
bounding boxes to identify objects.
5. What is Object localization? CO5,K1
Object localization seeks to identify the location of one or more objects in an
image, whereas object detection identifies all objects and their borders without much
focus on placement.
14. What is the difference between depth image and RGB image? CO5,K1
An RGB-D image provides a per-pixel depth information aligned with corresponding
image pixels. An image formed through depth information is an image channel in
which each pixel relates to a distance between the image plane and the
corresponding object in the RGB image.
The main purpose of the RGB color model is for the sensing, representation, and
display of images in electronic systems, such as televisions and computers, though it
has also been used in conventional photography.
18. What are the approaches used for dimensionality estimation in deep
learning? CO5,K1
Principle Component analysis.
Variational autoencoders
Random Projections
They are used in complex planning systems and also to find low cost paths around
obstacles. PRMs use a random sample of points on their map where a robot device
can possibly move and then the shortest path is calculated.
Part B – Q & A
Unit - V
PART B
1. Explain in detail on how the object is detected and classified using deep learning
concepts. (CO5,K3)
2. Detailed overview on building electric power grid for controllable energy resources
in deep learning concept. (CO5,K3)
3. Explain in detail on the process involved for prediction on historical time
dependent data using neural network. (CO5,K3)
8. Implement a model for object detection of traffic images using python. (CO5,K3)
Supportive online
Certification courses
(NPTEL, Swayam,
Coursera, Udemy, etc.,)
https://www.edx.org/learn/linear-algebra
https://online-learning.harvard.edu/course/linear-algebra?delta=0
1 Deep https://onlinecourses.nptel.ac.in/noc23_cs110/unit?
NPTEL Learning unit=17&assessment=226
2 Deep https://www.coursera.org/specializations/deep-
Learning learning
Coursera Specializa
tion
3 Deep https://www.udemy.com/course/deep-learning-
Learning with-python-and-keras/
with
Coursera Python
and Keras
Assessment Schedule
(Proposed Date & Actual
Date)
ASSESSMENT SCHEDULE
Name of the
S. No. Start Date Portion
Assessment
79 6
TEXT BOOKS:
1.Ian Goodfellow, Yoshua Bengio, Aaron Courville, ``Deep Learning'', MIT Press, 2016
3.Vance, William , Data Science: A Comprehensive Beginners Guide to Learn the Realms
4.Wani, M.A., Raj, B., Luo, F., Dou, D. (Eds.), Deep Learning Applications, Volume 3,Springer
Publications 2022
Mini Projects
a) Recommendation system
b) Credit Card Fraud Detection
c) Fake News Detection
d) Customer Segmentation
e) Sentiment Analysis
f) Recommender Systems
g) Emotion Recognition
h) Stock Market Prediction
i) Email classification
j) Tweets classification
k) Uber Data Analysis
l) Social Network Analysis
Thank you
Disclaimer:
This document is confidential and intended solely for the educational purpose of
RMK Group of Educational Institutions. If you have received this document through
email in error, please notify the system manager. This document contains proprietary
information and is intended only to the respective group / learning community as
intended. If you are not the addressee you should not disseminate, distribute or
copy through e-mail. Please notify the sender immediately by e-mail if you have
received this document by mistake and delete this document from your system. If
you are not the intended recipient you are notified that disclosing, copying,
distributing or taking any action in reliance on the contents of this information is
strictly prohibited.
82 6