0% found this document useful (0 votes)
19 views31 pages

Ad8552 ML Unit I

Uploaded by

saiprassad20
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
19 views31 pages

Ad8552 ML Unit I

Uploaded by

saiprassad20
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 31

SYLLABUS L T P C

3 0 0 3
AD8552 MACHINE LEARNING

UNIT I MACHINE LEARNING BASICS 8


Introduction to Machine Learning (ML) - Essential concepts of ML – Types of learning –
Machine learning methods based on Time – Dimensionality – Linearity and Non linearity –
Early trends in Machine learning – Data Understanding Representation and visualization.

UNIT II MACHINE LEARNING METHODS 11


Linear methods – Regression -Classification –Perceptron and Neural networks – Decision
trees –Support vector machines – Probabilistic models ––Unsupervised learning –
Featurization

UNIT III MACHINE LEARNING IN PRACTICE 9


Ranking – Recommendation System - Designing and Tuning model pipelines- Performance
measurement – Azure Machine Learning – Open-source Machine Learning libraries –
Amazon’s Machine Learning Tool Kit: Sagemaker

UNIT IV MACHINE LEARNING AND DATA ANALYTICS 9


Machine Learning for Predictive Data Analytics – Data to Insights to Decisions – Data
Exploration –Information based Learning – Similarity based learning – Probability based
learning – Error based learning – Evaluation – The art of Machine learning to Predictive
Data Analytics.

UNIT V APPLICATIONS OF MACHINE LEARNING 8


Image Recognition – Speech Recognition – Email spam and Malware Filtering – Online
fraud detection – Medical Diagnosis.
ACTIVITY BASED LEARNING

(MODEL BUILDING/PROTOTYPE)

S NO TOPICS

Cross word Puzzle for machine learning


1
Lecture Notes – Unit 1
UNIT I MACHINE LEARNING BASICS 8

Introduction to Machine Learning (ML) - Essential concepts of ML – Types of learning –


Machine learning methods based on Time – Dimensionality – Linearity and Non linearity – Early
trends in Machine learning – Data Understanding Representation and visualization.

1. Introduction to Machine Learning (ML)

 The term machine learning was first coined in the 1950s when Artificial Intelligence
pioneer Arthur Samuel built the first self-learning system for playing checkers. He
noticed that the more the system played, the better it performed.
 Fueled by advances in statistics and computer science, as well as better datasets and
the growth of neural networks, machine learning has truly taken off in recent years.
 Today, whether you realize it or not, machine learning is everywhere ‒ automated
translation, image recognition, voice search technology, self-driving cars, and beyond.
1.1 What Is Machine Learning?
• Machine learning (ML) is a branch of artificial intelligence (AI) that enables computers to
“self-learn” from training data and improve over time, without being explicitly
programmed. Machine learning algorithms are able to detect patterns in data and learn
from them, in order to make their own predictions. In short, machine learning
algorithms and models learn through experience.
1.2 History Machine Learning (ML)
2. Essential concepts of Machine Learning

There are various concepts that are spread all over the place in terms of categorizing them
with techniques or applications or implementations. However, they are at the heart of the
entire machine learning theory and its applications and need special attention.

2.1 Big Data and Not-So-Big Data


What Is Big Data?
When the size of data is large enough such that it cannot be processed on a single
machine, it is called as big data. Based on current (2018) generation of computers this
equates to something roughly more than 10 GB. From there it can go to hundreds of
petabytes (1 petabyte is 1000 terabytes and 1 terabyte is 1000
gigabytes) and more.
Why Should We Treat Big Data Differently?
Although the big data is separated from the not-so-big-data by simple size restrictions,
the handling of the data changes drastically when we move from not-so-big data to big
data. For processing not-so-big data, one need not worry about the location where each
bit of data is stored. As all the data can be loaded into memory of single machine and can
be accessed without additional overhead, all the typical numeric or string operations are
executed in predictable time.

2.2 Types of Learning

The machine learning algorithms are broadly classified into three types:
2.2.1 Supervised learning
 In Supervised learning, an AI system is presented with data which is labeled, which
means that each data tagged with the correct label.
 The goal is to approximate the mapping function so well that when you have new input
data (x) that you can predict the output variables (Y) for that data.

 As shown in the above example, we have initially taken some data and marked them
as ‘Spam’ or ‘Not Spam’. This labelled data is used by the training supervised model,
this data is used to train the model.Once it is trained we can test our model by testing
it with some test new mails and checking of the model is able to predict the right
output.

 Real-world applications of supervised learning are Risk Assessment, Fraud


Detection, Spam filtering, etc.
 Types of Supervised learning
 Classification: A classification problem is when the output variable is a
category, such as “red” or “blue” or “disease” and “no disease”.
 Regression: A regression problem is when the output variable is a real
value, such as “dollars” or “weight”.
Advantages and Disadvantages of Supervised Learning

Advantages:
•Since supervised learning work with the labelled dataset so we can have an exact idea about
the classes of objects.
•These algorithms are helpful in predicting the output on the basis of prior experience.
Disadvantages:
•These algorithms are not able to solve complex tasks.
•It may predict the wrong output if the test data is different from the training data.
•It requires lots of computational time to train the algorithm.

Applications of Supervised Learning Some common applications of Supervised Learning


are given below:
• Image Segmentation:
Supervised Learning algorithms are used in image segmentation. In this process, image
classification is performed on different image data with pre-defined labels.
• Medical Diagnosis
Supervised algorithms are also used in the medical field for diagnosis purposes. It is done by
using medical images and past labelled data with labels for disease conditions. With such a
process, the machine can identify a disease for the new patients.
•Fraud Detection - Supervised Learning classification algorithms are used for identifying
fraud transactions, fraud customers, etc. It is done by using historic data to identify the
patterns that can lead to possible fraud.
•Spam detection - In spam detection & filtering, classification algorithms are used. These
algorithms classify an email as spam or not spam. The spam emails are sent to the spam
folder.
•Speech Recognition - Supervised learning algorithms are also used in speech recognition.
The algorithm is trained with voice data, and various identifications can be done using the
same, such as voice-activated passwords, voice commands, etc.
2.2.2 Unsupervised Learning

In unsupervised learning, an AI system is presented with unlabeled, uncategorized data


and the system’s algorithms act on the data without prior training. The output is
dependent upon the coded algorithms. Subjecting a system to unsupervised learning is
one way of testing AI.

In the above example, we have given some characters to our model which are ‘Ducks’ and
‘Not Ducks’. In our training data, we don’t provide any label to the corresponding data.
The unsupervised model is able to separate both the characters by looking at the type of
data and models the underlying structure or distribution in the data in order to learn more
about it.

Types of Unsupervised learning


 Clustering: A clustering problem is where you want to discover the inherent
groupings in the data, such as grouping customers by purchasing behavior.
Some of the popular clustering algorithms are given below:
•K-Means Clustering algorithm
•DBSCAN Algorithm
•Principal Component Analysis
 Association: An association rule learning problem is where you want to discover rules
that describe large portions of your data, such as people that buy X also tend to buy Y.
• This algorithm is mainly applied in Market Basket analysis, Web usage mining,
continuous production, etc.Some popular algorithms of Association rule learning are
Apriori Algorithm, Eclat, FP-growth algorithm.
Advantages and Disadvantages of Unsupervised Learning Algorithm

Advantages:
•These algorithms can be used for complicated tasks compared to the supervised ones
because these algorithms work on the unlabeled dataset.
•Unsupervised algorithms are preferable for various tasks as getting the unlabeled dataset is
easier as compared to the labelled dataset.
Disadvantages:
•The output of an unsupervised algorithm can be less accurate as the dataset is not labeled,
and algorithms are not trained with the exact output in prior.
•Working with Unsupervised learning is more difficult as it works with the unlabeled dataset
that does not map with the output.

Applications of Unsupervised Learning


•Network Analysis: Unsupervised learning is used for identifying plagiarism and copyright
in document network analysis of text data for scholarly articles.
•Recommendation Systems: Recommendation systems widely use unsupervised learning
techniques for building recommendation applications for different web applications and e-
commerce websites.
•Anomaly Detection: Anomaly detection is a popular application of unsupervised learning,
which can identify unusual data points within the dataset. It is used to discover fraudulent
transactions.
•Singular Value Decomposition: Singular Value Decomposition or SVD is used to extract
particular information from the database. For example, extracting information of each user
located at a particular location.
2.2.3 Reinforcement Learning
A reinforcement learning algorithm, or agent, learns by interacting with its environment.
The agent receives rewards by performing correctly and penalties for performing
incorrectly. The agent learns without intervention from a human by maximizing its reward
and minimizing its penalty. It is a type of dynamic programming that trains algorithms using
a system of reward and punishment.

In the above example, we can see that the agent is given 2 options i.e. a path with water
or a path with fire. A reinforcement algorithm works on rewarding a system i.e. if the
agent uses the fire path then the rewards are subtracted and the agent tries to learn that
it should avoid the fire path. If it had chosen the water path or the safe path then some
points would have been added to the reward points, the agent then would try to learn
what path is safe and what path isn’t. It is basically leveraging the rewards obtained, the
agent improves its environment knowledge to select the next action.
An example of an algorithm used in RL is Markov Decision Process.
2.3 Machine Learning Methods Based on Time
Another way to slice the machine learning methods is to classify them based on
the type of data that they deal with. The systems that take in static labelled data are called
static learning methods. The systems that deal with data that is continuously changing with
time are called dynamic methods. Each type of method can be supervised, or unsupervised,
however, reinforcement learning methods are always.
dynamic.

2.3.1 Static Learning


Static learning refers to learning on the data that is taken as a single snapshot
and the properties of the data remain constant over time. Once the model is trained on the
data (either using supervised or unsupervised learning) the trained model can be applied to
similar data anytime in the future and the model will still be valid and will perform as
expected. Typical examples of this would be image classification of different animals.

2.3.2 Dynamic Learning


This is also called as time series based learning. The data in this type of
problems is time sensitive and keeps changing over time. Hence the model training is not a
static process, but the model needs to be trained continuously (or after every reasonable
time window) to remain effective. A typical example of such problems is weather forecasting,
or stock market predictions. A model trained a year back will be completely useless to predict
the weather for tomorrow, or predict the price of any stock for tomorrow.

2.4 Dimensionality

From the physical standpoint, dimensions are space dimensions: length,


width,and height. However, it is very common to have tens if not hundreds and more
dimensions when we deal with data for machine learning. In order to understand these
high dimensions, we need to look at the fundamental property of dimensions. The space
dimensions are defined such that each of the dimensions is perpendicular or orthogonal to
other two. This property of orthogonality is essential to have unique representation of all
the points in this 3-dimensional space.
For example if we setup the three coordinates as length, width,and height with
some arbitrary origin (The precise location of origin only changes the value of coordinates,
but does not affect the uniqueness property and hence any choice of origin is fine as long as
it remains unchanged throughout the calculation.)
The coordinates (0,0,0) mark the location of origin itself. The coordinates
(1,1,1)will mark a point space that is 1 unit away from the origin in each of the dimensions
and is unique. No other set of coordinates will mean the same location in space.

2.4.1 Curse of Dimensionality


Even if adding arbitrarily large number of dimensions is fine from mathematical
standpoint, there is one problem. With increase in dimensions the density of the data gets
diminished exponentially. For example if we have 1000 data points in training data, and data
has 3 unique features. Let’s say the value of all the features is within 1–10. Thus all these
1000 data points lie in a cube of size 10 × 10 × 10.Thus the density is 1000/1000 or 1
sample per unit cube. If we had 5 unique features instead of 3, then quickly the density of
the data drops to 0.01 sample per unit 5 dimensional cube.
The density of the data is important, as higher the density of the data, better is
the likelihood of finding a good model, and higher is the confidence in the accuracy of the
model. If the density is very low, there would be very low confidence in the trained model
using that data.
Hence, although high dimensions are acceptable mathematically, one needs to
be careful with the dimensionality in order to be able to develop a good ML model with high
confidence.
2.5 Linearity and Nonlinearity

• The concept of linearity and nonlinearity is applicable to both the data and the model that
built on top of it. Data is called as linear if the relationship between the input and output
is linear. when the value of input increases, the value of output also increases and vice
versa.
• All the models that use linear equations to model the relationship between input and
output are called as linear models. However, sometimes, by preconditioning the input or
output a nonlinear relationship between the data can be converted into linear relationship
and then the linear model can be applied on it.
For example if input and output are related with exponential relationship as y = 2𝑒 𝑥 The
data is clearly nonlinear. However, instead of building the model on original data, we can
build a model after applying a log operation. This operation transforms the original nonlinear
relationship into linear one as log y = 2𝑒 𝑥 . Then we build the linear model to predict log y
instead of y, which can then be converted to y by taking exponent.
There can also be cases where a problem can be broken down into multiple parts and linear
model can be applied to each part to ultimately solve a nonlinear problem. Below figures
show examples of converted linear and piecewise linear relationships,
respectively.

Examples of piecewise linear relationships between input and output


While in some cases the relationship is purely nonlinear and needs a proper nonlinear
model to map it . Below Figure shows examples of pure nonlinear relationships.

Examples of pure nonlinear relationships between input and output

Linear models are the simplest to understand, build, and interpret. Our brain is highly tuned for
linear models, as most of our experiences tend to have linear trends. All the models in the
theory of machine learning can handle linear data. Examples of purely linear models are linear
regression, support vector machines without nonlinear kernels, etc. Nonlinear models inherently
use some nonlinear functions to approximate the nonlinear characteristics of the data. Examples
of nonlinear models include neural networks, decision trees, probabilistic models based on
nonlinear distributions, etc.
2.6 Early Trends in Machine Learning

Before the machine learning started off commercially in true sense, there were few other
systems that were already pushing the boundary if routine computation. One such
notable application was Expert Systems.
2.6.1 Expert Systems
The definition by Alan Turin marks the beginning of the era where machine intelligence
was recognized and with that field of AI was born. However, in the early days (all the
way till 1980s), the field of Machine Intelligence or Machine Learning was limited to what
were called as Expert Systems or Knowledge based Systems. One of the leading
experts in the field of expert systems, Dr. Edward Feigenbaum,once defined as expert
system as,
Definition : Expert Systems An intelligent computer program that uses knowledge and
inference procedures to solve problems that are difficult enough to require significant
human expertise for the solution
Such systems were capable of replacing experts in certain areas. These
machines were programmed to perform complex heuristic tasks based on elaborate
logical operations. In spite of being able to replace the humans who are experts in the
specific areas, these systems were not “intelligent” in the true sense, if we compare
them with human intelligence.
The reason being the system were “hard-coded” to solve only a specific
type of problem and if there is need to solve a simpler but completely different problem,
these system would quickly become completely useless.
Nonetheless, these systems were quite popular and successful specifically
in areas where repeated but highly accurate performance was needed, e.g.,
diagnosis,inspection, monitoring, control
3. Data Understanding, Representation, and Visualization

 This chapter focusses on understanding, representing, and visualizing the data.


These steps together can be called as data preprocessing.
 With recent explosion of small devices that are connected to internet,the amount of data
that is being generated has increased exponentially. This data can be quite useful for
generating variety of insights if handled properly, else it can only burden the systems
handling it and slow down everything. The science that deals with general handling and
organizing and then interpreting the data is called as data science.

3.1 Understanding the Data


First step in building an application of artificial intelligence is to understand the data. The
data in raw form can come from different sources, in different formats. Some data can be
missing, some data can be mal-formatted, etc. It is the first task to get familiar with the
data. Clean up the data as necessary.
The step of understanding the data can be broken down into three parts:
1. Understanding entities
2. Understanding attributes
3. Understanding data types
3.1.1 Understanding Entities
In the field of data science or machine learning and artificial intelligence,
entities represent groups of data separated based on conceptual themes and/or data
acquisition methods. An entity typically represents a table in a database, or a flat file, e.g.,
comma separated variable (csv) file, or tab separated variable (tsv) file.Sometimes it is
more efficient to represent the entities using a more structured formats like svmlight.

Each entity can contain multiple attributes. The raw data for each application
can contain multiple such entities (Table 3.1).In case of Iris data, we have only one such
entity in the form of dimensions of sepals and petals of the flowers. However, if one is
trying to solve this classification problem and finds that the data about sepals and petals
alone is not sufficient, then he/she can add more information in the form of additional
entities. For example more information about the flowers in the form of their colors, or
smells or longevity of the trees that produce them, etc. can be added to improve the
classification performance.
Table 3.1 Sample from Iris data set containing 3 classes and 4 attributes
3.1.2 Understanding Attributes
Each attribute can be thought of as a column in the file or table. In case of Iris
data,the attributes from the single given entity are sepal length in cm, sepal width in cm,
petal length in cm, petal width in cm. If we had added additional entities like color,smell, etc.,
each of those entities would have their own attributes. It is important to note that in the
current data, all the columns are all features, and there is no ID column.

3.1.3 Understanding Data Types


Attributes in each entity can be of various different types from the storage and
processing perspective, e.g., string, integer valued, datetime, binary (“true”/“false”, or
“1”/“0”), etc. Sometimes the attributes can originate from completely different domains like
an image or a sound file, etc. Each type needs to be handled separately for generating a
feature vector that will be consumed by the machine learning algorithm.
Some attributes will have missing values. This missing data is typically replaced
with special characters, which should not be confused with any of the real values. In order to
process data with missing values, one can either fill those missing values with some default
values, or use an algorithm that can work with missing data.
In case of Iris data, all the attributes are real valued numerical and there is no
missing data. However, if we add additional entities like color, it would have enumerative type
string features like green, orange, etc.

3.2 Representation and Visualization of the Data

The last step in preprocessing the data showing how the data is distributed and
how it related to the output or class label. We live in 3-dimensional world, so any data that is
up to 3-dimensional, we can plot it and visualize it.
However, when there are more than 3-dimensions, it gets tricky. The Iris data,
for example, also has 4 dimensions. There is no way we can plot the full information in each
sample in a single plot that we can visualize.

There are couple of options in such cases:


1. Draw multiple plots taking 2 or three dimensions at a time.
2. Reduce the dimensionality of the data and plot upto 3 dimensions
Drawing multiple plots is easy, but it splits the information and it becomes harder to
understand how different dimensions interact with each other.
Reducing dimensionality is typically preferred method. Most common methods used to
reduce the dimensionality are:
1. Principal Component Analysis or PCA
2. Linear Discriminant Analysis or LDA

3.2.1 Principal Component Analysis or PCA


We can only visualize the data in maximum of 2 or 3 dimensions. However, it
is common practice to have the dimensionality of the data in tens or even hundreds. In
such cases, we can employ the algorithms just fine, as the mathematics on which they are

based scales perfectly fine for higher dimensions. Principal components are orthogonal

projections of data onto lower-dimensional space. Imagine we have a 2-dimensional


dataset. Each dimension can be represented as a feature column:

We can represent the same dataset as a scatterplot:


 The main aim of PCA is to find such principal components, which can describe the data
points with a set of principal components.

 The principal components are vectors, but they are not chosen at random. The first
principal component is computed so that it explains the greatest amount of variance in
the original features. The second component is orthogonal to the first, and it explains
the greatest amount of variance left after the first principal component.

 The algorithm can be used on its own, or it can serve as a data cleaning or data
preprocessing technique used before another machine learning algorithm.

 When PCA is used as part of preprocessing, the algorithm is applied to:

1. Reduce the number of dimensions in the training dataset.

2. De-noise the data. Because PCA is computed by finding the components which
explain the greatest amount of variance, it captures the signal in the data and omits
the noise.
Application of PCA

•PCA is used to visualize multidimensional data.


•It is used to reduce the number of dimensions in healthcare data.
•PCA can help resize an image.
•It can be used in finance to analyze stock data and forecast returns.
•PCA helps to find patterns in the high-dimensional datasets.

3.2.2 Linear Discriminant Analysis

Linear discriminant analysis is used as a tool for classification, dimension


reduction, and data visualization. It is also known as Normal Discriminant Analysis (NDA) or
Discriminant Function Analysis (DFA). This can be used to project the features of higher
dimensional space into lower-dimensional space in order to reduce resources and
dimensional costs.
It is also considered a pre-processing step for modeling differences in ML

and applications of pattern classification. Whenever there is a requirement to separate


two or more classes having multiple features efficiently, the Linear Discriminant Analysis
model is considered the most common technique to solve such classification problems.
For e.g., if we have two classes with multiple features and need to separate them
efficiently. When we classify them using a single feature, then it may show overlapping.

Overlapping

To overcome the overlapping issue in the classification process, we must increase the
number of features regularly.
Example:
 Let's assume we have to classify two different classes having two sets of data points in a
2-dimensional plane as shown below image:

 However, it is impossible to draw a straight line in a 2-d plane that can separate these
data points efficiently but using linear Discriminant analysis; we can dimensionally reduce
the 2-D plane into the 1-D plane. Using this technique, we can also maximize the
separability between multiple classes.

 Linear Discriminant analysis is used as a dimensionality reduction technique in machine


learning, using which we can easily transform a 2-D and 3-D graph into a 1-dimensional
plane.Let's consider an example where we have two classes in a 2-D plane having an X-Y
axis, and we need to classify them efficiently. As we have already seen in the above
example that LDA enables us to draw a straight line that can completely separate the two
classes of the data points. Here, LDA uses an X-Y axis to create a new axis by separating
them using a straight line and projecting data onto a new axis.Hence, we can maximize
the separation between these classes and reduce the 2-D plane into 1-D.
To create a new axis, Linear Discriminant Analysis uses the following criteria:
•It maximizes the distance between means of two classes.
•It minimizes the variance within the individual class.
Using the above two conditions, LDA generates a new axis in such a way that it can
maximize the distance between the means of the two classes and minimizes the variation
within each class.

Applications of LDA
 Face Recognition
 Medical
 Customer Identification
 Predictions

Difference between Linear Discriminant Analysis and PCA

Below are some basic differences between LDA and PCA:


 PCA is an unsupervised algorithm that does not care about classes and labels and only
aims to find the principal components to maximize the variance in the given dataset. At
the same time, LDA is a supervised algorithm that aims to find the linear discriminants
to represent the axes that maximize separation between different classes of data.
 LDA is much more suitable for multi-class classification tasks compared to PCA.
However, PCA is assumed to be an as good performer for a comparatively small sample
size.
 Both LDA and PCA are used as dimensionality reduction techniques, where PCA is first
followed by LDA.
Part A – Q & A
Unit - I
PART -A

S.N Question and Answer


o
Define Machine Learning.
1. Machine learning is a subfield of artificial intelligence, which is
broadly defined as the capability of a machine to imitate intelligent
human behavior.
2. List out Different Types of learning methods.
Supervised Learning, Unsupervised Learning, reinforcement
Learning
3. What is meant by supervised learning?
Supervised learning, also known as supervised machine learning, is
a subcategory of machine learning and artificial intelligence. It is
defined by its use of labeled datasets to train algorithms that to
classify data or predict outcomes accurately.
4. What is Unsupervised Learning?
Unsupervised learning is a type of machine learning in which
models are trained using unlabeled dataset and are allowed to act
on that data without any supervision.
5. Differentiate supervised and unsupervised machine learning.
In supervised machine learning, the machine is trained using
labeled data. Then a new dataset is given into the learning model
so that the algorithm provides a positive outcome by analyzing the
labeled data. For example, we first require to label the data which is
necessary to train the model while performing classification.
In the unsupervised machine learning, the machine is not trained
using labeled data and let the algorithms make the decisions
without any corresponding output variables.

6. Define Reinforcement Learning?


Reinforcement learning is a machine learning training method based
on rewarding desired behaviors and/or punishing undesired ones.
In general, a reinforcement learning agent is able to perceive and
interpret its environment, take actions and learn through trial and
error.
7. Where is supervised learning used?
Linear regression is a supervised learning technique typically used
in predicting, forecasting, and finding relationships between
quantitative data.
PART -A

S.No Question and Answer

8. Give Example for Unsupervised Learning.


ome examples of unsupervised learning algorithms include K-Means
Clustering, Principal Component Analysis and Hierarchical Clustering.

9. Give Example for Reinforcement Learning.


Reinforcement learning can be used in different fields such as
healthcare, finance, recommendation systems, etc. Playing games like
Go: Google has reinforcement learning agents that learn to solve
problems by playing simple games like Go, which is a game of strategy

10. List Out real time application of ML.


Image recognition
Speech recognition.
Medical diagnosis.
Statistical arbitrage
Predictive analytics
11. Define Data Science.
ata science is the domain of study that deals with vast volumes of data
using modern tools and techniques to find unseen patterns, derive
meaningful information, and make business decisions.

12. What is PCA?


Principal component analysis (PCA) is a technique for reducing the
dimensionality of such datasets, increasing interpretability but at the
same time minimizing information loss

13. What is the use of PCA in machine learning?


or Write Applications of PCA in Machine Learning.

It is used to reduce the number of dimensions in healthcare data. PCA


can help resize an image. It can be used in finance to analyze stock
data and forecast returns. PCA helps to find patterns in the high-
dimensional datasets
PART -A
S.N Question and Answer
o
14 Define Discriminant Analysis.
Discriminant analysis is a versatile statistical method often used by
market researchers to classify observations into two or more groups or
categories. In other words, discriminant analysis is used to assign
objects to one group among a number of known groups
15. Define Data preparation.
ata preparation (also referred to as “data preprocessing”) is the process
of transforming raw data so that data scientists and analysts can run it
through machine learning algorithms to uncover insights or make
predictions.
16. What is data visualization and representation?
Data visualization is the graphical representation of information and data
in a pictorial or graphical format(Example: charts, graphs, and maps).
Data visualization tools provide an accessible way to see and understand
trends, patterns in data, and outliers. you
must create a representation of the data to provide the model with a
useful vantage point into the data's key qualities.
17. List two application of PCA.
Some of the applications of Principal Component Analysis (PCA) are:
Spike-triggered covariance analysis in Neuroscience. Quantitative
Finance. Image Compression.

18. Write Disadvantage of supervised learning.


Computation time is vast for supervised learning.
Unwanted data downs efficiency.
Pre-processing of data is no less than a big challenge.
Always in need of updates.
Anyone can overfit supervised algorithms easily.

19. What is classifier in machine learning?


A classifier in a Machine Learning is a system that inputs a vector of
discrete or continuous feature values and outputs a single discrete
value, the class.

20. What is the main goal of supervised learning?


The goal of Supervised Learning is to come up with, or infer, an
approximate mapping function that can be applied to one or more input
variables, and produce an output variable or result. The training process
involves taking a supervised training data set with non features and a
label.
Part B – Questions
PART -B

S.No Question and Answer

1. Explain in detail about machine learning concepts with different


learning types.
2. Discuss with examples some useful applications of machine learning.

3. Differentiate supervised, unsupervised and reinforcement learning.

4. Explain in detail about data understanding and representation with


suitable examples.

5. Explain Principal Component Analysis in detail with example.

6. Explain PCA and Discriminative analysis with suitable example.

You might also like