0% found this document useful (0 votes)
105 views28 pages

Unit 1

The document discusses the differences between artificial intelligence, machine learning, and deep learning. It defines each term and provides examples to illustrate how they relate to each other, with machine learning and deep learning being subsets of artificial intelligence. Key aspects include machine learning using algorithms to build predictive models from data, and deep learning using vast amounts of data and complex algorithms in neural networks.

Uploaded by

Thanuja Malla
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
105 views28 pages

Unit 1

The document discusses the differences between artificial intelligence, machine learning, and deep learning. It defines each term and provides examples to illustrate how they relate to each other, with machine learning and deep learning being subsets of artificial intelligence. Key aspects include machine learning using algorithms to build predictive models from data, and deep learning using vast amounts of data and complex algorithms in neural networks.

Uploaded by

Thanuja Malla
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 28

AI vs Machine Learning vs Deep Learning

Artificial Intelligence, Machine Learning, and Deep Learning have become the most talked-about technologies in
today’s commercial world as companies are using these innovations to build intelligent machines and applications.
And although these terms are dominating business dialogues all over the world, many people have difficulty
differentiating between them. This blog will help you gain a clear understanding of AI, machine learning, and deep
learning and how they differ from one another.

Before jumping into the technicalities, let’s look at what tech influencers, industry personalities, and authors have to
say about these three concepts.

“AI doesn't have to be evil to destroy humanity – if AI has a goal and humanity just happens in the way, it will
destroy humanity as a matter of course without even thinking about it, no hard feelings.” – Elon Musk, Technology
Entrepreneur, and Investor.

“Artificial Intelligence, deep learning, machine learning — whatever you’re doing if you don’t understand it — learn
it. Because otherwise, you’re going to be a dinosaur within 3 years.” - Mark Cuban, American entrepreneur, and
television personality.

“In deep learning, the algorithms we use now are versions of the algorithms we were developing in the 1980s, the
1990s. People were very optimistic about them, but it turns out they didn’t work too well.” - Geoffrey Hinton, Father
of Deep Learning

The three terms are often used interchangeably, but they do not quite refer to the same things.

Here is an illustration designed to help us understand the fundamental differences between artificial intelligence,
machine learning, and deep learning.

Artificial Intelligence is the concept of creating smart intelligent machines.

Machine Learning is a subset of artificial intelligence that helps you build AI-driven applications.

Deep Learning is a subset of machine learning that uses vast volumes of data and complex algorithms to train a
model.
Now, let’s explore each of these technologies in detail.

What is Artificial Intelligence?


Artificial intelligence, commonly referred to as AI, is the process of imparting data, information, and human
intelligence to machines. The main goal of Artificial Intelligence is to develop self-reliant machines that can think
and act like humans. These machines can mimic human behavior and perform tasks by learning and problem-
solving. Most of the AI systems simulate natural intelligence to solve complex problems.

Let’s have a look at an example of an AI-driven product - Amazon Echo.

Amazon Echo is a smart speaker that uses Alexa, the virtual assistant AI technology developed by Amazon. Amazon
Alexa is capable of voice interaction, playing music, setting alarms, playing audiobooks, and giving real-time
information such as news, weather, sports, and traffic reports.

As you can see in the illustration below, the person wants to know the current temperature in Chicago. The person’s
voice is first converted into a machine-readable format. The formatted data is then fed into the Amazon Alexa
system for processing and analyzing. Finally, Alexa returns the desired voice output via Amazon Echo.

Now that you’ve been given a simple introduction to the basics of artificial intelligence, let’s have a look at
its different types.

Types of Artificial Intelligence


Reactive Machines - These are systems that only react. These systems don’t form memories, and they don’t use any
past experiences for making new decisions.

Limited Memory - These systems reference the past, and information is added over a period of time. The referenced
information is short-lived.

Theory of Mind - This covers systems that are able to understand human emotions and how they affect decision
making. They are trained to adjust their behavior accordingly.
Self-awareness - These systems are designed and created to be aware of themselves. They understand their own
internal states, predict other people’s feelings, and act appropriately.

Applications of Artificial Intelligence

 Machine Translation such as Google Translate

 Self Driving Vehicles such as Google’s Waymo

 AI Robots such as Sophia and Aibo

 Speech Recognition applications like Apple’s Siri or OK Google

Now that we have gone over the basics of artificial intelligence, let’s move on to machine learning and see how it
works.

What is Machine Learning?


Machine learning is a discipline of computer science that uses computer algorithms and analytics to build predictive
models that can solve business problems.

As per McKinsey & Co., machine learning is based on algorithms that can learn from data without relying on rules-
based programming.

Tom Mitchell’s book on machine learning says “A computer program is said to learn from experience E with respect
to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves
with experience E.”

So you see, machine learning has numerous definitions. But how does it really work?

How Does Machine Learning Work?


Machine learning accesses vast amounts of data (both structured and unstructured) and learns from it to predict the
future. It learns from the data by using multiple algorithms and techniques. Below is a diagram that shows how a
machine learns from data.

Now that you have been introduced to the basics of machine learning and how it works, let’s see the different types
of machine learning methods.

Types of Machine Learning


Machine learning algorithms are classified into three main categories:

1. Supervised Learning

In supervised learning, the data is already labeled, which means you know the target variable. Using this method of
learning, systems can predict future outcomes based on past data. It requires that at least an input and output variable
be given to the model for it to be trained.

Below is an example of a supervised learning method. The algorithm is trained using labeled data of dogs and cats.
The trained model predicts whether the new image is that of a cat or a dog.

Some examples of supervised learning include linear regression, logistic regression, support vector machines, Naive
Bayes, and decision tree.

2. Unsupervised Learning
Unsupervised learning algorithms employ unlabeled data to discover patterns from the data on their own. The
systems are able to identify hidden features from the input data provided. Once the data is more readable, the
patterns and similarities become more evident.

Below is an example of an unsupervised learning method that trains a model using unlabeled data. In this case, the
data consists of different vehicles. The purpose of the model is to classify each kind of vehicle.

Some examples of unsupervised learning include k-means clustering, hierarchical clustering, and anomaly detection.

3. Reinforcement Learning

The goal of reinforcement learning is to train an agent to complete a task within an uncertain environment. The
agent receives observations and a reward from the environment and sends actions to the environment. The reward
measures how successful action is with respect to completing the task goal.

Below is an example that shows how a machine is trained to identify shapes.

Examples of reinforcement learning algorithms include Q-learning and Deep Q-learning Neural Networks.

Machine Learning Processes


Machine Learning involves seven steps:
Machine Learning Applications

 Sales forecasting for different products

 Fraud analysis in banking

 Product recommendations

 Stock price prediction

Now that we’ve explored machine learning and its applications, let’s turn our attention to deep learning, what it is,
and how it is different from AI and machine learning.

What is Deep Learning?


Deep learning is a subset of machine learning that deals with algorithms inspired by the structure and function of the
human brain. Deep learning algorithms can work with an enormous amount of both structured and unstructured data.
Deep learning’s core concept lies in artificial neural networks, which enable machines to make decisions.
The major difference between deep learning vs machine learning is the way data is presented to the machine.
Machine learning algorithms usually require structured data, whereas deep learning networks work on multiple
layers of artificial neural networks.

This is what a simple neural network looks like:

The network has an input layer that accepts inputs from the data. The hidden layer is used to find any hidden
features from the data. The output layer then provides the expected output.

Here is an example of a neural network that uses large sets of unlabeled data of eye retinas. The network model is
trained on this data to find out whether or not a person has diabetic retinopathy.

Now that we have an idea of what deep learning is, let’s see how it works.

How Does Deep Learning Work?


1. Calculate the weighted sums.

2. The calculated sum of weights is passed as input to the activation function.

3. The activation function takes the “weighted sum of input” as the input to the function, adds a bias, and decides
whether the neuron should be fired or not.
4. The output layer gives the predicted output.

5. The model output is compared with the actual output. After training the neural network, the model uses the
backpropagation method to improve the performance of the network. The cost function helps to reduce the error
rate.

In the following example, deep learning and neural networks are used to identify the number on a license plate. This
technique is used by many countries to identify rules violators and speeding vehicles.
Types of Deep Neural Networks
Convolutional Neural Network (CNN) - CNN is a class of deep neural networks most commonly used for image
analysis.

Recurrent Neural Network (RNN) - RNN uses sequential information to build a model. It often works better for
models that have to memorize past data.

Generative Adversarial Network (GAN) - GAN are algorithmic architectures that use two neural networks to create
new, synthetic instances of data that pass for real data. A GAN trained on photographs can generate new
photographs that look at least superficially authentic to human observers.

Deep Belief Network (DBN) - DBN is a generative graphical model that is composed of multiple layers of latent
variables called hidden units. Each layer is interconnected, but the units are not.

Deep Learning Applications

 Cancer tumor detection

 Captionbot for captioning an image

 Music generation

 Image coloring

 Object detection
Types of Machine Learning
Machine learning is a subset of AI, which enables the machine to automatically
learn from data, improve performance from past experiences, and make
predictions. Machine learning contains a set of algorithms that work on a huge amount
of data. Data is fed to these algorithms to train them, and on the basis of training, they
build the model & perform a specific task.

These ML algorithms help to solve different business problems like Regression,


Classification, Forecasting, Clustering, and Associations, etc.

Based on the methods and way of learning, machine learning is divided into mainly four
types, which are:

1. Supervised Machine Learning


2. Unsupervised Machine Learning
3. Semi-Supervised Machine Learning
4. Reinforcement Learning

In this topic, we will provide a detailed description of the types of Machine Learning
along with their respective algorithms:

1. Supervised Machine Learning


As its name suggests, Supervised machine learning is based on supervision. It means in
the supervised learning technique, we train the machines using the "labelled" dataset,
and based on the training, the machine predicts the output. Here, the labelled data
specifies that some of the inputs are already mapped to the output. More preciously, we
can say; first, we train the machine with the input and corresponding output, and then
we ask the machine to predict the output using the test dataset.

Let's understand supervised learning with an example. Suppose we have an input


dataset of cats and dog images. So, first, we will provide the training to the machine to
understand the images, such as the shape & size of the tail of cat and dog, Shape of
eyes, colour, height (dogs are taller, cats are smaller), etc. After completion of
training, we input the picture of a cat and ask the machine to identify the object and
predict the output. Now, the machine is well trained, so it will check all the features of
the object, such as height, shape, colour, eyes, ears, tail, etc., and find that it's a cat. So, it
will put it in the Cat category. This is the process of how the machine identifies the
objects in Supervised Learning.

The main goal of the supervised learning technique is to map the input variable(x)
with the output variable(y). Some real-world applications of supervised learning
are Risk Assessment, Fraud Detection, Spam filtering, etc.

Categories of Supervised Machine Learning


Supervised machine learning can be classified into two types of problems, which are
given below:

o Classification
o Regression

a) Classification

Classification algorithms are used to solve the classification problems in which the
output variable is categorical, such as "Yes" or No, Male or Female, Red or Blue, etc.
The classification algorithms predict the categories present in the dataset. Some real-
world examples of classification algorithms are Spam Detection, Email filtering, etc.

Some popular classification algorithms are given below:

o Random Forest Algorithm


o Decision Tree Algorithm
o Logistic Regression Algorithm
o Support Vector Machine Algorithm
b) Regression

Regression algorithms are used to solve regression problems in which there is a linear
relationship between input and output variables. These are used to predict continuous
output variables, such as market trends, weather prediction, etc.

Some popular Regression algorithms are given below:

o Simple Linear Regression Algorithm


o Multivariate Regression Algorithm
o Decision Tree Algorithm
o Lasso Regression

Advantages and Disadvantages of Supervised


Learning
Advantages:

o Since supervised learning work with the labelled dataset so we can have an exact
idea about the classes of objects.
o These algorithms are helpful in predicting the output on the basis of prior
experience.

Disadvantages:

o These algorithms are not able to solve complex tasks.


o It may predict the wrong output if the test data is different from the training data.
o It requires lots of computational time to train the algorithm.

Applications of Supervised Learning


Some common applications of Supervised Learning are given below:

o Image Segmentation:
Supervised Learning algorithms are used in image segmentation. In this process,
image classification is performed on different image data with pre-defined labels.
o Medical Diagnosis:
Supervised algorithms are also used in the medical field for diagnosis purposes. It
is done by using medical images and past labelled data with labels for disease
conditions. With such a process, the machine can identify a disease for the new
patients.
o Fraud Detection - Supervised Learning classification algorithms are used for
identifying fraud transactions, fraud customers, etc. It is done by using historic
data to identify the patterns that can lead to possible fraud.
o Spam detection - In spam detection & filtering, classification algorithms are
used. These algorithms classify an email as spam or not spam. The spam emails
are sent to the spam folder.
o Speech Recognition - Supervised learning algorithms are also used in speech
recognition. The algorithm is trained with voice data, and various identifications
can be done using the same, such as voice-activated passwords, voice
commands, etc.

2. Unsupervised Machine Learning


Unsupervised learning is different from the Supervised learning technique; as its name
suggests, there is no need for supervision. It means, in unsupervised machine learning,
the machine is trained using the unlabeled dataset, and the machine predicts the output
without any supervision.

In unsupervised learning, the models are trained with the data that is neither classified
nor labelled, and the model acts on that data without any supervision.

The main aim of the unsupervised learning algorithm is to group or categories the
unsorted dataset according to the similarities, patterns, and differences. Machines
are instructed to find the hidden patterns from the input dataset.

Let's take an example to understand it more preciously; suppose there is a basket of


fruit images, and we input it into the machine learning model. The images are totally
unknown to the model, and the task of the machine is to find the patterns and
categories of the objects.

So, now the machine will discover its patterns and differences, such as colour difference,
shape difference, and predict the output when it is tested with the test dataset.
Categories of Unsupervised Machine Learning
Unsupervised Learning can be further classified into two types, which are given below:

o Clustering
o Association

1) Clustering

The clustering technique is used when we want to find the inherent groups from the
data. It is a way to group the objects into a cluster such that the objects with the most
similarities remain in one group and have fewer or no similarities with the objects of
other groups. An example of the clustering algorithm is grouping the customers by their
purchasing behaviour.

Some of the popular clustering algorithms are given below:

o K-Means Clustering algorithm


o Mean-shift algorithm
o DBSCAN Algorithm
o Principal Component Analysis
o Independent Component Analysis

2) Association

Association rule learning is an unsupervised learning technique, which finds interesting


relations among variables within a large dataset. The main aim of this learning algorithm
is to find the dependency of one data item on another data item and map those
variables accordingly so that it can generate maximum profit. This algorithm is mainly
applied in Market Basket analysis, Web usage mining, continuous production, etc.

Some popular algorithms of Association rule learning are Apriori Algorithm, Eclat, FP-
growth algorithm.

Advantages and Disadvantages of Unsupervised


Learning Algorithm
Advantages:
o These algorithms can be used for complicated tasks compared to the supervised
ones because these algorithms work on the unlabeled dataset.
o Unsupervised algorithms are preferable for various tasks as getting the unlabeled
dataset is easier as compared to the labelled dataset.

Disadvantages:

o The output of an unsupervised algorithm can be less accurate as the dataset is


not labelled, and algorithms are not trained with the exact output in prior.
o Working with Unsupervised learning is more difficult as it works with the
unlabelled dataset that does not map with the output.

Applications of Unsupervised Learning

o Network Analysis: Unsupervised learning is used for identifying plagiarism and


copyright in document network analysis of text data for scholarly articles.
o Recommendation Systems: Recommendation systems widely use unsupervised
learning techniques for building recommendation applications for different web
applications and e-commerce websites.
o Anomaly Detection: Anomaly detection is a popular application of unsupervised
learning, which can identify unusual data points within the dataset. It is used to
discover fraudulent transactions.
o Singular Value Decomposition: Singular Value Decomposition or SVD is used to
extract particular information from the database. For example, extracting
information of each user located at a particular location.

3. Semi-Supervised Learning
Semi-Supervised learning is a type of Machine Learning algorithm that lies
between Supervised and Unsupervised machine learning. It represents the
intermediate ground between Supervised (With Labelled training data) and
Unsupervised learning (with no labelled training data) algorithms and uses the
combination of labelled and unlabeled datasets during the training period.

Although Semi-supervised learning is the middle ground between supervised and


unsupervised learning and operates on the data that consists of a few labels, it mostly
consists of unlabeled data. As labels are costly, but for corporate purposes, they may
have few labels. It is completely different from supervised and unsupervised learning as
they are based on the presence & absence of labels.

To overcome the drawbacks of supervised learning and unsupervised learning


algorithms, the concept of Semi-supervised learning is introduced. The main aim
of semi-supervised learning is to effectively use all the available data, rather than only
labelled data like in supervised learning. Initially, similar data is clustered along with an
unsupervised learning algorithm, and further, it helps to label the unlabeled data into
labelled data. It is because labelled data is a comparatively more expensive acquisition
than unlabeled data.

We can imagine these algorithms with an example. Supervised learning is where a


student is under the supervision of an instructor at home and college. Further, if that
student is self-analysing the same concept without any help from the instructor, it
comes under unsupervised learning. Under semi-supervised learning, the student has to
revise himself after analyzing the same concept under the guidance of an instructor at
college.

Advantages and disadvantages of Semi-supervised


Learning
Advantages:

o It is simple and easy to understand the algorithm.


o It is highly efficient.
o It is used to solve drawbacks of Supervised and Unsupervised Learning
algorithms.

Disadvantages:

o Iterations results may not be stable.


o We cannot apply these algorithms to network-level data.
o Accuracy is low.

4. Reinforcement Learning
Reinforcement learning works on a feedback-based process, in which an AI agent
(A software component) automatically explore its surrounding by hitting & trail,
taking action, learning from experiences, and improving its performance. Agent
gets rewarded for each good action and get punished for each bad action; hence the
goal of reinforcement learning agent is to maximize the rewards.

In reinforcement learning, there is no labelled data like supervised learning, and agents
learn from their experiences only.

The reinforcement learning process is similar to a human being; for example, a child
learns various things by experiences in his day-to-day life. An example of reinforcement
learning is to play a game, where the Game is the environment, moves of an agent at
each step define states, and the goal of the agent is to get a high score. Agent receives
feedback in terms of punishment and rewards.

Due to its way of working, reinforcement learning is employed in different fields such
as Game theory, Operation Research, Information theory, multi-agent systems.

A reinforcement learning problem can be formalized using Markov Decision


Process(MDP). In MDP, the agent constantly interacts with the environment and
performs actions; at each action, the environment responds and generates a new state.

Categories of Reinforcement Learning


Reinforcement learning is categorized mainly into two types of methods/algorithms:

o Positive Reinforcement Learning: Positive reinforcement learning specifies


increasing the tendency that the required behaviour would occur again by adding
something. It enhances the strength of the behaviour of the agent and positively
impacts it.
o Negative Reinforcement Learning: Negative reinforcement learning works
exactly opposite to the positive RL. It increases the tendency that the specific
behaviour would occur again by avoiding the negative condition.

Real-world Use cases of Reinforcement Learning

o Video Games:
RL algorithms are much popular in gaming applications. It is used to gain super-
human performance. Some popular games that use RL algorithms
are AlphaGO and AlphaGO Zero.
o Resource Management:
The "Resource Management with Deep Reinforcement Learning" paper showed
that how to use RL in computer to automatically learn and schedule resources to
wait for different jobs in order to minimize average job slowdown.
o Robotics:
RL is widely being used in Robotics applications. Robots are used in the industrial
and manufacturing area, and these robots are made more powerful with
reinforcement learning. There are different industries that have their vision of
building intelligent robots using AI and Machine learning technology.
o Text Mining
Text-mining, one of the great applications of NLP, is now being implemented
with the help of Reinforcement Learning by Salesforce company.

Advantages and Disadvantages of Reinforcement


Learning
Advantages

o It helps in solving complex real-world problems which are difficult to be solved


by general techniques.
o The learning model of RL is similar to the learning of human beings; hence most
accurate results can be found.
o Helps in achieving long term results.

Disadvantage

o RL algorithms are not preferred for simple problems.


o RL algorithms require huge data and computations.
o Too much reinforcement learning can lead to an overload of states which can
weaken the results.
Challenges in Machine Learning
Although machine learning is being used in every industry and helps organizations
make more informed and data-driven choices that are more effective than classical
methodologies, it still has so many problems that cannot be ignored. Here are some
common issues in Machine Learning that professionals face to inculcate ML skills and
create an application from scratch.

1. Inadequate Training Data


The major issue that comes while using machine learning algorithms is the lack of
quality as well as quantity of data. Although data plays a vital role in the processing of
machine learning algorithms, many data scientists claim that inadequate data, noisy
data, and unclean data are extremely exhausting the machine learning algorithms. For
example, a simple task requires thousands of sample data, and an advanced task such as
speech or image recognition needs millions of sample data examples. Further, data
quality is also important for the algorithms to work ideally, but the absence of data
quality is also found in Machine Learning applications. Data quality can be affected by
some factors as follows:

o Noisy Data- It is responsible for an inaccurate prediction that affects the decision
as well as accuracy in classification tasks.
o Incorrect data- It is also responsible for faulty programming and results
obtained in machine learning models. Hence, incorrect data may affect the
accuracy of the results also.
o Generalizing of output data- Sometimes, it is also found that generalizing
output data becomes complex, which results in comparatively poor future
actions.
2. Poor quality of data
As we have discussed above, data plays a significant role in machine learning, and it
must be of good quality as well. Noisy data, incomplete data, inaccurate data, and
unclean data lead to less accuracy in classification and low-quality results. Hence, data
quality can also be considered as a major common problem while processing machine
learning algorithms.

3. Non-representative training data


To make sure our training model is generalized well or not, we have to ensure that
sample training data must be representative of new cases that we need to generalize.
The training data must cover all cases that are already occurred as well as occurring.

Further, if we are using non-representative training data in the model, it results in less
accurate predictions. A machine learning model is said to be ideal if it predicts well for
generalized cases and provides accurate decisions. If there is less training data, then
there will be a sampling noise in the model, called the non-representative training set. It
won't be accurate in predictions. To overcome this, it will be biased against one class or
a group.

Hence, we should use representative data in training to protect against being biased
and make accurate predictions without any drift.

4. Overfitting and Underfitting


Overfitting:

Overfitting is one of the most common issues faced by Machine Learning engineers and
data scientists. Whenever a machine learning model is trained with a huge amount of
data, it starts capturing noise and inaccurate data into the training data set. It negatively
affects the performance of the model. Let's understand with a simple example where we
have a few training data sets such as 1000 mangoes, 1000 apples, 1000 bananas, and
5000 papayas. Then there is a considerable probability of identification of an apple as
papaya because we have a massive amount of biased data in the training data set;
hence prediction got negatively affected. The main reason behind overfitting is using
non-linear methods used in machine learning algorithms as they build non-realistic data
models. We can overcome overfitting by using linear and parametric algorithms in the
machine learning models.
Methods to reduce overfitting:

o Increase training data in a dataset.


o Reduce model complexity by simplifying the model by selecting one with fewer
parameters
o Ridge Regularization and Lasso Regularization
o Early stopping during the training phase
o Reduce the noise
o Reduce the number of attributes in training data.
o Constraining the model.

Underfitting:

Underfitting is just the opposite of overfitting. Whenever a machine learning model is


trained with fewer amounts of data, and as a result, it provides incomplete and
inaccurate data and destroys the accuracy of the machine learning model.

Underfitting occurs when our model is too simple to understand the base structure of
the data, just like an undersized pant. This generally happens when we have limited data
into the data set, and we try to build a linear model with non-linear data. In such
scenarios, the complexity of the model destroys, and rules of the machine learning
model become too easy to be applied on this data set, and the model starts doing
wrong predictions as well.

Methods to reduce Underfitting:

o Increase model complexity


o Remove noise from the data
o Trained on increased and better features
o Reduce the constraints
o Increase the number of epochs to get better results.

5. Monitoring and maintenance


As we know that generalized output data is mandatory for any machine learning model;
hence, regular monitoring and maintenance become compulsory for the same. Different
results for different actions require data change; hence editing of codes as well as
resources for monitoring them also become necessary.

6. Getting bad recommendations


A machine learning model operates under a specific context which results in bad
recommendations and concept drift in the model. Let's understand with an example
where at a specific time customer is looking for some gadgets, but now customer
requirement changed over time but still machine learning model showing same
recommendations to the customer while customer expectation has been changed. This
incident is called a Data Drift. It generally occurs when new data is introduced or
interpretation of data changes. However, we can overcome this by regularly updating
and monitoring data according to the expectations.

7. Lack of skilled resources


Although Machine Learning and Artificial Intelligence are continuously growing in the
market, still these industries are fresher in comparison to others. The absence of skilled
resources in the form of manpower is also an issue. Hence, we need manpower having
in-depth knowledge of mathematics, science, and technologies for developing and
managing scientific substances for machine learning.

8. Customer Segmentation
Customer segmentation is also an important issue while developing a machine learning
algorithm. To identify the customers who paid for the recommendations shown by the
model and who don't even check them. Hence, an algorithm is necessary to recognize
the customer behavior and trigger a relevant recommendation for the user based on
past experience.

9. Process Complexity of Machine Learning


The machine learning process is very complex, which is also another major issue faced
by machine learning engineers and data scientists. However, Machine Learning and
Artificial Intelligence are very new technologies but are still in an experimental phase
and continuously being changing over time. There is the majority of hits and trial
experiments; hence the probability of error is higher than expected. Further, it also
includes analyzing the data, removing data bias, training data, applying complex
mathematical calculations, etc., making the procedure more complicated and quite
tedious.
10. Data Bias
Data Biasing is also found a big challenge in Machine Learning. These errors exist when
certain elements of the dataset are heavily weighted or need more importance than
others. Biased data leads to inaccurate results, skewed outcomes, and other analytical
errors. However, we can resolve this error by determining where data is actually biased
in the dataset. Further, take necessary steps to reduce it.

Methods to remove Data Bias:

o Research more for customer segmentation.


o Be aware of your general use cases and potential outliers.
o Combine inputs from multiple sources to ensure data diversity.
o Include bias testing in the development process.
o Analyze data regularly and keep tracking errors to resolve them easily.
o Review the collected and annotated data.
o Use multi-pass annotation such as sentiment analysis, content moderation, and
intent recognition.

11. Lack of Explainability


This basically means the outputs cannot be easily comprehended as it is programmed in
specific ways to deliver for certain conditions. Hence, a lack of explainability is also
found in machine learning algorithms which reduce the credibility of the algorithms.

12. Slow implementations and results


This issue is also very commonly seen in machine learning models. However, machine
learning models are highly efficient in producing accurate results but are time-
consuming. Slow programming, excessive requirements' and overloaded data take more
time to provide accurate results than expected. This needs continuous maintenance and
monitoring of the model for delivering accurate results.

13. Irrelevant features


Although machine learning models are intended to give the best possible outcome, if
we feed garbage data as input, then the result will also be garbage. Hence, we should
use relevant features in our training sample. A machine learning model is said to be
good if training data has a good set of features or less to no irrelevant features.

Training and Testing loss


In the context of machine learning, particularly in supervised learning, “training loss,” “validation
loss,” and “evaluation loss” are terms used to describe different types of loss functions that are
used during the training and evaluation of a model. Let’s define each term:
Training Loss:
 The training loss is a measure of how well a machine learning model is performing on the
training data.
 It is calculated during the training process and is used to update the model’s parameters
through techniques like gradient descent to minimize this loss.
 The training loss reflects how well the model is fitting the training data. It should decrease
over time as the model learns from the data.
 However, a very low training loss doesn’t necessarily mean the model will perform well on
new, unseen data, as it may have overfit the training data.
 In a classification problem, a common loss function is the cross-entropy loss (also known as
log loss or logistic loss).
 For a single data point (i), where the true label is denoted as y_i (0 or 1 for binary
classification, or a one-hot encoded vector for multi-class classification) and the predicted
probability of the positive class is denoted as p_i, the cross-entropy loss can be expressed as:
L_train(i) = -[y_i * log(p_i) + (1 — y_i) * log(1 — p_i)]
 The training loss (L_train) is typically computed as the average of this loss over all training
data points:
L_train = (1/N_train) * Σ L_train(i) for i in [1, N_train]
Validation Loss:
 The validation loss is a measure of how well a machine learning model is performing on a
separate dataset called the validation set.
 During training, a portion of the data (not used for training) is set aside as the validation set.
 The model’s performance on the validation set is evaluated periodically (after each epoch, for
example), and the validation loss is calculated.
 The validation loss helps assess how well the model generalizes to data it hasn’t seen during
training.
 It is used to monitor the model’s performance and to detect overfitting. If the validation loss
starts increasing while the training loss is decreasing, it’s a sign of overfitting.
 Similar to the training loss, the validation loss is computed using the cross-entropy loss
function.
 For each data point in the validation set, you calculate the loss in the same way as for the
training data.
 The validation loss (L_validation) is computed as the average of the losses over all validation
data points:
L_validation = (1/N_validation) * Σ L_validation(i) for i in [1, N_validation]
Evaluation Loss (Test Loss)
 The evaluation loss, sometimes referred to as the test loss, is a measure of the model’s
performance on a completely separate dataset that it has never seen before, known as the test
set.
 The test set is used to provide an unbiased estimate of the model’s generalization
performance.
 After the model has been trained and its hyperparameters have been tuned based on the
training and validation performance, it is evaluated on the test set to assess how well it will
perform in real-world scenarios.
 The evaluation loss on the test set helps determine the model’s readiness for deployment and
provides an estimate of its expected performance on unseen data.
 The evaluation loss, often referred to as the test loss, is calculated using the same loss
function (cross-entropy) as the training and validation losses.
 For each data point in the test set, you calculate the loss in the same way as for training and
validation.
 The evaluation loss (L_evaluation or L_test) is computed as the average of the losses over all
test data points:
L_evaluation = (1/N_test) * Σ L_evaluation(i) for i in [1, N_test]
In summary, training loss is used to optimize the model’s parameters during training, validation
loss helps monitor the model’s performance during training and detect overfitting, and evaluation
loss on the test set provides an unbiased estimate of the model’s generalization performance on
unseen data. It’s crucial to keep these different types of losses separate to ensure that the model’s
performance is accurately assessed and that it can make good predictions on new, real-world data.
Tradeoffs
Trade off in statistical learning : Statistical learning involves using data to make predictions,
classify objects, or gain insights about relationships between variables. There are several
tradeoffs inherent in the process of statistical learning, and understanding these tradeoffs is
crucial for making informed decisions when developing and applying machine learning models.
Here are some key tradeoffs in statistical learning: Bias-Variance Tradeoff: This is one of the
fundamental tradeoffs in statistical learning.

Bias refers to the error introduced by approximating a real-world problem, which may be
complex, by a simplified model. Variance refers to the model's sensitivity to small fluctuations in
the training data. Models with high bias tend to underfit the data, while models with high
variance tend to overfit the data. There's a balance to strike between bias and variance to achieve
good generalization to new, unseen data.

Model Complexity: Choosing the right level of model complexity is another tradeoff. Simple
models, like linear regression, may not capture intricate relationships in the data, while complex
models, like deep neural networks, may memorize noise and overfit. A more complex model can
better fit the training data, but it might struggle to generalize to new data points. Regularization
techniques can help control model complexity.

Underfitting and Overfitting: These are two sides of the same coin. Underfitting occurs when a
model is too simple to capture the underlying patterns in the data. Overfitting occurs when a
model is too complex and fits the noise in the data. Balancing between these two extremes is
crucial for building a model that performs well on both training and test data.

Feature Selection vs. Feature Engineering: Deciding which features to include in a model is a
tradeoff. Including irrelevant or redundant features can lead to overfitting, while excluding
important features can lead to underfitting. Feature engineering involves creating new features
from existing ones, which can improve the model's performance, but also introduces complexity
and the potential for overfitting.
Interpretability vs. Performance: In many cases, more complex models offer better performance,
but they can be harder to interpret. Simple models like linear regression are highly interpretable,
while complex models like deep neural networks might provide excellent results but lack
transparency in their decision-making process.

Computational Complexity: More complex models often require more computational resources
and time for training and inference. Choosing a model that strikes a balance between
computational efficiency and predictive power is important, especially in real-time or resource-
constrained applications.

Data Quality vs. Model Performance: The quality of the input data significantly impacts the
performance of the model. Garbage in, garbage out. Cleaning and preprocessing data can be
time-consuming, but it's crucial for the model's effectiveness.

Training Set Size: The size of the training dataset affects model performance. With a small
dataset, complex models can overfit, while with a large dataset, they might perform better.
However, collecting and managing large datasets can be resource-intensive.

Model Selection: Choosing the right algorithm or model architecture is important. Different
algorithms have different strengths and weaknesses, and the choice depends on the nature of the
problem, the available data, and the desired outcomes.

Tradeoff in Evaluation Metrics: Depending on the problem, different evaluation metrics might be
relevant. For example, precision and recall are tradeoffs in binary classification problems, where
increasing one often leads to a decrease in the other.

Understanding these tradeoffs and making informed decisions about them is a key skill for
practitioners in the field of statistical learning and machine learning. It involves a mix of domain
knowledge, experimentation, and understanding the nuances of the specific problem at hand.

You might also like