Unit 2 AI Project Cycle-1
Unit 2 AI Project Cycle-1
cycle
Computer Dept
Lesson
Objectives
Introduction to AI Project
Cycle
Problem Scoping
Data Acquisition
Data Exploration
Modelling
Evaluation
Tuesday, February 2, 20XX Sample Footer Text 2
Introduction to Project Cycle
4
1. Introduction
Let us assume that you must make a greeting card for your mother as it is
her birthday. You are very excited about it and have thought of many
ideas to execute the same. Let us look at some of the steps which you
might take to accomplish this task:
1. Look for some cool greeting card ideas from different sources. You
might go online and checkout some videos or you may ask someone
who has knowledge about it.
2. After finalising the design, you would make a list of things that are
required to make this card.
3. You will check if you have the material with you or not. If not, you could
get all the items required, ready for use.
4. Once you have everything with you, you would start making the card.
PROBLE DATA
M ACQUISITIO DATA MODELLING
EVALUATIO
EXPLORATION N
SCOPING N
For example, If you want to make an Artificially Intelligent system which can predict
the salary of any employee based on his previous salaries, you will feed the data of his
previous salaries into the machine. This is the data with which the machine can be
trained. Now, once it is ready, it will predict his next salary efficiently. The previous
salary data here is known as Training Data while the next salary prediction data set is
known as the Testing Data. For better efficiency of an AI project, the Training data
needs to be relevant and authentic.
In the previous example, if the training data was not of the previous salaries but of his
expenses, the machine would not have predicted his next salary correctly since the
whole training went wrong. Similarly, if the previous salary data was not authentic,
that is, February
Tuesday, it was not correct,
2, 20XX then
Sample too the prediction could have gone wrong. Hence….20For
Footer Text
Data Features
Data features refer to the type of data you want to
collect. In our previous example, data features would
be salary amount, increment percentage, increment
period, bonus, etc.
Fill them:
Now, let’s take various possibilities of these parameters and see in which case the elephant may be
spotted and in which case it may not. After looking through all the cases, we feed this data into the
machine along with the rules which tell the machine all the possibilities. The machine trains on this
data and now is ready to be tested.
Outlook=Overcas While testing the machine,
Temperature=Normal we tell the machine
Humidity=Nor that
Wind=Weak
t mal
Based on this testing dataset, now the machine will be able to tell if the elephant has been spotted
before or not and will display the prediction to us. This is known as a rule-based approach because we
fed the data along with rules to the machine and the machine after getting trained on them is now able
to predict answers for the same. A drawback/feature for this approach is that the learning is static. The
machine once trained, does not take into consideration any changes made in the original training
dataset. That is, if you try testing the machine on a dataset which is different from the rules and data
you fed it at the training stage, the machine will fail and will not learn from its mistake. Once trained,
the model cannot improvise itself based on feedbacks. Thus, machine learning gets introduced as an
extension to this as in that case, the machine adapts to change in data and rules and follows the
Tuesday, February 2, 20XX Sample Footer Text 34
updated path only, while a rule-based model does what it has been taught once.
Rule Based Approach
DRAWBACKS:
1.Learning is static:
The machine once trained, does not take into consideration any changes
made in the original training dataset. That is, at the training stage, the
machine will fail and will not learn from its mistake. if you try testing the
machine on a dataset which is different from the rules and data you fed it
The AI model gets trained on the data fed to it and then can design a model which is
adaptive to the change in data. That is, if the model is trained with X type of data and the
machine designs the algorithm around it, the model would modify itself according to the
changes which occur in the data so that all the exceptions are handled in this case.
For example, suppose you have a dataset comprising of 100 images of apples and
bananas each. These images depict apples and bananas in various shapes and sizes.
These images are then labelled as either apple or banana so that all apple images are
labelled ‘apple’ and all the banana images have ‘banana’ as their label. Now, the AI model
is trained with this dataset and the model is programmed in such a way that it can
distinguish between an apple image and a banana image according to their features and
can predict the label of any image which is fed to it as an apple or a banana. After
training, the machine is now fed with testing data. Now, the testing data might not have
similar images as the ones on which the model has been trained. So, the model adapts to
the features on which it has been trained and accordingly predicts if the image is of an
apple or banana. In this way, the machine learns by itself by adapting to the new data
Tuesday, February 2, 20XX Sample Footer Text 36
which is flowing in. This is the machine learning approach which introduces the
Learning Based Approach
The learning-based
approach can further be
divided into three parts:
• Supervised Learning
• Unsupervised Learning
• Reinforcement
Learning
Tuesday, February 2, 20XX Sample Footer Text 37
Supervised Learning
The dataset which is fed to the machine is labelled.
For example, you have a random data of 1000 dog images, and you
wish to understand some pattern out of it, you would feed this data
into the unsupervised learning model and would train the machine
onTuesday,
it. February
After training,
2, 20XX
the machine would produce patterns which it39
Sample Footer Text
Unsupervised learning models can be
further divided into two categories:
Clustering:
Refers to the unsupervised learning algorithm which can
cluster the unknown data according to the patterns or
trends identified out of it. The patterns observed might
be the ones which are known to the developer, or it might
even produce some unique patterns out of it.
Dimensionality Reduction:
We humans can visualise up to 3-Dimensions only but
according to a lot of theories and algorithms, there are
various entities which exist beyond 3-Dimensions. For
example, in Natural language Processing, the words are
N-Dimensional entities. Which means that we cannot
visualise them as they exist beyond our visualisation
ability. Hence, to make sense out of it, we need to reduce
their dimensions.
Tuesday, February 2, 20XX Here, dimensionality
Sample Footer Text reduction 40
Tuesday, February 2, 20XX Sample Footer Text 41
Neural Networks
Neural networks are loosely modelled after how
neurons in the human brain behave. It is
essentially a system of organizing machine
learning algorithms to perform certain tasks.
• Advantages:
• They can extract data features automatically
without needing the input of the programmer.
How Neural layer is further divided into several blocks called nodes.
• Next to it, are the hidden layers. Hidden layers are the
layers in which the whole processing occurs. Their name
essentially means that these layers are hidden and are not
visible to the user.
• The last hidden layer passes the final processed data to the
output layer which then gives it to the user as the final
output.
Tuesday, February 2, 20XX Sample Footer Text 43
Features of Neural Network