Al Project Cycle
Al Project Cycle
IT project starts with the design which refers to the process of thinking about
interface design and its sil aspects.
• The secod stage is develop which is the process of develpping the solution with
an interface design as well as coding.
The third stage is testing the application which is a process of checking all the
modules and functions are working as per requirement or not.
The final stage of IT project cycle is to deploy the application or software for use.
Al Project Cycle Stages
Al Class 10 Al Project Cycle Class has the following main stages:
Problem Scoping
Data Acquisition
Data Exploration
Data Modelling
EVALUTATIING
You need to acquire data which will become the base of your project as it will
help you in understanding what the parameters that are related to problem
scoping are.
● You go for data acquisition by collecting data from various reliable and
authentic sources. Since the data you collect would be in large quantities, you
can try to give it a visual image of different types of representations like graphs,
databases, flow charts, maps, etc. This makes it easier for you to interpret the
patterns which your acquired data follows.
● After exploring the patterns, you can decide upon the type of model you
would build to achieve the goal. For this, you can research online and select
various models which give a suitable output.
● You can test the selected models and figure out which is the most efficient
one.
● The most efficient model is now the base of your AI project and you can
develop your algorithm around it.
● Once the modelling is complete, you now need to test your model on some
newly fetched data. The results will help you in evaluating your model and
improving it.
● Finally, after evaluation, the project cycle is now complete and what you get is
your AI project.
What is Problem Scoping?
Whenever we are starting any work, certain problems always associated with the
work or process. Actually we are surrounded by problems! These problems can
be small or big, sometimes we ignore them, sometimes we need an urgent
solution otherwise your work will suffer.
Important aspects of problem scoping
The following are few key points:
When you start with an Al project or model you need to do problem scoping first.
It the process of figure out the problem and what are the solutions.
• The Al project must have problem statement with required clarity
Problem Scoping example
The entire process of finding a specific solution is known as problem scoping.
Themes for problem scoping
Now look around you and find the specific field and select the problem which you
want to resolve. The
fields may be
Agriculture
Education
Banking
Health
Security
Infrastructure Transportation
If we talk about agriculture, then the following aspects should be taken care:
Pest issues
Yield Rates
Sowing and harvesting pattern
To solve these problems effectively you need to understand them and work for
the solutions. To find a specific solution you must start with the 4Ws.
These 4Ws are:
make an Artificially Intelligent system which can predict the salary of any
employee based on his previous salaries, you would feed the data of his previous
salaries into the machine. This is the data with which the machine can be
trained. Now, once it is ready, it will predict his next salary efficiently. The
previous salary data here is known as Training Data while the next salary
prediction data set is known as the Testing Data
Features of Data
Data features refer to the type of data you want to collect
It should be relevant to the problem statement.
For example, while analyzing the weakness of a batsman, you will have data
features such as
• Speed of ball
type of dismissal
type of bowler
• Type of swing
Type of spin
number of balls faced etc.
Methods of Data Aqcuisition
The most common methods of data acquisition are:
Surveys: Through Google Forms, MS Teams Forms or any other interface
• Web Scrapping: Some software are Scarpy, Scrape hero Cloud, ParseHub,
OutHitHub, Visual Web
Ripper, Import.io
Sensors: to convert physical parameters to electrical signals, to convert sensor
signals into a form that can be converted to digital values and to convert
conditioned sensor signals to digital values
Cameras: To capture images
Observations: Way of gathering data by watching behavior, events, or noting
physical characteristics in their natural setting
API (Application Program Interface
Open Source Datasets for data acquision
list of opensource datasets
Lionbridge Al
Amazon Mechanical Turk
Label Box
Figure Eight
Kaggle
http://mospi.nic.in/data
A collection of data that is huge in volume, yet growing exponentially with time.
Big Data for Al
It is a data with so large size and complexity that none of traditional data
management tools can store it or process it efficiently.
Examples of Big Data
Stock Exchange
Social Media Websites
Youtube and web series platforms
Types of Big Data
There are three types of big data:
Training. Testing and Validation of Data
Training set: The data where the model is trained on
Validation set: Data the model has not been trained on and used to tune
hyperparameters
Test set: In principle the same like the validation set, just used at the final end
after the model has been tailored. \
Extracting private data can be an offence. One of the most reliable and authentic
sources of information, are the open-sourced websites hosted by the
government. These government portals have general information collected in
suitable format which can be downloaded and used wisely.
Data Exploration
Data Exploration refers to the techniques and tools used to visualize data
through complex statistical methods.
Need of data visualization
Quickly get a sense of the trends, relationships and patterns contained within the
data.
Define strategy for which model to use at a later stage.
Communicate the same to others effectively.
To visualise data, we can use various types of visual representations.
Data Visualization tools
Microsoft Excel
Tableau
Qlikview
DataWrapper
Google Data Studio
Modelling
Artificial Intelligence, or Al, refers to any technique that enables computers to
mimic human intelligence.
Machine Learning, or ML, enables machines to improve at tasks with experience.
The machine learns from its mistakes and takes them into consideration in the
next execution.
Deep Learning, or DL, enables software to train itself to perform tasks with vast
amounts of data. In deep learning, the machine is trained with huge amounts of
data which helps it into training itself around the data.
Al Modelling refers to developing algorithms, also called models which can be
trained to get intelligent outputs. That is, writing codes to make a machine
artificially intelligent.
Types of Al models
Rule-Based model refers to setting up rules and training the model accordingly. It
follows an algorithm or code to train, test and validate data.
Learning-based refer to identifying the data by its attributes and behaviour and
training the model accordingly. There is no code or algorithm to train, test and
validate the data. It learns from past behaviour and attributes received from
data.
Decision Tree
Decision tree builds classification or regression models in the form of a tree
structure
It breaks down a dataset into smaller and smaller subsets while at the same time
an associated decision tree is incrementally developed.
The final result is a tree with decision nodes and leaf nodes.
Types of learning
There are three types of learning:
• Supervised
Unsupervised
Reinforcement
1. Supervised Learning
• The dataset which is fed to the machine is labelled.
A label is some information which can be used as a tag for data.
For example, students get grades according to the marks they secure in
examinations.
These grades are labels which categorise the students according to their marks.
Classification
Where the data is classified according to the labels.
• The entries are divided in two classes normally.
The boundary condition is defined to classify.
Regression
Regression deals with continuous data.
For example, if we know the growth rate, we can predict the salary of someone
after a certain number of years.
Regression is linear as well as non-linear.
2. Unsupervised Learning
An unsupervised learning model works on unlabelled dataset.
This means that the data which is fed to the machine is random and there is a
possibility that the person who is training the model does not have any
information regarding it.
The unsupervised learning models are used to identify relationships, patterns
and trends out of the data which is fed into it.
It helps the user in understanding what the data is about and what are the major
features identified by the machine in it.
Culstering
Refers to the unsupervised learning algorithm which can cluster the unknown
data according to the patterns or trends identified out of it.
The patterns observed might be the ones which are known to the developer or it
might even come up with some unique patterns out of it
Dimensionality reduction
We humans are able to visualise upto 3-Dimensions only.
if we have a ball in our hand, it is 3-Dimensions right now.
But if we click its picture, the data transforms to 2-D.
Hence, to reduce the dimensions and still be able to make sense out of the data,
we use Dimensionality Reduction.
3. Reinforcement Learning
Reinforcement Learning is defined as a Machine Learning method that is
concerned with how software agents should take actions in an environment
Reinforcement Learning is a part of the deep learning method that helps you to
maximize some portion of the cumulative reward.