AI Study Note
AI Study Note
Chess-playing AIs, for example, are reactive systems that optimize the best
strategy to win the game. Reactive AI tends to be fairly static, unable to
learn or adapt to novel situations. Thus, it will produce the same output
given identical inputs.
2. Limited memory AI can adapt to past experience or update itself based on
own existence. Still in the realm of science fiction, some experts believe that
an AI will never become conscious or "alive".
What are the Applications of Artificial Intelligence?
Here is the list of the top 18 applications of AI (Artificial Intelligence):
1. AI Application in E-Commerce
Personalized Shopping
Artificial Intelligence technology is used to create recommendation engines
through which you can engage better with your customers. These
recommendations are made in accordance with their browsing history, preference,
and interests. It helps in improving your relationship with your customers and their
loyalty towards your brand.
AI-Powered Assistants
Virtual shopping assistants and chat bots help improve the user experience while
shopping online. Natural Language Processing is used to make the conversation
sound as human and personal as possible. Moreover, these assistants can have real-
time engagement with your customers. Did you know that on amazon.com, soon,
customer service could be handled by chat bots?
Fraud Prevention
Credit card frauds and fake reviews are two of the most significant issues that E-
Commerce companies deal with. By considering the usage patterns, AI can help
reduce the possibility of credit card fraud taking place. Many customers prefer to
buy a product or service based on customer reviews. AI can help identify and
handle fake reviews.
2. Applications of Artificial Intelligence in Education
Although the education sector is the one most influenced by humans, Artificial
Intelligence has slowly begun to seep its roots into the education sector as well.
Even in the education sector, this slow transition of Artificial Intelligence has
helped increase productivity among faculties and helped them concentrate more on
students than office or administration work.
Some of these applications in this sector include:
Administrative Tasks Automated to Aid Educators
Artificial Intelligence can help educators with non-educational tasks like task-
related duties like facilitating and automating personalized messages to students,
back-office tasks like grading paperwork, arranging and facilitating parent and
guardian interactions, routine issue feedback facilitating, managing enrollment,
courses, and HR-related topics.
Creating Smart Content
Digitization of content like video lectures, conferences, and textbook guides can be
made using Artificial Intelligence. We can apply different interfaces like
animations and learning content through customization for students from different
grades.
Artificial Intelligence helps create a rich learning experience by generating and
providing audio and video summaries and integral lesson plans.
Voice Assistants
Without even the direct involvement of the lecturer or the teacher, a student can
access extra learning material or assistance through Voice Assistants. Through this,
printing costs of temporary handbooks and also provide answers to very common
questions easily.
Personalized Learning
Using top AI technologies, hyper-personalization techniques can be used to
monitor students’ data thoroughly, and habits, lesson plans, reminders, study
guides, flash notes, frequency or revision, etc., can be easily generated.
3. Applications of Artificial Intelligence in Lifestyle
Artificial Intelligence has a lot of influence on our lifestyle. Let us discuss a few of
them.
Autonomous Vehicles
Automobile manufacturing companies like Toyota, Audi, Volvo, and Tesla use
machine learning to train computers to think and evolve like humans when it
comes to driving in any environment and object detection to avoid accidents.
Spam Filters
The email that we use in our day-to-day lives has AI that filters out spam emails
sending them to spam or trash folders, letting us see the filtered content only. The
popular email provider, Gmail, has managed to reach a filtration capacity of
approximately 99.9%.
Facial Recognition
Our favorite devices like our phones, laptops, and PCs use facial recognition
techniques by using face filters to detect and identify in order to provide secure
access. Apart from personal usage, facial recognition is a widely used Artificial
Intelligence application even in high security-related areas in several industries.
Recommendation System
Various platforms that we use in our daily lives like e-commerce,
entertainment websites, social media, video sharing platforms, like
YouTube, etc., all use the recommendation system to get user data and
provide customized recommendations to users to increase engagement. This
is a very widely used Artificial Intelligence application in almost all
industries.
astounding projects of all time. This project seeks to survey the whole night
sky every night, gathering over 80 terabytes of data in one go to study how
stars and galaxies in the cosmos change over time.
Clogged city streets are a key impediment to urban transportation all around the
world. Cities throughout the world have enlarged highways, erected bridges, and
established other modes of transportation such as train travel, yet the traffic
problem persists. However, AI advancements in traffic management provide a
genuine promise of changing the situation.
Intelligent traffic management may be used to enforce traffic regulations and
promote road safety. For example, Alibaba's City Brain initiative in China uses AI
technologies such as predictive analysis, big data analysis, and a visual search
engine in order to track road networks in real-time and reduce congestion.
Building a city requires an efficient transformation system, and AI-based traffic
management technologies are powering next-generation smart cities.
Ride-Sharing
AI-enabled route planning using predictive analytics may help both businesses and
people. Ride-sharing services already achieve this by analyzing numerous real-
world parameters to optimize route planning.
AI-enabled route planning is a terrific approach for businesses, particularly
logistics and shipping industries, to construct a more efficient supply network by
anticipating road conditions and optimizing vehicle routes. Predictive analytics in
route planning is the intelligent evaluation by a machine of a number of road usage
parameters such as congestion level, road restrictions, traffic patterns, consumer
preferences, and so on.
Cargo logistics companies, such as vehicle transport services or other general
logistics firms, may use this technology to reduce delivery costs, accelerate
delivery times, and better manage assets and operations.
18. AI in Automotive Industry
The importance of Artificial Intelligence and machine learning in the automotive
sector cannot be overstated. With Artificial Intelligence driving more applications
to the automotive sector, more businesses are deciding to implement Artificial
Intelligence and machine learning models in production.
Manufacturing
We all wish to have a pleasant journey in our vehicles. Artificial Intelligence can
also help with this. When driving, Artificial Intelligence (AI) may assist drivers in
remaining focused by decreasing distractions, analyzing driving behaviors, and
enhancing the entire customer experience. Passengers can benefit from customized
accessibility as well as in-car delivery services thanks to AI.
Inspections
Everyone desires a premium vehicle and experience. Wouldn't you prefer to know
if something is wrong with your automobile before it breaks down? In this
application, AI enables extremely accurate predictive monitoring, fracture
detection, and other functions.
Why Is AI Used?
People leverage the strength of Artificial Intelligence because the work they need
to carry out is rising on a daily basis. So automating everyday tasks is a smart idea.
This saves the organization's staff while also increasing output. Furthermore, the
organization may obtain competent individuals for the company's development
through the use of Artificial Intelligence.
Also, businesses nowadays believe that they wish to automate all regular and
ordinary tasks. And they believe they can automate such routine tasks using a
simple application. Because automation is becoming more widespread as data
science advances. This AI is most commonly used in online chat portals. You've
undoubtedly seen the welcome message when you visit the websites. After that, the
genuine talk generally begins.
LEARNING IN AI.
Learning is “a process that leads to change, which occurs as a result of experience
and increases the potential for improved performance and future learning.
efficient.
4. Improving the accuracy and speed of medical diagnoses.
Note: A decision tree can contain categorical data (YES/NO) as well as numeric
data.
The picture above depicts a decision tree that is used to classify whether a person is
Fit or Unfit.
The decision nodes here are questions like ‘’‘Is the person less than 30 years of age?’,
‘Does the person eat junk?’, etc.and the leaves are one of the two possible outcomes
viz. Fit and Unfit.
Looking at the Decision Tree we can say make the following decisions:
if a person is less than 30 years of age and doesn’t eat junk food then he is Fit, if a
person is less than 30 years of age and eats junk food then he is Unfit and so on.
The initial node is called the root node (colored in blue), the final nodes are called
the leaf nodes (colored in green) and the rest of the nodes are called intermediate or
internal nodes.
The root and intermediate nodes represent the decisions while the leaf nodes
represent the outcomes.
ID3 in brief
ID3 stands for Iterative Dichotomiser 3 and is named such because the algorithm
iteratively (repeatedly) dichotomizes (divides) features into two or more groups at
each step.
Invented by Ross Quinlan, ID3 uses a top-down greedy approach to build a
decision tree. In simple words, the top-down approach means that we start
building the tree from the top and the greedy approach means that at each iteration
we select the best feature at the present moment to create a node.
Most generally ID3 is only used for classification problems with nominal features
only.
Statistical Learning.
Statistical Learning is a set of tools for understanding data. These tools broadly
come under two classes: supervised learning & unsupervised learning.
Generally, supervised learning refers to predicting or estimating an output based on
one or more inputs.
Supervised learning: Supervised learning, as the name indicates, has the
presence of a supervisor as a teacher. Basically supervised learning is when we
teach or train the machine using data that is well-labeled. Which means some data
is already tagged with the correct answer. After that, the machine is provided with
a new set of examples (data) so that the supervised learning algorithm analyses the
training data (set of training examples) and produces a correct outcome from
labeled data.
For instance, suppose you are given a basket filled with different kinds of fruits.
Now the first step is to train the machine with all the different fruits one by one
like this:
If the shape of the object is rounded and has a depression at the top, is red in
color, then it will be labeled as –Apple.
If the shape of the object is a long curving cylinder having Green-Yellow
2. Logistic Regression
3. Classification
6. Decision Trees
Advantages:-
1. Supervised learning allows collecting data and produces data output from
previous experiences.
2. Helps to optimize performance criteria with the help of experience.
computation problems.
4. It performs classification and regression tasks.
Steps
Unsupervised learning
Unsupervised learning is the training of a machine using information that is neither
classified nor labeled and allowing the algorithm to act on that information without
guidance. Here the task of the machine is to group unsorted information according
to similarities, patterns, and differences without any prior training of data.
Unlike supervised learning, no teacher is provided that means no training will be
given to the machine. Therefore the machine is restricted to find the hidden
structure in unlabeled data by itself.
For instance, suppose it is given an image having both dogs and cats which it has
never seen.
Thus the machine has no idea about the features of dogs and cats so we can’t
categorize it as ‘dogs and cats ‘. But it can categorize them according to their
similarities, patterns, and differences, i.e., we can easily categorize the above
picture into two parts. The first may contain all pics having dogs in them and the
second part may contain all pics having cats in them. Here you didn’t learn
anything before, which means no training data or examples.
It allows the model to work on its own to discover patterns and information that
was previously undetected. It mainly deals with unlabeled data.
Unsupervised learning is classified into two categories of algorithms:
Clustering: A clustering problem is where you want to discover the inherent
discover rules that describe large portions of your data, such as people that
buy X also tend to buy Y.
Types of Unsupervised Learning:-
Clustering
1. Exclusive (partitioning)
2. Agglomerative
3. Overlapping
4. Probabilistic
Clustering Types:-
1. Hierarchical clustering
2. K-means clustering
3. Principal Component Analysis
4. Singular Value Decomposition
5. Independent Component Analysis
learning.
3. Capable of finding previously unknown patterns in data.
the discovery of novel and potentially useful patterns that may not be
apparent from the outset.
6. Low cost: Unsupervised learning is often less expensive than supervised
Problem-solving is commonly known as the method to reach the desired goal or find a
solution to a given situation. In computer science, problem-solving refers to artificial
intelligence techniques, including various techniques such as forming efficient
algorithms, heuristics, and performing root cause analysis to find desirable solutions.
Problem-solving in Artificial Intelligence usually refers to researching a solution to a
problem by performing logical algorithms, utilizing polynomial and differential equations,
and executing them using modeling paradigms. There can be various solutions to a
single problem, which are achieved by different heuristics. Also, some problems have
unique solutions. It all rests on the nature of the given problem.
1. Steps of problem solving in AI
2. AI methods of problem solving
Evolutionary Computation
This problem-solving method utilizes the well-known evolution concept. The theory of
evolution works on the principle of survival of the fittest. It states that the organism
which can cope well with their environment in a challenging or changing environment
and reproduce, their future generations gradually inherit the coping mechanism,
generating the diversity in new child organisms. In this way, the new organisms are not
mere copies of the old ones but have the mixes of characteristics that go along with that
harsh environment. Humans are the most prominent example of the evolution process
that has evolved and diversified because of the accumulation of favorable mutations
over countless generations.
In AI, the evolution concept refers to the trial and error technique:
1. Randomly change the old solution to come up with the new one. Does it give better
results? If yes, then keep and throw away the rest of the solutions. If not, then discard it.
2. Go to point 1.
Genetic Algorithms
The evolution theory is the basis of genetic algorithms. These algorithms use the direct
random search method. The developers calculate the fit function to cross the two fittest
options to create a favorable child. The developers gather the population data and then
evaluate each member to calculate everyone’s fitness. It is estimated by how well each
member fits with the desired requirement. Then the developers use various selection
methods to keep the best members. Some of the ways are the following:
1. Rank Selection
2. Tournament Selection
3. Steady Selection
4. Roulette Wheel Selection (Fitness Proportionate Selection)
5. Elitism (A strategy in evolutionary algorithms where the best one or more
solutions, called the elites, in each generation, are inserted into the next, without
undergoing any change)
Although genetic algorithms do not always work best, they do not break easily, and the
inputs change slightly. The developers commonly use genetic algorithms to generate a
high-level solution to optimization and search problems by relying on bio-inspired
operations such as mutation, crossover, and selection.
The problem-solving techniques help in improving the performance of programs. The
AI-based searching algorithms provide high precision and maximum accuracy to results.
These algorithms are faster than others in execution and offer multiple searching
methods depending upon the problem faced. Implementing heuristics allows the
devising to conceptually more straightforward algorithms with cheaper computational
costs compared to using optimal algorithms. Evolutionary computations also help in
optimization and search problems. Overall, these techniques are the basis for solving
high-level problems in AI such as chess algorithms, and hill-climbing problems.
Why Problem Solving is Important in AI?
Decoding any type of problem needs specific organized measures to be observed.
Identical is the matter of solving issues by AI. The following are the details:
Goal –In this phase as soon as a crisis appears, the AI agent puts a goal or a mark. This
needs the agent to thoroughly examine and clarify the issue. This is a vital action as if
the goal for the issue is poorly developed then all the actions carried out to achieve the
goal would be useless.
Problem Description–This is one major stage of problem-solving. Whenever a problem
occurs, then the agent chooses what measures must be carried out to run to the
developed goal. This is accomplished in the subsequent actions:
Describing the State–A state area can be described as a group of all the accurate
conditions in which an agent can be joined when discovering a key to the crisis.
o
Specifying Primary State–For an agent to begin cracking the issue, it
must begin from a state. The primary state from where the agent begins
performing is directed to the primary state.
Collect Details–Now the agent collects data and utilizes the data needed
by it to fix the issue. These details will be collected with one-time
incidents as well as present pieces of knowledge.
Designing the Changes–Some issues are undersized and so these can
be deciphered efficiently. But most of the time issues will be such where
sound planning and implementation are required. Hence this needs
appropriate data structures and managing processes well in advance.
Problem Formulation in AI
It is one of the basic stages of problem-solving that determines what measure should be
brought to fulfill the developed goal. Problem formulation is the stage in problem
description that is utilized to comprehend and choose a course of activity that must be
evaluated to reach a goal. If there is more than one method an agent can attain its
objective, then it generates intricacy in terms of truly reaching the goal as there would
be too numerous measures and courses that the AI entity can carry to achieve the goal
that it induces chaos and a tremendous decline in the efficiency. Problem formulation
can be accomplished in many stages such as the description of the initial condition of
the agent, choosing probable steps that the agent can bear, and design of transition
standards to define the efforts of the agent.
Problem-solving agents in artificial intelligence
Here the issue is split into sub-issues. The effects of the different measures carried out
in cracking the last sub-problem are delivered to the following sub-problem and the
integrated outcome of the sub-problems ushers to the definitive solution. This needs
appropriate planning and implementation of changes.
Testing with the Goal State–In this phase, the outcomes generated from the agent are
analogized with that of the objective state. If the goal has been achieved, then the
agents block any additional activities and the issue arrives at the final state. But if the
goal is not achieved then the agent persists to discover activities to run to the goal.
Estimating the Expense of Path carried–Whenever an agent carries a course to
decipher a situation it permits a numeric value (or price) to that course. These prices are
then estimated by utilizing a price function. The estimated consequence is hence
employed in the agent’s implementation action. The solution which is achieved with the
minimum or most subordinate price of the path is called the perfect solution.
Artificial Intelligence (AI) can make lots of difference to human intelligence and their life. AI
helps machines behave more like humans in:
Talking
Thinking
Learning
Planning
Understanding
AI is also known as machine/Computer intelligence and is the result of the human need to
perform complex calculations, and is achieved by high-speed computers. AI brings intelligence
to machines.
The purpose of making machines intelligent is to make a person's life better, improve
productivity, relieve people from complex tasks, and reach remote and dangerous places while
keeping humans safe.
Machine learning (ML) is a type of AI that allows software applications to predict outcomes
accurately without the necessity of complex programming. Most ML algorithms use historical
data as input to predict new output values.
Hence, we can say ML is a subset of AI, i.e., all ML algorithms are classified as being part of AI.
ML enables the machine to automatically learn from data, improve performance from past
experiences, and make predictions. ML contains a set of algorithms that work on a huge amount
of data. ML algorithms use these data to train, build the model and generate new output
predictions from the built model.
Deep Learning (DL) is a subset of ML. It was inspired by the working of human brains. The term
Deep Learning usually refers to Deep Artificial Neural Networks.
Artificial Neural Networks (ANN) is a depiction of natural neurons in the brain. ANNs are
mainly involved with deep learning algorithms and mimic the functionality of the human brain.
The ANN can be used on all types of ML algorithms based on its functionality.
Hence, DL effectively teaches computers to do what humans naturally do, say learning by
example. It uses Neural Networks (NNs) to extract higher-level data. NNs consist of 3 layers, say
the input layer, the hidden layer, and the output layer. Each successive layer uses the output of
the preceding layer as its input. For example, optical reading uses lower layers to identify edges
and higher layers to identify letters.
There are many types of ML that can be used in different applications based on the type of
datasets used. ML type is determined by different criteria like the type and amount of data, ML
model requirement, how to train the ML model, etc.
Supervised Learning
Unsupervised Learning
Semi-Supervised Learning
Reinforcement Learning
DEEP LEARNING
What is Deep Learning?
Deep learning is a type of machine learning that uses artificial neural networks to learn from data.
Artificial neural networks are inspired by the human brain, and they can be used to solve a wide variety of
problems, including image recognition, natural language processing, and speech recognition.
Once a deep learning algorithm has been trained, it can be used to make predictions on new data. For
example, a deep learning algorithm that has been trained to recognize images of dogs can be used to
identify dogs in new images.
As the network learns, the weights on the connections between the nodes are adjusted so that the network
can better classify the data. This process is called training, and it can be done using a variety of
techniques, such as supervised learning, unsupervised learning, and reinforcement learning.
Once a neural network has been trained, it can be used to make predictions with new data it’s received.
Data availability: It requires large amounts of data to learn from. For using deep learning it’s a
big concern to gather as much data for training.
Computational Resources: For training the deep learning model, it is computationally expensive
because it requires specialized hardware like GPUs and TPUs.
Time-consuming: While working on sequential data depending on the computational resource it
can take very large even in days or months.
Interpretability: Deep learning models are complex; it works like a black box. It is very difficult
to interpret the result.
Over fitting: when the model is trained again and again, it becomes too specialized for the
training data, leading to over fitting and poor performance on new data.
3. GRU
This abbreviation stands for Gated Recurrent Unit. It’s a type of LSTM. The major
difference is that GRU has fewer parameters than LSTM, as it lacks an output gate.
GRUs are used for smaller and less frequent datasets, where they show better performance.
7. Transformer
The Transformer is a powerful deep learning architecture that has significantly impacted the
field of natural language processing (NLP). It was first introduced in a 2017 paper by Google
researchers and has since become a cornerstone in various advanced language models.
Unlike traditional models that rely on Recurrent Neural Networks (RNNs) for sequential
information extraction, Transformers leverage self-attention mechanisms to understand
context and relationships between different elements in a sequence.
Key points about the Transformer architecture include:
Self-Attention Mechanism
Transformers apply self-attention to model relationships between all elements in a
sequence, allowing them to capture dependencies regardless of position.
Encoder-Decoder Structure
They consist of an encoder that processes the input sequence and a decoder that generates
the output sequence, both utilizing self-attention.
Parallel Processing
Transformers can process input sequences in parallel, enhancing computational
efficiency.
Evolution of NLP
Transformers have led to the development of advanced models like BERT, GPT, and
LaMDA, which excel in tasks such as language understanding, generation, and
translation.
Applications
Transformers are widely used in tasks like machine translation, text generation, question-
answering, and more, showcasing their versatility and effectiveness in handling
sequential data.
The Transformer’s ability to capture complex relationships in data, its parallel processing
capabilities, and its impact on various NLP tasks make it a fundamental architecture in
modern deep learning research, driving advancements in language understanding and
generation.
8. Generative Adversarial Networks (GANs)
Generative Adversarial Networks (GANs) are a powerful class of deep learning models used
for generative tasks, where they automatically learn and generate new data instances that
resemble the original dataset. GANs consist of two primary components:
1. Generator: The generator network creates new data instances, such as images, based on
random input. It aims to generate outputs that are realistic and indistinguishable from real
data.
2. Discriminator: The discriminator network acts as a classifier, distinguishing between real
data instances from the original dataset and fake data generated by the generator. It assigns a
probability score to each input, indicating the authenticity of the data.
Key points about GANs include:
Adversarial Training: GANs operate in a competitive manner where the generator and
discriminator are trained simultaneously. The generator aims to produce realistic outputs
to fool the discriminator, while the discriminator learns to differentiate between real and
generated data.
Applications: GANs have diverse applications, including image generation, virtual
reality, predictive imagery, text-based image generation, and more. They are particularly
useful for tasks requiring the creation of new data based on existing patterns.
Conditional GANs: Conditional Generative Adversarial Networks (cGANs) are a type
of GAN that generates outputs based on additional auxiliary information, enhancing the
control and specificity of the generated data.
GANs have revolutionized generative modeling by enabling the creation of high-quality, realistic
data that can be used in various domains such as image synthesis, content creation, and pattern
recognition. Their ability to learn complex patterns and generate new data has made them a
fundamental tool in the field of deep learning.