CLASS 10 AI Chapter 2
CLASS 10 AI Chapter 2
Data feature plays a crucial role in the Al Project Cycle.' Explain with the help of an example.
A data feature is a specific piece of information or characteristic of the data that we use to
teach an AI system. For example, if we want the AI to predict a person's height based on their
age and weight, "age" and "weight" are the data features. These features help the AI learn and
make accurate predictions. Choosing the right features is essential for the AI system to work
effectively and give correct answers.
Suppose you want to develop an AI-based recommendation system for an online shopping
website. The goal is to provide personalized product recommendations to users based on their
browsing and purchase history. You collect data for each user, and the features you gather
include:
1. Browsing History: The products the user has viewed or interacted with while exploring the
website.
2. Purchase History: The products the user has bought in the past.
3. Product Categories: The categories of products the user has shown interest in (e.g.,
electronics, clothing, books, etc.).
4. Reviews and Ratings: The user's feedback and ratings for products they have purchased.
The AI system will use these features to understand the user's preferences and interests, and
then suggest relevant products that they might be interested in purchasing.
Choosing the right data features is essential for the effectiveness of the recommendation
system.
Data exploration is an essential step before training an AI system. It involves looking closely
at the data to understand its quality, patterns, and relationships. By exploring the data, we can
identify and fix issues, select the most relevant information, and make the AI system more
accurate and effective. Without data exploration, the AI model may not work well and could
give wrong answers.
You use charts and graphs to see the data better. You also ask questions and test ideas you
have about the data
What is Data Exploration? Data exploration is the initial step in analyzing the data that will
be used to build an AI model. It involves understanding the data’s structure, patterns, and
characteristics.
Understanding Data: It helps in understanding what data is available and how it can
be used.
Identifying Patterns: Helps in spotting trends, patterns, and outliers in the data.
Data Cleaning: Identifies missing or inconsistent data that needs to be fixed.
Feature Selection: Helps in choosing the most relevant features for the model.
Visualizing Data: Using charts and graphs to see the data visually.
Summary Statistics: Calculating mean, median, mode, and standard deviation to get
a sense of data distribution.
Correlation Analysis: Checking how different features relate to each other.
1. Data Collection:
o Description: Gather data on house prices, sizes, locations, number of
bedrooms, etc.
o Example: A real estate company collects data on houses sold in the past year.
2. Data Exploration:
o Visualizing Data:
Example: Create a scatter plot to see the relationship between house
size and price.
Observation: Larger houses generally have higher prices, but there are
some exceptions.
o Summary Statistics:
Example: Calculate the average price of houses in different
neighbourhoods.
Observation: Houses in certain neighbourhoods are consistently more
expensive.
o Correlation Analysis:
Example: Check if the number of bedrooms correlates with house
price.
Observation: More bedrooms usually mean a higher price, but the
correlation is not as strong as with house size.
Summary:
3. 'In reinforcement learning, the machine is bound to learn from experiences only. Comment.
Reinforcement learning (RL) is a type of machine learning where a machine (like a computer
program) learns by interacting with its environment. Think of it like training a pet. Just like a
pet learns what to do by getting rewards (like treats) for good behavior and no rewards or
even punishments for bad behavior, a machine in RL learns what actions to take based on the
rewards it receives.
In simple terms, reinforcement learning is all about learning from experiences. The machine
tries different actions and learns which ones get the best rewards. Over time, it gets better at
making decisions that maximize these rewards. So, the machine isn't given a set of
instructions to follow. Instead, it figures out the best actions on its own by trying things out
and learning from what happens.
4. 'Rule-based system are static whereas systems based on learning approach are continuously
updated. Explain.
The AI project cycle is a process to create and use Artificial Intelligence (AI) systems
effectively.
The stages in a standard AI project cycle are :
1. **Problem Scoping:**
- This is the starting point of any AI project. You define the specific goals you want to
achieve with the AI system and identify the problems it will help solve. For example,
improving customer service, predicting sales, or automating repetitive tasks.
2. **Data Acquisition:**
- In this stage, you gather and collect the relevant data needed to train the AI system.
3. **Data Exploration:**
- Once you have the data, you explore and analyze it to find useful information and
patterns. This step helps you understand the data better
4. ** Data Modelling:**
- Now, you use the collected and analyzed data to train various AI models. The AI system
needs to learn from the data to make accurate predictions or decisions that align with the
goals set in the problem scoping stage.
5. **Evaluation:**
- In this stage, you assess the outputs or predictions generated by the trained AI models.
You evaluate their performance and select the best-suited AI system for deployment based on
how well they meet the defined goal
Problem scoping is the first step in any project or task. It's like understanding the problem
before finding a solution , Problem scoping is the process of defining and understanding a
problem before trying to solve it. It involves identifying the problem, understanding its
context, and setting clear objectives for the solution.
The 4W Framework is a simple and effective approach for problem scoping, which involves
asking four essential questions:
2. **Why is it a problem?**
- Understand the significance of the problem and its impact. Consider who is affected and
why it's crucial to find a solution.
3. **Who is involved?**
- Identify the stakeholders who are related to the problem, including those affected by it and
those who can contribute to the solution.
Using the 4W Framework, you gain a clear understanding of the problem's scope, context,
and importance, which is vital for devising an effective solution. By addressing these
questions, you lay a strong foundation for the rest of the project and ensure that the approach
aligns with the desired outcomes.
A neural network is a computational model inspired by the way biological neural networks in
the human brain process information. It consists of interconnected nodes, called neurons,
organized into layers. Each neuron receives input, processes it through an activation function,
and produces an output. Neural networks are capable of learning from data through a process
called training, where they adjust the strength of connections (weights) between neurons to
improve their ability to perform tasks such as pattern recognition, classification, regression,
and more complex decision-making tasks.
Neural networks, a type of artificial intelligence model inspired by the human brain, find
applications in various fields. Here are two notable examples:
1. Image Recognition: Neural networks are extensively used for tasks like image
classification and object detection. For instance, in autonomous vehicles, neural
networks can identify pedestrians, vehicles, and road signs from camera images in
real-time, enabling safe driving decisions.
2. Natural Language Processing (NLP): Neural networks are employed in NLP
applications such as language translation, sentiment analysis, and chatbots. They can
understand and generate human language, allowing machines to interact with users
more naturally and effectively.
"GIGO" means "Garbage In, Garbage Out." It's a saying that tells us if you put bad or wrong
information into a computer, you'll get bad or wrong results out. In other words, the quality of
what comes out depends on the quality of what you put in. This idea is important in computer
systems and AI because using incorrect or poor-quality data can lead to unreliable or
inaccurate outcomes. It reminds us to always use accurate and good-quality information for
better results.
Example: If you feed wrong or bad data into a system, it will produce incorrect results.
For instance, using outdated information for weather forecasting can lead to inaccurate
predictions.
Importance: It highlights the critical role of using accurate and reliable data to ensure the
reliability and correctness of computational outputs.
7. Name some of the sources from where you can get accurate and reliable data.
Obtaining accurate and reliable data is crucial for various applications. Here are some sources
where you can typically find such data:
Gather raw data for further Discover meaningful information within the data, such
Purpose analysis as trends, correlations, anomalies, or hidden insights
Data in original format (Raw Analysing data in detail, including visualization and
Nature of Data Data) statistical analysis
Examples recording customer orders Creating histograms, Pie Chart , Line chart
b) Data Modelling and Data Visualization
Classification is like sorting objects into predefined categories. You already know the labels
you want to assign, and you train a model to learn the rules for putting things in those
categories. For example, you might want to classify emails into "spam" or "not spam."
Clustering, on the other hand, is about finding natural groupings in data without knowing in
advance what those groups should be. It's like organizing a collection of items into piles
based on their similarities, even if you don't know the exact categories beforehand. For
instance, if you have a mix of different fruits, clustering would help you group similar ones
together, even if you didn't know the specific types of fruit beforehand.
In summary, classification is about assigning known labels, while clustering is about
discovering hidden patterns and grouping similar items together without predefined
categories. They serve different purposes and are used in different types of data analysis
tasks.
In simple terms, classification means sorting or grouping things based on their similarities or
characteristics. It's like putting things into different categories based on certain traits they
share. For example, sorting fruits into groups like apples, oranges, and bananas based on their
shape, colour, and type.
Let's consider an example to better understand this concept. Imagine you have a dataset
containing 100 images of pears and pomegranates. Your goal is to train a computer program
(model) to recognize whether an image shows a pear or a pomegranate. To do this, you
provide the model with a set of labelled examples. For instance, you show it images of pears
and tell it, "This is a pear," and then show it images of pomegranates and say, "This is a
pomegranate."
The model learns from these examples and adjusts its internal parameters to become better at
distinguishing between pears and pomegranates. After this training process, the model is
equipped to classify new images based on what it has learned. When you give it a new,
unseen image, it will predict whether it's a pear or a pomegranate.
Classification models can make decisions or predictions based on data without human
intervention. This can save time and resources in tasks that require categorization or labeling.
In the context of Neural Networks and AI, regression refers to a type of task where the goal is
to predict a numerical value based on input data. Unlike classification, which aims to
categorize data into distinct classes, regression is concerned with estimating a continuous
quantity.
For example, let's say you have data on the square footage and number of bedrooms of
houses, and you want to predict their sale prices. This is a regression task because the output
(sale price) is a numerical value.
In a neural network for regression, the network is designed to learn the relationship between
the input features (like square footage and number of bedrooms) and the target variable (sale
price). During training, the network adjusts its parameters to minimize the difference between
its predictions and the actual values in the training data.
Once trained, the neural network can then make accurate predictions about the target variable
for new, unseen data. This makes regression in neural networks a powerful tool for tasks like
price prediction, demand forecasting, and many other scenarios where we're interested in
estimating numerical values based on input features.
1. Limited Adaptability: These systems can't learn or change on their own. They rely
on rules set by humans and can't handle new situations that aren't in their rules. For
instance, a medical diagnosis system with fixed rules can't diagnose new diseases it
hasn't been programmed for.
2. Manual Updates: Whenever the rules need to change, it requires experts to manually
update them. This makes the systems less flexible and slower to adapt compared to AI
that can learn from new information.
3. Specific Tasks: They're good at tasks where the rules are clear, like determining if an
email is spam based on certain keywords. But they struggle with tasks that need
understanding context or making decisions based on uncertain information.
4. Not Scalable: As the number of rules grows or the problems get more complex,
managing and expanding rule-based systems becomes difficult. They work best for
simple, well-defined tasks in a narrow area.
In essence, while rule-based systems are useful for straightforward tasks with clear rules,
their inability to learn or handle new challenges limits them to narrow applications within AI.
What is the difference between supervised and unsupervised learning? Explain with example.
Artificial Neural Networks (ANNs) are inspired by the structure and function of the human
brain's neural networks. Here are some key similarities between artificial and human neural
networks:
1. Basic Structure: Both artificial and human neural networks consist of interconnected
nodes (artificial neurons or biological neurons) organized into layers. In ANNs, these
layers include input, hidden, and output layers, while in the human brain, neurons are
interconnected in complex networks.
2. Information Processing: Both networks process information through interconnected
nodes. In ANNs, each node receives inputs, applies a mathematical function
(activation function), and passes an output to the next layer. Similarly, biological
neurons receive signals (inputs), process them through electrochemical signals, and
transmit outputs to other neurons.
3. Learning and Adaptation: ANNs and human neural networks are capable of
learning and adaptation. In artificial networks, learning is achieved through
algorithms that adjust the strength (weights) of connections between neurons based on
training data. Human brains learn through experience, strengthening or weakening
connections (synapses) between neurons based on learning and memory formation.
4. Complexity and Parallel Processing: Both systems exhibit complex patterns of
connectivity and can process multiple pieces of information simultaneously (parallel
processing). ANNs can handle large amounts of data in parallel, similar to how the
brain processes vast amounts of information simultaneously through interconnected
neural pathways.
5. Functionality: Both types of networks are used for pattern recognition, decision-
making, and processing sensory information. ANNs excel in tasks such as image and
speech recognition, while human brains perform complex cognitive functions
including reasoning, emotions, and motor control.
While ANNs aim to mimic the structure and function of human neural networks, they are
simplified models designed to solve specific tasks efficiently. They represent a computational
approach to understanding and replicating aspects of biological intelligence, albeit at a
different level of complexity and scale compared to the human brain.
What do you understand by the term 'Decision Tree'? Explain with example.
Example:
Imagine you're deciding whether to go for a walk based on the weather:
Decision Node: The first node asks, "Is it raining?"
o Yes Branch: If it's raining, you might decide not to go for a walk.
o No Branch: If it's not raining, you might then ask, "Is it sunny?"
Yes Branch: If it's sunny, you decide to go for a walk.
No Branch: If it's not sunny, you might still go for a walk depending
on other factors like temperature.
In this example, the decision tree helps you make a sequence of decisions based on conditions
(rain and sun) to reach a final outcome (go for a walk or not).
Decision trees are used in various fields for decision-making and problem-
solving because they are easy to understand and visualize, making complex
decisions more manageable.
Age
/ \
Young Not Young
/ \
Income > Medium? Yes No
/ \
Gender = Female Yes No
/ \
Purchase No Purchase
Write 4 concern of Sustainable Development Goals for creating an Al project. (For example:
Eradication of poverty across the world.
Here are four concerns related to creating AI projects aligned with Sustainable Development
Goals (SDGs), explained simply:
1. Ethical Data Use: Make sure that the data used in AI projects is collected and used in
a fair and respectful way, especially when dealing with personal information like
health data (SDG 3 - Good Health and Well-being).
2. Avoiding Bias: Ensure that AI algorithms don't unfairly favour or disadvantage
certain groups of people, aiming for fair outcomes for everyone (SDG 4 - Quality
Education).
3. Environmental Impact: Try to reduce the environmental impact of AI technologies,
such as using less energy or creating less electronic waste from AI hardware (SDG 7 -
Affordable and Clean Energy).
4. Equal Access: Ensure that AI benefits reach everyone, especially those who might
not normally have access, to help reduce poverty and inequality (SDG 1 - No
Poverty).
The organization should accumulate biometric data related to the iris patterns of their
employees. This includes high-resolution images or scans of each employee's iris, which are
unique to each individual.
Biometric Details: Specific measurements and characteristics from these images that are
unique to each person's iris.
Employee Information: Details like employee IDs or names to link each iris scan to the
right person.
Attendance Records: Time-stamped logs of when employees use the system to check in
or out.
Rupa is running a food outlet and has an app for it. She needs a chatbot for the app which
would be able to give personalized chat experience to customers. What type of approach must
be used for creating a chatbot? (Rule-based/Learning-based)
Rupa should use a learning-based approach to create the chatbot for her food outlet app.
This means the chatbot will learn from customer interactions over time to provide
personalized responses and better understand what customers need and prefer. It's like
teaching the chatbot to get smarter as it interacts more with customers, making the
conversations more helpful and tailored to each person.
You are the coach of the football team. You have their data. It contains team players'
performance in tournaments, which is arranged in a tabular form. Now, you are required to
find the trends, patterns and connections in the datasets. How will you do this?
To find trends, patterns, and connections in the football team players' performance data, you
can follow these steps:
1. Data Exploration: First, look at the data to understand its structure and variables.
Identify what each column represents (e.g., player name, goals scored, assists, etc.).
2. Descriptive Statistics: Calculate basic statistics like averages, ranges, and
distributions for key metrics (e.g., average goals per game, total assists per player).
3. Visualization: Create graphs and charts (e.g., bar charts, scatter plots) to visualize
relationships between variables. For example, plot goals scored against assists to see
if there's a correlation.
4. Pattern Recognition: Look for recurring patterns or trends in the data. For instance,
do certain players perform consistently well in specific tournaments or against
particular opponents?
5. Statistical Analysis: Use statistical techniques like correlation analysis to quantify
relationships between variables. Determine if performance metrics (e.g., goals,
assists) are correlated with each other or with other factors like playing time or
position.
6. Machine Learning Techniques: Apply machine learning algorithms if you want to
predict future performance based on historical data or identify clusters of players with
similar performance profiles.
7. Contextual Analysis: Consider external factors that could influence performance,
such as weather conditions, team strategy changes, or player injuries.
By systematically exploring and analyzing the data, you can uncover valuable insights that
help optimize player strategies, improve team performance, and make informed decisions as a
coach.