0% found this document useful (0 votes)
57 views25 pages

Unit 3 - Decision Making Under Uncertainty in AI

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
57 views25 pages

Unit 3 - Decision Making Under Uncertainty in AI

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 25

Unit 3: Decision Making under Uncertainty in AI

3.1 Decision making under uncertainty refers to the process of making choices when
the outcomes are not fully predictable due to incomplete information, randomness, or
ambiguity. In artificial intelligence (AI), this involves making decisions in environments where
data or outcomes are uncertain, leading to challenges in predicting the best actions to take. For
example, a self-driving car deciding whether to change lanes in heavy traffic is making
decisions under uncertainty because it cannot perfectly predict the actions of nearby drivers.

AI systems face uncertainty in many situations, such as:

 Incomplete information: When the AI system doesn’t have access to all the data
needed to make a fully informed decision.
 Stochastic (random) environments: Where outcomes are probabilistic and have
inherent randomness.
 Dynamic environments: Where the state of the environment changes rapidly, and the
system can only partially observe it.

3.2 Approaches for Handling Uncertainty in AI

To make effective decisions in the face of uncertainty, AI systems use various techniques to
model, assess, and mitigate the risks associated with uncertain situations. Some common
methods include:

1. Probability and Bayesian Inference

 Probabilistic Models: Probability theory provides a foundation for reasoning under


uncertainty by assigning probabilities to different outcomes. Probabilistic models such
as Bayesian networks and Hidden Markov Models (HMM) are commonly used to
represent uncertain variables and their relationships.
 Bayesian Inference: Bayesian inference is used to update the probability of a
hypothesis as more evidence or data becomes available. By applying Bayes' theorem,
AI systems can adjust their beliefs about uncertain events based on prior information
and observed evidence.

1
Authored By: Dr. Anup Girdhar
Example: A medical diagnostic system may use Bayesian inference to update the probability
of a disease based on new test results.

2. Markov Decision Processes (MDPs)

 MDPs: An MDP is a mathematical framework for decision-making in situations where


outcomes are partly random and partly under the control of an agent. In an MDP, the
agent chooses actions that maximize the expected cumulative reward over time.
 Components of an MDP:
o States: All possible situations in which the agent can find itself.
o Actions: Choices available to the agent in each state.
o Transition Model: Probabilities of moving from one state to another after
taking an action.
o Reward Function: Immediate reward received after transitioning from one
state to another.

MDPs are useful for long-term planning in uncertain environments and are often solved using
Dynamic Programming or Reinforcement Learning techniques.

Example: A robot navigating a factory floor can use an MDP to determine the optimal path
while accounting for uncertainty in movement.

3. Utility Theory and Expected Utility

 Utility Theory: Utility theory is used to model preferences over different outcomes.
Utility values represent the satisfaction or value associated with each outcome, and the
goal is to maximize the expected utility.
 Expected Utility: This is the sum of the utilities of all possible outcomes, each
weighted by its probability of occurrence. By maximizing expected utility, AI systems
can make rational decisions even when outcomes are uncertain.

Example: In financial decision-making, an AI system might use utility theory to choose


investments that maximize the expected financial return, taking into account the risks and
potential rewards of each investment.

2
Authored By: Dr. Anup Girdhar
4. Decision Trees and Influence Diagrams

 Decision Trees: Decision trees are graphical representations of decision-making


processes. Each node represents a decision or uncertain event, and branches represent
possible actions or outcomes. Decision trees are useful for structuring decision
problems, evaluating the possible consequences of actions, and finding optimal
solutions.
 Influence Diagrams: An influence diagram is an extension of a decision tree that
incorporates probabilities and utilities in a compact graphical format. Influence
diagrams help AI systems visualize complex decision problems with multiple decision
points and uncertain events.

Example: A marketing AI system might use a decision tree to decide whether to launch a new
product, based on uncertain factors like market demand and competitor actions.

5. Reinforcement Learning (RL)

 RL: Reinforcement learning is a type of machine learning where an agent learns to


make decisions by interacting with an environment and receiving feedback in the form
of rewards or penalties. In uncertain environments, RL algorithms like Q-learning or
Deep Q-networks (DQN) help AI agents optimize their decisions based on cumulative
rewards.
 Exploration vs. Exploitation: RL techniques balance between exploring new actions
(to gain knowledge about uncertain environments) and exploiting known actions that
yield high rewards.

Example: In autonomous driving, an RL agent learns to navigate safely by interacting with a


simulated environment, adjusting its actions based on the uncertain behavior of other vehicles.

6. Fuzzy Logic

 Fuzzy Logic: Fuzzy logic allows AI systems to handle uncertainty by representing


information with degrees of truth rather than binary true/false values. This approach is
useful when dealing with vague or ambiguous data.

3
Authored By: Dr. Anup Girdhar
 Membership Functions: Fuzzy logic uses membership functions to represent the
degree to which an element belongs to a set, allowing AI systems to make decisions
based on partial truths.

Example: In temperature control systems, fuzzy logic allows a system to interpret input values
(like "slightly warm" or "very cold") and adjust the output accordingly, even under uncertain
conditions.

3.3 Comparison of Approaches


Approach Strengths Limitations
Probabilistic Handles various sources of Requires well-defined
Models uncertainty, supports updating probabilities, can be
beliefs with new evidence computationally expensive
MDPs Useful for planning and sequential Complexity grows with state and
decision-making action spaces
Utility Theory Allows rational decision-making Requires well-defined utilities
by quantifying preferences over for all outcomes
uncertain outcomes
Decision Trees Simple to understand and visualize, Becomes complex with large
useful for structured decision numbers of decisions
problems
Reinforcement Effective in environments where Requires large amounts of data,
Learning the agent can learn through can be slow to converge
interaction
Fuzzy Logic Deals with vagueness and Less precise, not suitable for all
ambiguity, interprets partial truths types of uncertainty

Example Code for Decision Making under Uncertainty (Bayesian Inference)

Here is an example using Bayesian Inference in Python to make a decision under uncertainty.
Suppose a diagnostic system is trying to determine the probability of a disease given a positive
test result.

4
Authored By: Dr. Anup Girdhar
#python
# Bayesian Inference example

# Given probabilities
P_disease = 0.01 # Prior probability of having the disease
P_positive_given_disease = 0.99 # Probability of a positive test if disease is present
P_positive_given_no_disease = 0.05 # Probability of a positive test if disease is absent

# Calculate P(Positive)
P_no_disease = 1 - P_disease
P_positive = P_positive_given_disease * P_disease + P_positive_given_no_disease * P_no_disease

# Calculate P(Disease | Positive) using Bayes' Theorem


P_disease_given_positive = (P_positive_given_disease * P_disease) / P_positive

print(f"Probability of having the disease given a positive test result: {P_disease_given_positive:.2%}")

Output:

#plaintext
Probability of having the disease given a positive test result: 16.67%

This result shows that even with a positive test result, the probability of actually having the
disease is only 16.67% due to the relatively low prior probability and test imperfections.

3.4 Inductive learning

It is a process in artificial intelligence where an agent generalizes from specific observations


or examples to form broader rules or patterns. This type of learning allows AI systems to make
decisions under uncertainty by leveraging patterns identified in training data to predict
outcomes in new, unseen situations.

5
Authored By: Dr. Anup Girdhar
What is Learning by Induction?

Inductive learning involves creating hypotheses or models based on specific training examples
and then using those models to make decisions about new data. It contrasts with deductive
reasoning, where the system starts with general rules and applies them to specific cases.

In decision-making under uncertainty, inductive learning helps AI systems infer probable


outcomes and make decisions by learning from past experiences and patterns in data. This
approach is especially important in dynamic environments where exact rules may not be
known, and the system must infer patterns and relationships from data.

How Learning by Induction Works

1. Collect Training Data: Gather observations or examples from the environment. These
can include labeled examples where the outcomes or decisions are known.
2. Model Creation: Use machine learning algorithms (e.g., decision trees, support vector
machines, neural networks) to build a model that generalizes from the training data.
3. Pattern Recognition: Identify patterns, correlations, and relationships in the data to
form decision rules or predictive models.
4. Hypothesis Testing: Validate the learned model on unseen data to ensure it generalizes
well.
5. Decision Making: Apply the learned model to make predictions and decisions when
faced with uncertainty in new situations.

Common Inductive Learning Algorithms for Decision Making

Several machine learning algorithms can be used for inductive learning and decision-making
under uncertainty:

1. Decision Trees

 How They Work: Decision trees split the data into branches based on feature values,
creating paths that lead to a decision (leaf node). The tree is trained using historical data
to identify which features and thresholds best split the data for accurate decision-
making.

6
Authored By: Dr. Anup Girdhar
 Example Use: A medical diagnosis system might use a decision tree to decide whether
a patient has a particular disease based on symptoms.
 Pros and Cons:
o Pros: Easy to interpret, handle both categorical and numerical data.
o Cons: Can overfit the training data, leading to poor generalization.

2. Random Forests

 How They Work: Random forests combine multiple decision trees, each trained on
different subsets of data, to make more robust and accurate predictions. The final
decision is based on the majority vote or average prediction from all trees.
 Example Use: Financial decision-making, where an AI needs to assess the risk of
approving a loan under uncertain conditions.
 Pros and Cons:
o Pros: Reduces overfitting, robust against noisy data.
o Cons: Less interpretable than a single decision tree.

3. Bayesian Networks

 How They Work: Bayesian networks represent a set of variables and their conditional
dependencies using a directed acyclic graph (DAG). They use Bayes' theorem to update
beliefs based on new evidence, making them well-suited for decision-making under
uncertainty.
 Example Use: A fault diagnosis system in engineering that determines the likelihood
of a machine failure given certain observed symptoms.
 Pros and Cons:
o Pros: Can handle uncertainty effectively, can be updated dynamically with new
data.
o Cons: Requires knowledge of conditional probabilities and structure.

4. Support Vector Machines (SVM)

 How They Work: SVMs classify data by finding the hyperplane that best separates
classes in a high-dimensional space. This is useful when making binary decisions under
uncertainty.

7
Authored By: Dr. Anup Girdhar
 Example Use: Spam detection systems that classify emails as spam or not based on the
content.
 Pros and Cons:
o Pros: Effective in high-dimensional spaces, works well with small datasets.
o Cons: Not ideal for larger datasets, can be sensitive to the choice of kernel.

5. Neural Networks

 How They Work: Neural networks consist of interconnected layers of nodes that
process data and learn complex patterns through backpropagation. They are highly
effective for decision-making in uncertain environments, especially when trained on
large datasets.
 Example Use: Predicting stock prices based on historical trends and external factors.
 Pros and Cons:
o Pros: Can model complex and nonlinear relationships, highly accurate for large
datasets.
o Cons: Requires significant computational resources, can be a "black box" with
less interpretability.

Handling Uncertainty in Inductive Learning

Inductive learning models can handle uncertainty by incorporating probability distributions and
learning from data that includes varying levels of reliability. Here are some techniques:

1. Ensemble Learning: Combining multiple models (e.g., using random forests or


gradient boosting) can improve robustness and reduce the impact of uncertainty by
averaging or voting on predictions.
2. Probabilistic Outputs: Algorithms like logistic regression and Bayesian networks
provide probabilistic outputs rather than deterministic ones, allowing the system to
express uncertainty in the form of confidence scores or probabilities.
3. Regularization: Techniques like L1 and L2 regularization in models help prevent
overfitting, ensuring that the model generalizes well to new data, which is crucial in
uncertain environments.

8
Authored By: Dr. Anup Girdhar
4. Cross-Validation: Cross-validation helps to evaluate how the model performs on
different subsets of the data, providing an estimate of its reliability and performance
under uncertainty.

Example Python Code for Decision Making Using Inductive Learning (Decision Tree)
from sklearn import tree

from sklearn.model_selection import train_test_split

from sklearn.metrics import accuracy_score

# Sample data: [features: [Symptom1, Symptom2], target: Disease (1 or 0)]

X = [[1, 1], [1, 0], [0, 1], [0, 0], [1, 1], [0, 0]]

y = [1, 0, 0, 0, 1, 0]

# Split data into training and test sets

X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.3, random_state=42)

# Create and train the decision tree classifier

clf = tree.DecisionTreeClassifier()

clf = clf.fit(X_train, y_train)

# Make predictions

y_pred = clf.predict(X_test)

# Calculate accuracy

accuracy = accuracy_score(y_test, y_pred)

print(f"Model Accuracy: {accuracy:.2%}")

9
Authored By: Dr. Anup Girdhar
# Visualize the decision tree (optional, requires graphviz)

tree.plot_tree(clf)

Explanation of the Code:

 Training a Decision Tree: The code trains a decision tree classifier using a small
dataset of symptoms and corresponding outcomes (diseases).
 Evaluating the Model: It splits the data into training and testing sets and calculates the
accuracy on the test data.
 Decision Tree Visualization: tree.plot_tree() visualizes the decision tree structure,
making it easier to interpret the learned decision rules.

3.5 Introduction to Neural Network Genetic Algorithms

Neural networks and genetic algorithms (GAs) are two powerful techniques in artificial
intelligence that can be combined to solve complex optimization and learning problems.
Understanding the basics of how these two methods work individually and in combination is
essential for leveraging their potential in real-world applications.

1. Neural Networks: A Quick Overview

Neural networks are a subset of machine learning models inspired by the structure and
function of the human brain. They consist of interconnected nodes (or neurons) organized into
layers:

 Input Layer: Receives the input data.


 Hidden Layers: Perform complex computations and transformations on the input data.
 Output Layer: Produces the final output, such as a classification or a predicted value.

Training a neural network involves finding the optimal weights for connections between
neurons to minimize a loss function, which represents the error between predicted and actual
outputs. This is typically done using gradient descent and backpropagation.

10
Authored By: Dr. Anup Girdhar
2. Basics of Genetic Algorithms

Genetic algorithms (GAs) are search heuristics inspired by the process of natural selection.
They are used for optimization problems where the solution space is vast, and traditional
optimization methods might not be effective.

Key Components of Genetic Algorithms:

 Population: A set of candidate solutions (individuals) to the problem.


 Chromosome: Representation of an individual solution, typically as a string or array
of parameters.
 Fitness Function: Measures the quality or "fitness" of each individual in the
population.
 Selection: Chooses the fittest individuals for reproduction based on their fitness scores.
 Crossover (Recombination): Combines parts of two parent chromosomes to produce
offspring with mixed characteristics.
 Mutation: Randomly alters parts of a chromosome to introduce diversity and explore
new areas of the solution space.
 Iteration: The process repeats for multiple generations, gradually evolving better
solutions.

Workflow of a Genetic Algorithm:

1. Initialization: Create an initial population of random solutions.


2. Evaluation: Use the fitness function to assess each individual’s quality.
3. Selection: Choose individuals for reproduction based on their fitness scores.
4. Crossover: Combine selected parents to create offspring.
5. Mutation: Apply random changes to offspring.
6. Replacement: Form a new population from the offspring and possibly some of the
previous generation’s best individuals.
7. Repeat: Continue the process until a stopping criterion (e.g., a maximum number of
generations or satisfactory fitness level) is met.

11
Authored By: Dr. Anup Girdhar
3. Combining Neural Networks with Genetic Algorithms

Neural networks and genetic algorithms can be combined in various ways to take advantage of
the strengths of both approaches. Here are some common use cases:

3.1 Optimizing Neural Network Hyperparameters

Hyperparameters like learning rate, number of hidden layers, number of neurons, and
activation functions can significantly impact a neural network's performance. Instead of using
traditional grid search or manual tuning, GAs can be used to optimize these hyperparameters.

Process:

1. Chromosome Representation: Represent the hyperparameters as a chromosome (e.g.,


[learning_rate, num_layers, num_neurons, activation_function]).

2. Fitness Function: Train the neural network with the given hyperparameters and
evaluate its performance (e.g., using accuracy or loss on a validation set).
3. GA Process: Use the standard GA workflow (selection, crossover, mutation) to evolve
better sets of hyperparameters over generations.

3.2 Training Neural Network Weights

GAs can also be used to optimize the weights of a neural network, either alone or in
combination with backpropagation. This approach can be useful when:

 The loss function is non-differentiable or noisy, making gradient descent less


effective.
 Exploring non-convex solution spaces, where GAs can help escape local minima.

Process:

1. Chromosome Representation: Flatten the neural network's weight matrices into a


single array representing an individual.
2. Fitness Function: Evaluate the individual by unrolling the weights back into the
network, running the network on the training data, and measuring its performance.
3. GA Process: Evolve the population to find weights that minimize the loss function.

12
Authored By: Dr. Anup Girdhar
3.3 Evolutionary Neural Architecture Search

Neural Architecture Search (NAS) is the automated process of designing neural network
architectures. GAs can be used to evolve architectures by representing different neural network
configurations as chromosomes.

Example:

 Each chromosome can represent different architectural decisions, such as the number
of layers, types of layers (e.g., convolutional or fully connected), and connections
between them.
 The GA evolves architectures by selecting, recombining, and mutating them, searching
for the best-performing architecture according to a fitness function (e.g., accuracy on a
validation set).

Advantages of Using Genetic Algorithms with Neural Networks

 Avoids Local Minima: GAs can explore a broader search space and are less likely to
get stuck in local minima compared to gradient-based methods.
 Non-Differentiable Optimization: GAs can be used for optimizing functions that are
non-differentiable or complex.
 Parallelism: GAs can evaluate multiple solutions simultaneously, making them
suitable for parallel computing environments.

Disadvantages

 Computational Cost: GAs can be computationally expensive, especially when


combined with training neural networks, due to the large number of fitness evaluations.
 Convergence Speed: GAs might take longer to converge to an optimal solution
compared to gradient-based methods.
 Complexity: Designing an effective fitness function and representation can be
challenging for more complex problems.

13
Authored By: Dr. Anup Girdhar
Basic Python Example of Genetic Algorithm for Neural Network Hyperparameter
Tuning
#python
import random

from sklearn.model_selection import train_test_split

from sklearn.datasets import load_iris

from sklearn.neural_network import MLPClassifier

from sklearn.metrics import accuracy_score

# Load dataset

data = load_iris()

X_train, X_test, y_train, y_test = train_test_split(data.data, data.target, test_size=0.3,


random_state=42)

# Define fitness function

def fitness_function(params):

learning_rate, num_neurons = params

model = MLPClassifier(hidden_layer_sizes=(num_neurons,), learning_rate_init=learning_rate,


max_iter=500)

model.fit(X_train, y_train)

predictions = model.predict(X_test)

return accuracy_score(y_test, predictions)

# Genetic Algorithm parameters

population_size = 10

generations = 5

mutation_rate = 0.1

14
Authored By: Dr. Anup Girdhar
# Initialize population

population = [(random.uniform(0.001, 0.1), random.randint(5, 100)) for _ in range(population_size)]

# Evolution process

for generation in range(generations):

print(f"Generation {generation + 1}")

# Evaluate fitness

fitness_scores = [fitness_function(individual) for individual in population]

print("Best fitness:", max(fitness_scores))

# Select the top individuals for crossover (e.g., top 50%)

top_individuals = [x for _, x in sorted(zip(fitness_scores, population),


reverse=True)][:population_size // 2]

# Crossover: create new individuals by averaging top individuals

offspring = []

for i in range(len(top_individuals) - 1):

p1, p2 = top_individuals[i], top_individuals[i + 1]

child = ((p1[0] + p2[0]) / 2, (p1[1] + p2[1]) // 2)

offspring.append(child)

# Mutation: randomly change some parameters in the offspring

for child in offspring:

if random.random() < mutation_rate:

15
Authored By: Dr. Anup Girdhar
child = (child[0] * random.uniform(0.9, 1.1), max(5, min(100, child[1] + random.randint(-5,
5))))

# Create new population

population = top_individuals + offspring

# Best individual

best_individual = max(zip(fitness_scores, population))[1]

print(f"Best individual: Learning rate={best_individual[0]}, Number of


neurons={best_individual[1]}")

Explanation:

 Fitness Function: Measures the performance (accuracy) of an MLPClassifier with the


given hyperparameters.
 Crossover: Combines parameters from the top individuals to create offspring.
 Mutation: Introduces slight random changes to maintain diversity and avoid premature
convergence.
 Population: Represents different hyperparameter sets, each evaluated by the fitness
function.

3.6 Introduction to Rough Sets

Rough Set Theory (RST) is a mathematical approach to dealing with uncertainty and
vagueness in data analysis. It was introduced by Zdzisław Pawlak in the early 1980s and
provides a framework for reasoning about data without requiring additional information, such
as probabilities or fuzzy memberships. Rough sets are particularly useful in situations where it
is challenging to define sharp boundaries between different sets due to incomplete or imprecise
information.

16
Authored By: Dr. Anup Girdhar
Basic Concepts of Rough Sets

1. Information System:
o An information system (or data table) is a collection of objects (rows)
described by attributes (columns). Each object is characterized by values for
each attribute.
o Formally, an information system IS=(U,A), where U is the universe of objects,
and A is a set of attributes.

17
Authored By: Dr. Anup Girdhar
Applications of Rough Set Theory

Rough sets are widely used in:

 Data mining and knowledge discovery: Identifying hidden patterns and extracting
rules from datasets.
 Feature selection: Reducing the number of attributes needed for analysis while
preserving important information.
 Decision support systems: Supporting decisions by analyzing uncertain or incomplete
data.
 Classification and pattern recognition: Building models that can handle vague or
imprecise data.

Strengths of Rough Sets

 No Need for External Parameters: Unlike other methods such as probability-based


or fuzzy logic, rough sets do not require additional parameters (e.g., membership
functions or prior probabilities).
 Data-driven: RST relies solely on the information available in the dataset.
 Rule Generation: RST can generate decision rules that are easy to interpret and
analyze.

Limitations of Rough Sets

 Handling Real-valued Attributes: RST is best suited for categorical data, and
handling continuous attributes requires discretization.
 Computational Complexity: Calculating the lower and upper approximations for large
datasets can be computationally intensive.
 Dependency on Data Quality: Like other data-driven techniques, the quality of
analysis depends on the completeness and accuracy of the data.

Example of Rough Set Analysis

Consider an information table with objects described by attributes:

18
Authored By: Dr. Anup Girdhar
Object Attribute 1 Attribute 2 Decision
1 A X Yes
2 A Y No
3 B X Yes
4 B Y No
5 A X Yes

Python Code for Basic Rough Set Analysis

Here's a simple Python implementation to compute indiscernibility relations and


approximations:

import pandas as pd

from itertools import combinations

# Sample data

data = {

'Attribute 1': ['A', 'A', 'B', 'B', 'A'],

'Attribute 2': ['X', 'Y', 'X', 'Y', 'X'],

'Decision': ['Yes', 'No', 'Yes', 'No', 'Yes']

19
Authored By: Dr. Anup Girdhar
df = pd.DataFrame(data)

# Function to compute indiscernibility relation

def indiscernibility_relation(df, attributes):

groups = df.groupby(attributes).groups

return {tuple(sorted(v)) for v in groups.values()}

# Function to compute lower and upper approximations

def approximations(df, decision_value):

positive_region = df[df['Decision'] == decision_value].index

ind_relation = indiscernibility_relation(df, df.columns[:-1])

lower_approx = set()

upper_approx = set()

for group in ind_relation:

if set(group).issubset(positive_region):

lower_approx.update(group)

if set(group).intersection(positive_region):

upper_approx.update(group)

20
Authored By: Dr. Anup Girdhar
return lower_approx, upper_approx

# Compute approximations for decision 'Yes'

lower, upper = approximations(df, 'Yes')

print("Lower Approximation:", lower)

print("Upper Approximation:", upper)

print("Boundary Region:", upper - lower)

Explanation of the Code:

 indiscernibility_relation: Groups objects based on the provided attributes to find


indiscernible sets.
 approximations: Computes the lower and upper approximations for a given decision
value.

Output Example:

Lower Approximation: {0, 2, 4}

Upper Approximation: {0, 2, 4, 3}

Boundary Region: {3}

This output indicates that objects 0, 2, and 4 are definitely in the "Yes" decision set, while
object 3 lies in the boundary region and cannot be classified definitively.

21
Authored By: Dr. Anup Girdhar
3.7 Case Studies of Applications of Uncertainty in AI

Handling uncertainty is critical in many real-world AI applications where decisions must be


made with incomplete, imprecise, or noisy data. Below are a few notable case studies
illustrating how uncertainty is managed across different industries using various AI techniques:

Case Study 1: Medical Diagnosis System

Application: AI-powered diagnostic systems for identifying diseases.

Challenge: Medical data often come with uncertainties due to variability in test results,
incomplete patient histories, and subjective observations. Making accurate diagnoses under
these conditions is difficult.

Approach:

 Bayesian Networks: A probabilistic graphical model was employed to represent relationships


among symptoms, diseases, and patient data. Bayesian inference allowed the system to update
the probability of a disease given new evidence (e.g., test results or symptom reports).
 Result: By incorporating conditional probabilities and dynamically updating them as new data
were added, the AI system achieved a higher diagnostic accuracy and provided doctors with
probability estimates of potential diagnoses, helping them make more informed decisions.

Outcome: The use of Bayesian networks improved the system's ability to handle uncertainty,
making it reliable for supporting decision-making in clinical environments.

Case Study 2: Autonomous Vehicle Navigation

Application: Self-driving cars and automated navigation systems.

Challenge: Autonomous vehicles must make real-time decisions in environments filled with
uncertainty due to unpredictable human behavior, sensor noise, and varying weather
conditions.

Approach:

22
Authored By: Dr. Anup Girdhar
 Markov Decision Processes (MDPs): MDPs were used to model the decision-making process
under uncertainty. The vehicle's AI system evaluated the probability of different outcomes
when deciding actions like lane changes or braking.
 Sensor Fusion: Multiple sensors (e.g., LiDAR, radar, cameras) provided overlapping data to
mitigate individual sensor inaccuracies. A probabilistic model combined this data to create a
more reliable perception of the environment.
 Reinforcement Learning (RL): The system was trained using RL algorithms that rewarded
safe navigation and penalized risky behaviors. The AI learned to adapt and make optimal
decisions despite uncertain factors like unexpected pedestrian movement.

Outcome: These techniques significantly reduced the vehicle's error rate and improved its
response to uncertain situations, ensuring safer navigation and better adherence to traffic rules.

Case Study 3: Financial Market Prediction

Application: AI systems for stock market prediction and algorithmic trading.

Challenge: The stock market is highly uncertain, influenced by myriad factors including
economic indicators, political events, and investor sentiment. Predicting price movements and
identifying profitable trading opportunities require managing this uncertainty.

Approach:

 Machine Learning Models: Neural networks and decision trees were trained on historical
price data, company performance metrics, and economic indicators to identify patterns and
predict price movements.
 Monte Carlo Simulation: Used to model and simulate different potential future market
scenarios, accounting for variability in asset returns and volatility.
 Portfolio Optimization: Techniques like stochastic optimization were employed to construct
diversified portfolios that could maximize returns while minimizing risk under uncertainty.

Outcome: By integrating machine learning models with simulation techniques, the AI system
improved the prediction of stock price movements and optimized trading strategies, resulting
in better financial performance and reduced risk exposure.

23
Authored By: Dr. Anup Girdhar
Case Study 4: Weather Forecasting and Climate Modeling

Application: AI systems for short-term weather forecasting and long-term climate modeling.

Challenge: Weather forecasting involves inherent uncertainties due to the chaotic nature of the
atmosphere and limitations in observing and simulating weather patterns.

Approach:

 Probabilistic Forecasting: Ensemble models were used to generate multiple weather forecast
simulations. Each simulation varied slightly in initial conditions to represent different potential
outcomes.
 Neural Networks with Uncertainty Estimation: Advanced deep learning models like
Bayesian neural networks incorporated uncertainty estimation in their predictions, providing
not only the most likely weather outcome but also a confidence interval for each forecast.
 Data Assimilation: Combined data from various sources (e.g., satellite, radar, and ground
stations) using statistical techniques that accounted for measurement errors and data gaps.

Outcome: The integration of probabilistic models and data assimilation significantly enhanced
the accuracy and reliability of weather forecasts. Users received not just a single deterministic
forecast but a range of potential weather outcomes with associated probabilities.

Case Study 5: Supply Chain Management

Application: AI-based systems for managing supply chains and inventory in the retail and
manufacturing industries.

Challenge: Supply chain management involves multiple uncertain factors, such as fluctuations
in demand, supplier delays, and transportation issues. Making precise decisions to maintain
optimal inventory levels and ensure timely deliveries is complex under these conditions.

Approach:

 Stochastic Modeling: AI systems used stochastic models to simulate different demand


scenarios and optimize inventory levels to minimize the risk of stockouts or overstocking.
 Reinforcement Learning: RL algorithms were employed to dynamically adjust inventory
policies based on real-time data, learning the best actions to take under uncertain demand and
supply conditions.

24
Authored By: Dr. Anup Girdhar
 Fuzzy Logic: Incorporated in some systems to handle linguistic uncertainty (e.g., "high
demand" or "short delay"), allowing for more flexible decision-making.

Outcome: These AI techniques helped companies better predict demand, plan logistics, and
maintain optimal inventory levels. The result was a reduction in operational costs and improved
customer satisfaction due to fewer supply chain disruptions.

Case Study 6: Customer Sentiment Analysis

Application: AI systems for analyzing customer feedback and sentiment on social media or
review platforms.

Challenge: Customer feedback often contains unstructured and subjective data, making it
difficult to derive precise conclusions. There is inherent uncertainty in interpreting sentiments
due to language nuances, sarcasm, and varying expressions.

Approach:

 Natural Language Processing (NLP): Sentiment analysis models, including recurrent neural
networks (RNNs) and transformer-based architectures (e.g., BERT), were trained on large
corpora of labeled data to classify sentiment as positive, negative, or neutral.
 Fuzzy Sentiment Analysis: Fuzzy logic was used to handle uncertain and ambiguous
expressions, such as "somewhat satisfied" or "very disappointed," assigning degrees of
sentiment rather than binary classifications.
 Uncertainty Estimation: The models were enhanced with techniques to quantify the
uncertainty of their predictions, providing confidence scores for each sentiment classification.

Outcome: The AI system achieved higher accuracy in sentiment classification and helped
businesses gain deeper insights into customer satisfaction. The addition of confidence scores
enabled businesses to identify feedback with high uncertainty, prompting further manual
review.

25
Authored By: Dr. Anup Girdhar

You might also like