Presentaton PPT Stock Prediction
Presentaton PPT Stock Prediction
Q2: What are the differences between AI, Machine Learning, and Deep Learning?
A2:
AI is the broader concept of creating intelligent machines.
Machine Learning is a subfield of AI that involves creating algorithms that can learn from
data.
Deep Learning is a subset of Machine Learning that utilizes artificial neural networks to
model complex patterns and representations.
Q3: What are the differences between Computational Intelligence and Machine Learning?
A3:
Computational Intelligence is a subfield of AI that emphasizes biologically-inspired methods,
such as neural networks, evolutionary algorithms, and fuzzy logic.
Machine Learning is a broader subfield that also includes statistical and mathematical
methods for learning from data.
Q4: What is data mining? What is knowledge engineering? What are the differences
between them?
A4:
Data mining is the process of discovering useful patterns and knowledge from large data
sets.
Knowledge engineering is the process of creating knowledge-based systems by encoding
expert knowledge in a structured form.
Data mining focuses on extracting knowledge from data, while knowledge engineering is
about encoding and representing knowledge.
Identification: Determine the problem and domain (e.g., diagnosing a medical condition).
Conceptualization: Define the main concepts and relationships (e.g., symptoms, diseases,
and their relationships).
Formalization: Represent the knowledge using formal methods (e.g., rules, ontologies).
Implementation: Create a knowledge-based system (e.g., a rule-based expert system for
diagnosis).
Evaluation: Assess the system's performance and correctness (e.g., comparing diagnoses
with expert opinions).
Maintenance: Update and refine the system as needed (e.g., incorporating new medical
knowledge).
Q8: Describe the 7 typical problem types addressed by intelligent systems and provide
examples to illustrate the types.
A8: The 7 problem types are:
Incompatible data: Data collected from different sources or formats. Solution: Data
integration or transformation.
Inconsistent data: Contradictory information in the data. Solution: Data cleaning, resolving
conflicts using domain knowledge or expert input.
Missing data: Absence of data for certain attributes. Solution: Data imputation, using
statistical methods or machine learning to estimate missing values.
Noisy data: Random errors or fluctuations in the data. Solution: Data smoothing, outlier
detection and removal, or noise-robust algorithms.
Imbalanced data: Unequal distribution of classes in the data. Solution: Resampling
techniques (oversampling or undersampling), cost-sensitive learning, or ensemble methods.
Q10: What are test cases? Give examples on how you would do it.
A10: Test cases are specific scenarios or inputs used to evaluate the functionality,
correctness, or performance of a system. Examples include providing input data to an AI
system and checking whether the output matches the expected result, or simulating user
interactions with a chatbot to test its response quality and accuracy.
Q11:
What is a prototype?
How is this useful in developing an AI system?
A11:
A prototype is an early, working version of a system, designed to demonstrate its core
functionality and gather feedback.
In AI development, prototyping helps to identify potential issues, refine the system's design,
and validate the system's usefulness before investing more resources in building the final
product.
Q12: Describe what are antecedents and consequent in rules? Give examples to illustrate
them.
A12: In rule-based systems, antecedents are the conditions that must be met for a rule to be
applied, while the consequent is the action or conclusion that follows. For example, in a
medical expert system, the rule could be "IF (symptom1 AND symptom2) THEN (diagnosis)",
where "symptom1 AND symptom2" is the antecedent, and "diagnosis" is the consequent.
Q13: List and describe the five members of the expert system development team.
A13:
Domain expert: Provides expertise in the specific subject matter.
Knowledge engineer: Translates domain knowledge into a formal representation for the
system.
Programmer: Implements the system using programming languages and tools.
Project manager: Oversees the project, coordinates the team, and ensures timelines and
objectives are met.
End-user: The person who will use the system, providing feedback and requirements to
ensure usability and relevance.
Q14:
Provide the basic structure of a rule-based expert system.
Describe each of the components in the structure.
Knowledge base: Contains the rules and facts representing domain knowledge.
Inference engine: Applies the rules to given facts or input data to derive conclusions.
Working memory: Temporarily stores facts and intermediate results during the inference
process.
User interface: Allows users to interact with the system, providing inputs and receiving
outputs.
Explanation facility (optional): Explains the system's reasoning and justifications for its
conclusions.
Q15:Describe the two common ways for conflict resolution in rule-based systems.
A15: The two common ways are:
Refraction: Prevents a rule from firing repeatedly by not allowing the same rule to fire again
until the working memory has changed.
Priority-based: Assigns a priority level to each rule, and when conflicts arise, the rule with
the highest priority is chosen.
Q17: What is the difference between knowledge-driven and data-driven AI systems? Give
examples for each.
A17: Knowledge-driven systems rely on predefined rules and expert knowledge, such as
rule-based expert systems for medical diagnosis. Data-driven systems learn from data, such
as a machine learning model for predicting house prices based on historical data.
Q20: List and describe the commonly used transfer functions (aka activation functions) used
in artificial neurons.
A20: Common activation functions include:
Q24: Explain how backpropagation network do the learning – you can use figures and steps
or algorithms to explain.
A24: Backpropagation is a supervised learning algorithm for neural networks. It involves the
following steps:
Update the weights using the computed gradients and a learning rate.
Repeat steps 2-5 for multiple epochs or until convergence.
Q25: In ANN, what is the learning rate? What could happen if the learning rate is too fast?
A25: The learning rate is a hyperparameter that controls the step size of weight updates
during training. If the learning rate is too high, the training may become unstable, leading to
oscillations or divergence instead of convergence to the optimal solution.
Q28: Given a set of available data, you normally divide them into 3 sets. What are the
Training, Testing, and Validation sets? How can you use them?
A28: Training set: Used to train the model. Testing set: Used to evaluate the model's
performance on unseen data. Validation set: Used to tune hyperparameters and select the
best model during training. The model is trained on the training set, its performance is
assessed on the validation set, and after selecting the best model, it is tested on the testing
set to estimate its performance on new data.
Q29: Provide the steps of how a Self-Organizing Map (SOM) do the learning.
A29: The steps for SOM learning are:
Q31: What is the main difference between Deep Learning and the Multi-layer Neural
Network?
A31:
Deep learning refers to neural networks with many hidden layers, allowing them to learn
hierarchical and complex representations.
Multi-layer neural networks can have multiple hidden layers,
but the term does not necessarily imply the depth or complexity of deep learning models.
Q32: What is Recurrent Neural Network (RNN) and how it performs training?
A32: RNN is a type of neural network designed for processing sequences of data. It contains
loops, allowing it to maintain a hidden state that can store information from previous time
steps. RNNs are trained using backpropagation through time (BPTT), an extension of
backpropagation that unfolds the network through time and computes gradients across
multiple time steps.
Q33: What are the decision steps in using Artificial Neural Networks for a problem?
A33: The decision steps include:
Defining the problem and identifying the input and output variables.
Selecting the appropriate network architecture (e.g., number of hidden layers, number of
neurons per layer).
Choosing the activation functions for neurons.
Determining the learning algorithm and its parameters (e.g., learning rate, batch size).
Splitting the data into training, validation, and testing sets.
Training the network using the learning algorithm.
Evaluating the network's performance on the validation and testing sets.
Tuning hyperparameters or adjusting the architecture to optimize performance.
Deploying the trained model for use in real-world applications.
Q34:
What is the difference between probability and fuzzy logic?
Give examples to show the difference.
A34: Probability deals with the likelihood of events, quantifying uncertainty in terms of
frequencies or degrees of belief. Fuzzy logic represents uncertainty using fuzzy sets, which
allow partial membership to multiple categories.
For example, in probability, an object might have a 70% chance of being red and a 30%
chance of being blue.
In fuzzy logic, the same object could have a 0.7 membership in the "red" fuzzy set and a 0.3
membership in the "blue" fuzzy set.
Q35: What is the difference between Boolean logic and fuzzy logic? Explain the difference
between them with an example.
A35: Boolean logic deals with binary values (true or false) and crisp categories, while fuzzy
logic allows for partial membership in categories, using continuous values between 0 and 1.
For example, in Boolean logic, an object is either tall or not tall.
In fuzzy logic, an object could have a membership of 0.8 in the "tall" fuzzy set, indicating it is
somewhat tall but not entirely.
Fuzzification: Converts crisp input values into fuzzy values using membership functions.
Inference: Combines fuzzy input values using fuzzy rules to produce fuzzy output values.
Defuzzification: Converts the fuzzy output values back into crisp values for decision-making
or control.
Q39: What is the difference between Mamdani and Sugeno fuzzy inference methods?
A39: Mamdani inference uses fuzzy sets for both input and output variables and typically
employs a defuzzification step to produce crisp output values.
Sugeno inference uses crisp functions (linear or constant) for output variables, simplifying
the defuzzification process and often resulting in more efficient computations.
They are well-suited for complex search spaces and global optimization problems.
They can explore multiple solutions simultaneously, reducing the likelihood of getting
trapped in local optima.
They are robust to noisy or incomplete data.
They can be applied to a wide range of problem domains with relatively little problem-
specific knowledge.
Q46: What is the difference between multi-objectives and single objective problems?
A46: Single objective problems have one optimization goal, such as minimizing cost or
maximizing profit. Multi-objective problems involve optimizing multiple objectives
simultaneously, often requiring trade-offs among conflicting objectives, such as minimizing
cost and maximizing performance.
Q47: What is Bayesian Networks and based on what theory does BN perform its reasoning
with?
A47:
Bayesian networks are graphical models that represent the probabilistic relationships
among a set of variables. They perform reasoning based on Bayesian probability theory,
allowing for the efficient computation of conditional probabilities and updating beliefs given
new evidence.
Q48: Describe the two main approaches of how Bayesian networks can be created.
A48: The two main approaches are:
Expert-driven: Domain experts provide the structure and parameters of the network based
on their knowledge and experience.
Data-driven: The structure and parameters of the network are learned from data using
statistical learning algorithms, such as the Expectation-Maximization (EM) algorithm.
Q49: In SVM, how does the algorithm find the separation between linear and non-linear
problems?
A49: Support Vector Machines (SVMs) can find separation between linear and non-linear
problems by using kernel functions. Kernel functions implicitly map the input data into a
higher-dimensional space, where a linear hyperplane can be used to separate the data.
Common kernel functions include the linear kernel, polynomial kernel, and radial basis
function (RBF) kernel.
Q50: What is the 'maximum margin' in SVM and why does it need to be maximized?
A50: The 'maximum margin' in SVM refers to the distance between the separating
hyperplane and the closest data points (support vectors) from each class. Maximizing the
margin helps to improve the generalization of the model, as it seeks to find the hyperplane
that best separates the data while keeping the largest possible distance from the support
vectors.
This approach tends to produce models that are less sensitive to noise and more robust to
new, unseen data.
Q51: In GAN, there are two deep networks, explain what they are and their purpose.
A51: In Generative Adversarial Networks (GANs), there are two deep networks, the
Generator and the Discriminator:
Generator: This network aims to generate realistic data samples by learning the underlying
data distribution. It takes random noise as input and outputs synthetic data samples.
Discriminator: This network is trained to distinguish between real data samples and
synthetic data samples generated by the Generator. Its goal is to correctly classify real
samples as real and synthetic samples as fake.
During training, the Generator and Discriminator play a two-player minimax game, where
the Generator tries to fool the Discriminator by producing increasingly realistic samples,
while the Discriminator tries to improve its ability to differentiate between real and fake
samples. As a result, the Generator learns to create highly realistic data samples.