0% found this document useful (0 votes)
178 views8 pages

Unit-2 Advanced Concepts of Modeling in AI - Question Answers

The document provides an overview of various machine learning paradigms including supervised, unsupervised, and reinforcement learning, along with their applications, challenges, and evaluation methods. It also discusses neural networks, classification and regression models, clustering, association rules, and deep learning techniques, highlighting key concepts, algorithms, and strategies for preventing overfitting. The information is structured in a question-and-answer format, making it accessible for understanding fundamental machine learning concepts.

Uploaded by

abhijith5f12698
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
178 views8 pages

Unit-2 Advanced Concepts of Modeling in AI - Question Answers

The document provides an overview of various machine learning paradigms including supervised, unsupervised, and reinforcement learning, along with their applications, challenges, and evaluation methods. It also discusses neural networks, classification and regression models, clustering, association rules, and deep learning techniques, highlighting key concepts, algorithms, and strategies for preventing overfitting. The information is structured in a question-and-answer format, making it accessible for understanding fundamental machine learning concepts.

Uploaded by

abhijith5f12698
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 8

Supervised Learning

Q1: What is supervised learning and how does it differ from other machine learning
approaches?
Supervised learning is a machine learning paradigm where an algorithm learns from labeled
training data to make predictions or decisions. The algorithm is "supervised" because it learns
from example inputs paired with their correct outputs. This differs from unsupervised learning
(which uses unlabeled data) and reinforcement learning (which learns through trial and error
with rewards/penalties).

Q2: What are some common applications of supervised learning?


Common applications include:
• Email spam filtering
• Sentiment analysis
• Image recognition
• Medical diagnosis
• Credit scoring
• Fraud detection
• Speech recognition
• Product recommendation systems

Q3: What are the key challenges in supervised learning?


Key challenges include:
• Acquiring sufficient labeled training data
• Overfitting (model performs well on training data but poorly on new data)
• Underfitting (model fails to capture the underlying pattern)
• Feature selection/engineering
• Handling class imbalance
• Managing computational complexity for large datasets
• Dealing with noisy or missing data
Unsupervised Learning

Q1: What is unsupervised learning and what types of problems is it used for?
Unsupervised learning is a type of machine learning where algorithms identify patterns in data
without labeled responses or guidance. It's used for:
• Discovering hidden structures in data
• Clustering similar data points
• Dimensionality reduction
• Anomaly detection
• Feature learning
• Density estimation
• Market basket analysis

Q2: Name and describe three common unsupervised learning algorithms.


1. K-means clustering: Partitions data into K distinct clusters based on distance to cluster
centers
2. Principal Component Analysis (PCA): Reduces dimensionality by transforming data into a
new coordinate system while preserving variance
3. DBSCAN (Density-Based Spatial Clustering of Applications with Noise): Identifies
clusters of arbitrary shape based on density, and can identify outliers

Q3: How do you evaluate the performance of unsupervised learning models?


Evaluating unsupervised learning models is challenging due to the absence of ground truth
labels, but common methods include:
• Internal validation metrics (silhouette coefficient, Davies-Bouldin index, etc.)
• Visual inspection of results
• Domain expert assessment
• Stability analysis (consistency of results with different initializations)
• Downstream task performance (using the unsupervised results in a supervised task)
• Information theoretical measures (entropy, mutual information)
Reinforcement Learning

Q1: What is reinforcement learning and what are its key components?
Reinforcement learning is a machine learning approach where an agent learns to make decisions
by performing actions in an environment to maximize cumulative rewards. Key components
include:
• Agent: The decision-maker or learner
• Environment: The world with which the agent interacts
• State: The current situation of the agent in the environment
• Action: What the agent can do in each state
• Reward: Feedback from the environment after an action
• Policy: The agent's strategy for selecting actions
• Value function: Estimation of future rewards
• Model: The agent's representation of the environment (optional)

Q2: What is the exploration-exploitation tradeoff in reinforcement learning?


The exploration-exploitation tradeoff refers to the challenge of balancing two competing
objectives:
• Exploration: Trying new actions or strategies to discover potentially better rewards
• Exploitation: Using known information to maximize immediate rewards
An effective reinforcement learning algorithm must balance these concerns. Too much
exploration may waste time on suboptimal actions, while too much exploitation may miss better
long-term strategies. Common approaches to managing this tradeoff include ε-greedy, softmax
action selection, and Upper Confidence Bound (UCB) algorithms.

Q3: What are some real-world applications of reinforcement learning?


Reinforcement learning applications include:
• Robotics and autonomous vehicles
• Game playing (AlphaGo, chess, poker)
• Resource management and scheduling
• Personalized recommendations
• Trading and financial portfolio management
• Smart grid management
• Healthcare treatment optimization
• Industrial process control
• Dialogue systems and chatbots
Neural Networks

Q1: What is a neural network and how does it work at a basic level?
A neural network is a computational model inspired by the human brain's structure. At a basic
level, it consists of interconnected nodes (neurons) organized in layers. Each neuron:
1. Receives input, either from data or other neurons
2. Applies weights to these inputs
3. Sums the weighted inputs
4. Passes this sum through an activation function
5. Outputs the result to the next layer
Through training (typically using backpropagation), the weights are adjusted to minimize
prediction error on training data.

Q2: Explain the purpose of activation functions in neural networks and name three
common ones.
Activation functions introduce non-linearity into neural networks, allowing them to learn
complex patterns and relationships in data. Without activation functions, neural networks would
be equivalent to linear regression regardless of their depth. Three common activation functions:
1. Sigmoid: Maps values to a range between 0 and 1; useful for binary classification output
layers but can suffer from vanishing gradient problems
2. ReLU (Rectified Linear Unit): Returns input if positive, otherwise returns zero;
computationally efficient and helps solve vanishing gradient problems
3. Tanh (Hyperbolic Tangent): Maps values to a range between -1 and 1; often used in
hidden layers of recurrent networks

Q3: What is backpropagation and why is it important for neural networks?


Backpropagation is an algorithm for training neural networks by calculating gradients of the loss
function with respect to the network's weights. It works by:
1. Forward pass: Computing predictions using current weights
2. Calculating error/loss between predictions and actual values
3. Backward pass: Calculating gradients by propagating the error backwards through the
network
4. Updating weights using gradients and an optimization algorithm (e.g., gradient descent)

Backpropagation is crucial because it efficiently computes these gradients, enabling neural


networks to learn from data and improve their performance over time.
Classification & Regression Models

Q1: What is the difference between classification and regression in machine learning?
Classification and regression are both supervised learning tasks but differ in their outputs:
• Classification predicts discrete class labels or categories (e.g., spam/not spam,
dog/cat/bird)
• Regression predicts continuous numerical values (e.g., house prices, temperature, stock
prices)
Classification evaluates performance using metrics like accuracy, precision, recall, and F1-score,
while regression uses metrics like mean squared error, mean absolute error, and R-squared.

Q2: Name and briefly describe three common classification algorithms.


1. Logistic Regression: Despite its name, it's a linear classification model that estimates
probabilities using a logistic function to predict discrete classes
2. Decision Tree: Creates a tree-like model of decisions based on feature values, with leaves
representing class labels
3. Support Vector Machine (SVM): Finds the optimal hyperplane that maximally separates
different classes in feature space, effective in high-dimensional spaces

Q3: What is the bias-variance tradeoff in the context of regression and classification
models?
The bias-variance tradeoff refers to the balance between two sources of error:
• Bias: Error from oversimplified assumptions in the model (underfitting)
• Variance: Error from sensitivity to fluctuations in the training data (overfitting)
High-bias models (e.g., linear regression) are simpler but may miss important patterns. High-
variance models (e.g., complex decision trees) can capture detailed patterns but may fit noise in
training data. The goal is to find a sweet spot that minimizes total error.
Techniques like regularization, ensembling, and cross-validation help manage this tradeoff.
Clustering & Association Models

Q1: What is clustering in machine learning and what are its main objectives?
Clustering is an unsupervised learning technique that groups similar data points together based
on their features or characteristics. Its main objectives include:
• Discovering natural groupings in data
• Reducing dimensionality for visualization
• Identifying outliers or anomalies
• Preprocessing for other algorithms
• Customer/market segmentation
• Document categorization
• Image segmentation

Q2: Compare and contrast K-means clustering and hierarchical clustering.


K-means clustering:
• Partitions data into K non-overlapping clusters
• Requires specifying K (number of clusters) in advance
• Iteratively assigns points to clusters and updates centroids
• Computationally efficient for large datasets (O(n))
• Sensitive to initial centroid selection
• Works best with spherical clusters of similar sizes

Hierarchical clustering:
• Creates a tree of clusters (dendrogram)
• Doesn't require specifying number of clusters beforehand
• Can be agglomerative (bottom-up) or divisive (top-down)
• Computationally more expensive (O(n²) or worse)
• Not sensitive to initialization
• Can handle irregular or nested cluster shapes
• Provides multiple partitioning levels for analysis

Q3: What are association rules and what is the Apriori algorithm used for?
Association rules are if-then statements that help discover relationships between seemingly
unrelated data in a database or other information repository. They take the form "if X, then Y"
where X and Y are itemsets.
The Apriori algorithm is a classic algorithm used for mining frequent itemsets for association
rule learning. It:
• Identifies items that frequently appear together in transactions
• Uses a "bottom-up" approach, extending frequent itemsets one item at a time
• Prunes candidates using the principle that all subsets of a frequent itemset must also be
frequent
• It is commonly used in market basket analysis to discover relationships between products
purchased together
• Can be applied to other domains like web usage mining, intrusion detection, and
bioinformatics
Deep Learning Models (ANN & CNN)

Q1: What are the key differences between traditional neural networks (ANNs) and
convolutional neural networks (CNNs)?
Key differences include:
Traditional Neural Networks (ANNs):
• Fully connected layers where each neuron connects to all neurons in adjacent layers
• Don't preserve spatial relationships in input data
• Prone to overfitting with image data due to high parameter count
• Typically used for tabular data, time series, or text

Convolutional Neural Networks (CNNs):


• Specialized for grid-like data (images, time series)
• Use convolutional layers that apply local filters across the input
• Leverage parameter sharing and spatial hierarchies
• Built-in translation invariance through pooling operations
• Significantly fewer parameters for equivalent tasks
• Automatically extract relevant features from raw data

Q2: Explain the concept of deep learning and why it has become so prominent in recent
years.
Deep learning refers to neural networks with multiple hidden layers (hence "deep") that can learn
hierarchical representations of data. Each layer learns increasingly abstract features, enabling the
model to automatically discover patterns without manual feature engineering.
Its prominence has grown due to:
1. Availability of large datasets (big data)
2. Increased computational power, especially GPUs and TPUs
3. Algorithmic improvements (better activation functions, normalization techniques, etc.)
4. Demonstrated breakthrough performance in computer vision, speech recognition, and
natural language processing
5. Transfer learning capabilities that allow leveraging pre-trained models
6. Open-source frameworks like TensorFlow and PyTorch that make implementation
accessible
7. Success in solving previously intractable problems across diverse domains

Q3: What are some techniques used to prevent overfitting in deep learning models?
Techniques to prevent overfitting in deep learning include:

1. Regularization methods:
• L1/L2 regularization (weight decay)
• Dropout (randomly deactivating neurons during training)
• Early stopping (halting training when validation performance starts degrading)
2. Data augmentation:
• Creating modified versions of training examples (rotations, flips, etc.)
• Adds diversity without requiring new data collection

3. Batch normalization:
• Normalizes layer inputs to stabilize training
• It has regularizing effects that improve generalization

4. Transfer learning:
• Starting with pre-trained models and fine-tuning for specific tasks

5. Reducing model complexity:


• Decreasing the number of layers of neurons
• Using simpler architecture when possible

6. Ensemble methods:
• Combining multiple models to reduce variance

7. Data collection:
• Gathering more training data when possible

You might also like