0% found this document useful (0 votes)
23 views12 pages

Lecture 4.1 AML

Uploaded by

Vivek Sreekar
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
23 views12 pages

Lecture 4.1 AML

Uploaded by

Vivek Sreekar
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 12

Advanced Machine

Learning with
TensorFlow
22TCSE532
Lecture_4.1
Model Evaluation and Hyper-parameter Tuning

A comprehensive overview of key concepts and


methodologies
Streamlining Workflows with Pipelines
• Pipelines are used to automate and streamline the process of applying
data preprocessing and model training in sequence.
• - Ensures consistency by processing data through the same steps each
time.
• - Combines multiple processing steps into a single, repeatable process.
• - Makes it easier to tune hyper-parameters since all steps are
encapsulated in a single entity.
K-Fold Cross-Validation
• K-fold cross-validation is a resampling technique used to evaluate
model performance.
• - The dataset is divided into K subsets (folds).
• - The model is trained on K-1 folds and tested on the remaining fold.
• - This process is repeated K times, with each fold serving as the test set
once.
• - Helps in reducing the variance associated with a single train-test split.
Advantages:
- Reduces the bias associated with a single train-test split.
- Provides a more robust estimate of model performance.

Choosing 'K':
- Common values for 'K' are 5 or 10. A higher 'K' increases
computational cost but provides a more stable performance
estimate.
Model Performance Measures
• Evaluating a model's performance requires various metrics:
• - Accuracy, Precision, Recall, F1-score for classification tasks.
• - Mean Absolute Error (MAE), Mean Squared Error (MSE) for
regression tasks.
• - ROC-AUC for probabilistic classifiers.
• - Helps in understanding the strengths and weaknesses of the
model.
Debugging with Learning and Validation Curves

• Learning curves help in understanding the performance of a


model as it learns.
• - Shows how the model's error rate changes with varying
training sizes.
• - Validation curves show model performance with changes in
hyper-parameters.
• - Useful in diagnosing underfitting and overfitting.
Fine-Tuning Models with Grid Search

• Grid search is used for hyper-parameter tuning to find the best


combination of parameters.
• - Exhaustively searches over a specified parameter grid.
• - Evaluates model performance for each parameter
combination using cross-validation.
• - Returns the parameter set that yields the best performance.
Performance Evaluation Metrics
• Metrics vary depending on the type of task:
• - Ranking Metrics: Mean Average Precision (MAP),
Normalized Discounted Cumulative Gain (NDCG).
• - Classification Metrics: Accuracy, Precision, Recall, ROC-
AUC.
• - Regression Metrics: MSE, RMSE, R-squared.
Bootstrapping and Jackknife Methods

• Resampling methods for statistical inference:


• - Bootstrapping: Generates multiple datasets by sampling with
replacement.
• - Jackknife: Involves systematically leaving out one observation
at a time from the sample set.
• - Used to estimate statistics like mean, variance, and
confidence intervals.
Hold-Out Validation
• A simple validation method for model evaluation:
• - The dataset is split into two parts: training set and test set.
• - Model is trained on the training set and evaluated on the
test set.
• - Does not use the entire dataset for training, hence prone
to variance.
Model Validation vs. Testing
• Validation and testing serve different purposes:
• - Validation is used to tune hyper-parameters and select the
best model.
• - Testing is performed on a separate dataset to assess the final
model's performance.
• - Testing provides an unbiased evaluation, while validation
helps optimize the model.

You might also like