0% found this document useful (0 votes)
5 views2 pages

Predictive Analytics Descriptive Answer Key

Uploaded by

Hruthik
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
5 views2 pages

Predictive Analytics Descriptive Answer Key

Uploaded by

Hruthik
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 2

Predictive Analytics - Descriptive Answer Key

1. Explain Linear regression and its importance in prediction.


Linear regression is a supervised learning algorithm that models the relationship between a
dependent variable (target) and one or more independent variables (predictors) using a
linear function. The equation is represented as:

Y = β0 + β1X1 + β2X2 + ... + βnXn + ε

Where:
- Y is the dependent variable.
- β0 is the intercept.
- β1, β2,..., βn are the coefficients of the predictors.
- X1, X2,..., Xn are the independent variables.
- ε is the error term.

Importance:
1. It is simple to implement and understand.
2. It helps identify trends and relationships in data.
3. It is widely used for predictive analysis in business, economics, and more.

2. What is Multiple Regression? Give an example.


Multiple regression is an extension of linear regression that involves more than one
independent variable. It is used to predict the value of a dependent variable based on
multiple predictors.

The general form of the multiple regression model is:


Y = β0 + β1X1 + β2X2 + ... + βnXn + ε

Example:
Suppose we want to predict house prices (Y) based on area (X1), number of bedrooms (X2),
and location rating (X3):
Price = 5000 + 300(Area) + 1500(Bedrooms) + 2000(Location)

Importance:
1. It improves accuracy by considering multiple predictors.
2. It helps analyze the impact of each predictor on the outcome.

3. Explain bias-variance tradeoff with a diagram.


The bias-variance tradeoff is a fundamental concept in machine learning that describes the
tradeoff between two sources of error:
- **Bias**: Error due to overly simplistic assumptions in the model. High bias leads to
underfitting.
- **Variance**: Error due to model sensitivity to small fluctuations in training data. High
variance leads to overfitting.

A good model minimizes both bias and variance to achieve optimal performance.

**Diagram:**

4. Explain ensemble methods and give an example of Boosting.


Ensemble methods combine multiple models to improve overall prediction accuracy.
Boosting is a specific ensemble technique where models are built sequentially, with each
new model focusing on correcting errors made by previous ones.

Example - AdaBoost (Adaptive Boosting):


1. Initialize equal weights for all observations.
2. Train a weak learner (e.g., decision tree) and evaluate errors.
3. Adjust weights to give more importance to misclassified data.
4. Repeat the process to combine learners for a strong model.

Boosting reduces bias and variance, making it powerful for classification tasks.

5. What is Support Vector Machines (SVM)? Explain with an example.


Support Vector Machines (SVM) is a supervised learning algorithm used for classification
and regression. It finds the optimal hyperplane that separates data points into distinct
classes with the maximum margin.

Key Concepts:
1. **Hyperplane**: A decision boundary separating classes.
2. **Support Vectors**: Data points closest to the hyperplane.
3. **Kernel Trick**: Allows SVM to classify nonlinear data by mapping it to higher-
dimensional space.

Example:
Classifying emails as spam or non-spam:
- Input: Features like word frequency, subject line keywords.
- SVM identifies the optimal boundary (hyperplane) separating spam and non-spam emails.

You might also like