Regression
Regression
• β0: This is the y-intercept, the value of Y when X is zero. It represents the
expected value of Y when X is zero.
• β1 : This is the slope coefficient, which represents the change in Y for a one-
unit change in X. It indicates the rate of change in Y for each unit change in X.
• ε : This is the error term, also known as the residual. It represents the
difference between the observed value of Y and the value predicted by the
regression equation. It captures all other factors that influence Y aside from X
and the intercept.
• Courtesy: https://gbhat.com/machine_learning/linear_regression.html
Residual and Regression Error
• Residuals: Residuals, also known as "errors," are the differences between the observed values of the
dependent variable and the values predicted by the regression model. In other words, the residual for
each data point is the vertical distance between the observed value and the value predicted by the
regression line. Mathematically, the residual for the ith observation is given by:
• Residuali = Observedi − Predictedi
• Regression Errors: Regression errors, on the other hand, refer to the differences between the true values
of the dependent variable and the values predicted by the regression model. These errors are inherent to
the relationship between the independent and dependent variables and cannot be reduced by improving
the model.
• Residuals are a key diagnostic tool in regression analysis. They are used to assess the goodness-of-fit of
the model, detect outliers, check for violations of assumptions, and identify patterns in the data that may
not be captured by the model.
• Regression errors represent the discrepancy between the observed data and the true underlying
relationship between the variables. While residuals are specific to the sample data and the fitted model,
regression errors are conceptual and relate to the population from which the sample is drawn
Assumptions
• Linearity: The relationship between the independent variables and the
dependent variable is linear. This means that changes in the independent
variables result in a constant change in the dependent variable.
Citation: https://www.superdatascience.com/blogs/assumptions-of-linear-regression