Fba 1
Fba 1
y = β0 + β1 x + ε Note:
• Parameters: The characteristics of the • The regression line in Panel A shows that
population, β0 and β1 the mean value of y is related positively to
• Random variable: Error term, ε. x, with larger values of E(y|x) associated
• The error term accounts for the with larger values of x.
variability in y that cannot be explained • In Panel B, the mean value of y is related
by the linear relationship between x and negatively to x, with smaller values of E(y|
y. x) associated with larger values of x.
• The parameter values are usually not
• In Panel C, the mean value of y is not
known and must be estimated using
sample data. related to x; that is, E(y|x) is the same for
every value of x.
II. Least Squares Method ➢ Interpretation of b1: If the length of a
A. Least Squares Estimates of driving assignment were 1 unit (1 mile)
thRegression Parameters longer, the mean travel time for that
B. Using Excel’s Chart Tools to driving assignment would be 0.0678 units
Compute the Estimated Regression (0.0678 hours, or approximately 4
Equation minutes) longer.
➢ Interpretation ofb0: If the driving distance
III. Least Squares Method for a driving assignment was 0 units (0
➢ Least squares method miles), the mean travel time would be
- A procedure for using sample data 1.2739 units (1.2739 hours, or
to find the estimated regression approximately 76 minutes)
equation.
➢ Determine the values of b0 and b1. ➢ Experimental region
➢ Interpretation of b0 and b1: - The range of values of the
• The slope b1 is the estimated independent variables in the data used
change in the mean of the to estimate the model.
dependent variable y that is - The regression model is valid only
associated with a one unit increase over this region.
in the independent variable x. ➢ Extrapolation
• The y-intercept b0 is the estimated - Prediction of the value of the
value of the dependent variable y dependent variable outside the
when the independent variable x is experimental region.
equal to 0. - It is risky.
➢ Least squares equation:
c. Stepwise procedure:
➢ The analyst establishes both a criterion for
allowing independent variables to enter the
model and a criterion for allowing
independent variables to remain in the
model.
➢ In the first step, the independent variable
that best satisfies the criterion for entering
the model is added.
➢ First, the remaining independent variables
not in the current model are evaluated, and
a. Backward elimination the one that best satisfies the criterion for
b. Forward selection entering is added to the model.
c. Stepwise selection ➢ Then, the independent variables in the
d. Best subsets current model are evaluated, and the one that
violates the criterion for remaining in the
model to the greatest degree is removed.
➢ The procedure stops when no independent
a. Forward selection procedure:
variables not currently in the model meet the
criterion for being added to the regression ➢ It is recommended to divide the original
model, and no independent variables sample data into training and validation sets.
currently in the model violate the criterion ➢ Training set
for remaining in the regression model. - The data set used to build the candidate
models that appear to make practical
sense.
➢ Validation set
d. Best-subsets procedure: - The set of data used to compare model
➢ Simple linear regressions for each of the performances and ultimately pick a
independent variables under consideration model for predicting values of the
are generated, and then the multiple dependent variable•
regressions with all combinations of two ➢ Holdout method
independent variables under consideration - The sample data are randomly divided
are generated, and so on. into mutually exclusive and collectively
➢ Once a regression has been generated for exhaustive training and validation sets.
every possible subset of the independent ➢ k-fold cross-validation
variables under consideration, an output - The sample data are randomly divided
that provides some criteria for selecting into k equal-sized, mutually exclusive,
regression models is produced for all and collectively exhaustive subsets
models generated. called fold, and k iterations are executed
➢ Leave-one-out cross-validation
- For a sample of n observations, an
B. Overfitting iteration consists of estimating the
➢ Results from creating an overly complex model on n – 1 observations and
model to explain idiosyncrasies in the evaluating the model on the single
sample data. observation that was omitted from the
➢ Results from the use of complex functional training data.
forms or independent variables that do not
have meaningful relationships with the
dependent variable.
➢ If a model is overfit to the sample data, it
will perform better on the sample data used
to fit the model than it will on other data
from the population.
➢ Thus, an overfit model can be misleading
about its predictive capability and its
interpretation.