6.estimators (C)
6.estimators (C)
Foundational concepts such as parameter estimation, bias and variance are useful
to formally characterize notions of generalization, underfitting and overfitting.
Point Estimation
The definition does not require that g return a value that is close to
the true θ or even that the range of g is the same as the set of allowable
values of θ.
Function Estimation
bias(θˆm ) = E(θˆm ) − θ
Var(θˆ) (5.45)
Where the random variable is the training set. Alternately, the square root
of the variance is called the standard error, denoted SE( ˆθ).
When we compute any statistic using a finite number of samples, our estimate
of the true underlying parameter is uncertain, in the sense that we could have
obtained other samples from the same distribution and their statistics would
have been different.
Bias measures the expected deviation from the true value of the function or
parameter.
Variance on the other hand, provides a measure of the deviation from the
expected estimator value that any particular sampling of the data is likely
to cause.
The most common way to negotiate this trade-off is to use cross-validation.
Alternatively, we can also compare the mean squared error (MSE) of the estimates:
Consistency
The number of data points m in our dataset increases, our point estimates
converge to the true value of the corresponding parameters.