Approach 1: Exact RBF
Approach 1: Exact RBF
.
.
.
.
.
Exact RBF
The first layer weights u are set to the training data; U=XT. That is the
gaussians are centered at the training data instances.
The spread is chosen as , where dmax is the maximum Euclidean
distance between any two centers, and N is the number of training data
points. Note that H=N, for this case.
The output of the kth RBF output neuron is then
Define:
If {xi}iN=1 are a distinct set of points in the d-dimensional space, then the
N by N interpolation matrix Φ with elements obtained from radial basis
functions is nonsingular, and hence can be inverted!
Note that the theorem is valid regardless the value of N, the choice of the
RBF (as long as it is an RBF), or what the data points may be, as long as
they are distinct!
.
.
.
Approach1
.
. (Cont.)
Gaussian RBFs are localized functions ! unlike the sigmoids used by MLPs
Using localized functions typically makes RBF networks more suitable for
function approximation problems.
Since first layer weights are set to input patterns, second layer weights are
obtained from solving linear equations, and spread is computed from the
data, no iterative training is involved !!!
Guaranteed to correctly classify all training data points!
However, since we are using as many receptive fields as the number of
data, the solution is over determined, if the underlying physical process
does not have as many degrees of freedom 🡺 Overfitting!
The importance of σ: Too small will
also cause overfitting. Too large will
fail to characterize rapid changes in
the signal.
.
.
Too many
Receptive Fields?
.
.
.