Lecture 2
Lecture 2
(parameter estimation)
2
Linear discrete inverse problems
Under-determined
Over-determined
Even-determined
3
Over-determined: Linear discrete inverse problem
But the data contain errors. Let’s assume these are independent and
normally distributed, then we weight each residual inversely by the
standard deviation of the corresponding (known) error distribution.
4
Over-determined: Linear discrete inverse problem
Compare with
We seek the model vector m which minimizes maximum likelihood
3 5
Over-determined: Linear discrete inverse problem
6
Linear discrete inverse problem: Least squares
Under-determined, N=1:
Even-determined, N=2:
r is the−
What= d G = m
prediction 0
error ?
We need to:
Assess the quality of the data fit.
Goodness of fit: Does the model fit the data to within
the statistical uncertainty of the noise ?
9
Goodness of fit
10
Why do we always assume errors are Gaussian ?
Probability density functions of 5 random variables
X 100000
Xi deviates
i The Central
Limit Theorem
11
Mathematical Background:
Probability distributions
Multivariate case
12
Mathematical Background:
Probability distributions
Expectation operator
Variance
Covariance
13
Mathematical Background:
Probability distributions
Multi-dimensional Gaussian
Covariance matrix
⎡ ⎤
σ1,1 σ1,2 σ1,3 σ1,4
Correlation matrix ⎢ ⎥
⎢ σ2,1 σ2,2 σ2,3 σ2,4 ⎥
⎢ ⎥
⎣ σ3,1 σ3,2 σ3,3 σ3,4 ⎦
σ4,1 σ4,2 σ4,3 σ4,4
14
Background: Chi-square distribution
If x follows a Normal distribution
p = 0.95 p = 0.05
The ratio of the variances of the two distributions follows a chi-square distribution.
Chi-square tables tell us the likelihood that the two sets of observables
come from the same distribution.
18
Goodness of fit
Exercise:
If I fit 7 data points with a straight line and get
what would you conclude ?
19
Goodness of fit
Fraud ! 20
Solution error
Once we have our least squares solution mLS and we know
that the data fit is acceptable, how do we find the likely errors
in the model parameters arising from errors in the data ?
mLS = G−g d
d0 → d + ²
m0LS → mLS + ²m
m0LS = G−g d0
mLS + ²m = G−g (d + ²)
The effect of adding noise to the data is to add noise to the solution
²m = G−g ²
The model noise is a linear combination of the data noise !
21
Solution error: Model Covariance
How to turn this into a probability distribution for the model errors ?
22
Solution error: Model Covariance
23
Solution error: Model Covariance
24
Confidence intervals by projection (1-D)
The model covariance is a symmetric M x M matrix
26
Confidence intervals by projection
The M-dimensional confidence ellipsoid can be projected onto any subset (or
combination) of Δ parameters to obtain the corresponding confidence ellipsoid.
Can you see that this procedure gives the same formula for the 1-D
case obtained previously ? 27
Recap: Goodness of fit and model covariance
Once a best fit solution has been obtained we test goodness of fit with a
chi-square test (assuming Gaussian statistics)
If the model passes the goodness of fit test we may proceed to evaluating
model covariance (if not then your data errors are probably too small)
Minimize
29
Monte Carlo error propagation
30
Monte Carlo error propagation
31
What if we do not know the errors on the data ?
CD = σ 2I
Independent data errors
So we can still estimate model errors using the calculated data errors
but we can no long claim anything about goodness of fit.
32
Example: Over-determined, Linear discrete inverse
problem
MATLAB exercise
Generate data with Gaussian noise for a linear regression
problem and invert for the best fitting gradient and intercept
2. Calculate data yi
4. Calculate G matrix
6. Plot the data, plot the data errors and plot the least squares solution
33
Model Resolution matrix
mest = G−g d
But we know
d = Gmtrue
and hence
The matrix R shows how the elements of mest are built from
linear combination of the true model, mtrue. Hence matrix R
measures the amount of blurring produced by the inverse
operator.
For the least squares solution we have
−1 −1
G−g = (GT CD G)−1GT CD ⇒R=I
5 34
Example: Model resolution in a tomographic
experiment
m = Rmtrue
If the calculated model resolution matrix looked like this
⎡ ⎤
0.75 −0.25 0.25 0.25
⎢ −0.25 0.75 0.25 0.25 ⎥
⎢
R=⎢
⎥
⎥ What units do the
⎣ 0.25 0.25 0.75 −0.25 ⎦ elements of R have ?
0.25 0.25 −0.25 0.75
Spike test
True model Recovered model
35
Data Resolution matrix
dpre = Gmest
But we know
36
Recap: Linear discrete inverse problems