Grad Lecture 3
Grad Lecture 3
Systematic Uncertainties
Incidental, estimated (bounded), systematic biases
incurred as a result of limited measurement precision
(also always present).
There is no universally applicable method for estimating/bounding*
systematic uncertainties. A typical approach often relies on independent
cross-checks, accounting for possible statistical limitations of calibration
procedures, knowledge about the experimental design and general
consistency arguments.
* Systematic errors that are “determined” become corrections!
or sys stat
Testing Models
Consider:
n
2 2
∑
χ ≡ gi σ=1
i=1
∫0
Where Γ(k) = (k − 1)! or Γ(z) = x z−1e −xdx
if k = positive integer for any real value z
1 − χ2
2
Note that: P( χ ,2) = e 2
2
(will come back to this)
( )
2
n χ
γ 2, 2 α
∫0
2
P( > χ , n) = 1 − where γ(z, α) = x z−1e −xdx
Γ( n2 )
Probability to be outside of √4
= 2 standard devia6ons
is 0.0455
Probability to be outside of
√1 = 1 standard devia6ons
is 0.3173
Pearson’s χ2 Test
So, for example, if we have a model, m, involving k free parameters
(determined by a fit to the data) that seeks to predict the values, x, of n
data points, each with normally distributed uncertainties, we can
construct the sum:
n
(xi − mi)2 for binned data
∑
S≡ (Poisson sta6s6cs)
i=1
σm
2
i
Note that, if you vary one of the model parameters from its best fit
value until χ2 increases by 1, this therefore represents the change
in the model parameter associated with 1 unit of variance in the fit
quality (i.e. the “1σ uncertainty” in the model parameter).
Example 1:
n 2
(xi − μ) Let’s find the
χ2 =
∑ σi2 value of µ that
minimises this
i=1
n n n
∂χ 2 (xi − μ) xi 1
∑ ∑
=−2 + 2μ
∑
=−2 =0
∂μ σi2 σ
i=1 i
2 σ
i=1 i
2
i=1
n xi
∑i=1 n
∑i=1 wi xi
σi2 1
μbest = n
= n
where wi ≡
σi2
∑i=1 12 ∑i=1 wi
σi
Example 2:
decay probability:
1 − tt
P(t) = e o
to
to = mean decay lifetime
normalised to total often approximate
number of events integral over bin with
observed in 10 days the average
DoF = 10 - 2 = 8
222Rn
mean lifetime
= 5.51 days
~1σ ~1σ
best fit
value
“Chi by eye”
100
80
60
40
30
0 1 2 3 4 5 6 7 8 9 10
Degrees of Freedom = 7 – 2 = 5
MoMdel
M o de
M el l
de el
ER ! ! oM d
Mo od
D M M M o
d
O R o o d e de l
M
de l
l
el
l
Model l o d
Simplest and
most predic6ve M
A theory is judged not
on what it can explain,
but on what it can
Test for Rejected with
reproducibly predict!
reproducible
Model
high confidence
predic6ons
We don’t prove models
to disprove Next simplest & correct; we reject those
Not$rejected$with$ most predic6ve models that are wrong!
high$confidence$
Rejected with
Test for
high confidence
reproducible
Don’t state that data are “consistent”
with a given model, but rather that they
are “not inconsistent.”
Student’s t
( Often misinterpreted as referring to being from or for “a student,” rather
than the fact that the name of the author happens to be “Student” )
( Except this was actually a pseudonym used by William Sealy Gosset in
his 1908 paper, who was couching himself as “a student”! )
σ rms of the full
σm =
distribution.
Recall that the rms deviation in the estimated
mean from a set of n samples is given by :
n
But what if we don’t know σ a priori and all
we have are the sampled estimators?
1 n
n∑
x̄ = xi
x̄ − μ
t≡ i=1
(s/ n )
where
n
1
s2 = (xi − x̄)2
(n − 1) ∑
i=1
Want to find the distribution of t
x̄ − μ
t≡
(s/ n )
ν = # degrees of freedom
( sx ) ( sy )
1 xi − x̄ yi − ȳ
n−1∑
rxy =
i=1
1 − r2
σr = DoF for 2 free
n−2 parameters in
linear fit
n−2
tr = r
1 − r2
Spearman Rank-Order Correlation Coefficient
A non-parametric test of correlation between two sets of data (i.e. linearity is not assumed)
Plot the cumulative fraction of events less than or equal to a particular value
of x as a function of x, along with the cumulative distribution for some model:
data x
fraction For example: Is this data Normally distributed?
≤x
point value
1 10.1 0.1
2 10.7 0.2
3 11.2 0.3
4 14.6 0.4
5 15.1 0.5
data
6 16.3 0.6
7 16.5 0.7
integral Gaussian for
8 18.8 0.8 mean=x and σ=sx
9 24.2 0.9
10 27.9 1.0
x̄ = 16.54
sx = 5.8
Equivalently:
A more clearly defined
span on the x-axis with
a visually simple model
D+ = 0.2 expectation that is
independent of the test
distribution
D- = 0.13
There are several statistics that can be used to assess the level of agreement:
∑( 2n )
D+ = maximum positive deviation from the model line n 2
2i − 1 1
W2 = yi − +
D- = maximum negative deviation from the model line
i=1
12n
( 2)
D = max(D+ , D-) 2
1
V = D+ + D- (Kuiper test) U 2 = W 2 − n ȳ −
In general, the probability distributions for these statistics need to be
determined by Monte Carlo calculations. However, for continuous variables
tested against a well-defined model distribution under the null hypothesis,
tables and approximate parameterisations exist to obtain p-values:
“High Tail”
Modified Approximate
Test Statistic Test Statistic Parameterisation
(T) (T*) for P(T* > z)
M.A. Stephens, Journal of the Royal Statistical Society. Series B (Methodological), Vol. 32, No. 1. (1970), pp. 115-122