FSMLecture6 - Statistics
FSMLecture6 - Statistics
1
Today, Lecture 6: Confidence Sequences
2
Confidence Intervals
Neyman and Pearson, 1930s
∀𝜃 ∈ Θ: 𝑃! (𝜃 ∈ CS",$&' ) ≥ 1 − 𝛼
• 𝑃! p! ≤ 𝛼 ≤ 𝛼
• Set CS",$&' ≔ {𝜃: p! > 𝛼} The 𝜃 you have not been able to reject
• …so for CS 𝐜",$&' : ∀𝜃 ∈ Θ: 𝑃! (𝜃 ∈ CS",$&'
𝐜
) ≤ 𝛼 so
∀𝜃 ∈ Θ: 𝑃! (𝜃 ∈ CS",$&' ) ≥ 1 − 𝛼 (*)
• 𝑃! p! ≤ 𝛼 ≤ 𝛼
∀𝜃 ∈ Θ: 𝑃! (𝜃 ∈ CS",$&' ) ≥ 1 − 𝛼
Even though 𝝁(𝟏) , 𝝁(𝟐) , … are all > −𝟏𝟎, every now and then we might observe an
experiment 𝑗 with 𝜇-(*) ≪ −10.
(say (ℓ 𝑌 * , 𝑟 𝑌 * = −12.3, −10.1 )
We might now tempted to conclude that for this experiment/medication, we are 95%
certain that
𝜇 * < −10
But this would be wrong: the ‘world’ we set up is such that 𝜇(*) is never < −10.
Similar to what we saw for hypothesis testing in the previous lecture, we cannot use CI’s to
give conditional conclusions: they only say something about long-run averages
Bayesian Credible Intervals
14
Example: Normal Location Family
5'
• But this just a normal distribution: 𝑊 ∣ 𝑋" = 𝑁 𝜇,̂ "
• Hence the Bayesian 1 − 𝛼 credible interval for a noninformative (high-
variance) prior is essentially indistiguingishable from a standard 1 −
𝛼 confidence interval!
16
Similarity in Form of Bayes 𝐂𝐫𝐈 and 𝐂𝐈
18
Dissimilarity in Meaning of Bayes 𝐂𝐫𝐈 and 𝐂𝐈
• Each sample 𝑌(-) = (𝑋 - ,$, … , 𝑋 - ,"! ) consists of data points with 𝑋 - ,. the
difference between two measurements of a patient’s blood pressure, one
before and one after taking medicatin of type 𝑗
• Research group 1 tries med. of type 1 (say, paracetamol) on sample
𝑌($) , research group 2 tries medication 2 (say, green tea) on 𝑌(,) , 𝑒𝑡𝑐.
• We assume 𝑋 - ,. ∼ 𝑁(𝜇(-) , 𝜎 ,) for some known 𝜎 ,
• Suppose that a medication for lowering blood pressure is considered
effective if 𝝁 ≤ −𝟏𝟎
• When analyzing standard CIs we considered the scenario that, while there
are many medications around, none of these achieves the goal 𝜇 ≤ −10.
So 𝝁(𝟏) , 𝝁(𝟐) , … are all > −𝟏𝟎.
19
Previous Thought Experiment, Continued
22
Standard CIs: invalid under optional stopping
• Just like the Neyman-Pearson tests on which they are often based,
standard CIs cannot handle (become invalid under) optional stopping
• Bayesian Credible Intervals can handle optional stopping if (there it is
agian) you really believe your prior, but if you choose it pragmatically
(which you usually do), they cannot
• This has to follow from the fact that standard CIs cannot handle
optional stopping, for the Bayesian CrI and the standard CI are
essentially the same with normal distributions
23
Z-test ⇒ Z-Confidence Interval
… always wider than Bayes credible posterior interval based on same prior
AV CI’s vs. Bayesian Credible Sets
• The AV confidence intervals are invariable wider than the standard ones.
But in fact we can considerably improve them so that they sometimes
(not always) are even tighter at some 𝑛 than the standard ones.
• We do this by taking the running intersection:
38
∗
The Running Intersection CI!,#$%
∗
• …CIF,GHI ≔ ⋂KLG..F CIK,GHI The 𝜃 you have not yet been able to reject at 𝑛
Running Intersection, Illustration
AV CIs for normal location family, revisited
I1 J .
CI",$&M = 𝜃: >𝛼
I𝑾𝜽 J .
• Still they use the prior in a different way, and they lead to very different
conclusions. Understanding this difference is important!
• Unfortunately many Bayesian statisticians don’t understand it…
44
Yellow: Bayes 95% credible interval based on noninformative
prior = standard confidence interval = 𝑋D ± 1.96/ 𝑛
Blue: 95% AV interval based on same prior:
Subjective and Objective, at same time:
luckiness
• E-Posteriors and the AV CIs they induce rely on a prior, just like
Bayesian posteriors…
…but they remain valid irrespective of prior you use
…suppose for example you have a pretty mistaken prior belief that 𝜃 =
0, with variance 0.5 …
Subjective and Objective, at same time:
luckiness
• The AV CIs induced by e-variables rely on a prior, just like Bayesian
credible intervals…
…but they remain valid irrespective of prior you use
49
Nonparametric Anytime-Valid CI
…recall from last week:
Testing the Mean of a Bounded Random Variable
Waudby-Smith and Ramdas, JRSS B, 2024, Orabona & Jun, IEEE Trans. Inf. Th., 2023
Last week we tested whether mean is 𝜇, now we use exact same technique
to make AV CI for 𝜇
Nonparametric Anytime-Valid CI
Variation:
$
We simply set: CI",$&' = 𝜇: 𝑆 3" < '
(")
For fixed 𝜆: 𝑆V, 3 ≔ ∏");$ 𝑠V, 3 (𝑋) ) is an e-variable
(") (")
Now put “prior” 𝑤[3] on Λ3 : 𝑆[3] ∶= ∫[ 𝑆V, 3 𝑤 3 𝜆 d𝜆
$
(")
Since 𝑆[3] is a mixture of e-variables, it is itself an e-variable
Variation
(")
For fixed 𝜆: 𝑆V, 3 ≔ ∏");$ 𝑠V, 3 (𝑋) ) is an e-variable
(") (")
Now put “prior” 𝑤[3] on Λ3 : 𝑆[3] ∶= ∫[ 𝑆V, 3 𝑤 3 𝜆 d𝜆
$
60
Where we stand and where we will go
61