Learning From Uniform Convergence
Learning From Uniform Convergence
Definition (ε-representative)
A training set S is called ε-representative (w.r.t. domain Z, hypothesis class ℋ,
loss function ℓ, and distribution D) if
∀ℎ ∈ ℋ: |𝐿𝑆 ℎ − 𝐿𝐷 (ℎ)| ≤ 𝜖
Theorem:
𝜖
Assume that training set S is -representative (w.r.t. domain Z, hypothesis
2
class ℋ, loss function ℓ, distribution D). Then, any output of 𝐸𝑅𝑀ℋ 𝑆 (i.e.,
any hs ∈ 𝑎𝑟𝑔min 𝐿𝑆 ℎ ) satisfies:
ℎ∈ℋ
𝐿𝐷 (ℎ𝑆 ) ≤ min 𝐿𝐷 ℎ + 𝜖
ℎ∈ℋ
Combine together:
𝜖 𝜖 𝜖 𝜖
𝐿𝐷 ℎ𝑠 ≤ 𝐿𝑆 ℎ𝑠 + ≤ 𝐿𝑆 ℎ + ≤ 𝐿𝐷 ℎ + +
2 2 2 2
⇓
𝐿𝐷 ℎ𝑠 ≤ 𝐿𝐷 ℎ + 𝜖
Uniform Convergence
❑ Next step: demonstrate that for any fixed hypotheses h the difference
|𝐿𝑠 ℎ − 𝐿𝐷 ℎ | is likely to be small
❑ Notice that 𝐿𝐷 ℎ is the expectation and 𝐿𝑠 ℎ the average value: the
random variable should not deviate too much from its expectation
❑ INTUITIVE IDEA from law of large numbers: if m is large the average
converges to the expectation
From theorem* (uniform convergence implies PAC learnable): ℋ is PAC learnable with
𝑈𝐶 𝜖 2ℋ
sample complexity 𝑚ℋ 𝜖, 𝛿 ≤ 𝑚ℋ (2 , 𝛿) ≤ 2log( )/𝜖 2
𝛿
(*) recall: