Semi-Supervised Learning: Lukas Tencer
Semi-Supervised Learning: Lukas Tencer
Lukas Tencer
PhD student @ ETS
Motivation
Image Similarity - Domain of origin
• Self-Training
• Help-Training
• Transductive SVM (S3VM)
• Multiview Algorithms
• Graph-Based Algorithms
• Generative Models
• …….
…..
…
• Advantages:
– Very simple and fast method
– Frequently used in NLP
• Disadvantages:
– Amplifies noise in labeled data
– Requires explicit definition of 𝑃 𝑦 𝑥
– Hard to implement for discriminative classifiers (SVM)
1 2
𝐽 𝜃 = 𝑤 + 𝑐1 𝐿(𝑦𝑖 𝑓𝜃 (𝑥𝑖 )) + 𝑐2 𝐿( 𝑓𝜃 (𝑥𝑖 ) )
2
𝑥𝑖 ∈𝑇 𝑥𝑖 ∈𝑈
• Methods:
– Local Combinatorial Search
– Standard unconstrained optimization solvers (CG, BFGS…)
– Continuation Methods
– Concave-Convex procedure (CCCP)
– Branch and Bound
• Advantages:
– Can be used with any SVM
– Clear optimization criterion, mathematically well
formulated
• Disadvantages:
– Hard to optimize
– Prone to local minima – non convex
– Only small gain given modest assumptions
• Advantages:
– Simple Method applicable to any classifier
– Can correct mistakes in classification between the 2
classifiers
• Disadvantages:
– Assumes conditional independence between features
– Natural split may not exist
– Artificial split may be complicated if only few eatures
• Advantages:
– Great performance if graph fits the tasks
– Can be used in combination with any model
– Explicit mathematical formulation
• Disadvantages:
– Problem if graph does not fit the task
– Hard to construct graph in sparse spaces
• Advantages:
– Nice probabilistic framework
– Instead of EM you can go full Bayesian and include
prior with MAP
• Disadvantages:
– EM find only local minima
– Makes strong assumptions about class distributions
Books to read:
• Semi-Supervised Learning – Chapelle, Schölkopf, Zien
• Introduction to Semi-Supervised Learning - Zhu, Oldberg,
Brachman, Dietterich