Introduction
Introduction
b = A
x ∈ Rn
m n linear equations about x
b = Ax, b ∈ Rm
Find the locations on the genome that influence a trait: e.g. cholesterol level
Model
y = Ax + z
(a) (b)
uz (t, z = 0) = f (t)
A(z) = 1 + α(z)
α small pertubation
Linear approximation
Z t
g(t) = f (t) + h(t) = f (t − τ )α0 (τ /2) dτ
0
f ? r = g.
UV 1
ser
0.5
CCD Camera 0
−kmax 0 kmax
Iout = H ? Iin
Magnetic resonance imaging (sketch)
(Powerful) magnetic field aligns nuclear magnetization
of (usually) hydrogen atoms in water in the body
RF fields systematically alter the alignment of this
magnetization → hydrogen nuclei produce a rotating
magnetic field detectable by the scanner
Make excitation strength space dependent
Goal is to recover proton density
Model
Discrete-time Fourier transform of f [t1 , t2 ], 0 ≤ t1 , t2 ≤ n − 1
fˆ(ω1 , ω2 ) =
X
f [t1 , t2 ]e−i2π(ω1 t1 +ω2 t2 )
(t1 ,t2 )
MR scan
y(ω1 , ω2 ) = fˆ(ω1 , ω2 ) + noise
for some frequencies (ω1 , ω2 )
Undersampling in magnetic resonance angiography
space frequency
50
100
20
15
150
10
200
5
0
250
50 100 150 200 250
−5
−10
−15
−20
−25
0 50 100 150 200 250 300
N -pixel image f
Underdetermined system
z is a noise term
Interested in m N samples
The Netflix problem
Netflix database
About a million users
About 25,000 movies
People rate movies
Sparsely sampled entries
The Netflix problem
Movies
Netflix database
× ×
About a million users × ×
About 25,000 movies
× ×
Users
People rate movies
× ×
Sparsely sampled entries ×
× ×
Challenge
Complete the “Netflix matrix”
? ? ? × ?
×
? ? × × ? ?
Matrix M ∈ Rn1 ×n2
? ? × ? ?
×
Observe subset of entries ? ? × ? ? ×
? ? ? ? ?
Wish to infer the missing entries ×
? ? × × ? ?
Underdetermined systems of linear equations
b = A
= *
If unknown is assumed to be
sparse (genomics example)
low-rank (Netflix)
then one can often find solutions to these problems by convex optimization
Sparsity in genomics
y = Ax (+z)
n ∼ 500, 000
m ∼ 1, 000
But number of nonzero components of x in the tens
x is said to be s-sparse if
kxk`0 = #{i : xi 6= 0} ≤ s
`0 minimization `1 minimization
min kxk`0 min kxk`1
s. t. Ax = b s. t. Ax = b
Combinatorially hard Linear program
Why an ! objective?
`1 norm is closest convex 1approximation to `0 quasi-norm
P
minimize i |xi |
subject to Ax = b
is equivalent to P
minimize i ti
subject to Ax = b
−ti ≤ xi ≤ ti
with variables x, t ∈ Rn
y = Ax + z
(a) (b)
min krk`1 s. t. f ?r =g
2.5
1.5
0.5
!0.5
!1
!1.5
0 50 100 150 200 250 300
original
`1 reconstruction from undersampled freq. data
2.5 2.5
2 2
1.5 1.5
1 1
0.5 0.5
0 0
!0.5 !0.5
!1 !1
!1.5 !1.5
0 50 100 150 200 250 300 0 50 100 150 200 250 300
0.5
2 2
0.4
1.5 1.5
0.3
1 1
0.2
0
0 0
!0.1
!0.5 !0.5
!0.2
!1 !1
!0.3
50
100
150
200
250
50 100 150 200 250
original
Example in MRI
image is not sparse
gradient is sparse
So minimize `1 norm of gradient subject to constraints
X
min kxkT V := |∇x(t1 , t2 )| s. t. Ax = b
t1 ,t2
Original Phantom (Logan−Shepp) Naive Reconstruction
50 50
100 100
150 150
200 200
250 250
50 100 150 200 250 50 100 150 200 250
50 50 50
? ? ? × ?
×
? ? × × ? ?
large number of entries
× ? ? × ? ?
only sees a few ? ? × ? ? ×
? ? ? ? ?
But unknwon matrix has low rank ×
? ? × × ? ?
i σi (X)
P
min rank(X) min
s. t. Xij = Mij (Mij observed) s. t. Xij = Mij (Mij observed)
JPEG: 150KB
A long established tradition
N >> M
N M
sample compress transmit/store
!
! sparse !
wavelet
transform
receive
M decompress
N
Sparsity in signal processing
1 megapixel image
Sparsity in signal processing
wavelet coefficients
16000
14000
12000
10000
8000
6000
4000
2000
−2000
−4000
0 2 4 6 8 10 12
5
x 10
⇓
3000
2000
1000
−1000
−3000
2 2.5 3 3.5 4 4.5 5 5.5 6 6.5 7
4
x 10
zoom in
1 megapixel image
Sparsity and wavelet “compression”
Take a mega-pixel image
1 Compute 1,000,000 wavelet coefficients
2 Set to zero all but the 25,000 largest coefficients
3 Invert the wavelet transform
yk = hak , xi, k = 1, . . . , m
e.g. ak ∼ N (0, I)
Reconstruct by `1 minimization
yk = hak , xi, k = 1, . . . , m
e.g. ak ∼ N (0, I)
Reconstruct by `1 minimization
• Signal is local,
signal is localmeasurements are global
measurements are global
• Each measurement picks up a little information about each component
0.5
!0.5
0 5 10 15 20 25 30 35 40 45 50
Classical viewpoint
Measure everything
(all the pixels, all the coefficients)
Store s-largest coefficients
distortion : kx − xs k`2
Nonadaptive sensing
Motivation:of compressible
Sampling signals
a Sparse Vector
concentrated vector incoherent measurements
1
0.5
!0.5
0 5 10 15 20 25 30 35 40 45 50
yk = hak , xi + σzk , k = 1, . . . , m
a ∼ N (0, I) (say)
zk iid N (0, 1) (say)
But data are always noisy...
Random sensing with a ∼ N (0, I)
yk = hak , xi + σzk , k = 1, . . . , m
a ∼ N (0, I) (say)
zk iid N (0, 1) (say)
sσ 2
kx̂ − xk2`2 . inf kx − xs k2`2 + log n
1≤s.s̄ m
= near-optimal bias-variance trade off
Course objectives
expose students to a novel active field of research
give tools to make theoretical and/or practical contributions