Introduction To Compressive Sampling
Introduction To Compressive Sampling
Emanuele Grossi
Outline
1 Introduction
Traditional data acquisition
Compressive data acquisition
2 Compressed sensing
Measurement protocol
Recovery procedure
Recovery conditions
Sensing matrices
3 Discussion
Connections with other fields
Numerical methods
Applications
Conclusions
Introduction Compressed sensing Discussion
Outline
1 Introduction
Traditional data acquisition
Compressive data acquisition
2 Compressed sensing
Measurement protocol
Recovery procedure
Recovery conditions
Sensing matrices
3 Discussion
Connections with other fields
Numerical methods
Applications
Conclusions
Introduction Compressed sensing Discussion
data transmit/store
sample compress
size n size k n
Sparsity/Compressibility
Many signals can be well-approximated by a sparse expansion in
terms of a suitable basis, i.e., by few of non-zero coefficients
Fourier transform
Sparsity/Compressibility
Wavelet transform
Sparsity/Compressibility
Wavelet transform
1100
1000
900
800
700
600
500
400
300
200
100 threshold
0
0 1 2 3 4 5 6
wavelet coefficients (sorted) x 10
6
Sparsity/Compressibility
Wavelet transform
Outline
1 Introduction
Traditional data acquisition
Compressive data acquisition
2 Compressed sensing
Measurement protocol
Recovery procedure
Recovery conditions
Sensing matrices
3 Discussion
Connections with other fields
Numerical methods
Applications
Conclusions
Introduction Compressed sensing Discussion
Why spend so much effort to acquire all the data when most of it
will be discarded?
†
E. J. Candès and T. Tao, “Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency
information,” IEEE Trans. Inorm. Theory, 2006; D. L. Donoho, “Compressed sensing,” IEEE Trans. Inorm. Theory, 2006.
Introduction Compressed sensing Discussion
data transmit/store
compressed sensing
size m = O(k ln n)
Outline
1 Introduction
Traditional data acquisition
Compressive data acquisition
2 Compressed sensing
Measurement protocol
Recovery procedure
Recovery conditions
Sensing matrices
3 Discussion
Connections with other fields
Numerical methods
Applications
Conclusions
Introduction Compressed sensing Discussion
CS recipe
signal to be sensed
basis vectors (not necessarily orthogonal)
into an n × n matrix Ψ = (ψ1 · · · ψn ),
e.g., spikes, sinusoids, wavelets, etc.
signal coefficients into an n-dimensional vector
Introduction Compressed sensing Discussion
Pn
p 1/p ,
kxkp = i=1 |xi | `p -norm, e.g.,
v
u n
uX
kxk2 = t |xi |2 , Euclidian
i=1
n
X
kxk1 = |xi |, gives the Manhattan distance
i=1
kxk0 = card i ∈ {1, . . . , n} : xi 6= 0 , number of
non-zero entries (little abuse: it is not a norm)
Introduction Compressed sensing Discussion
Pn
p 1/p ,
kxkp = i=1 |xi | `p -norm
Definitions:
x sparse if kxk0 n
x k-sparse, if kxk0 ≤ k ≤ n
best k-term approximation of x
Traditional compression
Measurement model
Take m linear measurements
s measurement y = Ax xk
compress sensor side
size n
process size m ≥ n size k n
xk ŝ
decompress receiver side
size k size n
CS intuition
If x is k-sparse, then it should have k degrees of freedom, not n
⇒ only k measurements or so are needed
Analogy with the 12 coin problem:
“Of 12 coins, one is counterfeit and weighs either
more or less than the others. Find the counterfeit coin
and say if its lighter or heavier with 3 weighings on a
balance scale.”
General problem: (3p − 3)/2 coins coins and p weighings
2nd weighing 1 3
Introduction Compressed sensing Discussion
CS intuition
2nd weighing 1 2 3 11 7 8 9 10
3rd weighing 1 4 7 10 2 5 8 12
Key points
counterfeit data is sparse
weigh the coins in suitably chosen batches
each measurement picks up little information
about many coins
Introduction Compressed sensing Discussion
CS protocol
s measurement y = Ax
sensor side
size n
process size m = O(k ln n)
y ŝ
reconstruct receiver side
size m size n
CS protocol
s measurement y = Ax
sensor side
size n
process size m = O(k ln n)
y ŝ
reconstruct receiver side
size m size n
Outline
1 Introduction
Traditional data acquisition
Compressive data acquisition
2 Compressed sensing
Measurement protocol
Recovery procedure
Recovery conditions
Sensing matrices
3 Discussion
Connections with other fields
Numerical methods
Applications
Conclusions
Introduction Compressed sensing Discussion
The equation Ax = y
Assume w.l.o.g. rank(A) = min{m, n}
m = n, determined system ⇒ solution = A−1 y
m > n, over-determined system; two cases
y ∈ I (A) ⇒ solution = AT A)−1 AT y = A† y
y∈/ I (A) (e.g., noisy measurements)
⇒ no solution; the Least-squares (LS) one is
Least-squares
min kxk2 , s.t. Ax = y
x
Ax = y
Problem
(P0 ) min kxk0 , s.t. Ax = y
x
Ax = y
Uniqueness
Proposition
If any 2k ≤ m columns of A are linearly independent, then any
k-sparse signal x can be recovered uniquely from Ax.
Proof: If not there would exists x1 , x2 such that Ax1 = Ax2 . This
implies A(x1 − x2 ) = 0, with x1 − x2 2k-sparse, and it is not
possible
Observation
If (A)i, j are Gaussian (or from other continuous distribution) i.i.d.
then the condition is satisfied w.p.1.
Introduction Compressed sensing Discussion
Computational complexity
p=2 p=1
|x|p
p=1/3
p=0
0 x
(
convex, if p ≥ 1
kxkpp is That’s why (P0 ) is hard!
non-convex, otherwise
Introduction Compressed sensing Discussion
Computational complexity
p=2 p=1
|x|p
p=1/3
p=0
0 x
Possible ways:
look for iterative algorithms: greedy algorithms
convex relaxation: use the convex norm with the lowest p
Introduction Compressed sensing Discussion
`1 regularization
Problem
(P1 ) min kxk1 , s.t. Ax = y
x
linear program (LP) in the real case, second order cone program
(SOCP) in the complex case
fast (polynomial time), accurate algorithms are available
Introduction Compressed sensing Discussion
`1 regularization
Problem
(P1 ) min kxk1 , s.t. Ax = y
x
Ax = y
`0 , `1 , and `2 together
{z : Az = y} {z : Az = y} {z : Az = y}
x̂ = x x x̂ = x
x̂
`0 `2 `1
x is k-sparse, and y = Ax
`0 , `1 , and `2 together
{z : Az = y} {z : Az = y} {z : Az = y}
x̂ = x x x̂ = x
x̂
`0 `2 `1
x is k-sparse, and y = Ax
`0 works if any 2k columns of A are linearly independent
`2 never works
`1 works if the condition on A is strengthened
Introduction Compressed sensing Discussion
Example
Reconstruction of a 512-long signal from 120 random measurements
1.5 4
3
1
2
0.5
1
0 0
ï1
ï0.5
ï2
ï1
ï3
ï1.5 ï4
ï5
0 100 200 300 400 500 0 100 200 300 400 500
Example
Reconstruction of a 512-long signal from 120 random measurements
5 5
4 4
1
3 3
2 0.5 2
1 1
0 0 0
ï1 ï1
ï2 ï0.5 ï2
ï3 ï3
ï1
ï4 ï4
ï5 ï5
0 100 200 300 400 500 0 100 200 300 400 500 0 100 200 300 400 500
Example
Reconstruction of a 256 × 256 image (= 65536 pixels) from 5481
measurements in the Fourier domain†
Example
Reconstruction of a 256 × 256 image (= 65536 pixels) from 5481
measurements in the Fourier domain†
†
E. J. Candès, J. Romberg, and T. Tao, “Robust uncertainty principles: exact signal reconstruction from highly incomplete
frequency information,” IEEE Trans. Inform. Theory, 2006.
Introduction Compressed sensing Discussion
Outline
1 Introduction
Traditional data acquisition
Compressive data acquisition
2 Compressed sensing
Measurement protocol
Recovery procedure
Recovery conditions
Sensing matrices
3 Discussion
Connections with other fields
Numerical methods
Applications
Conclusions
Introduction Compressed sensing Discussion
Theorem†
Let (a1 · · · an ) be the columns of A, normalized so that kai k2 = 1
∀ i, M = maxi6=j aTi aj , and y = Ax. If
†
D. L. Donoho and M. Elad, “Optimally sparse representation in general (nonorthogonal) dictionaries via `1
minimization,” Proc. Nat. Acad. Sci., 2003
Introduction Compressed sensing Discussion
Theorem†
√
Let y = Ax and x̂ be the solution of (P1 ). If δ2k < 2 − 1, then
kx − x̂k1 ≤ C0 kx − xk k1
√
kx − x̂k2 ≤ C0 kx − xk k1 / k
Noisy measurements
y = Ax + n
Problem
(P1 ) min kxk1 , s.t. kAx − yk2 ≤
x
Noisy measurements
Problem
(P1 ) min kxk1 , s.t. kAx − yk2 ≤
x
kAx − yk2 ≤
Introduction Compressed sensing Discussion
Stable recovery
Reconstruction error bounded by 2 terms:
same as in the noiseless case
proportional to the noise level
C0 and C1 are rather small, e.g., if δ = 0.25, then C0 ≤ 5.5 and
C1 ≤ 6
†
E. J. Candès, “The restricted isometry property and its implications for compressed sensing,” Compte Rendus de
l’Academie des Sciences, 2008.”
Introduction Compressed sensing Discussion
A has the null space property (NSP) of order k if, for some
γ ∈ (0, 1),
Theorem†
Let y = Ax. If A has NSP of order k, then x is the solution of (P1 ) ∀ x
k-sparse. Conversely, if x is the solution of (P1 ) ∀ x k-sparse, then A
has NSP of order 2k.
†
A. Cohen, W. Dahmen, and R. DeVore, “Compressed sensing and best k-term approximation,” J. Amer. Math. Soc., 2009.
Introduction Compressed sensing Discussion
Recovery conditions
Mutual coherence: easy to check but “coarse/pessimistic”
y = Ax ≡ Gy = GAx, but
A satisfies RIP ; GA satisfies RIP
m ≥ Ck(ln n/m + 1)
O(k ln n) measurements are sufficient to recover the signal with an
accuracy comparable to that attainable with direct knowledge of the k
largest coefficients
†
B. Kashin, “Diameters of some finite-dimensional sets and classes of smooth functions,” Izv. Akad. Nauk SSSR, Ser.
Mat., 1977; A. Y. Garnaev and E. D. Gluskin, “On widths of the Euclidean ball,” Sov. Math.–Dokl., 1984.
Introduction Compressed sensing Discussion
Outline
1 Introduction
Traditional data acquisition
Compressive data acquisition
2 Compressed sensing
Measurement protocol
Recovery procedure
Recovery conditions
Sensing matrices
3 Discussion
Connections with other fields
Numerical methods
Applications
Conclusions
Introduction Compressed sensing Discussion
Concentration inequality
The random matrix A satisfies the concentration inequality if, ∀ x and
∀ ∈ (0, 1),
P kAxk22 − kxk22 ≥ kxk22 ≤ 2e−nc
where c > 0
Introduction Compressed sensing Discussion
Theorem†
Let δ ∈ (0, 1). If A satisfies the concentration inequality, then there
exist constants c1 , c2 > 0 depending only on δ such that the restricted
isometry constant of A satisfies δk ≤ δ with probability exceeding
1 − 2e−c1 m provided that m ≥ c2 k ln n/k.
Observation: m ≥ c2 k ln n/k ⇒ m≥ c2
1+c2 k(ln n/m + 1)
†
R. Baraniuk, M. Davenport, R. DeVore, and M. Wakin, “A Simple Proof of the Restricted Isometry Property for Random
Matrices,” Constr. Approx., 2009
Introduction Compressed sensing Discussion
Random sensing
Examples
Two important cases satisfy the concentration inequality
Gaussian: (A)i, j ∼ N (0, 1/m) i.i.d.
√
Bernoully: (A)i, j ∼ B(1/2) i.i.d. with values ±1/ m
Introduction Compressed sensing Discussion
If Ψ orthogonal ⇒ Φ = AΨT
m ≥ Ck ln n/k
m ≥ Ck ln n/k
Outline
1 Introduction
Traditional data acquisition
Compressive data acquisition
2 Compressed sensing
Measurement protocol
Recovery procedure
Recovery conditions
Sensing matrices
3 Discussion
Connections with other fields
Numerical methods
Applications
Conclusions
Introduction Compressed sensing Discussion
y = Ax + n
noise
In this case
min kxk1 , s.t. kAx − yk22 ≤
x
†
S. S. Chen, D. L. Donoho, and M. A. Saunders, “Atomic decomposition by basis pursuit,” SIAM J. Sci. Comput., 1999.
Introduction Compressed sensing Discussion
Regularization
As the parameter λ varies over (0, ∞), its solution traces out the
optimal trade-off curve
†
A. E. Hoerl and R. W. Kennard, “Ridge regression: applications to nonorthogonal problems,” Technometrics, 1970.
Introduction Compressed sensing Discussion
Regularization
Bridge regression† is more general
min kAx − yk22 + λkxkpp
x
Variants of lasso
If more prior information (other than sparsity) is available on x, it can
be included in the optimization problem through proper penalizing
terms
The fused lasso† preserves local constancy when the regressors
are properly arranged
( n
)
X
min kAx − yk2 + λ1 kxk1 + λ2
2
|xi − xi−1 |
x
i=2
†
R. Tibshirani, M. Saunders, S. Rosset, J. Zhu, and K. Knight, “Sparsity and smoothness via the fused lasso,” J. Roy. Stat.
Soc. B, 2005.
Introduction Compressed sensing Discussion
Variants of lasso
The group lasso† promotes group selection
( k
)
X
min kAx − yk22 + λ kxi k2
x
i=1
Outline
1 Introduction
Traditional data acquisition
Compressive data acquisition
2 Compressed sensing
Measurement protocol
Recovery procedure
Recovery conditions
Sensing matrices
3 Discussion
Connections with other fields
Numerical methods
Applications
Conclusions
Introduction Compressed sensing Discussion
Convex programs
Algorithms
Can be all solved through standard convex optimization methods,
e.g., interior point methods (primal-dual, log-barrier, etc.)
general purposes solvers can handle small to medium size
problems
optimized algorithms (with fast matrix-vector operations) can
scale to large problems
Outline
1 Introduction
Traditional data acquisition
Compressive data acquisition
2 Compressed sensing
Measurement protocol
Recovery procedure
Recovery conditions
Sensing matrices
3 Discussion
Connections with other fields
Numerical methods
Applications
Conclusions
Introduction Compressed sensing Discussion
Applications
http://dsp.rice.edu/cscamera
Introduction Compressed sensing Discussion
http://dsp.rice.edu/cscamera
Introduction Compressed sensing Discussion
http://dsp.rice.edu/cscamera
Introduction Compressed sensing Discussion
http://dsp.rice.edu/cscamera
Introduction Compressed sensing Discussion
⇒
(1)
⇓ (3)
original image
original image
Angiogram with observations along 80 lines in the Fourier domain
and 16129 measurements†
†
E. J. Candès and J. Romberg, “Practical signal recovery from random projections,” SPIE Conf. on Wavelet App. in
Signal and Image Process. 2008
Introduction Compressed sensing Discussion
detail
Angiogram with observations along 80 lines in the Fourier domain
and 16129 measurements†
†
E. J. Candès and J. Romberg, “Practical signal recovery from random projections,” SPIE Conf. on Wavelet App. in
Signal and Image Process. 2008
Introduction Compressed sensing Discussion
p(t)
P
x(t) = `∈Λ a` e
2πif` t , multi-tone signal
Λ ⊂ {0, ±1, . . . , ±(W/2 − 1), W/2}, W/2 ∈ N,
card(Λ) = k W
sampling rate: R = O(k ln W) ⇒ no need for a high-rate ADC
†
J. A. Tropp, J. N. Laska, M. F. Duarte, J. K. Romberg, and R. G. Baraniuk, “Beyond Nyquist: efficient sampling of sparse
bandlimited signals,” IEEE Trans. Inform. Theory, 2010
Introduction Compressed sensing Discussion
p(t)
|X(f )| |X(f )P (f )| |Y (f )|
x(t)
mixing functions pi (t)
pm (t)
n/B ym (n)
LPF
B
−W/2 0 W/2
x(t)
mixing functions pi (t)
pm (t)
n/B ym (n)
LPF
B
Hardware implementation
†
M. Mishali and Y. C. Eldar, “From theory to practice: sub-Nyquist sampling of sparse wideband analog signals,” IEEE
Trans. Signal Process., 2010
Introduction Compressed sensing Discussion
| {z }
shifts of the
received multiplex y user’s signature noise n
channel response
Model: y = Ax + n, x sparse
of the user
1
10
−1
10
0
10
NMSE
NMSE
−2
10
−1
10
−3
10
−2
10
−4
10 −3
10
−5 −4
10 10
0 10 20 30 40 50 0 10 20 30 40 50
SNR SNR
P=10, N=15, K=10, S=5, SNR=20dB, ISR=0dB P=10, N=15, K=10, S=5, ISR=0dB, PFA=0.01
0
1 10
Lasso
0.9 LS
OMP
0.8
−1
10
0.7
0.6 Lasso
LS
PMD
−2
PD
0.5 OMP 10
0.4
0.3
−3
10
0.2
0.1
−4
0 10
0 0.2 0.4 0.6 0.8 1 0 10 20 30 40 50
PFA SNR
Outline
1 Introduction
Traditional data acquisition
Compressive data acquisition
2 Compressed sensing
Measurement protocol
Recovery procedure
Recovery conditions
Sensing matrices
3 Discussion
Connections with other fields
Numerical methods
Applications
Conclusions
Introduction Compressed sensing Discussion
Conclusions
CS resources http://dsp.rice.edu/cs
CS blog http://nuit-blanche.blogspot.com/
Software:
SparseLab http://sparselab.stanford.edu/
`1 -magic http://www.acm.caltech.edu/l1magic/
GPSR http://www.lx.it.pt/˜mtf/GPSR/
`1 LS http://www.stanford.edu/˜boyd/l1_ls/