Noisy Signal Recovery via Iterative Reweighted L1 ... - Semantic Scholar

Report 4 Downloads 142 Views
Introduction

L1-Minimization

Reweighted L1

Main Results

Noisy Signal Recovery via Iterative Reweighted L1-Minimization Deanna Needell UC Davis / Stanford University

Asilomar SSC, November 2009

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

What? Theorem 1 I have injured vocal chords and can’t speak loudly. I’m not nervous, it’s just the voice!

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Problem Background

Setup 1 2 3

Suppose x is an unknown signal in Rd . Design measurement matrix Φ : Rd → Rm . Collect noisy measurements u = Φx + e.    



u  = 

4 5

         x  + e       

Φ

Problem: Reconstruct signal x from measurements u Wait, isn’t this impossible? def

Assume x is s-sparse: kxk0 = | supp(x)| ≤ s ≪ d. D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Problem Background

Setup 1 2 3

Suppose x is an unknown signal in Rd . Design measurement matrix Φ : Rd → Rm . Collect noisy measurements u = Φx + e.    



u  = 

4 5

         x  + e       

Φ

Problem: Reconstruct signal x from measurements u Wait, isn’t this impossible? def

Assume x is s-sparse: kxk0 = | supp(x)| ≤ s ≪ d. D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Problem Background

Setup 1 2 3

Suppose x is an unknown signal in Rd . Design measurement matrix Φ : Rd → Rm . Collect noisy measurements u = Φx + e.    



u  = 

4 5

         x  + e       

Φ

Problem: Reconstruct signal x from measurements u Wait, isn’t this impossible? def

Assume x is s-sparse: kxk0 = | supp(x)| ≤ s ≪ d. D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Problem Background

Applications Compressive Imaging Computational Biology Medical Imaging Astronomy Geophysical Data Analysis Compressive Radar Many more (see www.dsp.ece.rice.edu)

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Problem Background

How can we reconstruct? Obvious way: Suppose the matrix Φ is one-to-one on the set of sparse vectors and e = 0. Set xˆ = argmin ||z||0

such that Φz = u.

Then xˆ = x! Bad news: This would require a search through numerically feasible.

D. Needell

d s

subspaces! Not

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Problem Background

How can we reconstruct? Obvious way: Suppose the matrix Φ is one-to-one on the set of sparse vectors and e = 0. Set xˆ = argmin ||z||0

such that Φz = u.

Then xˆ = x! Bad news: This would require a search through numerically feasible.

D. Needell

d s

subspaces! Not

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Problem Background

How else can we reconstruct?

Geometric Idea Minimizing the ℓ0 -ball is too hard, so let’s try a different one. Our favorites... Least Squares L1-Minimization (using Linear Programming) Which one?

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Problem Background

How else can we reconstruct?

Geometric Idea Minimizing the ℓ0 -ball is too hard, so let’s try a different one. Our favorites... Least Squares L1-Minimization (using Linear Programming) Which one?

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Problem Background

Which one?

x* x

x = x*

Figure: Minimizing the ℓ2 versus the ℓ1 balls. D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Problem Background

What do we assume about Φ? Restricted Isometry Property (RIP) The sth restricted isometry constant of Φ is the smallest δs such that (1 − δs )kxk2 ≤ kΦxk2 ≤ (1 + δs )kxk2

whenever kxk0 ≤ s.

For Gaussian or Bernoulli measurement matrices, with high probability δs ≤ c < 1 when m & s log d. Random Fourier and others with fast multiply have similar property. D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Problem Background

What do we assume about Φ? Restricted Isometry Property (RIP) The sth restricted isometry constant of Φ is the smallest δs such that (1 − δs )kxk2 ≤ kΦxk2 ≤ (1 + δs )kxk2

whenever kxk0 ≤ s.

For Gaussian or Bernoulli measurement matrices, with high probability δs ≤ c < 1 when m & s log d. Random Fourier and others with fast multiply have similar property. D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Results

Proven Results

L1-Minimization [Cand`es-Tao] Assume √ that the measurement matrix Φ satisfies the RIP with δ2s < 2 − 1. Then every s-sparse vector x can be exactly recovered from its measurements u = Φx as a unique solution to the linear optimization problem: xˆ = argmin ||z||1

D. Needell

such that Φz = u.

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Results

Numerical Results Percentage of flat signals recovered with BP, d=256 100 90

Percentage recovered

80

s=4 s=12

70

s=20 60

s=28 s=36

50 40 30 20 10 0

0

50

100 150 Measurements m

200

250

Figure: The percentage of sparse flat signals exactly recovered by Basis Pursuit as a function of the number of measurements m in dimension d = 256 for various levels of sparsity s. D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Results

What about noise? Noisy Formulation For a non-sparse vector x with noisy measurements u = Φx + e, xˆ = argmin ||z||1

such that kΦz − uk2 ≤ ε.

(1)

L1-Minimization [Cand`es-Romberg-Tao] Let Φ √ be a measurement matrix satisfying the RIP with δ2s < 2 − 1. Then for any arbitrary signal and corrupted measurements u = Φx + e with kek2 ≤ ε, the solution xˆ to (1) satisfies kx − xs k1 √ . kˆ x − xk2 ≤ Cs · ε + Cs′ · s √ Note: As δ2s → 2 − 1, Cs , Cs′ → ∞!! D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Results

What about noise? Noisy Formulation For a non-sparse vector x with noisy measurements u = Φx + e, xˆ = argmin ||z||1

such that kΦz − uk2 ≤ ε.

(1)

L1-Minimization [Cand`es-Romberg-Tao] Let Φ √ be a measurement matrix satisfying the RIP with δ2s < 2 − 1. Then for any arbitrary signal and corrupted measurements u = Φx + e with kek2 ≤ ε, the solution xˆ to (1) satisfies kx − xs k1 √ . kˆ x − xk2 ≤ Cs · ε + Cs′ · s √ Note: As δ2s → 2 − 1, Cs , Cs′ → ∞!! D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Results

What about noise? Noisy Formulation For a non-sparse vector x with noisy measurements u = Φx + e, xˆ = argmin ||z||1

such that kΦz − uk2 ≤ ε.

(1)

L1-Minimization [Cand`es-Romberg-Tao] Let Φ √ be a measurement matrix satisfying the RIP with δ2s < 2 − 1. Then for any arbitrary signal and corrupted measurements u = Φx + e with kek2 ≤ ε, the solution xˆ to (1) satisfies kx − xs k1 √ . kˆ x − xk2 ≤ Cs · ε + Cs′ · s √ Note: As δ2s → 2 − 1, Cs , Cs′ → ∞!! D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Results

What about noise? Noisy Formulation For a non-sparse vector x with noisy measurements u = Φx + e, xˆ = argmin ||z||1

such that kΦz − uk2 ≤ ε.

(1)

L1-Minimization [Cand`es-Romberg-Tao] Let Φ √ be a measurement matrix satisfying the RIP with δ2s < 2 − 1. Then for any arbitrary signal and corrupted measurements u = Φx + e with kek2 ≤ ε, the solution xˆ to (1) satisfies kx − xs k1 √ . kˆ x − xk2 ≤ Cs · ε + Cs′ · s √ Note: As δ2s → 2 − 1, Cs , Cs′ → ∞!! D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Results

Numerical Results Recovery Error from BP on Perturbed Measurements (by Gaussian noise), d=256 14

12 s=4 s=12

Average Error

10

s=20 s=28

8

s=36

6

4

2

0 60

80

100

120

140 160 180 Measurements m

200

220

240

Figure: The recovery error of L1-Minimization under perturbed measurements (kek2 = 0.5) as a function of the number of measurements m in dimension d = 256 for various levels of sparsity s. D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Results

What if we are close?

Suppose we recover xˆ ≈ x

Most likely, this means xˆi ≈ xi

In particular, xˆi is small/large when xi is small/large Weighted L1 xˆ

(2)

d X zi = argmin xˆi z i =1

D. Needell

such that

kΦz − uk2 ≤ ε

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Results

What if we are close?

Suppose we recover xˆ ≈ x

Most likely, this means xˆi ≈ xi

In particular, xˆi is small/large when xi is small/large Weighted L1 xˆ

(2)

d X zi = argmin xˆi z i =1

D. Needell

such that

kΦz − uk2 ≤ ε

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Results

What if we are close?

Suppose we recover xˆ ≈ x

Most likely, this means xˆi ≈ xi

In particular, xˆi is small/large when xi is small/large Weighted L1 xˆ

(2)

d X zi = argmin xˆi + a z i =1

D. Needell

such that

kΦz − uk2 ≤ ε

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Results

Weighted Geometry x = x* x

x*

Figure: The geometry of the weighted ℓ1 -ball.

Noise-free case: In cases where xˆ 6= x, we should have that xˆ(2) is closer to x, or even equal. Noisy case: This implies xˆ(2) should be closer to x than xˆ was. Can we repeat this again? D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Reweighted L1-Minimization

Reweighted ℓ1 -minimization (RWL1) Input: Measurement vector u ∈ Rm , stability parameter a Output: Reconstructed vector xˆ Initialize Set the weights wi = 1 for i = 1 . . . d. Approximate Solve the reweighted ℓ1 -minimization problem: xˆ = argmin xˆ∈Rd

d X i =1

wi xˆi subject to kΦˆ x − uk2 ≤ ε.

Update Reset the weights: wi =

D. Needell

1 . |ˆ xi | + a

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Reweighted L1-Minimization

Numerical Results

Sup norm errors after 1 iteration of (reweighted) BP

Sup norm errors after nine iterations of reweighted BP

1.4

1.4

1.2

1.2

1

1

0.8

0.8

0.6

0.6

0.4

0.4

0.2

0.2

0

0

100

200

300

400

500

0

0

100

200

300

400

500

Figure: ℓ∞ -norm error for reweighted L1 in the noise-free case

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Reweighted L1-Minimization

Numerical Results

Figure: Probability of reconstruction [Cand`es-Wakin-Boyd].

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Reweighted L1-Minimization

Numerical Results with noise Improvement by reweighted L1 after 9 iterations for d=256 m=128 s=30 with decreasing e 150

Reconstruction error for d=256 m=128 s=30 with decreasing e 6 One iteration 9 Iterations 5

100 Number of trials

Error

4

3

2

50

1

0

0

100

200

300

400

500

Trials

0 0.1

0.2

0.3

0.4 0.5 Improvement factor

0.6

0.7

0.8

Figure: Improvements in the ℓ2 reconstruction error using reweighted ℓ1 -minimization versus standard ℓ1 -minimization for sparse Gaussian signals.

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Reweighted L1-Minimization

Numerical Results with noise Reconstruction error for d=256 m=128 s=30 with decreasing e One iteration

5

Improvement by reweighted L1 after 9 iterations for d=256 m=128 s=30 with decreasing e 140

9 Iterations

4.5

120

4 100 Number of trials

Error

3.5 3 2.5

80

60

2 40 1.5 20

1 0.5

0

100

200

300

400

500

Trials

0 0.1

0.2

0.3

0.4

0.5 0.6 0.7 Improvement factor

0.8

0.9

1

Figure: Improvements in the ℓ2 reconstruction error using reweighted ℓ1 -minimization versus standard ℓ1 -minimization for sparse Bernoulli signals.

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Reweighted L1-Minimization

Observations

The noiseless case suggests that an ℓ∞ -norm bound may be required for RWL1 to succeed. In the noisy case it is clear that we cannot recover signal coordinates that are below some threshold. If each iteration of RWL1 improves the error, perhaps we should take a → 0. (Recall wi = |ˆxi 1|+a ).

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Reweighted L1-Minimization

Observations

The noiseless case suggests that an ℓ∞ -norm bound may be required for RWL1 to succeed. In the noisy case it is clear that we cannot recover signal coordinates that are below some threshold. If each iteration of RWL1 improves the error, perhaps we should take a → 0. (Recall wi = |ˆxi 1|+a ).

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Reweighted L1-Minimization

Observations

The noiseless case suggests that an ℓ∞ -norm bound may be required for RWL1 to succeed. In the noisy case it is clear that we cannot recover signal coordinates that are below some threshold. If each iteration of RWL1 improves the error, perhaps we should take a → 0. (Recall wi = |ˆxi 1|+a ).

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Main Results

Main Results

RWL1 - Sparse case [N]

√ Assume Φ satisfies the RIP with δ2s ≤ δ where δ < 2 − 1. Let x be an s-sparse vector with noisy measurements u = Φx + e where kek2 ≤ ε. Assume the smallest nonzero coordinate µ of x satisfies 4αε . Then the limiting approximation from reweighted µ ≥ 1−ρ ℓ1 -minimization satisfies

where C ′′ =

2α 1+ρ ,

ρ=



kx − xˆk2 ≤ C ′′ ε,

2δ 1−δ

and α =

D. Needell

√ 2 1+δ 1−δ .

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Main Results

Remarks

Without noise, this result coincides with previous results on L1. √ The key improvement: As δ → 2 − 1, C ′′ remains bounded.

The error bound is the limiting bound, but a recursive relation in the proof gives exact improvements per iteration. We show in practice it is attained quite quickly. For signals whose smallest non-zero coefficient µ does not satisfy the condition of the theorem, we may apply the theorem to those coefficients that do satisfy this requirement, and treat the others as noise...

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Main Results

Remarks

Without noise, this result coincides with previous results on L1. √ The key improvement: As δ → 2 − 1, C ′′ remains bounded.

The error bound is the limiting bound, but a recursive relation in the proof gives exact improvements per iteration. We show in practice it is attained quite quickly. For signals whose smallest non-zero coefficient µ does not satisfy the condition of the theorem, we may apply the theorem to those coefficients that do satisfy this requirement, and treat the others as noise...

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Main Results

Remarks

Without noise, this result coincides with previous results on L1. √ The key improvement: As δ → 2 − 1, C ′′ remains bounded.

The error bound is the limiting bound, but a recursive relation in the proof gives exact improvements per iteration. We show in practice it is attained quite quickly. For signals whose smallest non-zero coefficient µ does not satisfy the condition of the theorem, we may apply the theorem to those coefficients that do satisfy this requirement, and treat the others as noise...

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Main Results

Remarks

Without noise, this result coincides with previous results on L1. √ The key improvement: As δ → 2 − 1, C ′′ remains bounded.

The error bound is the limiting bound, but a recursive relation in the proof gives exact improvements per iteration. We show in practice it is attained quite quickly. For signals whose smallest non-zero coefficient µ does not satisfy the condition of the theorem, we may apply the theorem to those coefficients that do satisfy this requirement, and treat the others as noise...

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Main Results

Extension RWL1 - non-sparse extension [N]

√ Assume Φ satisfies the RIP with δ2s ≤ 2 − 1. Let x be an arbitrary vector with noisy measurements u = Φx + e where kek2 ≤ ε. Assume the smallest nonzero coordinate µ of xs satisfies 0 √1 µ ≥ 4αε 1−ρ , where ε0 = 1.2(kx − xs k2 + s kx − xs k1 ) + ε. Then the limiting approximation from reweighted ℓ1 -minimization satisfies kx − xˆk2 ≤ and kx − xˆk2 ≤

 4.1α  kx − xs/2 k1 √ +ε , 1+ρ s

 kx − xs k1 2.4α  √ kx − xs k2 + +ε , 1+ρ s

where ρ and α are as before. D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Main Results

Theoretical Results

3

2

1

0

(b) Number of iterations til convergence, M=10, e=0.1 6

Iterations til Convergence

Iterations til Convergence

(a) Number of iterations til convergence, M=10, e=0.01 4

0

0.1

0.2 δ value

10 8 6 4 2 0.05

0.1 0.15 δ value

0.2

3 2 1

0.25

0

0.1

0.2 δ value

0.3

(d) Number of iterations til convergence, M=10, e=1 30

Iterations til Convergence

Iterations til Convergence

12

0

4

0

0.3

(c) Number of iterations til convergence, M=10, e=0.5

0

5

25 20 15 10 5 0

0

0.02

0.04 δ value

0.06

0.08

Figure: Number of iterations required for theoretical error bounds to reach limiting theoretical error when (a) µ = 10, ε = 0.01, (b) µ = 10, ε = 0.1, (c) µ = 10, ε = 0.5, (d) µ = 10, ε = 1.0. D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Main Results

Recent work

Wipf-Nagarajan elaborate on convergence and show connections to reweighted ℓ2 -minimization. Wipf-Nagarajan also show that a non-separable variant has desirable properties. Xu-Khajehnejad-Avestimehr-Hassibi provide a theoretical foundation for the analysis of RWL1 and show that for a nontrivial class of signals, a variant of RWL1 indeed can improve upon L1 in the noiseless case.

D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Introduction

L1-Minimization

Reweighted L1

Main Results

Thank you

For more information E-mail: [email protected] Web: http://www-stat.stanford.edu/~dneedell References: Candes, Wakin, Boyd, “Enhancing sparsity by reweighted ℓ1 minimization”, J. Fourier Anal. Appl., 14 877-905. Needell, “Noisy signal recovery via iterative reweighted L1-minimization,” 2009. Wipf and Nagarajan, “Solving Sparse Linear Inverse Problems: Analysis of Reweighted ℓ1 and ℓ2 Methods,” J. of Selected Topics in Signal Processing, Special Issue on Compressive Sensing, 2010. Xu, Khajehnejad, Avestimehr, and Hassibi, “Breaking through the Thresholds: an Analysis for Iterative Reweighted ℓ1 Minimization via the Grassmann Angle Framework,” Proc. Allerton Conference on Communication, Control, and computing, Sept. 2009. D. Needell

Noisy Signal Recovery via Iterative Reweighted L1-Minimization