SlideShare a Scribd company logo
1 of 71
Download to read offline
Low Complexity
Regularization of
Inverse Problems
Cours #2
Recovery Guarantees
Gabriel Peyré
www.numerical-tours.com
Overview of the Course

• Course #1: Inverse Problems

• Course #2: Recovery Guarantees

• Course #3: Proximal Splitting Methods
Overview

• Low-complexity Regularization with Gauges

• Performance Guarantees

• Grid-free Regularization
Inverse Problem Regularization
Observations: y = x0 + w 2 RP .
Estimator: x(y) depends only on

observations y
parameter
Inverse Problem Regularization
Observations: y = x0 + w 2 RP .
Estimator: x(y) depends only on

observations y
parameter

Example: variational methods
1
x(y) 2 argmin ||y
x||2 + J(x)
x2RN 2
Data fidelity Regularity
Inverse Problem Regularization
Observations: y = x0 + w 2 RP .
Estimator: x(y) depends only on

observations y
parameter

Example: variational methods
1
x(y) 2 argmin ||y
x||2 + J(x)
x2RN 2
Data fidelity Regularity

Choice of : tradeo

Noise level
||w||

Regularity of x0
J(x0 )
Inverse Problem Regularization
Observations: y = x0 + w 2 RP .
Estimator: x(y) depends only on

observations y
parameter

Example: variational methods
1
x(y) 2 argmin ||y
x||2 + J(x)
x2RN 2
Data fidelity Regularity

Choice of : tradeo
No noise:

Noise level
||w||

0+ , minimize

Regularity of x0
J(x0 )

x? 2 argmin J(x)
x2RQ ,Kx=y
Inverse Problem Regularization
Observations: y = x0 + w 2 RP .
Estimator: x(y) depends only on

observations y
parameter

Example: variational methods
1
x(y) 2 argmin ||y
x||2 + J(x)
x2RN 2
Data fidelity Regularity

Choice of : tradeo
No noise:

Noise level
||w||

0+ , minimize

Regularity of x0
J(x0 )

x? 2 argmin J(x)
x2RQ ,Kx=y

This course:

Performance analysis.
Fast computational scheme.
Union of Linear Models for Data Processing
Union of models: T 2 T linear spaces.
Synthesis
sparsity:
T

Coe cients x

Image

x
Union of Linear Models for Data Processing
Union of models: T 2 T linear spaces.
Synthesis
sparsity:

Structured
sparsity:

T

Coe cients x

Image

x
Union of Linear Models for Data Processing
Union of models: T 2 T linear spaces.
Synthesis
sparsity:

Structured
sparsity:

T

Coe cients x

Image

x

D

Analysis
sparsity:
Image x

Gradient D⇤ x
Union of Linear Models for Data Processing
Union of models: T 2 T linear spaces.
Synthesis
sparsity:

Structured
sparsity:

T

Coe cients x

Analysis
sparsity:

Image

x

D

Low-rank:

Image x

Gradient D⇤ x

S1,·

Multi-spectral imaging:
Pr
xi,· = j=1 Ai,j Sj,·

S2,·

x

S3,·
Gauges for Union of Linear Models
Gauge:

J :R

N

!R

+

Convex
8 ↵ 2 R+ , J(↵x) = ↵J(x)
Gauges for Union of Linear Models
Gauge:

J :R

N

!R

+

Convex
8 ↵ 2 R+ , J(↵x) = ↵J(x)

Piecewise regular ball , Union of linear models (T )T 2T

x
J(x) = ||x||1 T
T = sparse
vectors
Gauges for Union of Linear Models
Gauge:

J :R

N

!R

+

Convex
8 ↵ 2 R+ , J(↵x) = ↵J(x)

Piecewise regular ball , Union of linear models (T )T 2T

T0

x0

x

J(x) = ||x||1 T
T = sparse
vectors
Gauges for Union of Linear Models
Gauge:

J :R

N

!R

Convex
8 ↵ 2 R+ , J(↵x) = ↵J(x)

+

Piecewise regular ball , Union of linear models (T )T 2T

T
T0

x0

x

J(x) = ||x||1 T
T = sparse
vectors

x

T0

x
|x1 |+||x2,3 ||

T = block
sparse
vectors

0
Gauges for Union of Linear Models
Gauge:

J :R

N

!R

Convex
8 ↵ 2 R+ , J(↵x) = ↵J(x)

+

Piecewise regular ball , Union of linear models (T )T 2T

T
T0

x0

x

J(x) = ||x||1 T
T = sparse
vectors

x
T

x
|x1 |+||x2,3 ||

T = block
sparse
vectors

0

x

0

J(x) = ||x||⇤

T = low-rank
matrices
Gauges for Union of Linear Models
Gauge:

J :R

N

!R

Convex
8 ↵ 2 R+ , J(↵x) = ↵J(x)

+

Piecewise regular ball , Union of linear models (T )T 2T

T
T0

x0

x

J(x) = ||x||1 T
T = sparse
vectors

T0

x
T

x
|x1 |+||x2,3 ||

T = block
sparse
vectors

0

x

x
x0

0

J(x) = ||x||⇤

T = low-rank
matrices

J(x) = ||x||1

T = antisparse
vectors
Subdifferentials and Models
@J(x) = {⌘  8 y, J(y) > J(x)+h⌘, y

xi}

|x|
Subdifferentials and Models
@J(x) = {⌘  8 y, J(y) > J(x)+h⌘, y

|x|

xi}

Example: J(x) = ||x||1
⇢
supp(⌘) = I,
@||x||1 = ⌘ 
8 j 2 I, |⌘j | 6 1
/

I = supp(x) = {i  xi 6= 0}

@J(x)
0

x
Subdifferentials and Models
@J(x) = {⌘  8 y, J(y) > J(x)+h⌘, y

|x|

xi}

Example: J(x) = ||x||1
⇢
supp(⌘) = I,
@||x||1 = ⌘ 
8 j 2 I, |⌘j | 6 1
/

I = supp(x) = {i  xi 6= 0}
Tx = {⌘  supp(⌘) = I}

Definition:

Tx = VectHull(@J(x))?

@J(x)
0

x

Tx
Subdifferentials and Models
@J(x) = {⌘  8 y, J(y) > J(x)+h⌘, y

|x|

xi}

Example: J(x) = ||x||1
⇢
supp(⌘) = I,
@||x||1 = ⌘ 
8 j 2 I, |⌘j | 6 1
/

@J(x)
0

I = supp(x) = {i  xi 6= 0}

Tx

Tx = {⌘  supp(⌘) = I}
ex = sign(x)

Definition:

Tx = VectHull(@J(x))?
⌘ 2 @J(x)

ex x

=)

ProjTx (⌘) = ex
Examples
`1 sparsity: J(x) = ||x||1
ex = sign(x)

x

@J(x)

x

0

Tx = {z  supp(z) ⇢ supp(x)}
Examples
`1 sparsity: J(x) = ||x||1
ex = sign(x)

Tx = {z  supp(z) ⇢ supp(x)}
P
Structured sparsity: J(x) = b ||xb ||
N (a) = a/||a||
ex = (N (xb ))b2B
Tx = {z  supp(z) ⇢ supp(x)}

x

@J(x)

x

0

x

@J(x)

x

0
Examples
`1 sparsity: J(x) = ||x||1
ex = sign(x)

Tx = {z  supp(z) ⇢ supp(x)}
P
Structured sparsity: J(x) = b ||xb ||
N (a) = a/||a||
ex = (N (xb ))b2B
Tx = {z  supp(z) ⇢ supp(x)}

x = U ⇤V ⇤
SVD:
Nuclear norm: J(x) = ||x||⇤
Tx = U A + BV ⇤  (A, B) 2 (Rn⇥n )2
ex = U V ⇤

x

@J(x)

x

0

x

@J(x)

x

0

x
@J(x)
Examples
`1 sparsity: J(x) = ||x||1
ex = sign(x)

Tx = {z  supp(z) ⇢ supp(x)}
P
Structured sparsity: J(x) = b ||xb ||
N (a) = a/||a||
ex = (N (xb ))b2B
Tx = {z  supp(z) ⇢ supp(x)}

x = U ⇤V ⇤
SVD:
Nuclear norm: J(x) = ||x||⇤
Tx = U A + BV ⇤  (A, B) 2 (Rn⇥n )2
ex = U V ⇤
I = {i  |xi | = ||x||1 }
Anti-sparsity: J(x) = ||x||1
Tx = {y  yI / sign(xI )}
ex = |I| 1 sign(x)

x

@J(x)

x

0

x

@J(x)

@J(x)

x

0

x
@J(x)

x

x

0
Overview

• Low-complexity Regularization with Gauges

• Performance Guarantees

• Grid-free Regularization
Dual Certificates
Noiseless recovery:

min J(x)

x= x0

(P0 )

x?

x=

x0
Dual Certificates
Noiseless recovery:

min J(x)

x= x0

(P0 )

Proposition:
x0 solution of (P0 ) () 9 ⌘ 2 D(x0 )

Dual certificates:

D(x0 ) = Im(

⇤

⌘

x?

)  @J(x0 )

@J(x0 )
x=

x0
Dual Certificates
Noiseless recovery:

min J(x)

(P0 )

x= x0

x?

Proposition:
x0 solution of (P0 ) () 9 ⌘ 2 D(x0 )

Dual certificates:
Proof:

(P0 )

()

D(x0 ) = Im(

min

2ker( )

8 (⌘, ) 2 @J(x0 ) ⇥ ker( ),
⌘ 2 Im(

⇤

)

=)

x0 solution

=)

h , ⌘i = 0

⇤

⌘

@J(x0 )
x=

)  @J(x0 )

J(x0 + )
J(x0 + ) > J(x0 ) + h , ⌘i
=)

8 , h , ⌘i 6 0

x0 solution.
=)

⌘ 2 ker( )? .

x0
Dual Certificates and L2 Stability
Tight dual certificates:
¯
D(x0 ) = Im( ⇤ )  ri(@J(x0 ))
ri(E) = relative interior of E
= interior for the topology of a↵(E)

⌘

x?

@J(x0 )
x=

x0
Dual Certificates and L2 Stability
Tight dual certificates:
¯
D(x0 ) = Im( ⇤ )  ri(@J(x0 ))
ri(E) = relative interior of E

⌘

x?

@J(x0 )
x=

x0

= interior for the topology of a↵(E)

Theorem:
¯
If 9 ⌘ 2 D(x0 ), for

[Fadili et al. 2013]

⇠ ||w|| one has ||x?

x0 || = O(||w||)
Dual Certificates and L2 Stability
Tight dual certificates:
¯
D(x0 ) = Im( ⇤ )  ri(@J(x0 ))
ri(E) = relative interior of E

⌘

x?

@J(x0 )
x=

x0

= interior for the topology of a↵(E)

Theorem:
¯
If 9 ⌘ 2 D(x0 ), for

[Fadili et al. 2013]

⇠ ||w|| one has ||x?

x0 || = O(||w||)

[Grassmair, Haltmeier, Scherzer 2010]: J = || · ||1 .
[Grassmair 2012]: J(x? x0 ) = O(||w||).
Dual Certificates and L2 Stability
Tight dual certificates:
¯
D(x0 ) = Im( ⇤ )  ri(@J(x0 ))
ri(E) = relative interior of E

⌘

x?

@J(x0 )
x=

x0

= interior for the topology of a↵(E)

Theorem:
¯
If 9 ⌘ 2 D(x0 ), for

[Fadili et al. 2013]

⇠ ||w|| one has ||x?

x0 || = O(||w||)

[Grassmair, Haltmeier, Scherzer 2010]: J = || · ||1 .
[Grassmair 2012]: J(x? x0 ) = O(||w||).

! The constants depend on N . . .
Compressed Sensing Setting
Random matrix:

2 RP ⇥N ,

i,j

⇠ N (0, 1), i.i.d.
Compressed Sensing Setting
Random matrix:

2 RP ⇥N ,

Sparse vectors: J = || · ||1 .

Theorem: Let s = ||x0 ||0 . If

i,j

⇠ N (0, 1), i.i.d.

[Rudelson, Vershynin 2006]
[Chandrasekaran et al. 2011]

P > 2s log (N/s)
¯
Then 9⌘ 2 D(x0 ) with high probability on

.
Compressed Sensing Setting
Random matrix:

2 RP ⇥N ,

Sparse vectors: J = || · ||1 .

Theorem: Let s = ||x0 ||0 . If

i,j

⇠ N (0, 1), i.i.d.

[Rudelson, Vershynin 2006]
[Chandrasekaran et al. 2011]

P > 2s log (N/s)
¯
Then 9⌘ 2 D(x0 ) with high probability on

Low-rank matrices: J = || · ||⇤ .

Theorem: Let r = rank(x0 ). If

.

[Chandrasekaran et al. 2011]

x0 2 RN1 ⇥N2

P > 3r(N1 + N2 r)
¯
Then 9⌘ 2 D(x0 ) with high probability on

.
Compressed Sensing Setting
Random matrix:

2 RP ⇥N ,

Sparse vectors: J = || · ||1 .

Theorem: Let s = ||x0 ||0 . If

i,j

⇠ N (0, 1), i.i.d.

[Rudelson, Vershynin 2006]
[Chandrasekaran et al. 2011]

P > 2s log (N/s)
¯
Then 9⌘ 2 D(x0 ) with high probability on

Low-rank matrices: J = || · ||⇤ .

Theorem: Let r = rank(x0 ). If

.

[Chandrasekaran et al. 2011]

x0 2 RN1 ⇥N2

P > 3r(N1 + N2 r)
¯
Then 9⌘ 2 D(x0 ) with high probability on

! Similar results for || · ||1,2 , || · ||1 .

.
Phase Transitions
THE THE GEOMETRYPHASE TRANSITIONS IN CONVEX OPTIMIZATION
GEOMETRY OF OF PHASE TRANSITIONS IN CONVEX OPTIMIZATION

J = || · ||1

1

100 100

J = || · ||⇤

1

900 900

75

75

P/N

50

P/N

600 600

50

300 300
25

25

0

0

0
00

25

25

50

50

75

s/N 1

75

100 100

0
00

0

0

10

10

p
r/ N 1

20

20

30

30

F IGURE Phase transitions for for linear inverse problems. [left] Recovery of sparse vectors. empirical
IGURE 2.2:2.2: Phase transitions linear inverse problems. [left] Recovery of sparse vectors. The The empiri
probability the `1 `1 minimization problem (2.6) identifies a sparse vector 0 2 100 given random line
robability thatthat the minimization problem (2.6) identifies a sparse vector x 0 2xR100Rgiven random linear

From [Amelunxen et al. 20013]
Minimal-norm Certificate
⌘ 2 D(x0 )

=)

⇢

⌘ = ⇤q
ProjT (⌘) = e

T = T x0
e = ex0
Minimal-norm Certificate
⌘ 2 D(x0 )

=)

⇢

⌘ = ⇤q
ProjT (⌘) = e

Minimal-norm pre-certificate: ⌘0 =

T = T x0
e = ex0

argmin
⌘=

⇤ q,⌘

T =e

||q||
Minimal-norm Certificate
⌘ 2 D(x0 )

=)

⇢

⌘ = ⇤q
ProjT (⌘) = e

Minimal-norm pre-certificate: ⌘0 =
Proposition:

One has

⌘0 = (

+
T

T = T x0
e = ex0

argmin
⌘=

)⇤ e

⇤ q,⌘

T =e

T

||q||

=

ProjT
Minimal-norm Certificate
⌘ 2 D(x0 )

=)

⇢

⌘ = ⇤q
ProjT (⌘) = e

Minimal-norm pre-certificate: ⌘0 =
Proposition:
Theorem:

One has

⌘0 = (

¯
If ⌘0 2 D(x0 ) and

+
T

T = T x0
e = ex0

argmin
⌘=

)⇤ e

⇤ q,⌘

T =e

T

||q||

=

ProjT

⇠ ||w||,

the unique solution x? of P (y) for y = x0 + w satisfies

Tx ? = T x 0

and ||x?

x0 || = O(||w||) [Vaiter et al. 2013]
Minimal-norm Certificate
⌘ 2 D(x0 )

=)

⇢

⌘ = ⇤q
ProjT (⌘) = e

Minimal-norm pre-certificate: ⌘0 =
Proposition:
Theorem:

One has

⌘0 = (

¯
If ⌘0 2 D(x0 ) and

+
T

T = T x0
e = ex0

argmin
⌘=

)⇤ e

⇤ q,⌘

T =e

T

||q||

=

ProjT

⇠ ||w||,

the unique solution x? of P (y) for y = x0 + w satisfies

Tx ? = T x 0

and ||x?

x0 || = O(||w||) [Vaiter et al. 2013]

[Fuchs 2004]: J = || · ||1 .
[Vaiter et al. 2011]: J = ||D⇤ · ||1 .
[Bach 2008]: J = || · ||1,2 and J = || · ||⇤ .
Compressed Sensing Setting
Random matrix:

2 RP ⇥N ,

Sparse vectors: J = || · ||1 .

Theorem: Let s = ||x0 ||0 . If

i,j

⇠ N (0, 1), i.i.d.
[Wainwright 2009]
[Dossal et al. 2011]

P > 2s log(N )

¯
Then ⌘0 2 D(x0 ) with high probability on

.
Compressed Sensing Setting
Random matrix:

2 RP ⇥N ,

i,j

Sparse vectors: J = || · ||1 .

⇠ N (0, 1), i.i.d.
[Wainwright 2009]
[Dossal et al. 2011]

Theorem: Let s = ||x0 ||0 . If

P > 2s log(N )

¯
Then ⌘0 2 D(x0 ) with high probability on

Phase
transitions:

L2 stability
P ⇠ 2s log(N/s)

vs.

.

Model stability
P ⇠ 2s log(N )
Compressed Sensing Setting
Random matrix:

2 RP ⇥N ,

i,j

Sparse vectors: J = || · ||1 .

⇠ N (0, 1), i.i.d.
[Wainwright 2009]
[Dossal et al. 2011]

Theorem: Let s = ||x0 ||0 . If

P > 2s log(N )

¯
Then ⌘0 2 D(x0 ) with high probability on

Phase
transitions:

L2 stability
P ⇠ 2s log(N/s)

vs.

.

Model stability

! Similar results for || · ||1,2 , || · ||⇤ , || · ||1 .

P ⇠ 2s log(N )
Compressed Sensing Setting
Random matrix:

2 RP ⇥N ,

i,j

Sparse vectors: J = || · ||1 .

⇠ N (0, 1), i.i.d.
[Wainwright 2009]
[Dossal et al. 2011]

Theorem: Let s = ||x0 ||0 . If

P > 2s log(N )

¯
Then ⌘0 2 D(x0 ) with high probability on

Phase
transitions:

L2 stability
P ⇠ 2s log(N/s)

vs.

.

Model stability

! Similar results for || · ||1,2 , || · ||⇤ , || · ||1 .

P ⇠ 2s log(N )

! Not using RIP technics (non-uniform result on x0 ).
1-D Sparse Spikes Deconvolution
⇥x =

xi (·

i)

x0

i

J(x) = ||x||1

Increasing :
reduces correlation.
reduces resolution.

x0
1-D Sparse Spikes Deconvolution
⇥x =

xi (·

x0

i)

i

J(x) = ||x||1

Increasing :
reduces correlation.
reduces resolution.

x0

||⌘0,I c ||1
2

1
0

10

20

I = {j  x0 (j) 6= 0}
||⌘0,I c ||1 < 1
()
¯
⌘0 2 D(x0 )
()
support recovery.
Overview

• Low-complexity Regularization with Gauges

• Performance Guarantees

• Grid-free Regularization
Support Instability and Measures
1
N

When N ! +1, support is not stable:

||⌘0,I c ||1

! c > 1.

N !+1

||⌘0,I c ||1

c
1
Unstable

Stable
Support Instability and Measures
1
N

When N ! +1, support is not stable:

||⌘0,I c ||1

! c > 1.

N !+1

Intuition: spikes wants to move laterally.
! Use Radon measures m 2 M(T), T = R/Z.

||⌘0,I c ||1

c
1
Unstable

Stable
Support Instability and Measures
1
N

When N ! +1, support is not stable:

||⌘0,I c ||1

! c > 1.

N !+1

Intuition: spikes wants to move laterally.
! Use Radon measures m 2 M(T), T = R/Z.

Extension of `1 : total variation
Z
||m||TV = sup
g(x) dm(x)
||g||1 61

T

Discrete measure: mx,a =

P

i

ai

One has ||mx,a ||TV = ||a||1

xi .

||⌘0,I c ||1

c
1
Unstable

Stable
Sparse Measure Regularization
Measurements: y =

8
< m0 2 M(T),
2
: M(T) ! L (T),
(m0 ) + w where
:
2
w 2 L (T).
Sparse Measure Regularization
Measurements: y =

8
< m0 2 M(T),
2
: M(T) ! L (T),
(m0 ) + w where
:
2
w 2 L (T).

Acquisition operator:
Z
(m)(x) =
'(x, x0 )dm(x0 ) where ' 2 C 2 (T ⇥ T)
T
Sparse Measure Regularization
Measurements: y =

8
< m0 2 M(T),
2
: M(T) ! L (T),
(m0 ) + w where
:
2
w 2 L (T).

Acquisition operator:
Z
(m)(x) =
'(x, x0 )dm(x0 ) where ' 2 C 2 (T ⇥ T)
T

Total-variation over measures regularization:
1
min
|| (m) y||2 + ||m||TV
m2M(T) 2
Sparse Measure Regularization
Measurements: y =

8
< m0 2 M(T),
2
: M(T) ! L (T),
(m0 ) + w where
:
2
w 2 L (T).

Acquisition operator:
Z
(m)(x) =
'(x, x0 )dm(x0 ) where ' 2 C 2 (T ⇥ T)
T

Total-variation over measures regularization:
1
min
|| (m) y||2 + ||m||TV
m2M(T) 2
! Infinite dimensional convex program.

! If dim(Im( )) < +1, dual is finite dimensional.

! If

is a filtering, re-cast dual as SDP program.
Fuchs vs. Vanishing Pre-Certificates
Measures:

1
2 ||
m2M

min

m

y||2 + ||m||TV

+1

1
Fuchs vs. Vanishing Pre-Certificates
Measures:

On a grid z:

1
2 ||
m2M

m

1
2 ||

za

min

min

a2RN

y||2 + ||m||TV

+1

zi

y||2 + ||a||1
1
Fuchs vs. Vanishing Pre-Certificates
1
2 ||
m2M

m

1
2 ||

za

min

Measures:

On a grid z:

min

a2RN

y||2 + ||m||TV

+1

zi

y||2 + ||a||1
1

For m0 = mz,a0 , supp(m0 ) = x0 , supp(a0 ) = I:

⌘F =

⇤

⇤,+
I

sign(a0,I )

⌘F
Fuchs vs. Vanishing Pre-Certificates
1
2 ||
m2M

m

1
2 ||

za

min

Measures:

On a grid z:

min

a2RN

y||2 + ||m||TV

+1

zi

y||2 + ||a||1
1

For m0 = mz,a0 , supp(m0 ) = x0 , supp(a0 ) = I:

⌘F =

⇤

⇤,+
I

sign(a0,I )
where

⌘F

⌘V =
x (a, b) =

⇤ +,⇤
x0

P

⌘V

sign(a0 ), 0

⇤

ai '(·, xi ) + bi '0 (·, xi )
i
Fuchs vs. Vanishing Pre-Certificates
1
2 ||
m2M

m

1
2 ||

za

min

Measures:

On a grid z:

min

a2RN

y||2 + ||m||TV

⌘F

+1

zi

y||2 + ||a||1
1

⌘V

For m0 = mz,a0 , supp(m0 ) = x0 , supp(a0 ) = I:

⌘F =

⇤

⇤,+
I

sign(a0,I )
where

Theorem: [Fuchs 2004]
If 8 j 2 I, |⌘F (xj )| < 1,
/

⌘V =
x (a, b) =

⇤ +,⇤
x0

P

sign(a0 ), 0

⇤

ai '(·, xi ) + bi '0 (·, xi )
i

then supp(a ) = supp(a0 )

(holds for ||w|| small enough and

⇠ ||w||)
Fuchs vs. Vanishing Pre-Certificates
1
2 ||
m2M

m

1
2 ||

za

min

Measures:

On a grid z:

min

a2RN

y||2 + ||m||TV

⌘F

+1

zi

y||2 + ||a||1
1

⌘V

For m0 = mz,a0 , supp(m0 ) = x0 , supp(a0 ) = I:

⌘F =

⇤

⇤,+
I

sign(a0,I )
where

Theorem: [Fuchs 2004]
If 8 j 2 I, |⌘F (xj )| < 1,
/

then supp(a ) = supp(a0 )

⌘V =
x (a, b) =

⇤ +,⇤
x0

P

sign(a0 ), 0

⇤

ai '(·, xi ) + bi '0 (·, xi )
i

Theorem: [Duval-Peyr´ 2013]
e
If 8 t 2 x0 , |⌘V (t)| < 1,
/
then m = mx ,a with
||x
x0 ||1 = O(||w||)

(holds for ||w|| small enough and

⇠ ||w||)
Numerical Illustration
0

Ideal low-pass filter: '(x, x ) =
+1

sin((2fc +1)⇡(x x0 ))
,
sin(⇡(x x0 ))

⌘F

Zoom

⌘F ⌘V
+1

1

fc = 6.

⌘V
Numerical Illustration
0

Ideal low-pass filter: '(x, x ) =
+1

sin((2fc +1)⇡(x x0 ))
,
sin(⇡(x x0 ))

⌘F

Zoom

⌘F ⌘V
+1

⌘V

1

Solution path

7! a

fc = 6.
Numerical Illustration
0

Ideal low-pass filter: '(x, x ) =
+1

sin((2fc +1)⇡(x x0 ))
,
sin(⇡(x x0 ))

⌘F

Zoom

⌘F ⌘V

fc = 6.

+1

⌘V

1

Discrete ! continuous:

Theorem: [Duval-Peyr´ 2013]
e
If ⌘V is valid, then a
is supported on pairs of
neighbors around supp(m0 ).

Solution path

7! a

(holds for

⇠ ||w|| small enough.
Conclusion
Gauges: encode linear models as singular points.
Conclusion
Gauges: encode linear models as singular points.

Performance measures

L2 error
model

di↵erent CS guarantees
Conclusion
Gauges: encode linear models as singular points.

Performance measures

L2 error
model

di↵erent CS guarantees
Specific certificate:
⌘ 0 , ⌘ F , ⌘V , . . .
Conclusion
Gauges: encode linear models as singular points.

Performance measures

L2 error
model

di↵erent CS guarantees
Specific certificate:
⌘ 0 , ⌘ F , ⌘V , . . .

Open problems:
– CS performance with arbitrary gauges.
– Approximate model recovery Tx? ⇡ Tx0 .
(e.g. grid-free recovery)

More Related Content

What's hot

Mesh Processing Course : Multiresolution
Mesh Processing Course : MultiresolutionMesh Processing Course : Multiresolution
Mesh Processing Course : MultiresolutionGabriel Peyré
 
Geodesic Method in Computer Vision and Graphics
Geodesic Method in Computer Vision and GraphicsGeodesic Method in Computer Vision and Graphics
Geodesic Method in Computer Vision and GraphicsGabriel Peyré
 
Learning Sparse Representation
Learning Sparse RepresentationLearning Sparse Representation
Learning Sparse RepresentationGabriel Peyré
 
Lecture 2: linear SVM in the dual
Lecture 2: linear SVM in the dualLecture 2: linear SVM in the dual
Lecture 2: linear SVM in the dualStéphane Canu
 
Actuarial Science Reference Sheet
Actuarial Science Reference SheetActuarial Science Reference Sheet
Actuarial Science Reference SheetDaniel Nolan
 
Mesh Processing Course : Mesh Parameterization
Mesh Processing Course : Mesh ParameterizationMesh Processing Course : Mesh Parameterization
Mesh Processing Course : Mesh ParameterizationGabriel Peyré
 
Open GL 04 linealgos
Open GL 04 linealgosOpen GL 04 linealgos
Open GL 04 linealgosRoziq Bahtiar
 
Signal Processing Course : Sparse Regularization of Inverse Problems
Signal Processing Course : Sparse Regularization of Inverse ProblemsSignal Processing Course : Sparse Regularization of Inverse Problems
Signal Processing Course : Sparse Regularization of Inverse ProblemsGabriel Peyré
 
Numerical Methods - Oridnary Differential Equations - 3
Numerical Methods - Oridnary Differential Equations - 3Numerical Methods - Oridnary Differential Equations - 3
Numerical Methods - Oridnary Differential Equations - 3Dr. Nirav Vyas
 
Interpolation with Finite differences
Interpolation with Finite differencesInterpolation with Finite differences
Interpolation with Finite differencesDr. Nirav Vyas
 
Lecture 1: linear SVM in the primal
Lecture 1: linear SVM in the primalLecture 1: linear SVM in the primal
Lecture 1: linear SVM in the primalStéphane Canu
 
Lecture3 linear svm_with_slack
Lecture3 linear svm_with_slackLecture3 linear svm_with_slack
Lecture3 linear svm_with_slackStéphane Canu
 
Finite difference method
Finite difference methodFinite difference method
Finite difference methodDivyansh Verma
 
A series of maximum entropy upper bounds of the differential entropy
A series of maximum entropy upper bounds of the differential entropyA series of maximum entropy upper bounds of the differential entropy
A series of maximum entropy upper bounds of the differential entropyFrank Nielsen
 
Classification with mixtures of curved Mahalanobis metrics
Classification with mixtures of curved Mahalanobis metricsClassification with mixtures of curved Mahalanobis metrics
Classification with mixtures of curved Mahalanobis metricsFrank Nielsen
 
Bregman divergences from comparative convexity
Bregman divergences from comparative convexityBregman divergences from comparative convexity
Bregman divergences from comparative convexityFrank Nielsen
 
Iceaa07 Foils
Iceaa07 FoilsIceaa07 Foils
Iceaa07 FoilsAntonini
 
Image Processing 3
Image Processing 3Image Processing 3
Image Processing 3jainatin
 

What's hot (20)

Mesh Processing Course : Multiresolution
Mesh Processing Course : MultiresolutionMesh Processing Course : Multiresolution
Mesh Processing Course : Multiresolution
 
Geodesic Method in Computer Vision and Graphics
Geodesic Method in Computer Vision and GraphicsGeodesic Method in Computer Vision and Graphics
Geodesic Method in Computer Vision and Graphics
 
Learning Sparse Representation
Learning Sparse RepresentationLearning Sparse Representation
Learning Sparse Representation
 
Lecture 2: linear SVM in the dual
Lecture 2: linear SVM in the dualLecture 2: linear SVM in the dual
Lecture 2: linear SVM in the dual
 
Actuarial Science Reference Sheet
Actuarial Science Reference SheetActuarial Science Reference Sheet
Actuarial Science Reference Sheet
 
Mesh Processing Course : Mesh Parameterization
Mesh Processing Course : Mesh ParameterizationMesh Processing Course : Mesh Parameterization
Mesh Processing Course : Mesh Parameterization
 
Open GL 04 linealgos
Open GL 04 linealgosOpen GL 04 linealgos
Open GL 04 linealgos
 
Signal Processing Course : Sparse Regularization of Inverse Problems
Signal Processing Course : Sparse Regularization of Inverse ProblemsSignal Processing Course : Sparse Regularization of Inverse Problems
Signal Processing Course : Sparse Regularization of Inverse Problems
 
Lecture5 kernel svm
Lecture5 kernel svmLecture5 kernel svm
Lecture5 kernel svm
 
Numerical Methods - Oridnary Differential Equations - 3
Numerical Methods - Oridnary Differential Equations - 3Numerical Methods - Oridnary Differential Equations - 3
Numerical Methods - Oridnary Differential Equations - 3
 
Interpolation with Finite differences
Interpolation with Finite differencesInterpolation with Finite differences
Interpolation with Finite differences
 
Lecture 1: linear SVM in the primal
Lecture 1: linear SVM in the primalLecture 1: linear SVM in the primal
Lecture 1: linear SVM in the primal
 
Lecture3 linear svm_with_slack
Lecture3 linear svm_with_slackLecture3 linear svm_with_slack
Lecture3 linear svm_with_slack
 
Finite difference method
Finite difference methodFinite difference method
Finite difference method
 
Image denoising
Image denoisingImage denoising
Image denoising
 
A series of maximum entropy upper bounds of the differential entropy
A series of maximum entropy upper bounds of the differential entropyA series of maximum entropy upper bounds of the differential entropy
A series of maximum entropy upper bounds of the differential entropy
 
Classification with mixtures of curved Mahalanobis metrics
Classification with mixtures of curved Mahalanobis metricsClassification with mixtures of curved Mahalanobis metrics
Classification with mixtures of curved Mahalanobis metrics
 
Bregman divergences from comparative convexity
Bregman divergences from comparative convexityBregman divergences from comparative convexity
Bregman divergences from comparative convexity
 
Iceaa07 Foils
Iceaa07 FoilsIceaa07 Foils
Iceaa07 Foils
 
Image Processing 3
Image Processing 3Image Processing 3
Image Processing 3
 

Similar to Low Complexity Regularization of Inverse Problems - Course #2 Recovery Guarantees

International journal of engineering and mathematical modelling vol2 no1_2015_1
International journal of engineering and mathematical modelling vol2 no1_2015_1International journal of engineering and mathematical modelling vol2 no1_2015_1
International journal of engineering and mathematical modelling vol2 no1_2015_1IJEMM
 
Modeling the Dynamics of SGD by Stochastic Differential Equation
Modeling the Dynamics of SGD by Stochastic Differential EquationModeling the Dynamics of SGD by Stochastic Differential Equation
Modeling the Dynamics of SGD by Stochastic Differential EquationMark Chang
 
Automatic Bayesian method for Numerical Integration
Automatic Bayesian method for Numerical Integration Automatic Bayesian method for Numerical Integration
Automatic Bayesian method for Numerical Integration Jagadeeswaran Rathinavel
 
A Szemeredi-type theorem for subsets of the unit cube
A Szemeredi-type theorem for subsets of the unit cubeA Szemeredi-type theorem for subsets of the unit cube
A Szemeredi-type theorem for subsets of the unit cubeVjekoslavKovac1
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Valentin De Bortoli
 
Density theorems for anisotropic point configurations
Density theorems for anisotropic point configurationsDensity theorems for anisotropic point configurations
Density theorems for anisotropic point configurationsVjekoslavKovac1
 
2.3 Operations that preserve convexity & 2.4 Generalized inequalities
2.3 Operations that preserve convexity & 2.4 Generalized inequalities2.3 Operations that preserve convexity & 2.4 Generalized inequalities
2.3 Operations that preserve convexity & 2.4 Generalized inequalitiesRyotaroTsukada
 
1 hofstad
1 hofstad1 hofstad
1 hofstadYandex
 
SPDE presentation 2012
SPDE presentation 2012SPDE presentation 2012
SPDE presentation 2012Zheng Mengdi
 
Integration techniques
Integration techniquesIntegration techniques
Integration techniquesKrishna Gali
 
Higher-order Factorization Machines(第5回ステアラボ人工知能セミナー)
Higher-order Factorization Machines(第5回ステアラボ人工知能セミナー)Higher-order Factorization Machines(第5回ステアラボ人工知能セミナー)
Higher-order Factorization Machines(第5回ステアラボ人工知能セミナー)STAIR Lab, Chiba Institute of Technology
 
Complex differentiation contains analytic function.pptx
Complex differentiation contains analytic function.pptxComplex differentiation contains analytic function.pptx
Complex differentiation contains analytic function.pptxjyotidighole2
 
A Szemerédi-type theorem for subsets of the unit cube
A Szemerédi-type theorem for subsets of the unit cubeA Szemerédi-type theorem for subsets of the unit cube
A Szemerédi-type theorem for subsets of the unit cubeVjekoslavKovac1
 
Algebric Functions.pdf
Algebric Functions.pdfAlgebric Functions.pdf
Algebric Functions.pdfMamadArip
 

Similar to Low Complexity Regularization of Inverse Problems - Course #2 Recovery Guarantees (20)

ch3.ppt
ch3.pptch3.ppt
ch3.ppt
 
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
 
International journal of engineering and mathematical modelling vol2 no1_2015_1
International journal of engineering and mathematical modelling vol2 no1_2015_1International journal of engineering and mathematical modelling vol2 no1_2015_1
International journal of engineering and mathematical modelling vol2 no1_2015_1
 
Modeling the Dynamics of SGD by Stochastic Differential Equation
Modeling the Dynamics of SGD by Stochastic Differential EquationModeling the Dynamics of SGD by Stochastic Differential Equation
Modeling the Dynamics of SGD by Stochastic Differential Equation
 
Automatic Bayesian method for Numerical Integration
Automatic Bayesian method for Numerical Integration Automatic Bayesian method for Numerical Integration
Automatic Bayesian method for Numerical Integration
 
A Szemeredi-type theorem for subsets of the unit cube
A Szemeredi-type theorem for subsets of the unit cubeA Szemeredi-type theorem for subsets of the unit cube
A Szemeredi-type theorem for subsets of the unit cube
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...
 
Density theorems for anisotropic point configurations
Density theorems for anisotropic point configurationsDensity theorems for anisotropic point configurations
Density theorems for anisotropic point configurations
 
2.3 Operations that preserve convexity & 2.4 Generalized inequalities
2.3 Operations that preserve convexity & 2.4 Generalized inequalities2.3 Operations that preserve convexity & 2.4 Generalized inequalities
2.3 Operations that preserve convexity & 2.4 Generalized inequalities
 
1 hofstad
1 hofstad1 hofstad
1 hofstad
 
SPDE presentation 2012
SPDE presentation 2012SPDE presentation 2012
SPDE presentation 2012
 
Integration techniques
Integration techniquesIntegration techniques
Integration techniques
 
QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...
QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...
QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...
 
Higher-order Factorization Machines(第5回ステアラボ人工知能セミナー)
Higher-order Factorization Machines(第5回ステアラボ人工知能セミナー)Higher-order Factorization Machines(第5回ステアラボ人工知能セミナー)
Higher-order Factorization Machines(第5回ステアラボ人工知能セミナー)
 
ML unit2.pptx
ML unit2.pptxML unit2.pptx
ML unit2.pptx
 
sada_pres
sada_pressada_pres
sada_pres
 
talk MCMC & SMC 2004
talk MCMC & SMC 2004talk MCMC & SMC 2004
talk MCMC & SMC 2004
 
Complex differentiation contains analytic function.pptx
Complex differentiation contains analytic function.pptxComplex differentiation contains analytic function.pptx
Complex differentiation contains analytic function.pptx
 
A Szemerédi-type theorem for subsets of the unit cube
A Szemerédi-type theorem for subsets of the unit cubeA Szemerédi-type theorem for subsets of the unit cube
A Szemerédi-type theorem for subsets of the unit cube
 
Algebric Functions.pdf
Algebric Functions.pdfAlgebric Functions.pdf
Algebric Functions.pdf
 

More from Gabriel Peyré

Adaptive Signal and Image Processing
Adaptive Signal and Image ProcessingAdaptive Signal and Image Processing
Adaptive Signal and Image ProcessingGabriel Peyré
 
Mesh Processing Course : Introduction
Mesh Processing Course : IntroductionMesh Processing Course : Introduction
Mesh Processing Course : IntroductionGabriel Peyré
 
Mesh Processing Course : Geodesics
Mesh Processing Course : GeodesicsMesh Processing Course : Geodesics
Mesh Processing Course : GeodesicsGabriel Peyré
 
Mesh Processing Course : Geodesic Sampling
Mesh Processing Course : Geodesic SamplingMesh Processing Course : Geodesic Sampling
Mesh Processing Course : Geodesic SamplingGabriel Peyré
 
Mesh Processing Course : Differential Calculus
Mesh Processing Course : Differential CalculusMesh Processing Course : Differential Calculus
Mesh Processing Course : Differential CalculusGabriel Peyré
 
Signal Processing Course : Theory for Sparse Recovery
Signal Processing Course : Theory for Sparse RecoverySignal Processing Course : Theory for Sparse Recovery
Signal Processing Course : Theory for Sparse RecoveryGabriel Peyré
 
Signal Processing Course : Presentation of the Course
Signal Processing Course : Presentation of the CourseSignal Processing Course : Presentation of the Course
Signal Processing Course : Presentation of the CourseGabriel Peyré
 
Signal Processing Course : Orthogonal Bases
Signal Processing Course : Orthogonal BasesSignal Processing Course : Orthogonal Bases
Signal Processing Course : Orthogonal BasesGabriel Peyré
 
Signal Processing Course : Fourier
Signal Processing Course : FourierSignal Processing Course : Fourier
Signal Processing Course : FourierGabriel Peyré
 
Signal Processing Course : Denoising
Signal Processing Course : DenoisingSignal Processing Course : Denoising
Signal Processing Course : DenoisingGabriel Peyré
 
Signal Processing Course : Compressed Sensing
Signal Processing Course : Compressed SensingSignal Processing Course : Compressed Sensing
Signal Processing Course : Compressed SensingGabriel Peyré
 
Signal Processing Course : Approximation
Signal Processing Course : ApproximationSignal Processing Course : Approximation
Signal Processing Course : ApproximationGabriel Peyré
 
Signal Processing Course : Wavelets
Signal Processing Course : WaveletsSignal Processing Course : Wavelets
Signal Processing Course : WaveletsGabriel Peyré
 
Sparsity and Compressed Sensing
Sparsity and Compressed SensingSparsity and Compressed Sensing
Sparsity and Compressed SensingGabriel Peyré
 
Optimal Transport in Imaging Sciences
Optimal Transport in Imaging SciencesOptimal Transport in Imaging Sciences
Optimal Transport in Imaging SciencesGabriel Peyré
 
An Introduction to Optimal Transport
An Introduction to Optimal TransportAn Introduction to Optimal Transport
An Introduction to Optimal TransportGabriel Peyré
 
A Review of Proximal Methods, with a New One
A Review of Proximal Methods, with a New OneA Review of Proximal Methods, with a New One
A Review of Proximal Methods, with a New OneGabriel Peyré
 

More from Gabriel Peyré (17)

Adaptive Signal and Image Processing
Adaptive Signal and Image ProcessingAdaptive Signal and Image Processing
Adaptive Signal and Image Processing
 
Mesh Processing Course : Introduction
Mesh Processing Course : IntroductionMesh Processing Course : Introduction
Mesh Processing Course : Introduction
 
Mesh Processing Course : Geodesics
Mesh Processing Course : GeodesicsMesh Processing Course : Geodesics
Mesh Processing Course : Geodesics
 
Mesh Processing Course : Geodesic Sampling
Mesh Processing Course : Geodesic SamplingMesh Processing Course : Geodesic Sampling
Mesh Processing Course : Geodesic Sampling
 
Mesh Processing Course : Differential Calculus
Mesh Processing Course : Differential CalculusMesh Processing Course : Differential Calculus
Mesh Processing Course : Differential Calculus
 
Signal Processing Course : Theory for Sparse Recovery
Signal Processing Course : Theory for Sparse RecoverySignal Processing Course : Theory for Sparse Recovery
Signal Processing Course : Theory for Sparse Recovery
 
Signal Processing Course : Presentation of the Course
Signal Processing Course : Presentation of the CourseSignal Processing Course : Presentation of the Course
Signal Processing Course : Presentation of the Course
 
Signal Processing Course : Orthogonal Bases
Signal Processing Course : Orthogonal BasesSignal Processing Course : Orthogonal Bases
Signal Processing Course : Orthogonal Bases
 
Signal Processing Course : Fourier
Signal Processing Course : FourierSignal Processing Course : Fourier
Signal Processing Course : Fourier
 
Signal Processing Course : Denoising
Signal Processing Course : DenoisingSignal Processing Course : Denoising
Signal Processing Course : Denoising
 
Signal Processing Course : Compressed Sensing
Signal Processing Course : Compressed SensingSignal Processing Course : Compressed Sensing
Signal Processing Course : Compressed Sensing
 
Signal Processing Course : Approximation
Signal Processing Course : ApproximationSignal Processing Course : Approximation
Signal Processing Course : Approximation
 
Signal Processing Course : Wavelets
Signal Processing Course : WaveletsSignal Processing Course : Wavelets
Signal Processing Course : Wavelets
 
Sparsity and Compressed Sensing
Sparsity and Compressed SensingSparsity and Compressed Sensing
Sparsity and Compressed Sensing
 
Optimal Transport in Imaging Sciences
Optimal Transport in Imaging SciencesOptimal Transport in Imaging Sciences
Optimal Transport in Imaging Sciences
 
An Introduction to Optimal Transport
An Introduction to Optimal TransportAn Introduction to Optimal Transport
An Introduction to Optimal Transport
 
A Review of Proximal Methods, with a New One
A Review of Proximal Methods, with a New OneA Review of Proximal Methods, with a New One
A Review of Proximal Methods, with a New One
 

Recently uploaded

How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonetsnaman860154
 
Top 5 Benefits OF Using Muvi Live Paywall For Live Streams
Top 5 Benefits OF Using Muvi Live Paywall For Live StreamsTop 5 Benefits OF Using Muvi Live Paywall For Live Streams
Top 5 Benefits OF Using Muvi Live Paywall For Live StreamsRoshan Dwivedi
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024The Digital Insurer
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfEnterprise Knowledge
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsEnterprise Knowledge
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024The Digital Insurer
 
Developing An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of BrazilDeveloping An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of BrazilV3cube
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slidevu2urc
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEarley Information Science
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountPuma Security, LLC
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Enterprise Knowledge
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Drew Madelung
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Miguel Araújo
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processorsdebabhi2
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking MenDelhi Call girls
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘RTylerCroy
 
Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Allon Mureinik
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Servicegiselly40
 

Recently uploaded (20)

How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 
Top 5 Benefits OF Using Muvi Live Paywall For Live Streams
Top 5 Benefits OF Using Muvi Live Paywall For Live StreamsTop 5 Benefits OF Using Muvi Live Paywall For Live Streams
Top 5 Benefits OF Using Muvi Live Paywall For Live Streams
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024
 
Developing An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of BrazilDeveloping An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of Brazil
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path Mount
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Service
 

Low Complexity Regularization of Inverse Problems - Course #2 Recovery Guarantees

  • 1. Low Complexity Regularization of Inverse Problems Cours #2 Recovery Guarantees Gabriel Peyré www.numerical-tours.com
  • 2. Overview of the Course • Course #1: Inverse Problems • Course #2: Recovery Guarantees • Course #3: Proximal Splitting Methods
  • 3. Overview • Low-complexity Regularization with Gauges • Performance Guarantees • Grid-free Regularization
  • 4. Inverse Problem Regularization Observations: y = x0 + w 2 RP . Estimator: x(y) depends only on observations y parameter
  • 5. Inverse Problem Regularization Observations: y = x0 + w 2 RP . Estimator: x(y) depends only on observations y parameter Example: variational methods 1 x(y) 2 argmin ||y x||2 + J(x) x2RN 2 Data fidelity Regularity
  • 6. Inverse Problem Regularization Observations: y = x0 + w 2 RP . Estimator: x(y) depends only on observations y parameter Example: variational methods 1 x(y) 2 argmin ||y x||2 + J(x) x2RN 2 Data fidelity Regularity Choice of : tradeo Noise level ||w|| Regularity of x0 J(x0 )
  • 7. Inverse Problem Regularization Observations: y = x0 + w 2 RP . Estimator: x(y) depends only on observations y parameter Example: variational methods 1 x(y) 2 argmin ||y x||2 + J(x) x2RN 2 Data fidelity Regularity Choice of : tradeo No noise: Noise level ||w|| 0+ , minimize Regularity of x0 J(x0 ) x? 2 argmin J(x) x2RQ ,Kx=y
  • 8. Inverse Problem Regularization Observations: y = x0 + w 2 RP . Estimator: x(y) depends only on observations y parameter Example: variational methods 1 x(y) 2 argmin ||y x||2 + J(x) x2RN 2 Data fidelity Regularity Choice of : tradeo No noise: Noise level ||w|| 0+ , minimize Regularity of x0 J(x0 ) x? 2 argmin J(x) x2RQ ,Kx=y This course: Performance analysis. Fast computational scheme.
  • 9. Union of Linear Models for Data Processing Union of models: T 2 T linear spaces. Synthesis sparsity: T Coe cients x Image x
  • 10. Union of Linear Models for Data Processing Union of models: T 2 T linear spaces. Synthesis sparsity: Structured sparsity: T Coe cients x Image x
  • 11. Union of Linear Models for Data Processing Union of models: T 2 T linear spaces. Synthesis sparsity: Structured sparsity: T Coe cients x Image x D Analysis sparsity: Image x Gradient D⇤ x
  • 12. Union of Linear Models for Data Processing Union of models: T 2 T linear spaces. Synthesis sparsity: Structured sparsity: T Coe cients x Analysis sparsity: Image x D Low-rank: Image x Gradient D⇤ x S1,· Multi-spectral imaging: Pr xi,· = j=1 Ai,j Sj,· S2,· x S3,·
  • 13. Gauges for Union of Linear Models Gauge: J :R N !R + Convex 8 ↵ 2 R+ , J(↵x) = ↵J(x)
  • 14. Gauges for Union of Linear Models Gauge: J :R N !R + Convex 8 ↵ 2 R+ , J(↵x) = ↵J(x) Piecewise regular ball , Union of linear models (T )T 2T x J(x) = ||x||1 T T = sparse vectors
  • 15. Gauges for Union of Linear Models Gauge: J :R N !R + Convex 8 ↵ 2 R+ , J(↵x) = ↵J(x) Piecewise regular ball , Union of linear models (T )T 2T T0 x0 x J(x) = ||x||1 T T = sparse vectors
  • 16. Gauges for Union of Linear Models Gauge: J :R N !R Convex 8 ↵ 2 R+ , J(↵x) = ↵J(x) + Piecewise regular ball , Union of linear models (T )T 2T T T0 x0 x J(x) = ||x||1 T T = sparse vectors x T0 x |x1 |+||x2,3 || T = block sparse vectors 0
  • 17. Gauges for Union of Linear Models Gauge: J :R N !R Convex 8 ↵ 2 R+ , J(↵x) = ↵J(x) + Piecewise regular ball , Union of linear models (T )T 2T T T0 x0 x J(x) = ||x||1 T T = sparse vectors x T x |x1 |+||x2,3 || T = block sparse vectors 0 x 0 J(x) = ||x||⇤ T = low-rank matrices
  • 18. Gauges for Union of Linear Models Gauge: J :R N !R Convex 8 ↵ 2 R+ , J(↵x) = ↵J(x) + Piecewise regular ball , Union of linear models (T )T 2T T T0 x0 x J(x) = ||x||1 T T = sparse vectors T0 x T x |x1 |+||x2,3 || T = block sparse vectors 0 x x x0 0 J(x) = ||x||⇤ T = low-rank matrices J(x) = ||x||1 T = antisparse vectors
  • 19. Subdifferentials and Models @J(x) = {⌘ 8 y, J(y) > J(x)+h⌘, y xi} |x|
  • 20. Subdifferentials and Models @J(x) = {⌘ 8 y, J(y) > J(x)+h⌘, y |x| xi} Example: J(x) = ||x||1 ⇢ supp(⌘) = I, @||x||1 = ⌘ 8 j 2 I, |⌘j | 6 1 / I = supp(x) = {i xi 6= 0} @J(x) 0 x
  • 21. Subdifferentials and Models @J(x) = {⌘ 8 y, J(y) > J(x)+h⌘, y |x| xi} Example: J(x) = ||x||1 ⇢ supp(⌘) = I, @||x||1 = ⌘ 8 j 2 I, |⌘j | 6 1 / I = supp(x) = {i xi 6= 0} Tx = {⌘ supp(⌘) = I} Definition: Tx = VectHull(@J(x))? @J(x) 0 x Tx
  • 22. Subdifferentials and Models @J(x) = {⌘ 8 y, J(y) > J(x)+h⌘, y |x| xi} Example: J(x) = ||x||1 ⇢ supp(⌘) = I, @||x||1 = ⌘ 8 j 2 I, |⌘j | 6 1 / @J(x) 0 I = supp(x) = {i xi 6= 0} Tx Tx = {⌘ supp(⌘) = I} ex = sign(x) Definition: Tx = VectHull(@J(x))? ⌘ 2 @J(x) ex x =) ProjTx (⌘) = ex
  • 23. Examples `1 sparsity: J(x) = ||x||1 ex = sign(x) x @J(x) x 0 Tx = {z supp(z) ⇢ supp(x)}
  • 24. Examples `1 sparsity: J(x) = ||x||1 ex = sign(x) Tx = {z supp(z) ⇢ supp(x)} P Structured sparsity: J(x) = b ||xb || N (a) = a/||a|| ex = (N (xb ))b2B Tx = {z supp(z) ⇢ supp(x)} x @J(x) x 0 x @J(x) x 0
  • 25. Examples `1 sparsity: J(x) = ||x||1 ex = sign(x) Tx = {z supp(z) ⇢ supp(x)} P Structured sparsity: J(x) = b ||xb || N (a) = a/||a|| ex = (N (xb ))b2B Tx = {z supp(z) ⇢ supp(x)} x = U ⇤V ⇤ SVD: Nuclear norm: J(x) = ||x||⇤ Tx = U A + BV ⇤ (A, B) 2 (Rn⇥n )2 ex = U V ⇤ x @J(x) x 0 x @J(x) x 0 x @J(x)
  • 26. Examples `1 sparsity: J(x) = ||x||1 ex = sign(x) Tx = {z supp(z) ⇢ supp(x)} P Structured sparsity: J(x) = b ||xb || N (a) = a/||a|| ex = (N (xb ))b2B Tx = {z supp(z) ⇢ supp(x)} x = U ⇤V ⇤ SVD: Nuclear norm: J(x) = ||x||⇤ Tx = U A + BV ⇤ (A, B) 2 (Rn⇥n )2 ex = U V ⇤ I = {i |xi | = ||x||1 } Anti-sparsity: J(x) = ||x||1 Tx = {y yI / sign(xI )} ex = |I| 1 sign(x) x @J(x) x 0 x @J(x) @J(x) x 0 x @J(x) x x 0
  • 27. Overview • Low-complexity Regularization with Gauges • Performance Guarantees • Grid-free Regularization
  • 28. Dual Certificates Noiseless recovery: min J(x) x= x0 (P0 ) x? x= x0
  • 29. Dual Certificates Noiseless recovery: min J(x) x= x0 (P0 ) Proposition: x0 solution of (P0 ) () 9 ⌘ 2 D(x0 ) Dual certificates: D(x0 ) = Im( ⇤ ⌘ x? ) @J(x0 ) @J(x0 ) x= x0
  • 30. Dual Certificates Noiseless recovery: min J(x) (P0 ) x= x0 x? Proposition: x0 solution of (P0 ) () 9 ⌘ 2 D(x0 ) Dual certificates: Proof: (P0 ) () D(x0 ) = Im( min 2ker( ) 8 (⌘, ) 2 @J(x0 ) ⇥ ker( ), ⌘ 2 Im( ⇤ ) =) x0 solution =) h , ⌘i = 0 ⇤ ⌘ @J(x0 ) x= ) @J(x0 ) J(x0 + ) J(x0 + ) > J(x0 ) + h , ⌘i =) 8 , h , ⌘i 6 0 x0 solution. =) ⌘ 2 ker( )? . x0
  • 31. Dual Certificates and L2 Stability Tight dual certificates: ¯ D(x0 ) = Im( ⇤ ) ri(@J(x0 )) ri(E) = relative interior of E = interior for the topology of a↵(E) ⌘ x? @J(x0 ) x= x0
  • 32. Dual Certificates and L2 Stability Tight dual certificates: ¯ D(x0 ) = Im( ⇤ ) ri(@J(x0 )) ri(E) = relative interior of E ⌘ x? @J(x0 ) x= x0 = interior for the topology of a↵(E) Theorem: ¯ If 9 ⌘ 2 D(x0 ), for [Fadili et al. 2013] ⇠ ||w|| one has ||x? x0 || = O(||w||)
  • 33. Dual Certificates and L2 Stability Tight dual certificates: ¯ D(x0 ) = Im( ⇤ ) ri(@J(x0 )) ri(E) = relative interior of E ⌘ x? @J(x0 ) x= x0 = interior for the topology of a↵(E) Theorem: ¯ If 9 ⌘ 2 D(x0 ), for [Fadili et al. 2013] ⇠ ||w|| one has ||x? x0 || = O(||w||) [Grassmair, Haltmeier, Scherzer 2010]: J = || · ||1 . [Grassmair 2012]: J(x? x0 ) = O(||w||).
  • 34. Dual Certificates and L2 Stability Tight dual certificates: ¯ D(x0 ) = Im( ⇤ ) ri(@J(x0 )) ri(E) = relative interior of E ⌘ x? @J(x0 ) x= x0 = interior for the topology of a↵(E) Theorem: ¯ If 9 ⌘ 2 D(x0 ), for [Fadili et al. 2013] ⇠ ||w|| one has ||x? x0 || = O(||w||) [Grassmair, Haltmeier, Scherzer 2010]: J = || · ||1 . [Grassmair 2012]: J(x? x0 ) = O(||w||). ! The constants depend on N . . .
  • 35. Compressed Sensing Setting Random matrix: 2 RP ⇥N , i,j ⇠ N (0, 1), i.i.d.
  • 36. Compressed Sensing Setting Random matrix: 2 RP ⇥N , Sparse vectors: J = || · ||1 . Theorem: Let s = ||x0 ||0 . If i,j ⇠ N (0, 1), i.i.d. [Rudelson, Vershynin 2006] [Chandrasekaran et al. 2011] P > 2s log (N/s) ¯ Then 9⌘ 2 D(x0 ) with high probability on .
  • 37. Compressed Sensing Setting Random matrix: 2 RP ⇥N , Sparse vectors: J = || · ||1 . Theorem: Let s = ||x0 ||0 . If i,j ⇠ N (0, 1), i.i.d. [Rudelson, Vershynin 2006] [Chandrasekaran et al. 2011] P > 2s log (N/s) ¯ Then 9⌘ 2 D(x0 ) with high probability on Low-rank matrices: J = || · ||⇤ . Theorem: Let r = rank(x0 ). If . [Chandrasekaran et al. 2011] x0 2 RN1 ⇥N2 P > 3r(N1 + N2 r) ¯ Then 9⌘ 2 D(x0 ) with high probability on .
  • 38. Compressed Sensing Setting Random matrix: 2 RP ⇥N , Sparse vectors: J = || · ||1 . Theorem: Let s = ||x0 ||0 . If i,j ⇠ N (0, 1), i.i.d. [Rudelson, Vershynin 2006] [Chandrasekaran et al. 2011] P > 2s log (N/s) ¯ Then 9⌘ 2 D(x0 ) with high probability on Low-rank matrices: J = || · ||⇤ . Theorem: Let r = rank(x0 ). If . [Chandrasekaran et al. 2011] x0 2 RN1 ⇥N2 P > 3r(N1 + N2 r) ¯ Then 9⌘ 2 D(x0 ) with high probability on ! Similar results for || · ||1,2 , || · ||1 . .
  • 39. Phase Transitions THE THE GEOMETRYPHASE TRANSITIONS IN CONVEX OPTIMIZATION GEOMETRY OF OF PHASE TRANSITIONS IN CONVEX OPTIMIZATION J = || · ||1 1 100 100 J = || · ||⇤ 1 900 900 75 75 P/N 50 P/N 600 600 50 300 300 25 25 0 0 0 00 25 25 50 50 75 s/N 1 75 100 100 0 00 0 0 10 10 p r/ N 1 20 20 30 30 F IGURE Phase transitions for for linear inverse problems. [left] Recovery of sparse vectors. empirical IGURE 2.2:2.2: Phase transitions linear inverse problems. [left] Recovery of sparse vectors. The The empiri probability the `1 `1 minimization problem (2.6) identifies a sparse vector 0 2 100 given random line robability thatthat the minimization problem (2.6) identifies a sparse vector x 0 2xR100Rgiven random linear From [Amelunxen et al. 20013]
  • 40. Minimal-norm Certificate ⌘ 2 D(x0 ) =) ⇢ ⌘ = ⇤q ProjT (⌘) = e T = T x0 e = ex0
  • 41. Minimal-norm Certificate ⌘ 2 D(x0 ) =) ⇢ ⌘ = ⇤q ProjT (⌘) = e Minimal-norm pre-certificate: ⌘0 = T = T x0 e = ex0 argmin ⌘= ⇤ q,⌘ T =e ||q||
  • 42. Minimal-norm Certificate ⌘ 2 D(x0 ) =) ⇢ ⌘ = ⇤q ProjT (⌘) = e Minimal-norm pre-certificate: ⌘0 = Proposition: One has ⌘0 = ( + T T = T x0 e = ex0 argmin ⌘= )⇤ e ⇤ q,⌘ T =e T ||q|| = ProjT
  • 43. Minimal-norm Certificate ⌘ 2 D(x0 ) =) ⇢ ⌘ = ⇤q ProjT (⌘) = e Minimal-norm pre-certificate: ⌘0 = Proposition: Theorem: One has ⌘0 = ( ¯ If ⌘0 2 D(x0 ) and + T T = T x0 e = ex0 argmin ⌘= )⇤ e ⇤ q,⌘ T =e T ||q|| = ProjT ⇠ ||w||, the unique solution x? of P (y) for y = x0 + w satisfies Tx ? = T x 0 and ||x? x0 || = O(||w||) [Vaiter et al. 2013]
  • 44. Minimal-norm Certificate ⌘ 2 D(x0 ) =) ⇢ ⌘ = ⇤q ProjT (⌘) = e Minimal-norm pre-certificate: ⌘0 = Proposition: Theorem: One has ⌘0 = ( ¯ If ⌘0 2 D(x0 ) and + T T = T x0 e = ex0 argmin ⌘= )⇤ e ⇤ q,⌘ T =e T ||q|| = ProjT ⇠ ||w||, the unique solution x? of P (y) for y = x0 + w satisfies Tx ? = T x 0 and ||x? x0 || = O(||w||) [Vaiter et al. 2013] [Fuchs 2004]: J = || · ||1 . [Vaiter et al. 2011]: J = ||D⇤ · ||1 . [Bach 2008]: J = || · ||1,2 and J = || · ||⇤ .
  • 45. Compressed Sensing Setting Random matrix: 2 RP ⇥N , Sparse vectors: J = || · ||1 . Theorem: Let s = ||x0 ||0 . If i,j ⇠ N (0, 1), i.i.d. [Wainwright 2009] [Dossal et al. 2011] P > 2s log(N ) ¯ Then ⌘0 2 D(x0 ) with high probability on .
  • 46. Compressed Sensing Setting Random matrix: 2 RP ⇥N , i,j Sparse vectors: J = || · ||1 . ⇠ N (0, 1), i.i.d. [Wainwright 2009] [Dossal et al. 2011] Theorem: Let s = ||x0 ||0 . If P > 2s log(N ) ¯ Then ⌘0 2 D(x0 ) with high probability on Phase transitions: L2 stability P ⇠ 2s log(N/s) vs. . Model stability P ⇠ 2s log(N )
  • 47. Compressed Sensing Setting Random matrix: 2 RP ⇥N , i,j Sparse vectors: J = || · ||1 . ⇠ N (0, 1), i.i.d. [Wainwright 2009] [Dossal et al. 2011] Theorem: Let s = ||x0 ||0 . If P > 2s log(N ) ¯ Then ⌘0 2 D(x0 ) with high probability on Phase transitions: L2 stability P ⇠ 2s log(N/s) vs. . Model stability ! Similar results for || · ||1,2 , || · ||⇤ , || · ||1 . P ⇠ 2s log(N )
  • 48. Compressed Sensing Setting Random matrix: 2 RP ⇥N , i,j Sparse vectors: J = || · ||1 . ⇠ N (0, 1), i.i.d. [Wainwright 2009] [Dossal et al. 2011] Theorem: Let s = ||x0 ||0 . If P > 2s log(N ) ¯ Then ⌘0 2 D(x0 ) with high probability on Phase transitions: L2 stability P ⇠ 2s log(N/s) vs. . Model stability ! Similar results for || · ||1,2 , || · ||⇤ , || · ||1 . P ⇠ 2s log(N ) ! Not using RIP technics (non-uniform result on x0 ).
  • 49. 1-D Sparse Spikes Deconvolution ⇥x = xi (· i) x0 i J(x) = ||x||1 Increasing : reduces correlation. reduces resolution. x0
  • 50. 1-D Sparse Spikes Deconvolution ⇥x = xi (· x0 i) i J(x) = ||x||1 Increasing : reduces correlation. reduces resolution. x0 ||⌘0,I c ||1 2 1 0 10 20 I = {j x0 (j) 6= 0} ||⌘0,I c ||1 < 1 () ¯ ⌘0 2 D(x0 ) () support recovery.
  • 51. Overview • Low-complexity Regularization with Gauges • Performance Guarantees • Grid-free Regularization
  • 52. Support Instability and Measures 1 N When N ! +1, support is not stable: ||⌘0,I c ||1 ! c > 1. N !+1 ||⌘0,I c ||1 c 1 Unstable Stable
  • 53. Support Instability and Measures 1 N When N ! +1, support is not stable: ||⌘0,I c ||1 ! c > 1. N !+1 Intuition: spikes wants to move laterally. ! Use Radon measures m 2 M(T), T = R/Z. ||⌘0,I c ||1 c 1 Unstable Stable
  • 54. Support Instability and Measures 1 N When N ! +1, support is not stable: ||⌘0,I c ||1 ! c > 1. N !+1 Intuition: spikes wants to move laterally. ! Use Radon measures m 2 M(T), T = R/Z. Extension of `1 : total variation Z ||m||TV = sup g(x) dm(x) ||g||1 61 T Discrete measure: mx,a = P i ai One has ||mx,a ||TV = ||a||1 xi . ||⌘0,I c ||1 c 1 Unstable Stable
  • 55. Sparse Measure Regularization Measurements: y = 8 < m0 2 M(T), 2 : M(T) ! L (T), (m0 ) + w where : 2 w 2 L (T).
  • 56. Sparse Measure Regularization Measurements: y = 8 < m0 2 M(T), 2 : M(T) ! L (T), (m0 ) + w where : 2 w 2 L (T). Acquisition operator: Z (m)(x) = '(x, x0 )dm(x0 ) where ' 2 C 2 (T ⇥ T) T
  • 57. Sparse Measure Regularization Measurements: y = 8 < m0 2 M(T), 2 : M(T) ! L (T), (m0 ) + w where : 2 w 2 L (T). Acquisition operator: Z (m)(x) = '(x, x0 )dm(x0 ) where ' 2 C 2 (T ⇥ T) T Total-variation over measures regularization: 1 min || (m) y||2 + ||m||TV m2M(T) 2
  • 58. Sparse Measure Regularization Measurements: y = 8 < m0 2 M(T), 2 : M(T) ! L (T), (m0 ) + w where : 2 w 2 L (T). Acquisition operator: Z (m)(x) = '(x, x0 )dm(x0 ) where ' 2 C 2 (T ⇥ T) T Total-variation over measures regularization: 1 min || (m) y||2 + ||m||TV m2M(T) 2 ! Infinite dimensional convex program. ! If dim(Im( )) < +1, dual is finite dimensional. ! If is a filtering, re-cast dual as SDP program.
  • 59. Fuchs vs. Vanishing Pre-Certificates Measures: 1 2 || m2M min m y||2 + ||m||TV +1 1
  • 60. Fuchs vs. Vanishing Pre-Certificates Measures: On a grid z: 1 2 || m2M m 1 2 || za min min a2RN y||2 + ||m||TV +1 zi y||2 + ||a||1 1
  • 61. Fuchs vs. Vanishing Pre-Certificates 1 2 || m2M m 1 2 || za min Measures: On a grid z: min a2RN y||2 + ||m||TV +1 zi y||2 + ||a||1 1 For m0 = mz,a0 , supp(m0 ) = x0 , supp(a0 ) = I: ⌘F = ⇤ ⇤,+ I sign(a0,I ) ⌘F
  • 62. Fuchs vs. Vanishing Pre-Certificates 1 2 || m2M m 1 2 || za min Measures: On a grid z: min a2RN y||2 + ||m||TV +1 zi y||2 + ||a||1 1 For m0 = mz,a0 , supp(m0 ) = x0 , supp(a0 ) = I: ⌘F = ⇤ ⇤,+ I sign(a0,I ) where ⌘F ⌘V = x (a, b) = ⇤ +,⇤ x0 P ⌘V sign(a0 ), 0 ⇤ ai '(·, xi ) + bi '0 (·, xi ) i
  • 63. Fuchs vs. Vanishing Pre-Certificates 1 2 || m2M m 1 2 || za min Measures: On a grid z: min a2RN y||2 + ||m||TV ⌘F +1 zi y||2 + ||a||1 1 ⌘V For m0 = mz,a0 , supp(m0 ) = x0 , supp(a0 ) = I: ⌘F = ⇤ ⇤,+ I sign(a0,I ) where Theorem: [Fuchs 2004] If 8 j 2 I, |⌘F (xj )| < 1, / ⌘V = x (a, b) = ⇤ +,⇤ x0 P sign(a0 ), 0 ⇤ ai '(·, xi ) + bi '0 (·, xi ) i then supp(a ) = supp(a0 ) (holds for ||w|| small enough and ⇠ ||w||)
  • 64. Fuchs vs. Vanishing Pre-Certificates 1 2 || m2M m 1 2 || za min Measures: On a grid z: min a2RN y||2 + ||m||TV ⌘F +1 zi y||2 + ||a||1 1 ⌘V For m0 = mz,a0 , supp(m0 ) = x0 , supp(a0 ) = I: ⌘F = ⇤ ⇤,+ I sign(a0,I ) where Theorem: [Fuchs 2004] If 8 j 2 I, |⌘F (xj )| < 1, / then supp(a ) = supp(a0 ) ⌘V = x (a, b) = ⇤ +,⇤ x0 P sign(a0 ), 0 ⇤ ai '(·, xi ) + bi '0 (·, xi ) i Theorem: [Duval-Peyr´ 2013] e If 8 t 2 x0 , |⌘V (t)| < 1, / then m = mx ,a with ||x x0 ||1 = O(||w||) (holds for ||w|| small enough and ⇠ ||w||)
  • 65. Numerical Illustration 0 Ideal low-pass filter: '(x, x ) = +1 sin((2fc +1)⇡(x x0 )) , sin(⇡(x x0 )) ⌘F Zoom ⌘F ⌘V +1 1 fc = 6. ⌘V
  • 66. Numerical Illustration 0 Ideal low-pass filter: '(x, x ) = +1 sin((2fc +1)⇡(x x0 )) , sin(⇡(x x0 )) ⌘F Zoom ⌘F ⌘V +1 ⌘V 1 Solution path 7! a fc = 6.
  • 67. Numerical Illustration 0 Ideal low-pass filter: '(x, x ) = +1 sin((2fc +1)⇡(x x0 )) , sin(⇡(x x0 )) ⌘F Zoom ⌘F ⌘V fc = 6. +1 ⌘V 1 Discrete ! continuous: Theorem: [Duval-Peyr´ 2013] e If ⌘V is valid, then a is supported on pairs of neighbors around supp(m0 ). Solution path 7! a (holds for ⇠ ||w|| small enough.
  • 68. Conclusion Gauges: encode linear models as singular points.
  • 69. Conclusion Gauges: encode linear models as singular points. Performance measures L2 error model di↵erent CS guarantees
  • 70. Conclusion Gauges: encode linear models as singular points. Performance measures L2 error model di↵erent CS guarantees Specific certificate: ⌘ 0 , ⌘ F , ⌘V , . . .
  • 71. Conclusion Gauges: encode linear models as singular points. Performance measures L2 error model di↵erent CS guarantees Specific certificate: ⌘ 0 , ⌘ F , ⌘V , . . . Open problems: – CS performance with arbitrary gauges. – Approximate model recovery Tx? ⇡ Tx0 . (e.g. grid-free recovery)