• Aucun résultat trouvé

Real versus complex null space properties for sparse vector recovery

N/A
N/A
Protected

Academic year: 2021

Partager "Real versus complex null space properties for sparse vector recovery"

Copied!
6
0
0

Texte intégral

(1)

HAL Id: inria-00539612

https://hal.inria.fr/inria-00539612

Submitted on 6 Feb 2011

HAL is a multi-disciplinary open access archive for the deposit and dissemination of sci-entific research documents, whether they are pub-lished or not. The documents may come from teaching and research institutions in France or

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires

Real versus complex null space properties for sparse

vector recovery

Simon Foucart, Rémi Gribonval

To cite this version:

Simon Foucart, Rémi Gribonval. Real versus complex null space properties for sparse vector recovery. Comptes rendus de l’Académie des sciences. Série I, Mathématique, Elsevier, 2010, 348 (15-16), pp.863-865. �10.1016/j.crma.2010.07.024�. �inria-00539612�

(2)

Real vs. Complex Null Space Properties for Sparse Vector

Recovery

Simon Foucart, R´emi Gribonval

Abstract

We identify and solve an overlooked problem about the characterization of underdeter-mined systems of linear equations for which sparse solutions have minimal �1-norm. This

characterization is known as the null space property. When the system has real coeffi-cients, sparse solutions can be considered either as real or complex vectors, leading to two seemingly distinct null space properties. We prove that the two properties actually coincide by establishing a link with a problem about convex polygons in the real plane. Incidentally, we also show the equivalence between stable null space properties which account for the stable reconstruction by �1-minimization of vectors that are not exactly sparse.

Nous identifions et r´esolvons un probl`eme li´e aux syst`emes sous-determin´es d’´equations lin´eaires, plus pr´ecis´ement `a la propri´et´e de leurs noyaux qui caract´erise le fait que les so-lutions parcimonieuses soient celles avec la plus petite norme �1. Quand les coefficients du

syst`eme sont r´eels, les solutions parcimonieuses peuvent ˆetre consid´er´es comme vecteurs r´eels ou complexes, ce qui conduit `a deux propri´et´es des noyaux a priori distinctes. Nous d´emontrons que ces deux propri´et´es sont en fait ´equivalentes en ´etablissant un lien avec un probl`eme sur les polygones convexes du plan r´eel. Accessoirement, nous prouvons aussi l’´equivalence entre des propri´et´es stables du noyau, lesquelles expliquent la stabilit´e de la reconstruction par minimisation �1de vecteurs qui ne sont pas exactement parcimonieux.

This note deals with the recovery of sparse vectors x from incomplete measurements y = Ax, where A is an m × N matrix with m � N. The interest in developing sparse data models for solving ill-posed inverse problems originates in the possibility of such a recovery in un-derdetermined situations. This is also the fundamental result underlying the recent field of Compressed Sensing, which aims at acquiring signals/images well below the Nyquist rate by exploiting their sparsity in an appropriate domain. It is well known by now that the recovery can be carried out by solving the convex optimization problem

(P1) minimize �z�1 subject to Az = y,

This work has been supported by the French National Research Agency (ANR) through the project ECHANGE (ANR-08-EMER-006).

(3)

provided some suitable conditions are satisfied by the matrix A — actually, by its null space. To be more precise, see [2, 3] for details, every vector x supported on a set S is the unique solution of (P1) with y = Ax if and only if

(1) �uS�1 <�uS�1 for all u ∈ ker A \ {0}.

The set S designates the complementary of the set S, and the notations uS and uS stand for the vectors whose entries indexed by S and S, respectively, equal those of u, while the other entries are set to zero. We are also interested in a strengthening of Property (1), namely

(2) �uS�1≤ ρ �uS�1 for all u ∈ ker A \ {0},

for some 0 < ρ < 1. This stable null space property is actually equivalent to the property that �z − x�1 ≤ 1 + ρ1 − ρ��z�1− �x�1+ 2�xS�1

whenever Az = Ax.

Note that the latter implies that z = x if x is supported on S and if z is a solution of (P1) with y = Ax. So far, we have deliberately been ambiguous about the underlying scalar field — often, this is not alluded at all. In fact, the above-mentioned equivalences are valid in the real and complex settings alike. However, since a real-valued measurement matrix A is also a complex-valued one, Properties (1) and (2) can be interpreted in two different ways. The real versions read (3) � j∈S |uj| <� �∈S |u�| �or ≤ ρ� �∈S

|u�|� for all u ∈ kerRA\ {0}, while the complex versions read, in view of kerCA = kerRA + i kerRA,

(4) � j∈S � v2 j + wj2< � �∈S � v2 � + w2� � or ≤ ρ� �∈S � v2 � + w2� �

for all (v, w) ∈ (kerRA)2\ {(0, 0)}. We are going to show that Properties (3) and (4) are identical. Thus, every complex vector supported on S is recovered by �1-minimization if and only if every real vector supported on S is recovered by �1-minimization — informally, real and complex �1-recoveries succeed simul-taneously. Before stating the theorem, we point out that, for a real measurement matrix, one may also recover separately the real and imaginary parts of a complex vector using two real �1-minimizations — which are linear programs — rather than recovering the vector directly using one complex �1-minimization — which is a second order cone program.

Theorem 1. For a measurement matrix A ∈ Rm×N and a set S ⊆ {1, . . . , N}, the real null space property (3) is equivalent to the complex null space property (4).

(4)

Proof. It is clear that (4) implies (3). Now, in order to handle null space properties and stable null space properties at once, we introduce the shorthand ‘≺’ to mean either ‘<’ or ‘≤ ρ’. We assume that (3) holds, and we consider (v, w) ∈ (kerRA)2\ {(0, 0)}. We suppose that v and w are linearly independent, for otherwise (4) is clear. By applying (3) to u = αv + βw, we have

� j∈S

|αvj+ βwj| ≺� �∈S

|αv�+ βw�| for all λ := (α, β) ∈ R2\ {0}.

If BSand BS denote the 2 × s and 2 × (N − s) matrices with columns bj := [vj, wj]�, j ∈ S, and b� := [v�, w�]�, � ∈ S, respectively, this translates into

(5) �BS�λ�1 ≺ �BS�λ�1 for all λ ∈ R

2\ {0}, in other words

(6) |�λ, BSµ�| = |�BS�λ, µ�| ≺ �BS�λ�1 for all λ ∈ R2\ {0} and all µ ∈ Rswith �µ�∞= 1. Let us observe that (5) implies the injectivity of B�

S, hence the surjectivity of BS. Thus, for µ ∈ Rs with �µ�∞ = 1, there exists ν ∈ RN−s such that B

Sν = BSµ. As a result of (6), we have |�B�

Sλ, ν�| ≺ �BS�λ�1 for all λ ∈ R

2 \ {0}. This means that the linear functional f defined on ran B�

S by f(η) := �η, ν� has norm �f�∗1 ≺ 1. By the Hahn–Banach theorem, we extend it to a linear functional �f defined on the whole of RN−s. The latter can be represented as �f (η) = �η, �ν�. The equality � �f1 = �f�∗1 translates into ��ν�∞ ≺ 1, while the identity

f (η) = f (η)for all η ∈ ran B�

S yields 0 = �B�Sλ,ν�− ν� = �λ, BSν�− BSν� for all λ ∈ R2, so that BSν = B� Sν = BSµ. In short, for any µ ∈ Rs with �µ�∞= 1, there exists �ν ∈ RN−s with ��ν�∞ ≺ 1 such that BSν = B Sµ. Therefore, the convex polygon CS := BS[−1, 1]s is strictly contained in the convex polygon CS := BS[−1, 1]N−s, respectively is contained in ρ CS. This intuitively implies that

(7) Perimeter(CS) ≺ Perimeter(CS).

In fact, the perimeter of a convex polygon C is the unique minimum of all the perimeters of compact convex sets containing the vertices of C. This can be seen by isolating the contribution to the perimeter of each angular sector originating from a point inside C and intercepting two consecutive vertices. One can also invoke Cauchy’s formula, see e.g. [6], for the perimeter of a compact convex set K as the integral of length of the projection of K onto a line of direction θ, namely Perimeter(K) =� π 0 � max (x,y)∈K �

x cos(θ) + y sin(θ)�− min (x,y)∈K

(5)

The convex polygons CS and of CS both take the form M[−1, 1]n, which may be viewed as the Minkowski sum of the line segments [−c1, c1], . . . , [−cn, cn], where c1, . . . , cnare the columns of the matrix M — in general dimension, Minkowski sums of line segments are called zonotopes. The perimeter of such convex polygons is explicitly given by 4(�c1�2+ · · · + �cn�2), see e.g. [5]. Thus, (7) reads 4� j∈S �bj�2 ≺ 4� �∈S �b��2. Up to the factor 4, this is the complex null space property (4).

Remark. Sparse recovery can also be achieved by �q-minimization for 0 < q < 1. Its success on a set S is characterized by the null space property �uS�q < �uS�q for all u ∈ ker A \ {0}, see [3]. The same ambiguity about its real or complex interpretation arises. The question whether the two notions coincide in this case is open.

Remark. The recovery of sparse complex vectors by �1-minimization can be viewed as the special case n = 2 of the recovery of jointly sparse real vectors by mixed �1,2-minimization. In this context, see [1] for details, every n-tuple (x1, . . . , xn) of vectors in RN, each of which supported on the same set S, is the unique solution of

minimize N � j=1 � z2

1,j+ · · · + z2n,j subject to Az1= Ax1, . . . , Azn= Axn, if and only if a mixed �1,2 null space property holds, namely

(8) � j∈S � u2 1,j+ · · · + u2n,j < � �∈S � u2

1,�+ · · · + u2n,� for all (u1, . . . , un) ∈ �

kerRA�n\{(0, . . . , 0)}. It is then natural to wonder whether the real null space property (3) implies the mixed �1,2 null space property (8) when n ≥ 3. This is also an open question.

Added in proof. Since the submission of this note, the two open questions raised in the remarks have been answered in the affirmative by Lai and Liu [4].

References

[1] E. van den Berg, M. P. Friedlander, Joint-sparse recovery from multiple measurements, preprint.

(6)

[2] D. Donoho, M. Elad, Optimally sparse representation in general (nonorthogonal) dictio-naries via �1minimization. Proc. Natl. Acad. Sci. USA 100(5) (2003), 2197–2202.

[3] R. Gribonval, M. Nielsen, Highly sparse representations from dictionaries are unique and independent of the sparseness measure, Appl. Comput. Harm. Anal., 22(3) (2007), 335–355.

[4] M.-J. Lai, Y. Liu, The null space property for sparse recovery from multiple measurement vectors, preprint.

[5] S. Markov, On quasivector spaces of convex bodies and zonotopes, Numer. Algorithms, 37 (2004), 263–274.

[6] F. Spitzer, H. Widom, The circumference of a convex polygon, Proceedings of the Ameri-can Mathematical Society, 12(3) (1961), 506–509.

Références

Documents relatifs

Results improvements on both TIMIT and Librispeech speech recognition tasks are reported with the introduction of a R2H encoder in a QLSTM architecture, with input made of 40

a perfect infinite crystal where the range of interaction is limited to the n th neighbours, we can start by writting the block matrix which concerns the atomic. rows

We also show the related but easier results that if G is a path, then for any matrix M e M(G) of corank 1, the null space kerM of M yields an embedding of G in the line; furthermore,

For this purpose, the tools of classical complex analysis like analytic sets, currents, plurisubhar- monic functions must be supplemented with tools coming from ergodic theory

Need the analog of Bernoulli periodic polynomials: some locally polynomial functions on alcoves that satisfy some equations. I will call them W not to mix them with the

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des

The second part amounts to show that the restriction of P s is surjective. This is proven in two steps. The proof relies on estimates of a certain di ff erential operator D λ acting

(2) Automorphisms that have positive entropy, also called hyperbolic type au- tomorphisms, can only occur on tori, K3 surfaces, Enriques surfaces, and (nonminimal) rational surfaces,