• Aucun résultat trouvé

On The Sample Complexity Of Sparse Dictionary Learning

N/A
N/A
Protected

Academic year: 2021

Partager "On The Sample Complexity Of Sparse Dictionary Learning"

Copied!
5
0
0

Texte intégral

(1)

HAL Id: hal-00990684

https://hal.inria.fr/hal-00990684

Submitted on 13 May 2014

HAL is a multi-disciplinary open access

archive for the deposit and dissemination of

sci-entific research documents, whether they are

pub-lished or not. The documents may come from

teaching and research institutions in France or

abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est

destinée au dépôt et à la diffusion de documents

scientifiques de niveau recherche, publiés ou non,

émanant des établissements d’enseignement et de

recherche français ou étrangers, des laboratoires

publics ou privés.

On The Sample Complexity Of Sparse Dictionary

Learning

Matthias Seibert, Martin Kleinsteuber, Rémi Gribonval, Rodolphe Jenatton,

Francis Bach

To cite this version:

Matthias Seibert, Martin Kleinsteuber, Rémi Gribonval, Rodolphe Jenatton, Francis Bach. On The

Sample Complexity Of Sparse Dictionary Learning. SSP 2014 - IEEE Workshop on Statistical Signal

Processing, Jun 2014, Jupiters, Gold Coast, Australia. �10.1109/SSP.2014.6884621�. �hal-00990684�

(2)

ON THE SAMPLE COMPLEXITY OF SPARSE DICTIONARY LEARNING

M. Seibert

1

, M. Kleinsteuber

1

, R. Gribonval

2

, R. Jenatton

3,4

, F. Bach

3,5

1

Department of Electrical Engineering and Information Technology, TU M¨unchen, Munich, Germany.

2

PANAMA Project-Team (INRIA & CNRS).

3

SIERRA Project-Team (INRIA Paris).

4

Centre de Math´ematiques Appliqu´ees - Ecole Polytechnique (CMAP).

5

Laboratoire d’informatique de l’´ecole normale sup´erieure (LIENS).

{m.seibert,kleinsteuber}@tum.de, gribonval@inria.fr, r.jenatton@criteo.com, francis.bach@ens.fr

ABSTRACT

In the synthesis model signals are represented as a sparse combinations of atoms from a dictionary. Dictionary learning describes the acquisition process of the underlying dictionary for a given set of training samples. While ideally this would be achieved by optimizing the expectation of the factors over the underlying distribution of the training data, in practice the necessary information about the distribution is not available. Therefore, in real world applications it is achieved by mini-mizing an empirical average over the available samples. The main goal of this paper is to provide a sample complexity es-timate that controls to what extent the empirical average de-viates from the cost function. This estimate then provides a suitable estimate to the accuracy of the representation of the learned dictionary. The presented approach exemplifies the general results proposed by the authors in [1] and gives more concrete bounds of the sample complexity of dictionary learn-ing. We cover a variety of sparsity measures employed in the learning procedure.

Index Terms— Dictionary learning, sample complexity, sparse coding

1. INTRODUCTION

The sparse synthesis model relies on the assumption that sig-nals x ∈ Rmcan be represented as a sparse combination of columns, or atoms, of a dictionary D ∈ Rm×d. As an equa-tion this reads as

x = Dα (1) where α ∈ Rdis the sparse coefficient vector.

The task of dictionary learning focuses on finding the best dictionary to sparsely represent a set of training samples

con-This work was partially supported Cluster of Excellence CoTeSys funded by the German DFG.

This work was partially supported by the EU FP7, SMALL project, FET-Open grant number 225913, and by the European Research Council, PLEASE project (ERC-StG-2011-277906)

catenated in the matrix X = [x1, . . . , xn]. The

correspond-ing sparse representations are stored in the coefficient matrix A = [α1, . . . , αn]. A common learning approach is to find a

solution to the minimization problem min D,A LX(D, A) (2) LX(D, A) := 2n1 kX − DAk2F+ 1 n n X i=1 g(αi). (3)

The function g in (3) serves as a measure of sparsity. Con-cretely, we consider scaled powers of the `p-norm, i.e.

g(α) := kα/λkqp (4)

for any p, q > 0 and the weighting parameter λ > 0.

In order to avoid trivial solutions, the learned dictionary D is generally forced to be an element of a constraint set D. In this paper we will focus on dictionaries with unit `2-norm

atoms, which is a commonly used constraint.

The vast amount of dictionary learning algorithms that take different approaches to the topic illustrates the popularity of the synthesis model. A probabilistic method is presented in [2]. Another famous representative is the K-SVD algorithm as proposed in [3] which is based on K-means clustering. Fi-nally, there are learning strategies that aim at learning dictio-naries with specific structures that enable fast computations, see e.g. [4, 5].

Assuming that these training samples are drawn according to some distribution, the goal of dictionary learning is to find a dictionary D?for which the expected value of the cost func-tion (3) is minimal. In practice the distribufunc-tion of the available training samples is unknown and therefore only an empirical minimizer ˆD can be obtained. The sample complexity results which we derive in this paper contribute to understand how accurately this empirical minimizer approximates D?.

We assume the training signals to be drawn according to a distribution in the ball with unit radius, i.e. the distribution is an element of the set

(3)

Our results are based on the work [1] where a more gen-eral framework of matrix factorizations has been considered. Our stringent setting here allows for more concrete bounds on the sample complexity.

Previous state of the art sample complexity results are pre-sented in [6, 7]. These results are restricted to the indicator function for `0 and `1-norms. These works also cover the

case of fast rates which we will not consider here.

2. PROBLEM STATEMENT & NOTATION Matrices are denoted by boldface capital letters such as X, vectors will be represented as boldface small letters, e.g. x. Scalars will be slanted letters like n, N . The ith entry of a vector α is denoted by αi. Finally, sets are denoted in

black-letter such as D.

The sparse representation of a given signal x according to a given dictionary can be found by solving the optimization problem arg min α∈Rd 1 2kx − Dαk 2 2+ g(α). (6)

The quality of how well a signal can be sparsely coded for a dictionary is evaluated via the function

fx(D) := inf α∈Rd 1 2kx − Dαk 2 2+ g(α). (7)

For a set of signals X the overall quality of the sparse repre-sentation is measured via

FX(D) := inf

A LX(D, A) (8)

with LXas defined in (3). This measure is equal to the mean

of the quality of all samples, i.e. FX(D) = 1nPifxi(D).

Our goal is to provide a bound for the generalization error of the empirical minimizer, i.e.

sup

D∈D

|FX(D) − Ex∼Pfx(D)| ≤ η(n, m, d, L) (9)

which depends on the number of samples n used in the learn-ing process, the size of the samples m, the number of dictio-nary atoms d, and a certain Lipschitz constant L for FX.

3. SAMPLE COMPLEXITY RESULTS

The general strategy will be to first find a Lipschitz constant for the functions FX and Efx. In combination with the

as-sumed underlying probability distribution and the structure of the dictionary this will allow us to provide an upper bound for the sample complexity.

3.1. Lipschitz Property

In this section we provide Lipschitz constants for the function FX(D). For the `p-penalty function the H¨older inequality

yields kαk1= d X i=1 |αi| ≤  d X i=1 |αi|p 1/p d X i=1 11/(1−1/p) 1−1/p = d1−1/p· kαkp

for 1 ≤ p < +∞. To cover quasi-norms with 0 < p < 1, we employ the estimate kαk1 ≤ kαkp which provides the

overall inequality

kαk1≤ d(1−1/p)+· kαkp, (10)

where the function (·)+: R → R+0 is defined as (t)+ =

max{t, 0}. Thus, we get the two estimates

kαkp≤ t ⇒ kαk1≤ d(1−1/p)+· t, (11)

kα/λkq

p≤ t ⇒ kαk1≤ λ · d(1−1/p)+· t1/q (12)

which will become of use in the following discussion. The matrix norm kAk1→2 := maxikαik2is used for the rest of

this paper while the subscript is omitted to improve readabil-ity. We also make use of the corresponding dual norm which is defined as kAk? := supU,kUk≤1hA, UiF for an

appropri-ately sized matrix U and the Frobenius inner product h·, ·iF.

For  > 0 we define the nonempty set of matrices A that are “-near solutions” as

A(X, D) := {A : αi∈ Rd, Lxi(D, αi) ≤ fxi(D) + }.

Proposition 1. The set A0is not empty and is bounded.

Proof. The function k·kq

pis non-negative and coercive. Thus,

LX(D, A) is non-negative and limk→∞LX(D, Ak) = ∞

whenever limk→∞kAkk = ∞. Therefore, the function A 7→

LX(D, A) has bounded sublevel sets. Moreover, since

pow-ers of the `p-norm are continuous, then so is A 7→ LX(D, A)

and thus attains its infimum value.

Next, note that for any D0the inequality FX(D0) − FX(D) ≤ LX(D)kD0− Dk + CX(D)kD0− Dk2 (13) holds with LX(D) := inf >0A∈Asup  1 nk(X − DA)A >k ?, (14) CX(D) := inf >0A∈Asup  1 2n n X i=1 kαik21. (15)

A detailed derivation of these parameters can be found in [1]. Proposition 2. There exist upper bounds for LX(D) and

CX(D) which are independent of the used dictionary.

Proof. For A = [α1, . . . , αn] ∈ A0we have 1

2kxi− Dαik 2

(4)

which immediately results in the estimates 0 ≤ kαi/λkqp≤ fxi(D) ≤ L(D, 0) = 1 2kxik 2 2, (16) kxi− Dαik2≤ p 2fxi(D) ≤ kxik2. (17)

Furthermore, the above in combination with (12) lets us bound the `1-norm of α via

kαk1≤ λ · d(1−1/p)+ 12kxk22

1/q

. (18) This yields the upper bound CX(D) ≤ CXwith

CX:= 2n1 n X i=1 λ · d(1−1/p)+ 1 2kxik 2 2 1/q (19)

for (15). In order to provide an upper bound for LX(D) which

is independent of the dictionary D we first note that h(X − DA)A>, Ui ≤Xkxi− Dαik2· kαik1· kUk.

By using the definition of the dual norm and the estimate de-veloped above, we obtain the upper bound LX(D) ≤ LX

with LX:= n1 n X i=1 kxik2· λ · d(1−1/p)+ 12kxik22 1/q (20)

which concludes the proof.

Proposition 2 allows us to rewrite (13) as |FX(D0) − FX(D)| kD0− Dk ≤ LX·  1 +CX LXkD 0− Dk, (21)

which implies that the function FXis uniformly locally

Lips-chitz with constant LX.

Lemma 3. The function FX is globally Lipschitz with

con-stantLX, i.e.

|FX(D0) − FX(D)| ≤ LXkD0− Dk (22)

for anyX and any D, D0∈ D.

Proof. Let  > 0 be arbitrary but fixed. For kD0− Dk ≤ LX/CXwe have

|FX(D0) − FX(D)| ≤ (1 + )LXkD0− Dk. (23)

If the bound on the distance between D and D0does not hold, we construct the sequence Di := D + ki(D0 − D), i =

0, . . . , k such that kDi+1− Dik ≤ LX/CX. This enables

us to show that the bound (23) holds for any D, D0. Note that there are no restrictions on D, D0, as the derived Lipschitz constant LX is independent of the constraint set D. Since

 > 0 is chosen arbitrarily, (22) follows.

3.2. Probability Distribution

As mentioned in the introduction we consider probability dis-tributions within the unit ball. In order to obtain meaning-ful results the distribution according to which the samples are

drawn has to fulfill two properties. First, we need to con-trol the Lipschitz constant LXfor signals drawn according to

the distribution when the number of samples n is large. This quantity will be measured by the function

Λn(L) := P (LX> L) . (24)

Furthermore, for a given D we need to control the concentra-tion of the empirical average FX(D) around its expectation.

This is measured via Γn(γ) := sup

D∈D

P (|FX(D) − Ex∼Pfx(D)| > γ) . (25)

For the considered distribution these quantities are well con-trolled, as can be seen in the following.

Proposition 4. For P ∈ P as defined in (5) we have Λn(L) =

0 for L ≥ d(1−1/p)+(1/2)1/q, and

Γn



τ /√8≤ 2 exp(−nτ2), ∀ 0 ≤ τ < +∞. (26)

Proof. The function evaluations yi = fxi(D) are

indepen-dent random variables. For samples drawn according to a dis-tribution within the unit sphere it immediately follows that they are bounded by 0 ≤ yi ≤ kxik22/2 ≤ 1/2. Using

Ho-effding’s Inequality we get

P " 1 n n X i=1 yi− E[yi] ! ≥ cτ # ≤ exp −8c22 and thus Γn(τ / √

8) ≤ 2 exp(−nτ2). Furthermore, due to

the chosen set of viable probability distributions P, we have LX ≤ λ · d(1−1/p)+(1/2)1/qand hence Λn(L) = 0 for L ≥

λ · d(1−1/p)+(1/2)1/q.

3.3. Role of the Constraint Set

In order to provide a sample complexity bound, it is necessary to take the structure of the set to which the learned dictionary is constrained into account. Of particular interest is the cov-ering number of the concerning set. For more information on covering numbers, the interested reader is referred to [8].

We will confine the discussion to the set of dictionaries with unit norm atoms, i.e. each dictionary column is an ele-ment of the sphere. It is well known that the covering number of the sphere is upper bounded by N(Sm−1) ≤ 1 + 2

m

. By using the metric k · k1→2 this is readily extended to the

product of unit spheres

N(D(m, d)) ≤ 1 + 2 md ≤ 3  md . (27) 3.4. Main Result

Theorem 5. For a given 0 ≤ t < ∞ and the Lipschitz constant L > λ · d(1−1/p)+(1/2)1/q, we have

sup

D∈D

(5)

with probability at least1 − 2e−t. The bound is defined as η(n, m, d, L) := 2 r β log n n + s β + t/√8 n (29) with the driving constant

β := md8 · maxnlog6√8L, 1o. (30) The parameter controlling the sample complexity is de-pendent on the size of the dictionary, the determined Lipschitz constant, and the number of samples. It is also dependent on the underlying distribution of the samples, which is an arbi-trary distribution in the unit ball in the examined case. Better estimates may hold for fixed probability distributions. Proof. First, note that Efx is Lipschitz with constant L >

LX for the considered case. Let , γ > 0 be given. The

constraint set D can be covered by and -net with Nelements

Dj. For a fixed dictionary D there exists an index j for which

kD − Djk ≤  and we can write

|FX(D) − Efx(D)| ≤|FX(D) − FX(Dj)|

+ sup

j

|FX(Dj) − Efx(Dj)|

+ |Efx(Dj) − Efx(D)|.

By using the concentration assumption (25) and the Lipschitz continuity of FXand Efxthis implies

sup

D

|FX(D) − Ex∼Pfx(D)| ≤ 2L + γ (31)

except for probability at most N· Γn(γ). Since the above

holds for and , γ > 0, we specify the constants  := 2L1 q β log n n , τ := √1 n q md log 3 + t

with γ := τ /√8 which fulfill the conditions 0 <  < 1 and 0 ≤ τ < ∞. Given these parameters we get

N· Γn(τ /

8) = 2e−t. (32) For the final estimate, recall that due to the definition of β the inequalities

log√6L β



≤ log6√8L≤ 8β/(md) and log n ≥ 1 hold. This allows us to provide the estimate

2L + τ /√8 ≤ 2 q β log n n + q β+t/√8 n (33)

which concludes the proof of Theorem 5.

In order to illustrate the results, we will discuss a short example.

Example 1. The general assumption is that the learned dic-tionary is an element of D(m, d) and the training samples

are drawn according to a distribution in the unit ball. Let the sparsity promoting function be defined as the `p-norm with

0 < p < 1. Then Theorem 5 holds with the sample complexity driving constant

β = md8 · log(3√8). (34) 4. CONCLUSION

Based on the general framework introduced in [1] we pro-vide a sample complexity result for learning dictionaries with unit `2-norm atoms. Powers of the `p-norm as a penalty in

the learning process while the samples are drawn according to a distribution within the unit ball. In general, we can say that the sample complexity bound η exhibits the behavior η ∝ q

log n

n with high probability. The sample complexity results

achieved in this paper recover those of [6, 7] for a different choice of sparsity measure. We suspect that the achieved re-sults can be further improved by utilizing Rademacher’s com-plexity. This approach will be discussed in future work.

5. REFERENCES

[1] R. Gribonval, R. Jenatton, F. Bach, M. Kleinsteuber, and M. Seibert, “Sample complexity of dictionary learn-ing and other matrix factorizations,” pre-print, 2014, http://arxiv.org/abs/1312.3790.

[2] B.A. Olshausen and D.J. Field, “Sparse coding with an overcomplete basis set: A strategy employed by VI?,” Vi-sion Research, vol. 37, no. 23, pp. 3311–3326, 1997. [3] M. Aharon, M. Elad, and A. Bruckstein, “K-SVD:

An algorithm for designing overcomplete dictionaries for sparse representation,” IEEE Transactions on Signal Pro-cessing, vol. 54, no. 11, pp. 4311–4322, 2006.

[4] R. Rubinstein, M. Zibulevsky, and M. Elad, “Double sparsity: Learning sparse dictionaries for sparse signal approximation,” IEEE Transactions on Signal Process-ing, vol. 58, no. 3, pp. 1553–1564, 2010.

[5] S. Hawe, M. Seibert, and M. Kleinsteuber, “Separable dictionary learning,” in IEEE Conference on Computer Vision and Pattern Recognition, 2013.

[6] A. Maurer and M. Pontil, “K-dimensional coding schemes in hilbert spaces,” IEEE Transactions on Infor-mation Theory, vol. 56, no. 11, pp. 5839–5846, 2010. [7] D. Vainsencher, S. Mannor, and A.M. Bruckstein, “The

sample complexity of dictionary learning,” Journal of Machine Learning Research, vol. 12, pp. 3259–3281, 2011.

[8] M. Anthony and P.L. Bartlett, Neural network learning: Theoretical foundations, Cambridge University Press, 2009.

Références

Documents relatifs

Gramm: causative suffix combination indicating that the subject makes or causes the object to carry out the action denoted by the verb.. Make

The algorithm for this “direct” optimization is based on a proximal gradient descent and it has been derived by casting the dictionary learning problem into a recent non-convex

Then a low-rank matrix constraint which models the clutter is applied in order to decompose the useful signal into sparse coefficients in a blind source separation framework..

In this paper, we consider a probabilistic model of sparse signals, and show that, with high probability, sparse coding admits a local minimum around the reference dictionary

The aim of this paper is to present such an application of the ParCoLab parallel corpus, a searchable trilingual 11 million token electronic database of aligned texts in

The new results, which close the above-mentioned gap between the lower bound and the upper bound, guarantee that no learning method, given the generative model of the MDP, can

the above-mentioned gap between the lower bound and the upper bound of RL, guarantee that no learning method, given a generative model of the MDP, can be significantly more

For instance, in the Dolev–Yao model it is not possible to take into account attacks based on algebraic properties of cryptographic operators, like exclusive or, in presence of