• Aucun résultat trouvé

Approximate tensor diagonalization by invertible transforms

N/A
N/A
Protected

Academic year: 2021

Partager "Approximate tensor diagonalization by invertible transforms"

Copied!
6
0
0

Texte intégral

(1)

HAL Id: hal-00435893

https://hal.archives-ouvertes.fr/hal-00435893

Submitted on 11 Feb 2015

HAL is a multi-disciplinary open access

archive for the deposit and dissemination of

sci-entific research documents, whether they are

pub-lished or not. The documents may come from

teaching and research institutions in France or

abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est

destinée au dépôt et à la diffusion de documents

scientifiques de niveau recherche, publiés ou non,

émanant des établissements d’enseignement et de

recherche français ou étrangers, des laboratoires

publics ou privés.

Approximate tensor diagonalization by invertible

transforms

Mikael Sorensen, Pierre Comon, Sylvie Icart, Luc Deneire

To cite this version:

Mikael Sorensen, Pierre Comon, Sylvie Icart, Luc Deneire. Approximate tensor diagonalization by

invertible transforms. European Signal Processing Conference, Aug 2009, Glasgow, United Kingdom.

pp.4. �hal-00435893�

(2)

APPROXIMATE TENSOR DIAGONALIZATION BY INVERTIBLE TRANSFORMS

Mikael Sørensen, Pierre Comon, Sylvie Icart and Luc Deneire

Laboratoire I3S, CNRS/UNSA

Les Algorithmes - Euclide-B, 06903 Sophia Antipolis, France phone: +33-4 9294 2751

{sorensen, pcomon, icart, deneire}@i3s.unice.fr

ABSTRACT

Multilinear techniques are increasingly used in Signal Pro-cessing and Factor Analysis. In particular, it is often of interest to transform a tensor into another that is as diagonal as possible or to simultaneously transform a set matrices into a set of matrices that are close to diagonal. In this paper we propose a parameterization of the general linear group. Based on this parameterization Jacobi-type procedures for congru-ent diagonalization and PARAFAC decomposition problems are developed. Comparisons with an existing congruent di-agonalization algorithm is reported.

1. INTRODUCTION

In signal processing the problem of congruent diago-nalization of a set of matrices is a frequently occuring problem. For instance, it occurs in Blind Source Separa-tion (BSS) of instantaneous mixtures [10], [12], [15], in analytical constant modulus algorithms for blind sepa-ration of communication signals [14] and in frequency estimation [11]. It also appears in under-determined BSS [8], [9], e.g. via a PARallel FACtor (PARAFAC) decomposition method, such as described in [7].

The Jacobi-like sweeping procedure for tensors was introduced in [4]. We propose in this paper Jacobi-like procedures that proceed pairwise, and which yield one or several nonsingular matrices at each step of the iter-ative congruent diagonalization. In the Tensor

Approx-imate Diagonalization (TAD) problem, each nonsingular

matrix acts on every mode of the tensor, such as to min-imize the sum of squares of the off-diagonal entries. Similary, In the Tensor Approximate Fitting (TAF) prob-lem, each nonsingular matrix acts on every mode of the tensor, so as to minimize the fit. On the other hand, in the Joint Approximate Congruent Diagonalization problem (JACD) we subsequently address, a single nonsingular matrix is sought so as to minimize the sum of squares of the off-diagonal entries of all the matrices of interest. The general framework we introduce is usable in at least three different problems: first the JACD of Her-mitian matrices, should they be positive definite as in [13] or not; second, the JACD of invertible matrices of general form; third, the TAD and TAF problems.

This work has been partially supported by contract ANR-06-BLAN-0074 ”Decotes”. The work of Mikael Sørensen is supported by the EU by a Marie-Curie Fellowship (EST-SIGNAL program : http://est-signal.i3s.unice.fr) under contract No MEST-CT-2005-021175.

1.1 Notations and Definitions

Let vectors, matrices and tensors be denoted by lower case boldface, upper case boldface and upper case calligraphic letters respectively. Let ◦ and " denote the outer and Khatri-Rao product respectively and let det (·), (·)T, (·)∗, (·)H, (·)†, Re{·} and Im{·} denote the de-terminant, transpose, conjugate, conjugate-transpose, Moore-Penrose pseudoinverse, real part and imaginary part of a matrix, respectively. Furthermore, GLm(C),

SLm(C), SOm(C) and Sm(C) will denote the set of m × m

nonsingular matrices, nonsingular matrices with de-terminant equal to one, unitary matrices with determi-nant equal to one, and Hermitian matrices, respectively. Moreover, let A ∈ CI×J, then Vec (A) ∈ CIJ will denote a column vector with the property (Vec(A))i+(j−1)I=(A)ij.

Similary, let A ∈ CI×I, then Vecd(A) ∈ CIdenotes a col-umn vector with the property (Vecd (A))i=(A)ii. Finally,

let A ∈ CI×J, then Dk(A) ∈ CJ×Jdenotes the diagonal

ma-trix holding row k of A on its diagonal.

The Frobenius norm of a tensor X ∈ CI×J×Kis defined to be 'X'F= ! " i,j,k # # #Xijk # # # 2 .

A rank-1 tensor X ∈ CI×J×Kis equal to the outer prod-uct of some vectors a ∈ CI, b ∈ CJ, c ∈ CK such that (X)ijk=(a)i(b)j(c)k. The rank of a tensor X is equal to

the minimal number of rank-1 tensors needed to com-pose X. Assume that the rank of X is R, then it can be written as X = R " r=1 ar◦ br◦ cr,

where {ar} ⊂ CI, {br} ⊂ CJand {cr} ⊂ CK. This

decompos-tion is sometimes referred to as the PARAFAC decom-position. If we stack the vectors {ar}, {br} and {cr} into

the matrices A = $ a1, · · · , aR % ∈ CI×R B = $ b1, · · · , bR % ∈ CJ×R C = $ c1, · · · , cR % ∈ CK×R

and let X(k)∈ CI×J denote the matrix such that& X(k)'

ij=

(X)ijk. Then

(3)

The rest of the paper is organized as follows. The proposed parameterization of the general linear group is introduced in section 2. Then pairwise procedures for the TAF problem is discussed in section 3. Next a pairwise procedure for the JACD problems of a set of Hermitian matrices are described in section 4. Fi-nally the proposed JACD algorithm is compared to an existing one in section 5.

2. PARAMETERIZATION OF GLm(C)

Let A ∈ GLm(C), then

A = ΛA,

where A ∈ SLm(C) and Λ = m

(

det (A) Im. Next we

pro-pose a normalized QL factorization for any nonsingular matrix.

Define the QL factorization A = QL, where Q ∈ SOm(C) is unitary and L lower triangular. We obtain

the product

A = ΛQL,

where det (L) = 1. Assume that A ∈ GL2(C), then the

following parameterization is possible:

A = ) (2 det(A) 0 0 (2det (A) * ) c seiφ −seiφ c * ) a 0 b 1a * ,

where a, b ∈ C, c = cos(θ), s = sin(θ) and θ, φ ∈ R. In fact, A being nonsingular, we necessarily have that a ! 0. Based on this factorization, we propose to factorize a matrix A ∈ GLk(C) into a product of elementary matrices

that differ from the Identity only in four places. This idea has been extensively used for parameterizing uni-tary matrices with the so-called Givens rotations [5]. Here we extend the idea to general invertible matrices of GLk(C). Consider the elementary matrix

A[p, q] = Λ[p, q]Q[p, q]L[p, q], where A[p, q]mn=                    1 if m = n and m " {p, q} App if m = n = p Aqq if m = n = q Apq if m = p and n = q Aqp if m = q and n = p 0 otherwise and Λ[p, q]mn =              1 if m = n and m " {p, q} k ( det/A[p, q]0 if m = n = p k ( det/A[p, q]0 if m = n = q 0 otherwise Q[p, q]mn =                    1 if m = n and m " {p, q} cos(θ) if m = n = p cos(θ) if m = n = q sin(θ)eiφ if m = p and n = q − sin(θ)e−iφ if m = q and n = p 0 otherwise L[p, q]mn =                1 if m = n and m " {p, q} a if m = n = p 1 a if m = n = q b if m = q and n = p and m> n 0 otherwise

It can be shown that any A ∈ GLm(C) can be

fac-torized as the product of the above three kinds of ele-mentary matrices, viz, diagonal, lower triangular and unitary or orthogonal, e.g. we can setφ = 0 in Q[p, q]. Hence we propose to factorize a matrix A ∈ GLm(C) into

a product of the above parameterization Givens-type matrices A = S 1 s=1 m−1 1 p=1 m 1 q=p+1 A[p, q, s],

where S is equal to number of sweeps executed by the iterative procedure and the third index indicate the sweep number.

3. TENSOR APPROXIMATE FITTING Given a tensor T ∈ CI×J×K with rank R satisfying the inequalities R ≤ min(IJ, K) and R(R−1) ≤ I(I −1)J(J −1)/2. Then under mild conditions it was shown in [7] that PARAFAC decomposition problem can be reformulated as a simultaneous matrix diagonalization problem of a set of R complex symmetric matrices of dimension R×R. Moreover, given a tensor T ∈ CI×J×K with rank R ≤ min(I, J, K) then by a dimension reduction step as explained in [6] the PARAFAC decomposition problem can be reformulated to the estimation of the parame-ters of a tensor X ∈ CR×R×K. This is typically done by

minimizing the function

f (U, V, W) = K " k=1 'X(k)− UDk(W) VT'2F = K " k=1

'Vec&X(k)'−(V " U) Vecd(Dk(W))'2F.

(1) The optimal Dk(W) can be expressed as follows:

Dk(W) = diag

&

(V " U)†Vec&X(k)'', ∀k. (2) Inserting (2) into (1) we obtain

fW(U, V) =

K

"

k=1

(4)

By assuming that X(k)= UDk(W) VTfor all k and U, V ∈

GLR(C), then obviously Dk(W) = diag

&

U−1X(k)VT'. (3) Inserting (3) into (1) we obtain the cost function

g (U, V) = K " k=1 'X(k)− Udiag&U−1X(k)VT'VT'2F(4) = g (UΛU, VΛV),

where ΛU and ΛVare arbitrary nonsingular diagonal

matrices. A first approach to reduce the number of unknowns would be to use (2). A simpler approach would be to make use of (4).

Under the assumption that Dk(W) is nonsingular

for some k, say k = 1, then as proposed in [6] another simplification can be considered which will result in a different optimization problem. This will yield a joint diagonalization problem by a congruence transform

Y(k) = X(k)X(1)−1= UDk(W)D1(W)−1U−1. (5)

By the same argument as before we can from (5) obtain the following cost function

h (U) = K " k=1 'Y(k)− Udiag&U−1Y(k)U'U−1'2F (6) = h (UΛU),

where ΛUis an arbitrary nonsingular diagonal matrix.

Due to the invariance property of the cost functions (4) and (6) the term Λ[p, q] vanish and in the lower trian-gular matrices we can set a = 1. Furthermore, since

h (UΠ) = h (U), where Π is an arbitrary permutation

ma-trix the parameterization A[p, q] could be reduced to

A[p, q]mn =            1 if m = n a if m = p and n = q b if m = q and n = q 0 otherwise where 1 − ab ! 0 and a, b ∈ C.

Remark that minimizing fW(U, V), g (U, V) and h (U)

will in general yield different solutions, unless the ten-sor considered is exactly diagonalizable by invertible transforms. An outline of a general sweeping proce-dure addressing this kind of problems can be seen on algorithm 1.

4. JOINT APPROXIMATE CONGRUENT DIAGONALIZATION OF HERMITIAN MATRICES

Various BSS problems are often addressed by min-imizing the off-diagonal terms of a set of Hermitian matrices. Let {T(k)}K

k=1⊂ Sm(C) be the matrices we

at-tempt to diagonalize. Then the minimization of the cost function

J (U) = K

"

k=1

'UT(k)UH− diag&UT(k)UH''2F (7)

Algorithm 1Outline of sweeping procedure for TAF. Initialize: U = Im, V = Im

Step 1: Repeat until convergence forp = 1 to m − 1 do

forq = p + 1 to m do

calculate optimal AU[p, q] and AV[p, q]

U ← U AU$p, q% , V ← VAV$p, q%

end for end for

Step 2: Check if algorithm has converged. If not, then go to Step 1.

Algorithm 2Outline of sweeping procedure for mini-mization of the off-term elements of a set of Hermitian matrices.

Initialize: U = Im

Step 1: Repeat until convergence forp = 1 to m − 1 do

forq = p + 1 to m do

calculate Q[p, q] according to subsection 4.1. T(k)← Q$p, q%HT(k)Q$p, q%

U ← Q$p, q%HU

calculate L[p, q] according to subsection 4.2. T(k)← L$p, q% T(k)L$p, q%H

U ← L$p, q% U end for

end for

Step 2: Check if algorithm has converged. If not, then go to Step 1.

have been applied in [10], [15]. This can be interpreted as an extension of the JADE algorithm [2] to the case of general invertible transforms.

The updating rules are T(k) ← A$p, q% T(k)A$p, q%H and U ← A$p, q% U. Let us parameterize A$p, q% ∈ Cm×m

as A$p, q% = (m

det/A[p, q]0Q $p, q% L $p, q%. Then since

J (αU) = α2α∗2J (U), ∀α ∈ C, (8) we can infer from equation (8) that the scalar

m (

det/A[p, q]0 does not contribute to any diagonaliza-tion of the matrices {T(k)}. Hence we will only consider the term A[p, q] = Q$p, q% L $p, q% ∈ SLm(C). Notice that

by this variant of the proposed parameterizarion Q$p, q% must be a unitary matrix.

Due to the computational complexity of the prob-lem, we calculate and update only one matrix at a time: either the unitary matrix or the triangular one. An out-line of the proposed sweeping procedure can be seen on algorithm 2.

4.1 Algebraic Solution for the Unitary Matrix The algebraic solution for the unitary Jacobi-subproblem was solved in [2], [3] known as the JADE problem and therefore the JADE algorithm can be used to solve the problem. Let U = Q[p, q], then due to the in-variance property of the Frobenius norm wrt. any uni-tary operator, the uniuni-tary Jacobi-subproblem is equiva-lent to maximizing the quadratic form

(5)

f/Q[p, q]0 = K " k=1 'diag&Q[p, q]HT(k)Q[p, q]''2F= sT1 2        2γ1 Re{γ2} Im{γ2}

Re{γ2} γ5+ 2Re{γ4} 2Im{γ4}

Im{γ2} 2Im{γ4} γ5− 2Re{γ4}

       s, where

s = 8cos (2θ) , sin (2θ) cos&φ', sin (2θ)sin&φ'9T γ1 = K " k=1 # # # #T (k) pp # # # # 2 + # # # #T (k) qq # # # # 2 γ2 = K " k=1 T(k)∗qq T(k)pq+ T(k)∗qp T(k)qq− T(k)∗pp T(k)pq− T(k)∗qp T(k)pp γ3 = K " k=1 2Re{T(k)∗pp T(k)qq} + # # # #T (k) pq # # # # 2 + # # # #T (k) qp # # # # 2 γ4 = K " k=1 T(k)∗qp T(k)pq γ5 = γ1+γ3

Hence the problem amounts to find an eigenvector as-sociated to the largest eigenvalue of the above matrix. 4.2 Algebraic Solution for the Lower Triangular Ma-trix

Let U = L[p, q] and α = ab, then

J/L[p, q]0 = β2αα∗+β1α + β∗1α∗+β0 ! f (α, α∗), where β2 = K " k=1 2 # # # #T (k) pp # # # # 2 β1 = K " k=1 & T(k)ppT(k)∗qp + T(k)ppT(k)pq ' β0 = K " k=1 2 # # # #T (k) qp # # # # 2

Let us apply the formal partial derivatives [1] , then the stationary points of f (α, α∗) can be found to be

∂ f (α, α) ∂α∗ = β2α + β ∗ 1= 0 ⇔α = − β∗1 β2,

which exists whenβ2!0. Furthermore, when it exists

it is a global minimum since ∂2f (α, α)

∂α∗∂α = β2> 0.

When a solution exists the pair a = 1 and b = −β ∗ 1

β2 will be used. On the other hand, when no solution exists then we make use of a = 1 and b = 0.

5. COMPUTER RESULTS

The comparison of the proposed algorithm, here called SL, and the ALS method by Li [10], called FAJD, will be on the test data {Tk}5k=1⊂ Sm(C), where Tk= UDkUH+

βEk, U ∈ Cm×m, Ek∈ Sm(C) and Ek∈ Sm(C) ,β ∈ R and Dk

are diagonal matrices with (Dk)ii∈ C.

In order to avoid degenerate solutions Li proposed to add the term log (det (U)) to the functional J (U). This additional term will penalize ill-conditioned solutions consisting of small singular values. A measure on how well-conditioned a given solution is, is given by the ratio

κ(U) = σmax(U) σmin(U),

where σmax(U) and σmin(U) denotes the maximal and

minimal singular values of U respectively.

In the SL method an iteration is equal to one sweep while in the FAJD method an iteration corresponds to an update of all the parameters of the transform matrix.

Let α = :5 k=1'UDkUH'2F :5 k=1'βEk'2F .

Then the mean and median convergence of the algo-rithms over 50 trials after 100 iterations whenα varies from 0 to 0.9 with a hop size 0.1 can be seen on figure 1 and 2 respectively. Furthermore the mean and me-dian condition numbers of the solutions can be seen on figures 3 and 4 respectively.

6. SUMMARY

A new parameterization of the general linear group has been proposed. Based on this parameterization sweep-ing algorithms for PARAFAC estimation and congruent diagonalization of a set of matrices have been proposed. Comparison with an existing congurent diagonaliza-tion algorithm shows promising results.

10−1 100 0 0.05 0.1 0.15 0.2

α

J

FAJD SL

Figure 1: The mean diagonalization of a set of indefi-nite matrices on the form Tk= UDkUH+βEkwhenβ is

(6)

10−1 100 0 0.05 0.1 0.15 0.2

α

J

FAJD SL

Figure 2: The median diagonalization of a set of indef-inite matrices on the form Tk= UDkUH+βEkwhenβ is

varying. 10−1 100 1 1.5 2 2.5 3

α

κ

FAJD SL

Figure 3: The mean condition number of the diagonal-ization matrix U whenα is varying.

10−1 100 1 1.5 2 2.5 3

α

κ

FAJD SL

Figure 4: The median condition number of the diago-nalization matrix U whenα is varying.

REFERENCES

[1] D.H. Brandwood , “A Complex Gradient Operator and its Applications in Adaptive Antenna Theory,”

IEE Proceedings-F and H, vol. 130, pp. 11–16 . 1983.

[2] J.-F. Cardoso and A. Souloumiac , “Blind Beamform-ing for non-Gaussian signals,” IEE ProceedBeamform-ings-F, vol. 140, pp. 362–370 . 1993.

[3] J.-F. Cardoso and A. Souloumiac , “Jacobi angles for simultaneous diagonalization,” SIAM J. Matrix

Anal. Appl., vol. 17, pp.161–164 . 1996.

[4] P. Comon, Independent Component Analysis, in Higher Order Statistics, pp. 29–38. Elsevier, 1992. [5] G. H. Golub and C. F. Van Loan, Matrix computations.

The John Hopkins University Press, 1989.

[6] L. De Lathauwer and B. De Moor and J. Vander-walle , “Computation of the Canonical Decomposi-tion by means of a Simultaneous Generalized Schur Decomposition,” SIAM J. Matrix Anal. Appl., vol. 26, pp. 295–327. 2004.

[7] L. De Lathauwer , “A Link between the Canonical Decomposition in multilinear algebra and Simul-taneous Matrix Diagonalization,” SIAM J. Matrix

Anal. Appl., vol. 28, pp. 642–666. 2006.

[8] L. De Lathauwer and J. Castaing and J.-F. Car-doso, “Fourth-Order Cumulant-Based Blind Identi-fication of Underdetermined Mixtures,” IEEE Trans.

Signal Processing, vol. 55, pp. 2965–2973. 2007.

[9] L. De Lathauwer and J. Castaing, “Blind Identifica-tion of Underdetermined Mixtures by Simultaneous Matrix Diagonalization,” IEEE Trans. Signal

Process-ing, vol. 56, pp. 1096–1105. 2008.

[10] X.-L. Li and X.-D. Zhang, “Nonorthogonal joint diagonalization Free of Degenerate Solution ,” IEEE

Trans. Signal Processing, vol. 55, pp. 1803–1814. 2007.

[11] O. J. Micka and A. J. Weiss, “Estimating Frequen-cies of Exponentials in Noise Using Joint Diagonal-ization,” IEEE Trans. Signal Processing, vol. 47, pp. 341–348. 1999.

[12] D.-T. Pham and J.-F. Cardoso, “Blind Separation of Instantaneous Mixtures of non Stationary Sources,”

IEEE Trans. Signal Processing, vol. 49, pp. 1837–1848.

2001.

[13] D.-T. Pham, “Joint Approximate Diagonalization of Positive Definite Hermitian Matrices,” SIAM J.

Matrix Anal. Appl., vol. 22, pp. 1136–1152. 2001.

[14] A.-J. Van Der Veen and A. Paulraj, “An Analytical Constant Modulus Algorithm,” IEEE Trans. Signal

Processing, vol. 44, pp. 1136–1155. 1996.

[15] A. Ziehe and P. Laskov and G. Nolte and K.-L. M ¨uller, “A Fast Algorithm for Joint Diagonaliza-tion with Non-orthogonal TransformaDiagonaliza-tions and its Application to Blind Source Separation,” Journal of

Figure

Figure 1: The mean diagonalization of a set of indefi- indefi-nite matrices on the form T k = UD k U H + βE k when β is varying.
Figure 2: The median diagonalization of a set of indef- indef-inite matrices on the form T k = UD k U H + βE k when β is varying

Références

Documents relatifs

Abstract—In this paper, we propose a successive convex approximation framework for sparse optimization where the nonsmooth regularization function in the objective function is

PRICIPAL AXIS FACTORING ( ﻙـﺤﻤ ﻡﺩﺨﺘـﺴﺍﻭ ﺜﻜﺃ ﻥﻤﺎﻜﻟﺍ ﺭﺫﺠﻟﺍ ﺔـﻘﻴﺭﻁﺒ ﺩـﻤﺎﻌﺘﻤ ﺭﻴﻭﺩـﺘ ﻙـﻟﺫ ﺏـﻘﻋﺃﻭ ،ﺎﻬﺠﺍﺭﺨﺘـﺴﺍ ﻡـﺘ ﻲـﺘﻟﺍ لـﻤﺍﻭﻌﻠﻟ ﺢﻴﺤـﺼ ﺩﺤﺍﻭ ﻥﻤ ﺭ ﺱﻜﺎﻤﺭﻓ .) VARIMAX (

A degassed toluene suspension of each generation dendrimer, 6, 8, and 10 with free amino groups anchored on solid support (Fmoc group was first removed from 6, 8, and 10 by

Automatic speech recognition (ASR) performance in difficult reverberant and noisy conditions has improved tremendously in the last decade [1–5]. This can be attributed to advances

/ La version de cette publication peut être l’une des suivantes : la version prépublication de l’auteur, la version acceptée du manuscrit ou la version de l’éditeur.. For

Using micro/mesoporous zeolites with well-defined mesoporosity, we show that adsorption at a given pressure can be described as a simple linear combination of the adsorbed

Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of

Lane 3, UGT2B-like enzyme purified from 3 mg of granule matrix extract; lane 4, total deglycosylation of the purified UGT2B-like enzyme from 3 mg of granule matrix extract; lane