• Aucun résultat trouvé

Concept Lattice Generation by Singular Value Decomposition

N/A
N/A
Protected

Academic year: 2022

Partager "Concept Lattice Generation by Singular Value Decomposition"

Copied!
9
0
0

Texte intégral

(1)

Singular Value Decomposition

Petr Gajdoˇs, Pavel Moravec, and V´aclav Sn´aˇsel

Department of Computer Science, VˇSB – Technical University of Ostrava 17. listopadu 15, 708 33 Ostrava–Poruba, Czech Republic {petr.gajdos, pavel.moravec, vaclav.snasel}@vsb.cz

Singular Value Decomposition

Petr Gajdoˇs, Pavel Moravec, and V´aclav Sn´aˇsel

Department of Computer Science, VˇSB – Technical University of Ostrava 17. listopadu 15, 708 33 Ostrava–Poruba, Czech Republic {petr.gajdos, pavel.moravec, vaclav.snasel}@vsb.cz

Abstract. Latent semantic indexing (LSI) is an application of numeri- cal method called singular value decomposition (SVD), which discovers latent semantic in documents by creating concepts from existing terms.

The application area is not limited to text retrieval, many applications such as image compression are known. We propose usage of SVD as a possible data mining method and lattice size reduction tool. We offer in this paper preliminary experiments to support usability of proposed method.

Keywords:LSI, SVD, FCA, concept lattice

1 Introduction

We are studying the possible usage of k-reduced singular value decomposition (rank-k SVD) for reduction of concept lattice size. This method is well known in the area of Information Retrieval under the name latent semantic indexing (LSI), where it has been used for discovery of latent dependencies between terms (or documents). We would like to apply this approach in the area of formal concept analysis(FCA). The main goal of this paper is to provide primary results obtained with this method.

We will first mention FCA and describe the rank-kSVD decomposition and LSI method. In second chapter we will mention our proposed approach and two parameters which help us to tune it together with an example. In third chap- ter we will show preliminary experiment on small set of objects and attributes from real-life data. In conclusion, we will mention the expected outcome of our research.

1.1 Formal Concept Analysis

Definition 1. Aformal contextC:= (G, M, I)consists of two setsGandM and relationI between GandM. The elements of G are called the objects and the elements ofM are called the attributes of the context. In order to express that

c V. Sn´aˇsel, R. Bˇelohl´avek (Eds.): CLA 2004, pp. 102–110, ISBN 80-248-0597-9.

VˇSB – Technical University of Ostrava, Dept. of Computer Science, 2004.

(2)

an object g is in a relation I with an attribute m, we write gIm or(g, m)∈I and read it as “the objectg has the attribute m”. The relation I is also called the incidence relation of the context.

Definition 2. for a setA⊂Gof object we define A0 ={m∈M |gIm f or all g∈A}

(the set of attributes common to the objects inA). Correspondingly, for a setB of attributes we define

B0 ={g∈G|gIm f or all m∈B}

(the set of objects which have all attributes in B).

Definition 3. A formal concept of the context (G, M, I) is a pair (A, B)with A⊆G,B⊆M,A0 =B andB0 =A. We call Athe extent and B the intent of the concept(A, B).B(G, M, I)denotes the set of all concepts of context(G, M, I)

Definition 4. The concept latticeB(G, M, I)is a complete lattice in which in- fimum and supremum are given by:

^

t∈T

(At, Bt) =

\

t∈T

At,

[

t∈T

Bt 00

_

t∈T

(At, Bt) =

[

t∈T

At

00 ,\

t∈T

Bt

.

We refer to [4].

1.2 Singular Value Decomposition

Singular value decomposition (SVD) is well known because of its application in information retrieval –Latent semantic indexing (LSI) [1, 2]. SVD is especially suitable in its variant for sparse matrices (Lanczos [5]).

Theorem 1 (Singular value decomposition). Let A is an n×m rank-r matrix. Be σ1≥. . . ≥σr eigenvalues of a matrix √

AAT. Then there exist or- thogonal matrices U = (u1, . . . , ur)andV = (v1, . . . , vr), whose column vectors are orthonormal, and a diagonal matrixΣ =diag(σ1, . . . , σr). The decomposi- tionA=U ΣVT is calledsingular value decompositionof matrixAand numbers σ1, . . . , σr are singular values of the matrixA. Columns of U (or V)are called left (or right) singular vectorsof matrixA.

(3)

Now we have a decomposition of original matrixA. It is not needed to say, that the left and right singular vectors are not sparse. We have at mostrnonzero singular numbers, where rank r is the smaller of the two matrix dimensions.

However, we would not conserve much memory by storing the term-by-document matrix this way. Luckily, because the singular values usually fall quickly, we can take only k greatest singular values and corresponding singular vector co- ordinates and create ak-reduced singular decomposition ofA.

Definition 5. Let us havek,0< k < rand singular value decomposition of A A=U ΣVT = (UkU0)

Σk 0 0 Σ0

VkT V0T

We callAk=UkΣkVkT ak-reduced singular value decomposition (rank-kSVD).

In information retrieval, if every document is relevant to only one topic, we obtain a latent semantics – semantically related words (and documents) will have similar vectors in the reduced space. For an illustration of rank-k SVD see Figure 1, the grey areas determine firstk coordinates from singular vectors, which are being used.

Fig. 1.k-reduced singular value decomposition

Theorem 2 (Eckart-Young).Among alln×mmatricesCof rank at mostk Ak is the one, that minimises||Ak−A||2F =P

i,j

(Ai,j−Cw,j)2.

Because rank-kSVD is the best rank-kapproximation of original matrixA, any other decomposition will increase the sum of squares of matrixA−Ak.

The SVD is hard to compute and once computed, it reflects only the de- composition of original matrix. The recalculation of SVD is expensive, so it is impossible to recalculate SVD every time new rows or columns are inserted. The SVD-Updating[6] is a partial solution, but since the error slightly increases with inserted rows and columns, if the updates happen frequently, the recalculation of SVD may be needed soon or later.

Note: From now on, we will assume rank-k singular value decomposition when speaking about SVD.

(4)

2 Proposed Usage of Singular Value Decomposition

As we already mentioned, our goals are the ability to build concept lattice on binary data which may contain noise and the reduction of node count in concept lattice.

The singular value decomposition may be applied on incidence matrix of objects and attributes (features)A. Since rank-kSVD is known to remove noise by ignoring small differences between row and column vectors ofA – they will correspond to small singular values, which we drop by the choice of k– it can be used for data mining.

Because the SVD is mainly used on real numbers, we must solve a problem that the resulting decomposition will also provide real numbers, regardless to the fact that the input matrix is binary. The matricesUk andVk don’t allow us to create the reduced concept lattice directly, but we can build the reconstructed matrixAk. It will not contain binary values, so we will have to choose a threshold tdiscriminating between ones and zeros.

This done, we have two parameters of our method – the reduced dimension kand the thresholdt, where generally lowerkwill result in more unified values and less noise (but if used incorrectly it could ignore important attributes to some extent, too) and highertin less attributes of objects and thus in less nodes in concept lattice.

a1a2 a3 a4

o1 1 0 1 0 o2 1 1 0 1 o3 0 1 1 0 o4 0 1 0 1 o5 1 1 1 1 o6 0 0 1 1

a) b)

Fig. 2.Example 1: a) incidence matrix, b) concept lattice

Moreover, we can say that SVD creates equivalence classes of nodes in original lattice, merging them into one node in the modified lattice.

The value ofkin information retrieval was experimentally determined as sev- eral tens or hundreds (e.g. 50–250), exact value ofkfor our method is currently

(5)

not known and has yet to be determined, for example by observing the singular numbers, as their value should drop quickly from some point.

On the other hand we will not encounter the problem with updating the SVD, since the incidence matrix is usually small enough for fast calculation and the lattice has to be reconstructed every time the incidence matrix changes.

Example 1. To exemplify the method, suppose there are six objects with four attributes, as shown in figure 2. We calculated SVD withk3 and 2 and chosen thresholdst1= 0.1, t2= 0.6. The results are shown in figures 3 and 4.

a) b)

Fig. 3.Example 1: Generated concept lattices fork= 3 and a)t= 0.1, b)t= 0.6

As one can see, lowerkled to a higher reduction as well as higher threshold values, which was expected. In case ofk= 2 the reduction seem to be to high, leading to a concept lattice where only two groups of object and attributes exist.

The question, whether this method creates a viable reduction of concept lattice has to yet to be evaluated in the future.

You can also see, that i.e. nodes in example 1 labeled o3 and o6 correspond to one node in reduced lattice in figure 3 a).

There are other methods of lattice size reduction like constraining the con- cept lattice by attribute dependencies [3], but they require table of relationships between attributes. We are trying to construct dependencies between attributes automatically by the usage of SVD.

3 Experimental Results

We used a data set of Bacterial Taxonomy often used for clustering tests, which was given by Rataj & Schindler. Data are presented for six bacteria species, most

(6)

a) b)

Fig. 4.Example 1: Generated concept lattices fork= 2 and a)t= 0.1, b)t= 0.6

having data for more than one strain and 16 phenotypic characters1(0 – absent, 1 – present). The incidence matrix is shown in table 1.

Table 1.Original Bacteria incidence matrix

H2S MAN LYS IND ORN CIT URE ONP VPT INO LIP PHE MAL ADO ARA RHA

ecoli1 0 1 1 1 0 0 0 1 0 0 0 0 0 0 1 1

ecoli2 0 1 0 1 1 0 0 1 0 0 0 0 0 0 1 0

ecoli3 1 1 0 1 1 0 0 1 0 0 0 0 0 0 1 1

styphi1 0 1 1 0 0 0 0 0 0 0 0 0 0 0 1 0

styphi2 0 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0

styphi3 1 1 1 0 0 0 0 0 0 0 0 0 0 0 1 0

kpneu1 0 1 1 1 0 1 1 1 1 1 0 0 0 1 1 1

kpneu2 0 1 1 1 0 1 1 1 1 1 0 0 1 0 1 1

kpneu3 0 1 1 1 0 1 1 1 1 1 0 0 1 1 1 1

kpneu4 0 1 1 1 0 1 1 1 0 1 0 0 1 1 1 1

kpneu5 0 1 1 1 0 1 0 1 1 1 0 0 1 1 1 1

pvul1 1 0 0 1 0 1 1 0 0 0 0 1 0 0 0 0

pvul2 1 0 0 1 0 0 0 0 0 0 0 1 0 0 0 0

pvul3 1 0 0 1 0 0 1 0 0 0 0 1 0 0 0 0

pmor1 0 0 1 1 1 0 1 0 0 0 0 1 0 0 0 0

pmor2 0 0 0 1 1 0 0 0 0 0 0 1 0 0 0 0

smar 0 1 1 0 1 1 0 1 1 0 1 0 0 0 0 0

1 For full specie names see e.g. http://149.170.199.144/multivar/ca eg.htm#Example2

(7)

Fig. 5.Original Bacteria concept lattice

The original concept lattice is shown in figure 5, whilst reduced ones in figures 7 and 6. Some modified lattices, i.e. one with k= 9 andt= 0.5 correspond to original concept lattice.

Fig. 6.Modified Bacteria concept lattice,k= 5, t= 0.85

(8)

Fig. 7.Modified Bacteria concept lattice,k= 9, t= 0.85

One can see, that the modified lattices are quite smaller than original one, the question whether this reduction is a good one is still open.

4 Conclusion and Future Work

In our work we have shown new approach to lattice size reduction and creation from data which may contain some noise through the usage ofk-reduced singular decomposition. Our method brings some promising results and should be tested on larger data.

We have yet to determine, if the resulting lattice is a suitable one and what are the main properties of this reduction technique. The other problem is a good choice ofkandtand decision, if they are fixed values or functions of the highest and lowest values produced by SVD.

The proposed method could lead to the construction of a new technique of lattice traversal, but it has yet to be determined, whether the complexity would not be too high for its usage.

We would like to compare our method with results of Palack´y University team, led by Doc. Belohlavek, in the area of attribute dependencies [3] and equivalence relations on concept lattices.

(9)

References

1. M. Berry and M. Browne. Understanding Search Engines, Mathematical Modeling and Text Retrieval. Siam, 1999.

2. M. Berry, S. Dumais, and T. Letsche. Computation Methods for Intelligent Infor- mation Access. InProceedings of the 1995 ACM/IEEE Supercomputing Conference, 1995.

3. R. Bˇelohl´avek, V. Sklen´aˇr, and J. Zacpal. Concept Lattices Constrained by At- tribute Dependencies. InProceedings of Dateso 2004, CEUR Workshop Proceedings, volume 98, pages 56–66, 2004.

4. B. Ganter and R. Wille. Formal Concept Analysis. Springer-Verlag Berlin Heidel- berg, 1999.

5. R. M. Larsen. Lanczos bidiagonalization with partial reorthogonalization. Technical report, University of Aarhus, 1998.

6. G. W. O’Brien. Information Management Tools for Updating an SVD-Encoded Indexing Scheme. Technical Report ut-cs-94-258, The University of Tennessee, Knoxville, USA, December, 1994.

Références

Documents relatifs

The plenary keynotes and papers cover the wide spectrum of the topics related to Natural fibres and materials, Mechanical performances of bio-based building materials and

DICTIONARY LEARNING WITH SVD The algorithm used to learn a union of L orthonormal bases (ONB) is described in Table 3.. To understand the rationale behind the use of the SVD in Step

When using SPIO particles for liver imaging, it is useful to add a T1-weighted sequence to T2-weighted images, thereby providing addi- tional information for lesion

Furthermore, the TSVD is used to analyze the small-angle X-ray scattering patterns of colloidal dispersions of hematite spindles and magnetotactic bacteria in the presence of

The very high spatially loca- lized odds ratios observed suggest that stereotaxically normalized imaging data are suitable to predict whether a patient with thal- amic stroke

Results: Using simulated data, we show that PCRR and PCIG models, using chromosome-wise SVD of a core sample of individuals, are appropriate for genomic prediction in a

In this paper, we present fast algorithms for computing the full and the rank-truncated HOSVD of third-order structured (symmetric, Toeplitz and Hankel) tensors.. These algorithms

Unité de recherche INRIA Rennes, Irisa, Campus universitaire de Beaulieu, 35042 RENNES Cedex Unité de recherche INRIA Rhône-Alpes, 655, avenue de l’Europe, 38330 MONTBONNOT ST