• Aucun résultat trouvé

Latent Space Exploration Using Generative Kernel PCA

N/A
N/A
Protected

Academic year: 2022

Partager "Latent Space Exploration Using Generative Kernel PCA"

Copied!
2
0
0

Texte intégral

(1)

Latent Space Exploration Using Generative Kernel PCA

David Winant1, Joachim Schreurs1, and Johan Suykens1

KU Leuven, Department of Electrical Engineering (ESAT),

STADIUS Center for Dynamical Systems, Signal Processing and Data Analytics, Kasteelpark Arenberg 10, B-3001 Leuven, Belgium

{david.winant,joachim.schreurs,johan.suykens}@kuleuven.be

1 Introduction

Latent spaces provide a representation of data by embedding the data into an underlying vector space. Exploring these spaces allows for deeper insights in the structure of the data distribution and the relationships between data points.

The focus of this paper will be on how the synthesis of new data with generative kernel PCA [4] can help with understanding the latent features extracted from a dataset.

2 Generative Kernel PCA

Kernel PCA, as first described in [3], is a well-known feature extractor method often used for denoising and dimensionality reduction of datasets. In [5], kernel PCA was cast within the framework of Restricted Kernel Machines (RKMs) which allows for an interpretation in terms of hidden and visible units similar to Restricted Boltzmann Machines (RBMs) [1]. This connection between kernel PCA and RBMs was later used to explore a generative mechanism for the kernel PCA [4]. In practice generative kernel PCA works as follows: first kernel PCA is performed to find the hidden features of the dataset. After choosing an initial hidden unit as starting point, the values for each component of the hidden unit can be varied to explore the latent space. The corresponding newly generated data point in the input space is estimated using the kernel smoother approach.

3 Latent space exploration

The goal of this paper is to explore the latent feature space extracted by kernel PCA, in an effort to interpret the components. This has led to the development of a Matlab tool which can be used to visualise the latent space of the kernel PCA method along its principal components. Along with the newly generated point, a partial visualisation of the latent space projected onto two principal Copyright c2019 for this paper by its authors. Use permitted under Creative Com- mons License Attribution 4.0 International (CC BY 4.0).

(2)

2 D. Winant et al.

components is shown. Continuously varying the values of the components of the selected hidden unit allows for the exploration of the extracted latent space by visualising the resulting variation in the input space. In Fig. 1 an example use case for the MNIST handwritten digits dataset [2] is shown.

-0.06 -0.04 -0.02 0 0.02 0.04 0.06 0.08

-0.1 -0.08 -0.06 -0.04 -0.02 0 0.02 0.04 0.06 0.08 0.1

(a)

(b)

Fig. 1: Latent space exploration for 1000 digits 0 and 1 from the MNIST dataset using generative kernel PCA. (a) Latent space projected on the first two principal components. (b) Generated digits along the directions A and B. The generated digits in the top row allow for the interpretation of the first component as transi- tioning from 1 to 0, while the bottom row indicates that the second component smoothly rotates the digit. This explains the limited variation of the hidden units for the digits zero along this direction as they are largely invariant under rotations.

4 Conclusion

Generative kernel PCA can be used for exploring the latent space. Gradually moving in the feature space allows for the interpretation of components and consequently additional insight into the underlying latent space. For this purpose a Matlab tool has been developed which can easily be used for additional datasets as well as aid interpretation in the context of novelty detection.

References

1. Hinton, G.E.: A practical guide to training restricted Boltzmann machines. In: Neu- ral networks: Tricks of the trade, pp. 599–619. Springer (2012)

2. LeCun, Y., Bottou, L., Bengio, Y., Haffner, P., et al.: Gradient-based learning ap- plied to document recognition. Proceedings of the IEEE86(11), 2278–2324 (1998) 3. Sch¨olkopf, B., Smola, A., M¨uller, K.R.: Kernel principal component analysis. In:

International conference on artificial neural networks. pp. 583–588. Springer (1997) 4. Schreurs, J., Suykens, J.A.K.: Generative kernel pca. ESANN 2018 pp. 129–134

(2018)

5. Suykens, J.A.K.: Deep restricted kernel machines using conjugate feature duality.

Neural computation29(8), 2123–2163 (2017)

Références

Documents relatifs

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des

Keywords: Wasserstein space; Geodesic and Convex Principal Component Analysis; Fréchet mean; Functional data analysis; Geodesic space;.. Inference for family

The reason we consider this particular form is that it corresponds to Maximum Variance Unfolding (MVU), a well-known method for graph drawing proposed by Weinberger, Sha, Zhu, and

In Europe, European Space Agency (ESA) supported several preliminary studies enabling the de finition of priorities to prepare for interplanetary- manned exploration missions, such

Intermediate MAP Score Curated Gene Hit Rate Curated Chemical Hit Rate Curated Disease Hit Rate Curated Action Hit Rate Text Mined Genes Curated Genes Gene Match Explanation Text

To address this, Kulisch proposed to use an internal accumulator large enough to cover the full exponent range of floating-point.. With it, sums and dot products become

Accordingly, the heat kernel on CP N−1 is expressed as a series of normal- ized Jacobi polynomials (P n N −2,0 /P n N −2,0 (1)) n≥0 which for each n, gives the n-th reproducing

All the same conclusions reached in this section with respect to the limits of groups of type 1 and the lower bound on diameter of the resulting space forms apply