• Aucun résultat trouvé

A k-nearest neighbor approach for functional regression

N/A
N/A
Protected

Academic year: 2021

Partager "A k-nearest neighbor approach for functional regression"

Copied!
10
0
0

Texte intégral

(1)

HAL Id: hal-01292692

https://hal.archives-ouvertes.fr/hal-01292692

Submitted on 23 Mar 2016

HAL is a multi-disciplinary open access

archive for the deposit and dissemination of sci-entific research documents, whether they are pub-lished or not. The documents may come from

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de

A k-nearest neighbor approach for functional regression

Thomas Laloë

To cite this version:

Thomas Laloë. A k-nearest neighbor approach for functional regression. Statistics and Probability Letters, Elsevier, 2008, 78 (10), pp.1189-1193. �10.1016/j.spl.2007.11.014�. �hal-01292692�

(2)

A

k-NEAREST NEIGHBOR APPROACH

FOR FUNCTIONAL REGRESSION

Thomas LALOË

Institut de Mathématiques et de Modélisation de Montpellier UMR CNRS 5149, Equipe de Probabilités et Statistiques

Université Montpellier II, Cc 051

Place Eugène Bataillon, 34095 Montpellier Cedex 5, France tlaloe@math.univ-montp2.fr

Abstract

Let(X, Y ) be a random pair taking values in H × R, where H is an infinite dimensional separable Hilbert space. We establish weak con-sistency of a nearest neighbor-type estimator of the regression function of Y on X based on independent observations of the pair (X, Y ). As a general strategy, we propose to reduce the infinite dimension of H by considering only the first d coefficients of an expansion of X in an orthonormal system of H, and then to perform k-nearest neighbor regression in Rd. Both the dimension and the number of neighbors are automatically selected from the observations using a simple data-dependent splitting device.

1

Introduction

Regression is the problem of predicting a variable from some observation. An observation is usually supposed to be a collection of numerical measurements represented by a d-dimensional vector. However, in many real-life problems, input data items are in the form of random functions (speech recordings,

(3)

problem into the general class of functional data analysis. Even though in practice such observations are observed at discrete sampling points, the chal-lenge in this context is to infer the data structure by exploiting the infinite-dimensional nature of the observations. The last few years have witnessed important developments in both the theory and practice of functional data analysis, and many traditional data analysis tools have been adapted to han-dle functional inputs. The book of Ramsay and Silverman [6] provides a comprehensive introduction to the area. For an updated list of references, we refer the reader to Cérou and Guyader [3], Rossi and Villa [7], and Tuleau [8].

In the present paper, we consider the functional regression setting, where the goal is to predict a scalar response Y from some infinite-dimensional observa-tions X. More precisely, we will denote by (X, Y ) a random pair taking values in Z = H × R, where H is an infinite dimensional separable Hilbert space. Throughout the document, we will denote by ρ the (unknown) distribution of (X, Y ), and by ρX the marginal distribution of X. Based on n

indepen-dent copies (X1, Y1), . . . , (Xn, Yn) of (X, Y ), we introduce an estimator fn of

the regression function fρ(x) = E[Y |X = x] which is designed as follows:

First, we reduce the dimension of H by considering the first d coefficients of an expansion of each observation in a orthonormal system of H; Second, we perform k-nearest neighbor regression (see Györfi, Kohler, Krzyzak, and Walk [5])in Rd. We select simultaneously the dimension d and the number

of neighbors k by a data-splitting device. Our main result states weak con-sistency of the resulting estimator, thereby extending the general strategy introduced by Biau, Bunea, and Wegkamp [2] in the context of classification (ie, when Y takes its values in a finite set).

The paper is organised as follows. We start in Section 2.1 by introducing some notation. Then, in Section 2.2, we present the construction of the estimator, and state its weak consistency. Proof are collected in Section 3.

(4)

2

Consistent functional regression

2.1

Notation

We let the symbols h.|.i and k.k denote the inner product and the associated norm on H, respectively, and we let (φj)j≥1be a complete orthonormal system

of H (Akhiezer and Glazman [1]). For each observation Xi, we set Xij =

hXi|φji. We know that Xi = ∞ X j=1 Xijφj,

where the consistency holds in the L2 sense.

Introduce H(d), the finite-dimensional vector space spanned by the functions

{φ1, φ2, . . . , φd}, and let, for each Xi,

Xi(d) =

d

X

j=1

Xijφj.

Finally, denote by fρ and fρ,dthe regression functions in H and H(d),

respec-tively, and by σ2

ρ and σρ,d2 their respective L2 errors. More precisely, we have

fρ(x) = E[Y |X = x], σ2ρ =

R

Z(y − fρ(x))

2dρ(x, y), and the same in H(d)× R

for fρ,d and σρ,d2 . Throughout the document, we suppose that Y < ∞ a.s.,

and all the integrals are to be understood over ρ or ρX.

2.2

k-nearest neighbor in H

(d)

Let us first formally define our k-nearest neighbor type estimator. To this aim, we consider the sequence (X1(d), Y1), . . . , (Xn(d), Yn) where the

observa-tions have been projected onto H(d). For x in H(d), we reorder the data:

 X(1)(d)(x), Y(1)(x)  , . . . ,X(n)(d)(x), Y(n)(x)  ,

according to the increasing Euclidean distances kX(d)

i − xk of the X (d) i to x.

In other words, X(d)

(i)(x) is the i-th nearest neighbor of x amongst X (d) j . If

(5)

neighbor estimator of fρis then defined (Györfi, Kohler, Krzyzak, and Walk [5]) as fn,k,d(x) = 1 k k X i=1 Y(i)(x). (1)

To select simultaneously the dimension d and the number of neighbors k, we suggest the following data-splitting device. First we split the data into a training set {(Xi, Yi), i ∈ Iℓ} of length ℓ, and a validation set {(Xj, Yj), j ∈ Jm}

of length m, with m+ℓ = n (ℓ and m possibly function of n). For each d ≥ 1, 1 ≤ k ≤ ℓ, we construct a k-nearest neighbor estimator based on the training set. Second we use the validation set to select ˆd and ˆk as follows:

( ˆd, ˆk) ∈ arg min d≥1,1≤k≤ℓ " 1 m X j∈Jm  Yj − fℓ,k,d(Xj(d)) 2 +√λd m # . (2)

Here, the term λd/√m is a given penalty term which tends to infinity with

d to prevent overfitting.

This method, which is computationnaly simple, leads to the estimator ˆ

fn(x) := fℓ,ˆk, ˆd(x( ˆd)), (3)

which has an error

E( ˆfn) = Z Z  y − ˆfn(x) 2 dρ(x, y) = Z H  ˆf n(x) − fρ(x) 2 dρX(x) + σρ2 .

The estimator fn satisfies the following oracle inequality:

Proposition 2.1 Let M be a positive constant such that (Y − fℓ,k,d(x))2 ≤

M a.s., and suppose that

∆ := ∞ X d=1 e−2(λd/M )2 < ∞. (4)

Then there exists a constant c > 0, only depending on ∆ and M , such that, for every integer ℓ > 1/∆ and m = n − ℓ,

(6)

E Z H  ˆfn(x) − fρ(x)2 ≤ inf d≥1  (σρ,d2 − σρ2) + inf 1≤k≤ℓ  E Z H(d)  fℓ,k,d(x) − fρ,d(x) 2 +√λd m  +cr ln ℓ m . (5) The term σ2

ρ,d− σρ2 may be viewed as the price to be paid for using a finite

dimensional approximation of the observations, and it converges to zero by Lemma 3.1 below. The term inf1≤k≤ℓ E

R

H(d)(fℓ,k,d(x) − fρ,d(x))2



converges also to zero by Lemma 3.2. Since the infimum is taken over all d ≥ 1, weak convergence of ˆfn(x) to fρ(x) is ensured.

Theorem 2.1 Under the assumption (4) and

lim n→∞ℓ = ∞, limℓ→∞k = ∞, limℓ→∞ k ℓ = 0, and limn→∞ ln ℓ m = 0, ˆ

fn weakly converges to fρ, i.e.

E Z

H

 ˆfn(x) − fρ(x)2

→ 0 as n → ∞.

Pratically speaking, as discussed in Biau, Bunea, and Wegkamp [2], choosing the penalty in (2) is not an easy task. Indeed, an abusive penalisation of high dimensions can mask helpful information. For a more involved discussion about the penalty choice, and experimental results, we refer the reader to Tuleau [8], who shows that that adding a penalty term improves the stability of the selected dimension d.

3

Proof

Proof of Proposition 2.1 Let 



(d)2

(7)

and ˆ L(k, d) = 1 m X j∈Jm  Yj − fℓ,k,d(Xj(d)) 2 .

We have to minimize ˆL(k, d) + λd/m in k and d.

Fix ε > 0. For every d ≥ 1 and every k satisfying 1 ≤ k ≤ ℓ, we may write

P  L(ˆk, ˆd) − ˆL(k, d) > λd m + ε  ≤ P  L(ˆk, ˆd) − ˆL(ˆk, ˆd) > λdˆ m + ε  , since, by definition of (ˆk, ˆd), ˆ L(ˆk, ˆd) + √λdˆ m ≤ ˆL(k, d) + λd √ m . Therefore, P n L(ˆk, ˆd) − ˆL(k, d) > λd/√m + ε o ≤ ∞ P d=1 ℓ P k=1P n L(k, d) − ˆL(k, d) > λd/√m + ε o (by the union bound)

= ∞ P d=1 ℓ P k=1EP n L(k, d) − ˆL(k, d) > λd/√m + ε | (Xi, Yi), i ∈ Iℓ o ≤ ∞ P d=1 ℓ expn− 2[(λd/√m) + ε]2 × (m/M2) o (by Hoeffding’s inequality)

≤ ℓe−2mε2/M2 P∞ d=1 e−2(λd/M )2 = ∆ℓe−2mε2/M2 , where ∆ = P∞ d=1e −2(λd/M )2

. Since, for every d ≥ 1 and k with 1 ≤ k ≤ ℓ,

EL(ˆk, ˆd) ≤ E ˆL(k, d) + √λd m + Z ∞ 0 P  L(ˆk, ˆd) − ˆL(k, d) > √λd m + ε  dε , we obtain, for every u > 0,

EL(ˆk, ˆd) ≤ E ˆL(k, d) + λd m + u + ∆ℓ Z ∞ u e−2mε2/M2 dε . Note that

(8)

Z ∞ u e−2mε2/M2dε ≤ 1 2 Z ∞ u  2 + M 2 2mε  e−2mε2/M2dε = −12 M 2 2mεe −2mε2/M2 ∞ u = M 2 4mue −2mu2/M2 . Whence, choosing u = Mpln(∆ℓ)/2m, we obtain

EL(ˆk, ˆd) ≤ E ˆL(k, d) + √λd m + M r ln(∆ℓ) 2m + M 2p2m ln(∆ℓ) . Since k et d are arbitrary,

EL(ˆk, ˆd) ≤ inf d≥1,1≤k≤ℓE ˆL(k, d) + λd √ m + M r ln(∆ℓ) 2m + M 2p2m ln(∆ℓ) . The fact that EˆL(k, d) = EL(k, d) for each fixed k,d leads to the inequality (5). 

Proof of Theorem 2.1 will rely on the following lemma.

Lemma 3.1 We have

σρ,d2 − σρ2 → 0 when d → ∞ .

Proof of Lemma 3.1 :

σρ,d2 − σ2ρ = EhY − EY |X(d)i2

− EhY − EY |Xi2 = EhY − EY |Xi2+ EhEY |X − EY |X(d)i2

− EhY − EY |Xi2 = EhEY |X − EY |X(d)i2

.

(9)

E[Y|X(d)] → E[Y |X] in the L2 sense as d → ∞. 

Lemma 3.2 Assume that k → ∞ and k/ℓ → 0 as ℓ → ∞. Then, for any fixed d, E Z H(d)  fℓ,k,d(x) − fρ,d(x) 2 → 0 as ℓ → ∞

Proof of Lemma 3.2 See Györfi, Kohler, Krzyzak, and Walk [5], Theorem 6.1, page 88. 

We are now in a position to prove Theorem 2.1.

Proof of Theorem 2.1 Fix ε > 0. By Lemma 3.1, we know there exist d0 such that σρ,d− σρ2 ≤ ε for all d ≥ d0. Then, by Lemma 3.2, we have

E Z H(d0)  fℓ,k,d0(x) − fρ,d0(x) 2 → 0 as ℓ → ∞. Finally by Proposition 2.1 we have :

E Z H  ˆf n(x) − fρ(x)  ≤ inf d≥1  (σ2 ρ,d− σρ2) + inf 1≤k≤ℓ  E Z H(d)  fℓ,k,d(x) − fρ,d(x) 2 + √λd m  + cr ln ℓ m ≤ (σ2 ρ,d0 − σ 2 ρ) + inf 1≤k≤ℓ  E Z H(d0)  fℓ,k,d0(x) − fρ,d0(x) 2 + λd0 √ m + c r ln ℓ m ≤ ε + o(1), as n → ∞.

(10)

References

[1] N.I. AKHIEZER and I.M. GLAZMAN (1961). Theory of linear operators in Hilbert space, Frederick Ungar Publishing Co., New York.

[2] G.BIAU, F. BUNEA, and M.H. WEGKAMP (2005). Functional classifi-cation in Hilbert Spaces, IEEE Transactions on Information Theory, Vol. 51, pp. 2163-2172.

[3] F. CEROU and A. GUYADER (2005). Nearest neighbor classification in infinite dimension, Research Report INRIA, RR 5536.

[4] F. CUCKER and S. SMALE (2001). On the mathematical foundations of learning, bulletin (New Series) of the american mathematical society, Vol. 39, pp. 1-49.

[5] L. GYÖRFI, M. KOHLER, A. KRZYZAK, and H. WALK (2002). A distribution free theory of nonparametric regression, Springer Verlag, New York.

[6] J.O. RAMSAY and B.W. SILVERMAN (2002). Functional data analysis, Springer, New-York.

[7] F. ROSSI and N. VILLA (2006). Support Vector Machine For Functional Data Classification, Neurocomputing Vol 69, pp. 730-742.

[8] C. TULEAU (2005). Sélection de variables pour la discrimination en grande dimension, classification de données fonctionnelles, PhD thesis, University Paris XI.

Références

Documents relatifs

Unité de recherche INRIA Rhône-Alpes : 655, avenue de l’Europe - 38334 Montbonnot Saint-Ismier (France) Unité de recherche INRIA Rocquencourt : Domaine de Voluceau - Rocquencourt -

Section 3 is devoted to the term induced by the dimension- reduction method: in a general setting, we propose and study the performances (convergence and rate) of a numerically

In this paper, we show a direct equivalence between the embedding computed in spectral clustering and the mapping computed with kernel PCA, and how both are special cases of a

The main contribution of the present work is two-fold: (i) we describe a new general strategy to derive moment and exponential concentration inequalities for the LpO estimator

In this paper, we propose a simple algorithm for overcoming these problems by editing the existing training dataset, and adapting this edited set with different

Starting from an initial partition with as many clusters as objects in the dataset, the method, called EK-NNclus, iteratively applies the EK-NN rule to change the class labels, until

The main contribution of the present work is two-fold: (i) we describe a new general strategy to derive moment and exponential concentration inequalities for the LpO estimator

Index Terms — Nonparametric estimation, Regression function esti- mation, Affine invariance, Nearest neighbor methods, Mathematical statistics.. 2010 Mathematics Subject