Functional linear model

https://doi.org/10.1016/S0167-7152(99)00036-XGet rights and content

Abstract

In this paper, we study a regression model in which explanatory variables are sampling points of a continuous-time process. We propose an estimator of regression by means of a Functional Principal Component Analysis analogous to the one introduced by Bosq [(1991) NATO, ASI Series, pp. 509–529] in the case of Hilbertian AR processes. Both convergence in probability and almost sure convergence of this estimator are stated.

Introduction

Classical regression models, such as generalized linear models, may be inadequate in some statistical studies: it is the case when explanatory variables are digitized points of a curve. Examples can be found in different fields of application such as chemometrics (Frank and Friedman, 1993), linguistic (Hastie et al., 1995) and many other areas (see Hastie and Mallows, 1993; Ramsay and Silverman, 1997, among others).

In this context, Frank and Friedman (1993) describe and compare different estimation procedures – Partial Least Squares, Ridge Regression and Principal Component Regression – which take into account both the number of explanatory variables (which may exceed the sample size) and the high correlation between these variables. On the other hand, several authors (see below) have developped models which allow to describe the “functional” nature of explanatory variables.

Formally, the above situation can be described through the following functional linear model. Let Y be a real random variable (r.r.v.) and X=(X(t),t∈[0,1]) be a continuous-time process defined on the same space (Ω,A,P). Assuming that E(01X2(t)dt)<∞, the dependence between X and Y is expressed asY=01ψ(t)X(t)dt+ε,where ψ is a square integrable function defined on [0,1] and ε is an r.r.v. independent of X with zero mean and variance equal to σ2.

Hastie and Mallows (1993) introduce an estimator of function ψ based on the minimization of a cubic spline criterion and Marx and Eilers (1996) use a smooth basis of B-splines and then introduce a difference penalty in the log-likelihood in order to derive a P-splines estimator of ψ.

Alternatively, model (1) can be generalized to the case where X is a random variable valued in a real separable Hilbert space H and the relation between X and Y can now be written asY=Ψ(X)+ε,where Ψ is an element of H′, and H′ is the space of R-valued continuous linear operators defined on H.

Following ideas from Bosq (1991) in the case of ARH processes, we propose in Section 2 below, an estimator of the operator Ψ. This estimator is based on the spectral analysis of the empirical second moment operator of X, which is then inverted in the space spanned by kn eigenvectors associated with the kn greatest eigenvalues. The main results are stated in Section 3, that is convergence in probability and almost sure convergence for this estimator. Computational aspects for the method are discussed in Section 4 through a simulation study. A sketch of the proofs are given in Section 5 (detailed proofs may be found in Cardot et al., 1998).

Section snippets

Definition of estimator

The inner product and norm in H are, respectively, denoted by 〈.,.〉H and ∥.∥H and the usual norm ∥.∥H in H′ is defined as∀T∈H′,∥T∥H′=sup∥x∥H=1|Tx|,and satisfies∀T∈H′,∥T∥H′=i∈N(Tei)21/2,where (ei)i∈N is an orthonormal basis in H. Assuming that the Hilbertian variable X satisfiesE[∥X∥2H]=Ω∥X(ω)∥2HdP(ω)<+∞,we define (cf. Grenander, 1963), from Riesz's Theorem, the second moment operator Γ of X byΓ(x)=E(X⊗HX(x))=E(〈X,x〉HX),∀x∈H.The operator Γ is nuclear (and therefore is an Hilbert–Schmidt

Main results

In order to state the main results of the paper, let us introduce the following condition:(H0)λ̂1>λ̂2>⋯>λ̂kn>0a.s.,which insures almost surely that Π̂knΓnΠ̂kn is regular and its eigenvectors are identifiable. Let us note j)j∈N the sequence of decreasing eigenvalues of Γ and let us defineaj=22λ1−λ2ifj=1,aj=22minj−1−λjj−λj+1)ifj≠1.

Theorem 3.1

Suppose that (H0) and the following hypotheses are satisfied:(H1)λ12>⋯>0,(H2)E∥X∥H4<+∞,(H3)limn→+∞kn4=+∞,(H3)limn→+∞kn2(j=1knaj)2=+∞.ThenΨ̂kn−Ψ∥H′n→+∞0in

A simulation study

We have simulated samples (Xi,Yi),i=1,…,n, from model (1) in which X(t) is a Brownian motion defined on [0,1], ε is normal with mean 0 and variance 0.2 var (Ψ(X)). The Hilbert space H is L2[0,1] and the eigenelements of the covariance operator of X are known to be (see Ash and Gardner, 1975):λj=1(j−0.5)2π2,Vj(t)=2sin{(j−0.5)πt},t∈[0,1],j=1,2,…In that case, assumptions (H3) (respectively (H6)) on the sequence of eigenvalues in Theorem 3.1 (respectively Theorem 3.2) are fulfilled provided that the

Proof of theorems

Let (Vj)j∈N be a sequence of orthonormal eigenvectors associated with j)j∈N and let us define in H the operator Ψkn as the “theoretical” version of Ψ̂knΨkn=ΔΠknknΓΠkn)−1,where Πkn is the orthogonal projection onto the space Hkn spanned by V1,…,Vkn. First of all, let us remark that∥Ψ−Ψ̂knH′⩽∥Ψ−ΨknH′+∥ΨknΨ̂knH′.We have for the first term on the right side of inequality (8)∥Ψ−ΨknH′2=j=1|(Ψ−Ψkn)(V′j)|2=j>kn|Ψ(V′j)|2,whereV′j=(signV̂j,VjH)Vj,j⩾1.Since ΨH′, we get∥Ψ−ΨknH′n→+∞0.

We

References (14)

There are more references available in the full text version of this article.

Cited by (0)

View full text