The inferences for semiparametric models with functional data are investigated. We propose an integral least-squares technique for estimating the parametric components, and the asymptotic normality of the resulting integral least-squares estimator is studied. For the nonparametric components, a local integral least-squares estimation method is proposed, and the asymptotic normality of the resulting estimator is also established. Based on these results, the confidence intervals for the parametric component and the nonparametric component are constructed. At last, some simulation studies and a real data analysis are undertaken to assess the finite sample performance of the proposed estimation method.
1. Introduction
In the recent literature, there has been increased interest in regression modeling for functional data, where both the predictor and response are random functions. Compared with the discrete multivariate analysis, functional data analysis can take into account the smoothness of the high dimensional covariates and can suggest some new approaches to the problems that have not been discovered before. Examples of functional data can be found in different application fields such as biomedicine, economics, and archaeology (see Ramsay and Silverman [1]). Furthermore, the statistical analysis for the regression model with functional data also has been considered by many authors. For example, Ramsay and Silverman [2] studied the linear regression model with functional data. Ait-Saïdi et al. [3] proposed a cross-validated estimation procedure for the single-functional index model. Ferraty et al. [4] and Chen et al. [5] considered the inferences for single and multiple index functional regression models by using the functional projection pursuit regression technology. In addition, Ferraty and Vieu [6] and Rachdi and Vieu [7] considered the nonparametric regression modeling for functional data. More works for the functional data analysis can be found in [8–10] and among others.
However, the linear functional model, which assumes that the model satisfies the linear relationship between the response and the covariates, may be too restrictive. Then, the semiparametric model with functional data is a useful extension of functional linear regression models and functional nonparametric regression models. More specifically, let X(t), Z(t), and Y(t) be continuous random functions of index t; then, the semiparametric regression model with functional data has the following structure:
(1)Y(t)=X(t)Tθ(t)+Z(t)Tβ+ϵ(t),
where Y(t) is the response variable, X(t) is the p×1 covariate vector, Z(t) is the q×1 covariate vector, β=(β1,…,βq)T is a vector of unknown parameters, θ(t)=(θ1(t),…,θp(t))T is a vector of unknown function of t, and ϵ(t) is a zero-mean stochastic process. Here, without loss of generality, we assume that index t ranges over a nondegenerate compact interval such as [0,1].
Because the samples of response and covariate are functions of index t, the Euclidean distance cannot measure the distance between Y(t) and X(t). Hence, the ordinary least-squares method cannot be implemented directly. Recently, the studies for such semiparametric regression model with functional data have been considered by many papers (see Aneiros-Pérez and Vieu [11], Shin [12], and Lian [13]). In this paper, we provide additional positive results of the inferences for semiparametric models with functional data and extend the application literature of the classical least-squares technology. More specifically, we propose an integral least-squares method for estimating the parametric components and the nonparametric component. Furthermore, the asymptotic normalities of the integral profile least-squares estimators are studied. Some simulation studies and a real data application imply that the proposed method is workable.
The rest of this paper is organized as follows. In Section 2, we introduce the integral least-squares based estimation procedure for the parametric components and the nonparametric components. The asymptotic distributions of these estimators are also derived under some regularity conditions. In Section 3, some simulations and a real data analysis are carried out to assess the performance of the proposed estimation method. The technical proofs of all asymptotic results are provided in the Appendix.
2. Estimation and Asymptotic Distributions
Suppose {Xi(t),Zi(t),Yi(t)}i=1n is a random sample of size n. From (1), we have that
(2)Yi(t)=XiT(t)θ(t)+Zi(t)Tβ+ϵi(t),i=1,…,n.
In this paper, we assume that, for given t, (Yi(t),Xi(t),Zi(t)) and ϵi(t) are i.i.d. for different i and E{εi(t)∣Xi(t),Zi(t)}=0.
Now, we in turn model (2); for t in a small neighborhood of t0, θk(t) can be locally approximated by a linear function
(3)θk(t)=θk(t0)+θk′(t0)(t-t0)≡ak+bk(t-t0),hhhhhhhhhhhhhhhhhhhhhhhhhhhhhk=1,…,p.
For given β, applying the integral least-squares method, we can get the weighted local integral least-squares estimator of {θ1(t),…,θp(t)} by minimizing
(4)∑i=1n∫01{Yi(s)-ZiT(s)β-∑k=1p[ak+bk(s-t)]Xik(s)}2hhh×Kh(t-s)ds,
where Kh(·)=h-1K(·/h), K(·) is a kernel function, h is a bandwidth, and Xik(s) denote the kth component of Xi(s). Let(5)Dxy(t)=(∑i=1n∫01Xi(s)Yi(s)Kh(t-s)ds∑i=1n∫01Xi(s)Yi(s)t-shKh(t-s)ds),Dxz(t)=(∑i=1n∫01Xi(s)ZiT(s)Kh(t-s)ds∑i=1n∫01Xi(s)ZiT(s)t-shKh(t-s)ds),Dxx(t)=(∑i=1n∫01Xi(s)XiT(s)Kh(t-s)ds∑i=1n∫01Xi(s)XiT(s)t-shKh(t-s)ds∑i=1n∫01Xi(s)XiT(s)t-shKh(t-s)ds∑i=1n∫01Xi(s)XiT(s)(t-sh)2Kh(t-s)ds).Then, the solution to (4) is given by
(6)θ~(t)=(Ip,0p)Dxx(t)-1[Dxy(t)-Dxz(t)β],
where Ip is p×p identity matrix and 0p is p×p zero matrix. Substituting (6) into (2), and by a simple calculation, we have
(7)Y˘i(t)=Z˘iT(t)β+ϵi(t),
where
(8)Z˘i(t)=Zi(t)-μ^(t)TXi(t),Y˘i(t)=Yi(t)-XiT(t)g^(t),(9)μ^(t)=(Ip,0p)Dxx(t)-1Dxz(t),g^(t)=(Ip,0p)Dxx(t)-1Dxy(t).
Applying the integral least-squares technology to linear model (7), we can get the integral least-squares estimator of β, say β^, by minimizing
(10)∑i=1n∫01{Y˘i(t)-Z˘iT(t)β}2dt.
Let
(11)Γ^=1n∑i=1n∫01Z˘i(t)Z˘i(t)Tdt.
If the matrix Γ^ is invertible, β^ can be given by
(12)β^=Γ^-11n∑i=1n∫01Z˘i(t)Y˘i(t)dt.
Let
(13)Γ=E{∫01[Z(t)-μT(t)X(t)]⊗2dt},(14)B=E(∫01{[Z(t)-μ(t)TX(t)]ϵ(t)}dN(t))⊗2,
where A⊗2=AAT, μ(t)=Ψ-1(t)Φ(t), Φ(t)=E{X(t)ZT(t)∣t}, and Ψ(t)=E{X(t)XT(t)∣t}. The following result states the asymptotic normality of β^.
Theorem 1.
Suppose that conditions (C1)–(C5) in the Appendix hold; then, one has
(15)n(β^-β)⟶LN(0,Σ),
where Σ=Γ-1BΓ-1.
In order to construct the confidence interval of β by Theorem 1, we give the estimator of Σ, say Σ^=Γ^-1B^(β^)Γ^-1, where Γ^ is defined in (11) and
(16)B^(β^)=1n∑i=1nη^i(β^)η^iT(β^).
Invoking ∥β^-β∥=Op(n-1/2), with the similar argument to Lemma A.6, we can prove that Σ^ is a consistent estimator of Σ. Thus, by Theorem 1, we have
(17)Σ^-1/2n(β^-β)⟶LN(0,Iq),
where Iq is an identity matrix of order q. Therefore, the confidence region of β can be constructed by using (17).
Furthermore, substituting β^ into (6), we can get the integral least-squares estimator of θ(t) as
(18)θ^(t)=(Ip,0p)Dxx(t)-1[Dxy(t)-Dxz(t)β^].
We state the asymptotic normality of θ^(t) in the following theorem.
Theorem 2.
Suppose that conditions (C1)–(C5) in the Appendix hold. For given t0, then one has
(19)nh{θ^(t0)-θ(t0)-b(t0)}⟶LN(0,V(t0)),
where b(t0)=(1/2)θ′′(t0)h2∫01t2K(t)dt, V(t0)=Ψ-1(t0)σ2(t0)∫01K2(t)dt.
Let
(20)Ψ^(t0)=1nh∑i=1n∫01Xi(t)Xi(t)TKh(t-t0)dt,b^(t0)=1nh∑i=1n∫01[θ^(t)-θ^(t0)]Kh(t-t0)dt,σ^2(t0)=1nh∑i=1n∫01[Yi(t)-Zi(t)Tβ^-Xi(t)Tθ^(t)]2hhhhhhhh×Kh(t-t0)dt.
Using the law of large numbers, with the similar argument to Lemma A.2 and Proposition 4.1 of Xue and Zhu [14] and Lemma 1 of Wu et al. [15], it can be shown that σ^2(t0), Ψ^(t0), and b^(t0) are consistent estimators of σ2(t0), Ψ(t0), and b(t0), respectively. Finally, we can obtain the estimator V^(t0) by substituting Ψ-1(t0) and σ2(t0) in V(t0) with Ψ^-1(t0) and σ^2(t0), respectively. It can be shown that b^(t0) and V^(t0) are the consistent estimators of b(t0) and V(t0), respectively. By Theorem 2, we have that
(21)V^(t0)-1/2nh{θ^(t0)-θ(t0)-b^(t0)}⟶LN(0,Ip),
where Ip is the unit matrix of order p.
Using (21), a pointwise confidence interval for θr(t0) can be given by
(22)θ^r(t0)-(nh)-1/2b^r(t0)±Zα/2(nh)-1/2ν^r(t0),hhhhhhhhhhhhhhhhhhhhhhhhhhr=1,…,p,
where b^r(t0) is the rth component of b^(t0), ν^r(t0) is the (r,r)th element of V^(t0), and Zα/2 is the 1-α/2 quantile value of the standard normal distribution.
3. Numerical Results
In this section, we conduct several simulation experiments to illustrate the finite sample performances of the proposed method and consider a real data set analysis for further illustration.
3.1. Simulation Studies
To evaluate the performance of the proposed method, we consider the following model:
(23)Y(t)=sin(2πt)X(t)+2.5Z(t)+ϵ(t),
where β=2.5 and θ(t)=sin(2πt). To perform the simulation, we generated n=30,50,100 samples, respectively. The covariates X(t) and Z(t) are generated according to the model
(24)X(t)=X01(0.5-t)2+X02,Z(t)=Z01(0.5-t)3+Z02,
where X01~U(5,7), X02~N(0,1.5), Z01~N(0,5), and Z02~N(0,1), respectively.
We use the Epanechnikov kernel function K(t)=0.75(1-t2)+ and use the cross-validation method to determine bandwidth h. Let θ^-i(·) and β^-i be the integral least-squares estimators of θ(·) and β, respectively, which are computed with all of the measurements but not the ith observation. Define the integral least-squares cross-validation function
(25)CV(h)=1n∑i=1n∫01{Yi(t)-Xi(t)Tθ^-i(t)-Zi(t)Tβ^-i}2.
The cross-validation bandwidth is the one that minimizes (25); that is,
(26)h^CV=argminhCV(h).
For the parametric component β, the average and standard deviation of the estimator β^, based on 1000 simulations, are reported in Table 1. In addition, the average length and coverage probability of the confidence interval (β^L,β^U), with a nominal level 1-α=95%, are computed with 1000 simulation runs. The results are also summarized in Table 1.
The estimators and the 95% confidence intervals for β.
n
β^
SD
(βL,βU)
Length
Coverage
30
2.472
0.166
(2.326, 2.618)
0.292
0.937
50
2.491
0.136
(2.389, 2.594)
0.205
0.945
100
2.502
0.096
(2.441, 2.562)
0.121
0.951
For the nonparametric component θ(t), the average pointwise confidence intervals, based on 1000 simulations, with a nominal level 1-α=95% are presented in Figure 1, and the corresponding coverage probabilities are presented in Figure 2.
The pointwise confidence intervals for θ(t).
The corresponding coverage probabilities.
Table 1 shows that, for the parametric component, our method can give a shorter confidence interval and the corresponding coverage probability is close to real nominal level. Figures 1 and 2 show that the average interval length decreases as the sample size increases, while the corresponding coverage probability increases. In addition, we can see that, for the nonparametric component, the proposed estimation method works well except for boundary points.
3.2. Application to Spectrometric Curves Data
In this section, we present an application of the proposed estimation method to spectrometric curves data. This original data comes from a quality control problem in the food industry. This data set concerns a sample of finely chopped meat, and each food sample contains finely chopped pure meat with different fat, protein, and moisture (water) contents. The sample size of this data set is n=240, and, for each food sample, the functional data consist of 100 channel spectrum of absorbances, which were recorded on the Tecator Infratec Food and Feed Analyzer working in the wavelength range 850–1050 nm by the near infrared transmission (NIT) principle. Because of the fineness of the grid, we can consider each subject as a continuous curve. Thus, each spectrometric analysis can be summarized by some continuous curves giving the observed absorbance as function of the wavelength. More details of the data can be found in Ferraty and Vieu [16].
The aim is to find the relationship between the percentage of fat content Y(t) and the corresponding percentages of protein content X1(t), the moisture content X2(t), and the spectrometric curve t. The results, obtained by Aneiros-Pérez and Vieu [11], indicate that there is a strong linear relationship between the fat content and the protein and moisture contents, but the spectrometric curve t has a functional effect on the fat content. Hence, we consider the following semiparametric model:
(27)Y(t)=θ(t)+β1X1(t)+β2X2(t)+ε.
We computed the estimators of the parametric components β1 and β2 and the nonparametric component θ(·) by using the proposed integral least-squares method. The results for the parametric components are reported in Table 2, and the results for the nonparametric components are reported in Figure 3, where the solid curve is the estimator of θ(·) and the dashed curve is the pointwise confidence interval of θ(·). From Table 2, we can see that there is a significant negative correlation relationship between the fat content and the protein and moisture contents. In addition, Figure 3 indicates that the baseline function θ(·) really varies over the spectrometric curve t. This finding basically agrees with what was discovered by Aneiros-Pérez and Vieu [11].
The estimators and the 95% confidence intervals for β.
β
Estimator
Confidence interval
Interval length
β1
-1.758
(-1.907,-1.609)
0.298
β2
-0.661
(-0.743,-0.579)
0.164
The estimator and 95% pointwise confidence interval for θ(t).
AppendixProof of Theorems
For convenience and simplicity, let c denote a positive constant which may be different value at each appearance throughout this paper. Before we prove our main theorems, we list some regularity conditions which are used in this paper.
The bandwidth satisfies h=cn-1/5, for some constant c>0.
The kernel K(·) is a symmetric probability density function, and ∫t4K(t)dt<∞.
θ(t), Φ(t), and Ψ(t) are twice continuously differentiable on (0,1).
sup0≤t≤1E(ϵ(t)4∣t)<∞, sup0≤t≤1E(Xr(t)4∣t)<∞, and E(ϵ(t)4∣t) and E(Xr(t)4∣t) are continuous at t, r=1,…,p, where Xr(t) is the rth component of X(t).
For given t, Ψ(t) is positive definite matrix.
The proofs of Theorems 1 and 2 rely on the following lemmas.
Lemma A.1.
Let a1,a2,…,an and b1,b2,…,bn(b1≥b2≥,…,≥bn) be two sequences of real numbers. Let Sk=∑i=1kai; then, one has
(A.1)|∑i=1naibi|≤cmax1≤i≤n|bi|max1≤i≤n|Si|.
Lemma A.2.
Let ei, i=1,…,n, be a sequence of multi-independent random variate with E(ei)=0 and E(ei2)<c<∞. Then, one has
(A.2)max1≤k≤n|∑i=1kei|=Op(nlogn).
Further, let (j1,j2,…,jn) be a permutation of (1,2,…,n). Then, we have
(A.3)max1≤k≤n|∑i=1keji|=Op(nlogn).
Lemma A.3.
Let (X1,Y1),…,(Xn,Yn) be i.i.d. random vectors, where Yi is scalar random variables. Further, assume that E|Y1|s<∞, supx∫|y|sf(x,y)dy<∞, where f(·,·) denotes the joint density of (X,Y). Let K(·) be a bounded positive function with a bounded support, satisfying a Lipschitz condition. Then,
(A.4)supx|1n∑i=1n{Kh(Xi-x)Yi-E[Kh(Xi-x)Yi]}|=Op({log(1/h)nh}1/2)
provided that n2δ-1h→∞ for some δ<1-s-1.
Lemma A.4.
Suppose that conditions (C1)–(C5) hold. Then, one has
(A.5)sup0<t<1∥μ^(t)-Ψ-1(t)Φ(t)∥=Op(Cn),sup0<t<1∥g^(t)-Ψ-1(t)Φ(t)β-θ(t)∥=Op(Cn),
where μ^(t), g^(t) are defined in (9) and Cn={log(1/h)/nh}1/2+h2.
Proof.
Let
(A.6)Snl(t)=∑i=1n∫01Xi(s)Xi(s)T(s-th)lKh(s-t)ds,hhhhhhhhhhhhhhhhhhhhhhhhhl=0,1,2;
then, a simple calculation yields
(A.7)E{Snl(t)}=nΨ(t)∫01slK(s)ds+o(1),l=0,1,2.
Note that
(A.8)Dxx(t)=(Sn0(t)Sn1(t)Snl(t)Sn2(t));
together with (A.6), (A.7), by using Lemma A.3, we obtain
(A.9)Dxx(t)=nΨ(t)⊗(100∫01s2K(s)ds){1+Op(Cn)},
uniformly for t∈(0,1), where ⊗ is the Kronecker product. By using the same argument, we have
(A.10)Dxz(t)=nΦ(t)⊗(1,0)T{1+Op(Cn)}
uniformly for t∈(0,1). Combining (A.9), (A.10), and (9) yields
(A.11)μ^(t)=Ψ-1(t)Φ(t){1+Op(Cn)},
uniformly for t∈(0,1). With the similar proof to (A.11), we have
(A.12)∑i=1n∫01Xi(s)Yi(s)(s-th)lKh(t-s)ds=∑i=1n∫01Xi(s)[XiT(s)θ(s)+ZiT(s)β+ϵi(s)]×(s-th)lKh(t-s)ds=n{Ψ(t)θ(t)+Φ(t)β}∫01slK(s)ds{1+Op(Cn)},hhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhl=0,1.
Combining (A.12) with (5), we can prove that
(A.13)Dxy(t)=n[Ψ(t)θ(t)+Φ(t)β]⊗(1,0)T{1+Op(Cn)},
uniformly for t∈(0,1).
Invoking (A.9) and (A.13), it is easy to show that
(A.14)g^(t)=(Ip,0p)[Dxx(t)]-1Dxy(t)=[θ(t)+Ψ-1(t)Φ(t)β]{1+Op(Cn)}.
This completes the proof of Lemma A.4.
Lemma A.5.
Suppose that conditions (C1)–(C5) hold. Then, one has that
(A.15)1n∑i=1nη^i(β)⟶LN(0,B),
where B is defined by (14) and η^i(β)=∫01Z˘i(t)[Y˘i(t)-Z˘i(t)Tβ]dt.
Proof.
By (8), a simple calculation yields
(A.16)Z˘i(t)=[Zi(t)-μ(t)TXi(t)]+[μ(t)-μ^(t)]TXi(t),Y˘i(t)-Z˘i(t)Tβ=Xi(t)T[θ(t)-g^(t)+μ^(t)β]+εi(t),
where μ(t)=Ψ(t)-1Φ(t). Then, we have
(A.17)η^i(β)=∫01[Zi(t)-μ(t)TXi(t)]εi(t)dt+∫01[μ(t)-μ^(t)]TXi(t)εi(t)dt+∫01[Zi(t)-μ(t)TXi(t)]Xi(t)T[θ(t)-g^(t)+μ^(t)β]dt+∫01[μ(t)-μ^(t)]TXi(t)Xi(t)T[θ(t)-g^(t)+μ^(t)β]dt≡Ji1+Ji2+Ji3+Ji4.
Hence, we get
(A.18)1n∑i=1nη^i(β)=1n∑i=1nJi1+1n∑i=1nJi2+1n∑i=1nJi3+1n∑i=1nJi4≡J1+J2+J3+J4.
Note that Ji1, i=1,…,n, are i.i.d. It is easy to show that E(Ji1)=0, Var(Ji1)=B+o(1). Using the central limit theorem, we have
(A.19)J1⟶LN(0,B).
Hence, to prove this lemma, we only need to prove Jν→P0, ν=2,3,4. Now, we deal with J2. Let brs(t) be the (r,s) component of [μ(t)-μ^(t)]T and let ais(t) be the sth component of Xi(t)ϵi(t), s=1,…,p. Moreover, let
(A.20)∫01brs0(t)ais0(t)dt=max1≤s≤p{∫01brs(t)ais(t)dt}.
Let J2,r be the rth component of J2. By Lemmas A.1–A.4, we obtain
(A.21)|J2,r|=1n|∑i=1n∑s=1p∫01brs(t)ais(t)dt|≤pn|∑i=1n∫01brs0(t)ais0(t)dt|≤cnsup0≤t≤1|brs0(t)|max1≤k≤n|∑i=1k∫01ais0(t)dt|=cnOp(Cn)Op(nlogn)=op(1).
That is, J2→P0. By Lemma A.4, a simple calculation yields
(A.22)sup0<t<1∥θ(t)-g^(t)+μ^(t)β∥=Op(Cn).
Together with E{[Zi(t)-μ(t)TXi(t)]Xi(t)T}=0, and similar to the proof of J2, we have J3→P0. In addition, by Lemma A.4, we have ∥J4∥≤Op(nCn2)=op(1). This completes the proof of Lemma A.5.
Lemma A.6.
Suppose that conditions (C1)–(C5) hold; then, one has
(A.23)Γ^⟶PΓ,
where Γ^ and Γ are defined in (11) and (13), respectively.
Proof.
Combining (8) and (11), a simple calculation yields
(A.24)Γ^=1n∑i=1n∫01[Zi(t)-μ(t)TXi(t)][Zi(t)-μ(t)TXi(t)]Tdt=1n∑i=1n∫01[Zi(t)-μ(t)TXi(t)]Xi(t)T[μ(t)-μ^(t)]dt=1n∑i=1n∫01[μ(t)-μ^(t)]TXi(t)[Zi(t)T-Xi(t)Tμ(t)]dt=1n∑i=1n∫01[μ(t)-μ^(t)]TXi(t)Xi(t)T[μ(t)-μ^(t)]dt≡A1+A2+A3+A4.
By the law of large numbers, we can derive that A1→PΓ. We now show A2→P0. From Lemma A.4, we have that
(A.25)sup0<t<1∥μ^(t)-μ(t)∥=Op(Cn).
Together with E{[Zi(t)-μ(t)TXi(t)]Xi(t)T}=0, using the similar argument to Lemma A.5, we can prove A2→P0. Similarly, we can prove that A3→P0. In addition, by (A.25), it is easy to show that ∥A4∥=Op(Cn2). This completes the proof.
Proof of Theorem 1.
From (12), we can derive that
(A.26)β^=Γ^-11n∑i=1n∫01Z˘i(t)Y˘i(t)dt=Γ^-11n∑i=1n∫01Z˘i(t)Z˘i(t)Tβdt+Γ^-11n∑i=1n∫01Z˘i(t)[Y˘i(t)-Z˘i(t)Tβ]dt=β+Γ^-11n∑i=1nη^i(β).
Together with Lemmas A.5 and A.6, it is easy to show that
(A.27)n(β^-β)=Γ^-11n∑i=1nη^i(β)⟶LN(0,Σ).
This completes the proof of Theorem 1.
Proof of Theorem 2.
For t in a small neighborhood of t0, such that |t-t0|≤h, by using Taylor expansion, we can get
(A.28)θ(t)=θ(t0)+θ′(t0)(t-t0)+12θ′′(t0)(t-t0)2+Op(h3).
Together with (5) and (1), we can derive that
(A.29)Dxy(t0)=Dxx(t0)(θ(t0)hθ′(t0))+12h2θ′′(t0)Txx(t0)+Dxz(t0)β+Dxϵ(t0)+Op(nh3),
where
(A.30)Txx(t0)=(∑i=1n∫01Xi(t)XiT(t)(t-t0h)2Kh(t-t0)dt∑i=1n∫01Xi(t)XiT(t)(t-t0h)3Kh(t-t0)dt),Dxϵ(t0)=(∑i=1n∫01Xi(t)ϵi(t)Kh(t-t0)dt∑i=1n∫01Xi(t)ϵi(t)(s-t0h)Kh(t-t0)dt).
Invoking (18), (A.9), and (A.29), it is easy to show that
(A.31)θ^(t0)=θ(t0)+12h2θ′′(t0)(Ip,0p)Dxx-1(t0)Txx(t0)+(Ip,0p)Dxx-1(t0)Dxz(t0)(β^-β)+(Ip,0p)Dxx-1(t0)Dxϵ(t0)+Op(h3).
Let Δn(t0)=n-1Dxϵ(t0); we can prove that E{Δn(t0)}=0 and Var{Δn(t0)}=(nh)-1Δ(t0), where
(A.32)Δ(t0)=σ2(t0)f(t0)Ψ(t0)⊗(∫01K2(t)dt∫01tK2(t)dt∫01tK2(t)dt∫01t2K2(t)dt).
By the central limit theorem, we have that
(A.33)nhΔn(t0)⟶LN(0,Δ(t0)).
Combining this with (A.9), we can prove that
(A.34)nh(Ip,0p)Dxx-1(t0)Dxϵ(t0)=(Ip,0p)nDxx-1(t0)nhΔn(t0)⟶LN(0,V(t0)).
Using the similar argument to (A.9) in Lemma A.4, we can get
(A.35)Txx(t)=nΨ(t0)h2⊗(∫01t2K(t)dt,∫01t3K(t)dt)T×{1+Op(Cn)}.
Together with (A.9) and β^-β=Op(n-1/2), it is easy to prove that
(A.36)12h2θ′′(t0)(I0,0p)Dxx-1(t0)Txx(t0)=b(t0)+Op(h2Cn),(I0,0p)Dxx-1(t0)Dxz(t0)(β^-β)=Op(n-1/2).
Invoking (A.31), (A.34), and (A.36), a simple calculation yields
(A.37)nh(θ^(t0)-θ(t0)-b(t0))⟶LN(0,V(t0)).
This completes the proof of Theorem 2.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This paper is supported by the National Natural Science Foundation of China (11101119), the Higher-Education Reform Project of Guangxi (Grant no. 2014JGA209), and the Training Program for Excellent Young Teachers in Guangxi Universities.
RamsayJ. O.SilvermanB. W.2002Berlin, GermanySpringerSpringer Series in Statistics10.1007/b98886MR1910407RamsayJ. O.SilvemanB. W.1997New York, NY, USASpringerAit-SaïdiA.FerratyF.VieuP.Cross-validated estimations in the single-functional index model200842647549410.1080/02331880801980377MR24651292-s2.0-54349129142FerratyF.GoiaA.SalinelliE.VieuP.Functional projection pursuit regression201322229332010.1007/s11749-012-0306-2MR30622592-s2.0-84877924893ChenD.HallP.MüllerH.Single and multiple index functional regression models with nonparametric link20113931720174710.1214/11-AOS882MR2850218FerratyF.VieuP.Nonparametric models for functional data, with application in regression, time-series prediction and curve discrimination2004161-211112510.1080/10485250310001622686MR2053065ZBLl1049.620392-s2.0-0842264210RachdiM.VieuP.Nonparametric regression for functional data: automatic smoothing parameter selection200713792784280110.1016/j.jspi.2006.10.001MR2323791ZBLl051823522-s2.0-34248189605EscabiasM.AguileraA.ValderramaM.Functional PLS logit regression model200751104891490210.1016/j.csda.2006.08.0112-s2.0-34247379877YaoF.MüllerH.WangJ.Functional linear regression analysis for longitudinal data20053362873290310.1214/009053605000000660MR22531062-s2.0-19744369661ZhangJ. T.ChenJ.Statistical inferences for functional data20073531052107910.1214/009053606000001505MR23416982-s2.0-49749105599Aneiros-PérezG.VieuP.Semi-functional partial linear regression200676111102111010.1016/j.spl.2005.12.0072-s2.0-33846409110ShinH.Partial functional linear regression2009139103405341810.1016/j.jspi.2009.03.001MR2549090ZBLl1168.623582-s2.0-67649921716LianH.Functional partial linear model201123111512810.1080/10485252.2010.500385MR2780819ZBLl059305982-s2.0-79952474151XueL.ZhuL.Empirical likelihood for a varying coefficient model with longitudinal data200710247864265410.1198/016214507000000293MR23708582-s2.0-34250718576WuC. O.ChiangC. T.HooverD. R.Asymptotic confidence regions for kernel smoothing of a varying-coefficient model with longitudinal data1998934441388140210.2307/2670054MR16666352-s2.0-0032283833FerratyF.VieuP.2006New York, NY, USASpringerMR2229687