Movatterモバイル変換


[0]ホーム

URL:


Next Article in Journal
Existence Results for Coupled Implicit \({\psi}\)-Riemann–Liouville Fractional Differential Equations with Nonlocal Conditions
Next Article in Special Issue
On Some New Ostrowski–Mercer-Type Inequalities for Differentiable Functions
Previous Article in Journal
Truncated Fractional-Order Total Variation for Image Denoising under Cauchy Noise
Previous Article in Special Issue
Pólya–Szegö Integral Inequalities Using the Caputo–Fabrizio Approach
 
 
Search for Articles:
Title / Keyword
Author / Affiliation / Email
Journal
Article Type
 
 
Section
Special Issue
Volume
Issue
Number
Page
 
Logical OperatorOperator
Search Text
Search Type
 
add_circle_outline
remove_circle_outline
 
 
Journals
Axioms
Volume 11
Issue 3
10.3390/axioms11030102
Font Type:
ArialGeorgiaVerdana
Font Size:
AaAaAa
Line Spacing:
Column Width:
Background:
Article

K-Nearest Neighbor Estimation of Functional Nonparametric Regression Model under NA Samples

1
College of Mathematics and Physics, Anqing Normal University, Anqing 246133, China
2
Department of Mathematics, Brunel University, London UB8 3PH, UK
*
Author to whom correspondence should be addressed.
Submission received: 10 January 2022 /Revised: 16 February 2022 /Accepted: 21 February 2022 /Published: 25 February 2022
(This article belongs to the Special IssueCurrent Research on Mathematical Inequalities)

Abstract

:
Functional data, which provides information about curves, surfaces or anything else varying over a continuum, has become a commonly encountered type of data. The k-nearest neighbor (kNN) method, as a nonparametric method, has become one of the most popular supervised machine learning algorithms used to solve both classification and regression problems. This paper is devoted to the k-nearest neighbor (kNN) estimators of the nonparametric functional regression model when the observed variables take values from negatively associated (NA) sequences. The consistent and complete convergence rate for the proposed kNN estimator is first provided. Then, numerical assessments, including simulation study and real data analysis, are conducted to evaluate the performance of the proposed method and compare it with the standard nonparametric kernel approach.

    1. Introduction

    Functional data analysis (FDA) is a branch of statistics that analyzes data providing information about curves, surfaces or anything else varying over a continuum. In its most general form, under an FDA framework, each sample element of functional data is considered to be a random function.
    Popularized by Ramsay and Silverman [1,2], statistics for functional data analysis have attracted considerable research interest because of its wide applications in many practical fields, such as medicine, economics and linguistics. For an introduction to the topics, we can refer to the monographs of Ramsay and Silverman [3] for parametric models, and Ferraty and Vieu [4] for nonparametric models.
    In this paper, the following functional non-parametric regression model is considered.
    Y=m(χ)+ϵ,(1)
    whereY is a scalar response variable,χ is a covariate taking value in a subsetSF of an infinite-dimensional functional spaceF endowed with a semi-metricd(·,·).m(·) is the unknown regression operator fromSF toR, and the random errorϵ satisfiesE(ϵ|χ)=0,a.s.
    For the estimation of model (1), Ferraty and Vieu [5] investigated the classical functional Nadaraya-Watson (N-W) kernel type estimator ofm(·) and obtained the asymptotic properties with rates in the case ofα-mixing functional data. Ling and Wu [6] studied the modified N-W kernel estimate and derived the asymptotic distribution for strong mixing functional time series data, Baíllo and Grane [7] proposed a functional local linear estimate based on the local linear idea. In this paper, we focus on the k-nearest neighbors (kNN) method for regression model (1). The kNN method, as one of the most simple and traditional nonparametric techniques, is often used as a nonparametric classification method. The kNN method was first developed by Evelyn Fix and Joseph Hodges in 1951 [8] and then expanded by Thomas Cover [9]. In our kNN regression, the input consists of the k-closest training examples in a dataset, whereas the output is the property value for the object. This value is the average of the values of the k-nearest neighbors. Under independent samples, research in kNN regression mostly focuses on the estimation of the continuous regression functionm(χ). For example, Burba et al. [10] investigated the kNN estimator based on the idea of the local adaptive bandwidth of functional explanatory variables. The papers [11,12,13,14,15,16,17,18], and others, obtained the asymptotic behavior of nonparametric regression estimators for functional data in independent and dependent cases. Further, Kudraszow and Vieu [19] obtained asymptotic results for a kNN generalized regression estimator when the observed variables take values in an abstract space. Kara-Zaitri et al. [20] provided an asymptotic theory for several different target operators and some simulated experiences, including regression, conditional density, conditional distribution and hazard operators. However, functional observations often behave with correlation, including satisfying some form of negative dependence or negative association.
    Negatively associated (NA) sequences were introduced by Joag-Dev and Proschan in [21]. Random variables{Yi}1in are said to be NA, if for every pair of disjoint subsetsA,B{1,2,,n},
    Cov(f(Yi,iA)g(Yj,jB))0,
    or equivalently,
    E(f(Yi,iA),g(Yj,jB))E(f(Yi,iA))E(g(Yj,jB)),
    wheref andg are coordinatewise non-decreasing, such that this covariance exists. An infinite sequence{Yn}n1 is NA if every finite subcollection is NA.
    For example, if{Yi}1in follows permutation distributions, where{Y1,Y2,,Yn}={y1,y2,,yn} always andy1y2yn aren real numbers, then{Yi}1in is NA.
    Whereas kNN regression under NA sequences has not been explored in the literature, in this paper, we extend the kNN estimation of functional data from the case of independent samples to NA sequences.
    Let a pair{(χi,Yi)}i=1,,n be a sample of NA pairs in(χ,Y), which is a random vector valued in theF×R.(F,d) is a semi-metric space,F is not necessarily of the finite dimension and we do not suppose the existence of a density for the functional random variableχ. For a fixedχF, the closed ball withχ as the center andϵ as the radius is denoted as:
    d:B(χ,ϵ)={χF|d(χ,χ)ϵ}.
    The kNN regression estimator [10] is defined as follows:
    m^kNN(χ)=i=1nYiK(Hn,k(χ)1d(χi,χ))i=1nK(Hn,k(χ)1d(χi,χ)),χF,
    whereK(·) is the kernel function supported on[0,).Hn,k(χ) is a positive random variable that depends on(χ1,χ2,,χn) and is defined by:
    Hn,k(χ)=min{hR+:i=1nIB(χ,h)χi=k},
    obviously, the kNN estimator can be seen as an expansion to a random locally adaptive neighborhood of the traditional kernel method [5] defined as:
    m^n(χ)=i=1nYiK(hn(χ)1d(χi,χ))i=1nK(hn(χ)1d(χi,χ)),χF,
    wherehn(χ) is a sequence of positive real numbers such ashn(χ)0 a.s.n.
    This paper is organized as follows. The main results of our paper about the asymptotic behavior of the kNN estimators using a data-driven random number of neighbors are given inSection 2.Section 3 illustrates the numerical performance of the proposed method, including nonparametric functional regression analysis of the sea level surface temperature (SST) data for the El Nin˜o area (0–100 S, 800–900 W). The technical proofs are postponed toSection 4. Finally,Section 5 is devoted to comments on the results and to related perspectives for the future.

    2. Assumptions and Main Results

    In this section, we focus on the asymptotic property of the kNN regression estimator and need to state the convergence rate of an estimator.
    One says that the rate of almost complete convergence of a sequence{Yn,n1} toY is of orderun if only if for anyϵ>0,
    n=1P(|YnY|>ϵun)<,
    and we writeYnY=Oa.co.(un)(see for instance [5]). By the Borel-Cantelli lemma, this implies thatYnYun0 almost surely, so almost complete convergence is a stronger result than almost sure convergence.
    Our results are stated under some mild assumptions we gather below for easy references. Throughout the paper, we will denote byC,C1,C some positive generic constants, which may be different in various places.
    Assumption 1.
    ϵ>0,P(χB(χ,ϵ))=φχ(ϵ)>0 andφχ(·) is a continuous function, and strictly monotonically increasing at the origin withφχ(0)=0.
    Assumption 2.
    There exist a functionϕ(·)0 and a bounded functionf(·)>0 such that:
    (i) 
    Fϕ(0)=0, andlimϵϕ(ϵ)=0.
    (ii) 
    limϵϕ(uϵ)ϕ(ϵ)=0, for anyu[0,1].
    (iii) 
    τ>0 such thatsupχSFφχ(ϵ)ϕ(ϵ)f(χ)=O(ϵτ),ϵ0.
    Assumption 3.
    K(t) is a nonnegative bounded kernel function with support [0, 1], and ifK(1)>0, the derivativeK(t) exists on [0, 1] satisfying:
    <C<K(t)<C<,fort[0,1].
    Assumption 4.
    m(·) is a bounded Lipschitz operator with order β onSF, and there existsβ>0 such that:
    χ1,χ2SF,m(χ1)m(χ2)Cd(χ1,χ2)β.
    Assumption 5.
    m2,E|Y|mX=χ=δm(χ)<C withδm(·) continuous onSF.
    Assumption 6.
    Kolmogorov’s ϵ-entropy ofSF satisfies:
    n=1exp(1ω)ΨSFlognn<,forsomeω>1.
    Forϵ>0, the Kolmogorov’s ϵ-entropy of some setSFF is defined byΨSF=log(Nϵ(SF)), whereNϵ(SF) is the minimal number of open balls, which can coverSF withχ1,χ2,,χNϵ(SF) as the center and ϵ as the radius inF.
    Remark 1.
    Assumption 1, Assumption 2((i)–(iii)) and Assumption 4 are the standard assumptions for small ball probability and regression operators in nonparametric FDA, see Kudraszow and Vieu [19]. Assumption 2(ii) will play a key role in the methodology particularly when we compute the asymptotic variance and permit it to be explicit in Ling and Wang [6]. Assumption 2(iii) shows that the small ball probability can be written as the product of the two independent functionsϕ(·) andf(·), which has been used many times in Masry [11], Laib and Louani [12] and other literatures. Assumption 5 is standard in the nonparametric setting and concerns the existence of the conditional moments in Masry [11] and Burba [10], which aims to obtain the rate of uniform almost complete convergence. Assumption 6 assumes the Kolmogorov’s ϵ-entropy condition, which we will use in the following proof of the rate of uniform almost complete convergence.
    Theorem 1.
    Under Assumptions 1–6, suppose that sequence{kn,n1} satisfiesknn0,n,log2nkn<ΨSFlognn<knlogn and0<C1<knlog2n<C2<, for n large enough, then we have:
    supχSFm^kNN(χ)m(χ)=Oa.co.ϕ1knnβ+sn2ΨSFlognnn2.
    Remark 2.
    The Theorem extends the kNN estimation result of Theorem 2 in Kudraszow and Vieu [19] from the independent case to the NA mixed dependent case, and obtains the same convergence rate under the assumptions. Second, the almost complete convergence rate of the prediction operator is divided into two parts, one part affected by strong mixing and Kolmogorov’sϵ-entropy, and the other part depends on the smoothness of the regression operator and smoothness parameter k.
    Corollary 1.
    Under the condition of the Theorem, we have:
    supχSFm^kNN(χ)m(χ)=Oa.s.ϕ1knnβ+sn2ΨSFlognnn2.
    Corollary 2.
    Under the condition of the Theorem, we have:
    supχSFm^kNN(χ)m(χ)=OPϕ1knnβ+sn2ΨSFlognnn2.

    3. Simulation

    3.1. A simulation Study

    In this section, we aim at illustrating the performance of the nonparametric functional regression model and we will make a comparison with traditional kernel density estimation methods. We consider the nonparametric functional regression model:
    Yi=m(χi)+εi,
    wherem(χi)=0π5χi(t)dt2,εi is distributed according toN(0,0.05), the functional curveχi(t) is generated in the following way:
    χi(t)=ait3+arctanbitπ5,t0,π5,i=1,2,,n.
    whereaiN0,π10,i=1,2,,n,b1,b2,,bnNn(0,Σ),0 represents zero vector and the covariance matrix is defined as:
    =1+θ2θ0000θ1+θ2θ0000θ1+θ20000001+θ2θ0000θ1+θ2θ0000θ1+θ2n×n,0<θ<1.
    By the definition of NA, it can be seen that(b1,b2,,bn) is an NA vector for eachn3 with a finite moment of any order (see Wu and Wang [22]).
    We choose casually thatθ=0.4, the sample sizesn asn=330, t takes 1000 equispaced values in[0,π5]. We carry out the simulation of the curveχ(t) for the 330 samples (seeFigure 1).
    We consider the Epanechnikov kernel given byK(u)=34(1u2)I[0,1](u), and the semi-metricsd(·,·) based on derivatives of orderq.
    d(χi,χj)=0π5χi(q)(t)χj(q)(t)2dt,χi,χjF,q={0,1,2,}.
    Our purpose is to compare the mean square error (MSE) of the kNN method with the NW kernel approach on finite simulated datasets. In the finite sample simulation, the following steps are followed.
    Step 1: We take 300 curves to construct the training samplesχi,Yii=1300, and the other 30 constitute the test samples{χi,Yi}i=301330.
    Step 2: In the training sample, the parametersk andh in the kNN method and NW kernel method are automatically selected based on the cross-validation method, respectively.
    Step 3: Based on the MSE standard (see [4] for details), we obtain that the respective semi-metric parametersq in both the kNN method and the NW method takesq=1.
    Step 4: The response valuesY^ii=301330 andY˜ii=301330 of the test sampleYii=301330 are calculated by using the kNN method and the NW method, respectively, and their MSE and scatter plots against the true value{Yi}i=301330 are represented byFigure 2.
    As we can see inFigure 2, the MSE of the kNN method is much smaller than that of the NW method, and the scattered points inFigure 2 are more densely distributed around the linear functiony=x, which shows that the kNN method has a better fit and higher prediction accuracy for the NA dependent functional samples.
    The kNN method and NW method were used to conduct 100 independent replicated experiments at sample sizes ofn=200,300,500,800, respectively. AMSE was calculated for both methods at different sample sizes using the following equation.
    AMSE=1100j=1100130i=n30nY¯iYi2,Y¯i=Y^i,Y˜i,n=200,300,500,800
    As can be seen fromTable 1, the AMSE of the kNN method is much smaller than that of the NW kernel method when the sample size is fixed atn=200,300,500,800, respectively; when the estimation method is fixed, the AMSE of the two estimation methods have the same trend—they both decrease as the sample size increases. However, the decreasing speed of the kNN method is significantly faster than that of the NW kernel method.

    3.2. A Real Study

    This section applies the proposed kNN regression analysis of the data, which consist of the sea level surface temperature (SST) for the ElNin˜o area (0–100 S, 800–900 W) for a total of 31 years from 1 January 1990 to 31 December 2020. The data are available online at the website:https://www.cpc.ncep.noaa.gov/data/indices/ (accessed on 1 January 2022). More relevant discussions of these data can be found in Ezzahrioui et al. [13,14], Delsol et al. [23], and Ferraty et al. [24] The 1618 weekly SST data from the original data were preprocessed and averaged by month to obtain 372 monthly average SST discrete data.Figure 3 displays the decomposition of the multiplicative time series of the monthly SST.
    Figure 4 shows that the monthly average SST in El Nin˜o regions from 1990 to 2020 had a clear seasonal variation, and the monthly trend of SST can also clearly be observed from the seasonal index plot of the monthly mean SST.
    The main factors affecting the temperature variation can be generally summarized as seasonal factors and random fluctuations. If the seasonal factor is removed, the SST should be left with only random fluctuations, i.e., the values fluctuate up and down at some mean value. At the same time, if the effect of random fluctuations is not considered, the SST is left with only the seasonal factor, i.e., the SST will have similar values in the same month in different years.
    The following steps implement the kNN regression estimation method for the analysis of the SST data and display the comparison with the NW sum estimation method inFigure 5.
    Step 1: Transform 372 months (31 years) of SST data{Zi,i=1,,372} into functional data.
    Step 2: Divide the 31 samples of data(χj,Yj(s))j=1,,31 into two parts: 30 training samples of data(χj,Yj(s))j=1,,30 for model fitting and 1 test sample of data(χ31,Y31(s)) for prediction assessment.
    Step 3: Here, the functional principal component analysis (FPCA) is applicable to semi-measures for rough curves such as SST data (see Chapter 3 of Ferraty et al. [25] for the methodology). A quadratic kernel function used inSection 3.1 is used in kNN regression.
    Step 4: The SST values(Y^31(s),s=1,,12) for 12 months in 2020 are predicted by the kNN method and the NW method, respectively, along with obtaining their MSEs for both methods.
    Then, in step 1, we split the discrete monthly average temperature data of 372 months into 31 years of temperature profiles and express them asχi={Zi(t),12(j1)<t<12j},i=1,,31. Therefore, the response variable can be expressed asYj(s)={Z12j+s,s=1,,12},j=1,,30. Thus,(χj,Yj(s))j=1,,30 is the sample set of dependent function type with a sample size of 30, whereχj is the function type data, andYj(s) is a real value.
    In Step 3, the choice of parametersq for the kNN method and NW method is performed via computation of cross-validation in R, which givesq=3 andq=1 for the kNN regression method and NW method, respectively. The selection of parametersk andh is similar toSection 3.1.
    FromFigure 5, which compares the MSE values calculated by the two methods, it can be seen that the MSE of the kNN method is much smaller than that of the NW method. Further, noting that the degree of fit between the curves fitted by the two methods to the true curve (dotted line), the predicted curves by two methods are generally closer to the true curve, indicating that the prediction effect of both methods is very good. However, a closer look reveals that the predicted values of the kNN method obviously have better fitting at the inflection points of the curves, such as January, February, March, November and December, which fully reflect the fact that the kNN method pays more attention to the local variation than the NW method when processing the data like this, including the abnormal or extreme distribution of the response variable.

    4. Proof of Theorem

    In order to prove the main results, we give some lemmas. Let(Ai,Bi)i=1,2,,n ben random pairs valued in(Ω×R,A×B(R)), where(Ω,A) is a general measurable space. LetSΩ be a fixed subset ofΩ,G(·,(χ,·)):R×(SΩ×Ω)R+ be a measurable function, fort,tR,
    (L0):ttG(t,z)G(t,z),zSΩ×Ω,
    Dn(χ)nN is a sequence of random real variables (r.r.v.), andc(·):SΩR is a nonrandom function such thatsupχSΩc(χ)<. ForχSΩ,nN/{0}, we define:
    cn,χ(t)=i=1nBiG(t,(χ,Ai))i=1nG(t,(χ,Ai)).
    Lemma 1
    ([10]).Letun(χ)nN be a decreasing positive real sequence satisfyinglimnun=0. For any increasing sequencesβn(0,1) andβn1=O(un), there exist two real random sequences{Dn(βn,χ)}nN and{Dn+(βn,χ)}nN such that:
    (L1) 
    Dn(βn,χ)Dn+(βn,χ),nN,χSΩ,
    (L2) 
    IDn(βn,χ)Dn(βn,χ)Dn+(βn,χ),χSΩ1, a.co.n,
    (L3) 
    supχSΩi=1nGDn(βn,χ)i=1nGDn+(βn,χ)βn=Oa.co.(un),
    (L4) 
    supχSΩcn,χDn(βn,χ)c(χ)=Oa.co.(un),
    (L5) 
    supχSΩcn,χDn+(βn,χ)c(χ)=Oa.co.(un),
    then, we have:
    supχSΩcn,χ(Dn(βn,χ))c(χ)=Oa.co.(un).
    The proof of Lemma 1 is not presented here because it follows, step by step, the same argument in Burba et al. [10], Kudraszow and Vieu [19].
    Lemma 2
    ([26]).Let{Xn,nN} be an NA random sequence with zero mean, and there exists a positive constantck,k=1,2,,n such that|Xk|ck, letSn=X1+X2++Xn. For anyϵ>0, we get:
    P(Snnϵ)expn2ϵ22i=1nci2,
    and
    P(|Sn|nϵ)2expn2ϵ22i=1nci2.
    Lemma 3.
    Suppose that Assumptions 1–6 hold, andhn(χ)0 a.s.n in model (3) satisfying:
    limφχ(Hn,k(χ))φχ(hn(χ))=0,
    0<C1hninfχSFhn(χ)supχSFhn(χ)C2hn<,
    and for n large enough,
    log2nnϕ(hn)<ΨSFlognn<nϕ(hn)logn,
    0<C1<nϕ(hn)log2n<C2<,
    then we have:
    supχSFm^n(χ)m(χ)=Oa.co.hnβ+Oa.co.sn2ΨSF(ϵ)n2,
    whereϵ=lognn.
    Proof of Lemma 3. 
    In order to simplify the proof, we introduce some notations in this article. ForχSF, letk(χ)=argmink=1,2,,Nϵ(SF)d(χ,χk),sn2=maxsn,12,sn,22,sn,32,sn,42 be the mixed operator covariance,
    sn,12=i=1nj=1nCov(Yiui,Yjuj),sn,22=i=1nj=1nCov(vi,vj),
    sn,32=i=1nj=1nCov(ui,uj),sn,42=i=1nj=1nCov(wi,wj),
    whereui=IBχkχ,C2hn(χ)+ϵ,0<hn0,
    vi=YiKhn(χk)1d(χk,χi)EKhn(χk)1d(χk,χ1)EYiK(hn(χk)1d(χk,χi)EKhn(χk)1d(χk,χ1),
    wi=Khn(χk)1d(χk,χi)EKhn(χk)1d(χk,χ1)EKhn(χk)1d(χk,χi)EKhn(χk)1d(χk,χ1).
    For the fixedχSF in model (3), we have the decomposition as follows:
    m^n(χ)m(χ)=m^2n(χ)m^1n(χ)mn(χ)=1m^1n(χ)[m^2n(χ)Em^n(χ)]+1m^1n(χ)[Em^n(χ)mn(χ)]+mn(χ)m^1n(χ)[1m^1n(χ)].
    where:
    m^1n(χ)=i=1nKhn(χ)1d(χ,χi)nEKhn(χ)1d(χ,χ1),m^2n(χ)=i=1nYiKhn(χ)1d(χ,χi)nEKhn(χ)1d(χ,χ1).
    It suffices to prove the three following results in order to establish (9),
    supχSFEm^2n(χ)m(χ)=Oa.co.hnβ,
    supχSFm^2n(χ)Em^2n(χ)=Oa.co.sn2ΨSF(ϵ)n2,
    supχSFm^1n(χ)1=supχSFm^1n(χ)Em^1n(χ)=Oa.co.sn2ΨSF(ϵ)n2.
    As to the Equation (10). ForχSF, by the Equation (6) and Assumption 4, it follows that:
    Em^2n(χ)m(χ)=Ei=1nYiKhn(χ)1d(χ,χi)nEKhn(χ)1d(χ,χ1)m(χ)=EY1Khn(χ)1d(χ,χi)EKhn(χ)1d(χ,χ1)m(χ)=m(χ1)EKhn(χ)1d(χ,χi)m(χ)EKhn(χ)1d(χ,χ1)EKhn(χ)1d(χ,χ1)=m(χ1)m(χ)=Oa.co.hnβ.
    Then, we need to show the Equation (11). In fact, we have the decomposition as follows:
    supχSFm^2n(χ)Em^2n(χ)supχSFm^2n(χ)m^2n(χk(χ))+supχSFm^2n(χk(χ))Em^2n(χk(χ))+supχSFEm^2n(χk(χ))Em^2n(χ)=:I1+I2+I3.
    ForI1, by Assumption 3, it is easily seen that:
    0<C1<EKhn(χ)1d(χ,χi)<C2<,
    thus,
    I1=supχSFi=1nYiKhn(χ)1d(χ,χi)nEKhn(χ)1d(χ,χ1)i=1nYiKhn(χk(χ))1d(χk(χ),χi)nEKhn(χk(χ))1d(χk(χ),χ1)=supχSF1ni=1nYiKhn(χ)1d(χ,χi)EKhn(χ)1d(χ,χ1)Khn(χk(χ))1d(χk(χ),χi)EKhn(χk(χ))1d(χk(χ),χ1)CsupχSF1ni=1nYiKd(χ,χi)hn(χ)Kd(χk(χ),χi)hn(χk(χ))IBχ,hn(χ)Bχk(χ),hn(χ)(χi)CsupχSF1ni=1nYiIBχk(χ),C2hn(χ)+ϵ(χi),
    forη>0, we have:
    PI1>ηsn,12ΨSF(ϵ)n2PCsupχSF1ni=1nYiIBχk(χ),C2hn(χ)+ϵ(χi)>ηsn,12ΨSF(ϵ)n2CNϵ(SF)maxkχ1,χ2,,χNϵ(SF)Pi=1n|Yi|IBχk(χ),C2hn(χ)+ϵ(χi)>ηsn,12ΨSF(ϵ).
    According to (4) in Lemma 2 and Assumption 6, we have:
    Pi=1n|Yi|IBχk(χ),C2hn(χ)+ϵ(χi)>ηsn,12ΨSF(ϵ)expηsn,12ΨSF(ϵ)2i=1nci2exp1ηsn,122i=1nci2ΨSF(ϵ)<.
    Hence, it follows that:
    I1=Oa.co.sn,12ΨSF(ϵ)n2.
    ForI2, similar to the proof ofI1, forη>0, we have:
    PI2>ηsn,22ΨSF(ϵ)n2=PsupχSFi=1nYiKdχk(χ),χihnχk(χ)nEKdχk(χ),χ1hnχk(χ)Ei=1nYiKdχk(χ),χihnχk(χ)nEKdχk(χ),χ1hnχk(χ)>ηsn,22ΨSF(ϵ)n2=PsupχSF1ni=1nYiKdχk(χ),χihnχk(χ)EKdχk(χ),χ1hnχk(χ)EYiKdχk(χ),χihnχk(χ)EKdχk(χ),χ1hnχk(χ)>ηsn,22ΨSF(ϵ)n2Nϵ(SF)maxkχ1,χ2,,χNϵ(SF)P1ni=1nYiKdχk(χ),χihnχk(χ)EKdχk(χ),χ1hnχk(χ)EYiKdχk(χ),χihnχk(χ)EKdχk(χ),χ1hnχk(χ)>ηsn,22ΨSF(ϵ)n2CNϵ(SF)maxkχ1,χ2,,χNϵ(SF)P1ni=1nYiIBχk(χ),C2hn(χ)+ϵ(χi)>ηsn,22ΨSF(ϵ)n2.
    Thus,
    I2=Oa.co.sn,22ΨSF(ϵ)n2.
    Finally, forI3, we can getI3EsupχSFm^2nχk(χ)m^2n(χ). The proof process is similar toI1, and we can obtain:
    I3=Oa.co.sn,12ΨSF(ϵ)n2.
    Therefore, combining the Equations (13)–(15), the Equation (11) can be established.
    Similarly, we may prove the Equation (12). Hence, the proof of Lemma 3 is completed. □
    Proof of Theorem 1. 
    According to Lemma 1, letSΩ=SF,Ai=χi,Bi=Yi,G(t,(χ,Ai))=K(t1d(χ,χi)),Dn(χ)=Hn,k(χ),cn,χ(χ)=m^kNN(χ),c(χ)=m^n(χ). Letβn(0,1) be an increasing sequence such thatβn1=O(un), whereun=ϕ1knnβ+sn2ΨSFlognnn2 is a decreasing positive real sequence such thatlimnun=0 andhn=ϕ1knnβ. LetDn(βn,χ)nN andDn+(βn,χ)nN be two real random sequences such that:
    φχDn(βn,χ)=φχ(hn(χ))βn12,
    φχDn+(βn,χ)=φχ(hn(χ))βn12,
    Firstly, we verify the conditions(L4) and(L5) in Lemma 1. ByφχDn(βn,χ) andβn1=O(un), it is easy to follow that the local bandwidthDn(βn,χ) satisfies the condition (5). Combininghn=ϕ1knnβ with Assumption 2, it follows thathn(χ) satisfies the condition (6). Letkn=nϕ(hn), from Assumption 2(i) we obtain thatknn=nϕ(hn)n=nϕ(hn) is satisfied. Hence, according to the conditions of the Theorem, the Equations (7) and (8) in Lemma 3 hold. Thus, by Lemma 3, we have:
    supχSFcn,χDn(βn,χ)c(χ)=Oa.co.ϕ1knnβ+sn2ΨSFlognnn2=Oa.co.(un).
    Similarly, forDn+(βn,χ), we can also get:
    supχSFcn,χDn+(βn,χ)c(χ)=Oa.co.ϕ1knnβ+sn2ΨSFlognnn2=Oa.co.(un).
    Secondly, checking the conditions(L1) and(L2) in Lemma 1, and combining (16) and (17) withβn(0,1), it is clearly followed that:
    φχDn(βn,χ)φχ(hn(χ))φχDn+(βn,χ),
    By Assumption 1 we get:
    Dn(βn,χ)hn(χ)Dn+(βn,χ).
    According to (5) and (18), forn, we have:
    φχDn(βn,χ)φχHn,χ(χ)φχDn+(βn,χ),
    That is:
    φχDn(βn,χ)φχDn(χ)φχDn+(βn,χ),
    Therefore, by Assumption 1 we can get:
    Dn(βn,χ)Dn(χ)Dn+(βn,χ),
    Thus,
    IDn(βn,χ)Dn(βn,χ)Dn+(βn,χ),χSF1,a.co.n.
    (L2) is checked.
    Finally, we establish the condition(L3) in Lemma 1. Similar to Kudraszow and Vieu [19], we denote:
    f*(χ,hn(χ))=EKhn(χ)1d(χ,χ1),χSF.
    and let:
    F1=f*χ,Dn(βn,χ)f*χ,Dn+(βn,χ),F2=m^1nχ,Dn(βn,χ)m^1nχ,Dn+(βn,χ)1,F3=f*χ,Dn+(βn,χ)f*χ,Dn(βn,χ)βn1.
    Then,(L3) can be decomposed as follows:
    supχSFi=1nGDn(βn,χ)i=1nGDn+(βn,χ)βn=m^1nχ,Dn(βn,χ)m^1nχ,Dn+(βn,χ)f*χ,Dn(βn,χ)f*χ,Dn+(βn,χ)f*χ,Dn+(βn,χ)f*χ,Dn(βn,χ)f*χ,Dn(βn,χ)f*χ,Dn+(βn,χ)βn=f*χ,Dn(βn,χ)f*χ,Dn+(βn,χ)m^1nχ,Dn(βn,χ)m^1nχ,Dn+(βn,χ)f*χ,Dn+(βn,χ)f*χ,Dn(βn,χ)βnf*χ,Dn(βn,χ)f*χ,Dn+(βn,χ)m^1nχ,Dn(βn,χ)m^1nχ,Dn+(βn,χ)1+f*χ,Dn+(βn,χ)f*χ,Dn(βn,χ)βn1|F1||F2|+|F1||F3|.
    By Assumption 3, it is followed that:
    supχSF|F1|C,
    and forχSF,m^1n(χ)=i=1nKhn(χ)1d(χ,χi)nEKhn(χ)1d(χ,χ1), refering to Ferraty et al. [25], we have:
    supχSFm^1n(χ)1=Oa.co.sn2ΨSFlognnn2,
    Therefore,
    supχSF|F2|=supχSFm^1nχ,Dn(βn,χ)m^1nχ,Dn+(βn,χ)1=supχSFm^1nχ,Dn(βn,χ)1+1m^1nχ,Dn+(βn,χ)m^1nχ,Dn+(βn,χ)supχSFm^1nχ,Dn(βn,χ)1+supχSFm^1nχ,Dn+(βn,χ)1infχSFm^1nχ,Dn+(βn,χ)=Oa.co.sn2ΨSFlognnn2.
    Moreover, forF3, according to Lemma 1 in Ezzahrioui and Ould-said [13] and Assumption 2(iii), there existsτ>0, forχSF,
    f*χ,hn(χ)=ϕ(hn(χ))τf(χ)+Oϕ(hn(χ))hn(χ)β=τφ(hn(χ))+Oϕ(hn)hnβ,
    byφDn(βn,χ)φDn+(βn,χ)=βn,supχSF|F3|=Oϕ(hn)hnβ=Oβnϕ1(knn)β holds. Hence, forβn1, it follows that
    supχSF|F3|=Oϕ1(knn)β.
    Combining (19)–(22), we obtain:
    supχSFi=1nGDn(βn,χ)i=1nGDn+(βn,χ)βn=Oa.co.(un).
    (L3) is established.
    Thus, the conditions(L1)(L5) in Lemma 1 have been established. By Lemma 1, we can get:
    supχSFm^KNN(χ)m(χ)=Oa.co.ϕ1knnβ+sn2ΨSFlognnn2.
    The proof of the Theorem 1 is completed. □

    5. Conclusions and Future Research

    Functional data analysis deals with the analysis and theory of data that are in the form of functions, images and shapes, or more general objects. In a way, correlation is really the heart of data science. The correlation between variables may be complicated, from simply independent toα-mixing or something else, such as negatively associated (NA). The kNN method, as one of the nonparametric methods, is very useful in statistical estimation and machine learning. While regression analysis of functional data under many variable correlated cases, except NA sequences, has been explored. This paper builds a kNN regression estimator of the functional regression model. In particular, we obtain the almost complete convergence rate of kNN estimation. Some simulated experiments and real data analyses illustrate the feasibility and the finite-sample behavior of the method. Further work includes introducing the kNN machine learning algorithm for functional data analysis and kNN high-dimensional modeling with NA sequences.

    Author Contributions

    Conceptualization, X.H. and J.W.; methodology, X.H.; software, J.W.; writing—original draft preparation, X.H. and J.W.; writing—review and editing, K.Y. and L.W.; visualization, K.Y.; supervision, X.H.; project administration, K.Y.; funding acquisition, K.Y. All authors have read and agreed to the published version of the manuscript.

    Funding

    This research was funded by the National Social Science Foundation (Grant No. 21BTJ040).

    Institutional Review Board Statement

    Not applicable.

    Informed Consent Statement

    Not applicable.

    Data Availability Statement

    https://www.cpc.ncep.noaa.gov/data/indices/ (accessed on 9 January 2022).

    Acknowledgments

    The authors are most grateful to the Editor and anonymous referee for carefully reading the manuscript and for valuable suggestions which helped in improving an earlier version of this paper. This research was funded by the National Social Science Foundation (Grant No. 21BTJ040).

    Conflicts of Interest

    The authors declare no conflict of interest in this paper.

    Abbreviations

    The following abbreviations are used in this manuscript:
    NANegatively Associated
    kNNk-Nearest Neighbor

    References

    1. Ramsay, J.; Dalzell, C. Some Tools for Functional Data Analysis.J. R. Stat. Soc. Ser. B Methodol.1991,53, 539–561. [Google Scholar] [CrossRef]
    2. Ramsay, J.; Silverman, B.Functional Data Analysis; Springer: New York, NY, USA, 1997. [Google Scholar]
    3. Ramsay, J.; Silverman, B.Functional Data Analysis, 2nd ed.; Springer: New York, NY, USA, 2005. [Google Scholar]
    4. Ferraty, F.; Vieu, P.Nonparametric Functional Data Analysis; Springer: New York, NY, USA, 2006. [Google Scholar]
    5. Ferraty, F.; Vieu, P. Nonparametric Models for Functional Data, with Application in Regression, Time Series Prediction and Curve Discrimination.J. Nonparametr. Stat.2004,16, 111–125. [Google Scholar] [CrossRef]
    6. Ling, N.X.; Wu, Y.H. Consistency of Modified Kernel Regression Estimation with Functional Data.Statistics2012,46, 149–158. [Google Scholar] [CrossRef]
    7. Baíllo, A.; Grané, A. Local Linear Regression for Functional Predictor and Scalar Response.J. Multivar. Anal.2009,100, 102–111. [Google Scholar] [CrossRef] [Green Version]
    8. Fix, E.; Hodges, J. Discriminatory Analysis. Nonarametric Discrimination: Consistency Properties.Inter. Stat. Re.1989,57, 238–247. [Google Scholar] [CrossRef]
    9. Altman, N.S. An introduction to kernel and nearest-neighbor nonparametic regression.Am. Stat.1992,46, 175–185. [Google Scholar]
    10. Burba, F.; Ferraty, F.; Vieu, P. k-Nearest Neighbour Method in Functional Nonparametric Regression.J. Nonparametr. Stat.2009,21, 453–469. [Google Scholar] [CrossRef]
    11. Masry, E. Nonparametric Regression Estimation for Dependent Functional Data: Asymptotic Normality.Stoch. Pro2005,115, 155–177. [Google Scholar] [CrossRef] [Green Version]
    12. Laib, N.; Louani, D. Rates of strong consistencies of the regression function estimator for functional stationary ergodic data.J. Stat. Plan. Inference2011,141, 359–372. [Google Scholar] [CrossRef]
    13. Ezzahrioui, M.; Ould-Sad, E. Asymptotic Normality of a Nonparametric Estimator of the Conditional Mode Function for Functional Data.J. Nonparametr. Stat.2008,20, 3–18. [Google Scholar] [CrossRef]
    14. Ezzahriouia, M.; Ould-Said, E. Some Asymptotic Results of a Nonparametric Conditional Mode Estimator for Functional Time-Series data.Stat. Neerl.2010,64, 171–201. [Google Scholar] [CrossRef]
    15. Horvath, L.; Kokoszka, P.Inference for Functional Data with Applications; Springer: New York, NY, USA, 2012. [Google Scholar]
    16. Ling, N.X.; Wang, C.; Ling, J. Modified Kernel Regression Estimation with Functional Time Series data.Stat. Probab. Lett.2016,114, 78–85. [Google Scholar] [CrossRef]
    17. Abdelmalek, G.; Abdelhak, C. Strong uniform consistency rates of the local linear estimation of the conditional hazard estimator for functional data.Int. J. Appl. Math. Stat.2020,59, 1–13. [Google Scholar]
    18. Mustapha, M.; Salim, B.; Ali, L. The consistency and asymptotic normality of the kernel type expectile regression estimator for functional data.J. Multivar. Anal.2021,181. [Google Scholar] [CrossRef]
    19. Kudraszow, N.L.; Vieu, P. Uniform Consistency of kNN Regressors for Functional Variables.Stat. Probab. Lett.2013,83, 1863–1870. [Google Scholar] [CrossRef]
    20. Kara, L.Z.; Laksaci, A.; Rachdi, M.; Vieu, F. Data-driven kNN Estimation in Nonparametric Functional Data Analysis.J. Multivar. Anal.2017,153, 176–188. [Google Scholar] [CrossRef]
    21. Joag-Dev, K.; Proschan, F. Negative Association of Random Variables with Application.Ann. Stat.1983,11, 286–295. [Google Scholar] [CrossRef]
    22. Yi, W.; Wang, X.; Sung, S.H. Complete Moment Convergence for Arrays of Rowwise Negatively Associated Random Variables and its Application in Non-parametric Regression Model.Probab. Eng. Inf. Sci.2017,32, 37–57. [Google Scholar]
    23. Delsol, L. Advances on Asymptotic Normality in Nonparametric Functional Time Series Analysis.Statistics2009,43, 13–33. [Google Scholar] [CrossRef]
    24. Ferraty, F.; Rabhi, A.; Vieu, P. Conditional Quantiles for Dependent Functional Data with Application to the Climatic El Niño Phenomenon.Sankhyā Indian J. Stat.2005,67, 378–398. [Google Scholar]
    25. Ferraty, F.; Laksaci, A.; Tadj, A.; Vieu, P. Rate of Uniform Consistency for Nonparametric Estimates with Functional Variables.J. Stat. Plan. Inference2010,140, 335–352. [Google Scholar] [CrossRef]
    26. Christofides, T.C.; Hadjikyriakou, M. Exponential Inequalities for N-demimartingales and Negatively Associated Random Variables.Stat. Probab. Lett.2009,79, 2060–2065. [Google Scholar] [CrossRef] [Green Version]
    Axioms 11 00102 g001 550
    Figure 1. Curve-sample with sample size ofn = 330.
    Figure 1. Curve-sample with sample size ofn = 330.
    Axioms 11 00102 g001
    Axioms 11 00102 g002 550
    Figure 2. Prediction effects of the two estimation methods. (a) kNN estimation method. (b) NW estimation method.
    Figure 2. Prediction effects of the two estimation methods. (a) kNN estimation method. (b) NW estimation method.
    Axioms 11 00102 g002
    Axioms 11 00102 g003 550
    Figure 3. Monthly mean SST factor decomposition fitting comprehensive output diagram.
    Figure 3. Monthly mean SST factor decomposition fitting comprehensive output diagram.
    Axioms 11 00102 g003
    Axioms 11 00102 g004 550
    Figure 4. Time series curve of SST in El Nin˜o during 31 years.
    Figure 4. Time series curve of SST in El Nin˜o during 31 years.
    Axioms 11 00102 g004
    Axioms 11 00102 g005 550
    Figure 5. Forecast value of SST in 2020 by KNN method and NW method.
    Figure 5. Forecast value of SST in 2020 by KNN method and NW method.
    Axioms 11 00102 g005
    Table
    Table 1. The AMSE of the predicted response variables of the two methods under different sample sizes.
    Table 1. The AMSE of the predicted response variables of the two methods under different sample sizes.
    n200300500800
    kNN- AMSE0.06230.04700.03270.0291
    NW- AMSE0.27640.25930.23290.2117
    Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

    © 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).

    Share and Cite

    MDPI and ACS Style

    Hu, X.; Wang, J.; Wang, L.; Yu, K. K-Nearest Neighbor Estimation of Functional Nonparametric Regression Model under NA Samples.Axioms2022,11, 102. https://doi.org/10.3390/axioms11030102

    AMA Style

    Hu X, Wang J, Wang L, Yu K. K-Nearest Neighbor Estimation of Functional Nonparametric Regression Model under NA Samples.Axioms. 2022; 11(3):102. https://doi.org/10.3390/axioms11030102

    Chicago/Turabian Style

    Hu, Xueping, Jingya Wang, Liuliu Wang, and Keming Yu. 2022. "K-Nearest Neighbor Estimation of Functional Nonparametric Regression Model under NA Samples"Axioms 11, no. 3: 102. https://doi.org/10.3390/axioms11030102

    APA Style

    Hu, X., Wang, J., Wang, L., & Yu, K. (2022). K-Nearest Neighbor Estimation of Functional Nonparametric Regression Model under NA Samples.Axioms,11(3), 102. https://doi.org/10.3390/axioms11030102

    Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further detailshere.

    Article Metrics

    No
    No

    Article Access Statistics

    For more information on the journal statistics, clickhere.
    Multiple requests from the same IP address are counted as one view.
    Axioms, EISSN 2075-1680, Published by MDPI
    RSSContent Alert

    Further Information

    Article Processing Charges Pay an Invoice Open Access Policy Contact MDPI Jobs at MDPI

    Guidelines

    For Authors For Reviewers For Editors For Librarians For Publishers For Societies For Conference Organizers

    MDPI Initiatives

    Sciforum MDPI Books Preprints.org Scilit SciProfiles Encyclopedia JAMS Proceedings Series

    Follow MDPI

    LinkedIn Facebook X
    MDPI

    Subscribe to receive issue release notifications and newsletters from MDPI journals

    © 1996-2025 MDPI (Basel, Switzerland) unless otherwise stated
    Terms and Conditions Privacy Policy
    We use cookies on our website to ensure you get the best experience.
    Read more about our cookieshere.
    Accept
    Back to TopTop
    [8]ページ先頭

    ©2009-2025 Movatter.jp