1 Introduction

As we known that semiparametric regression models (or partially linear models) rely on a dimension reduction assumption, while being still flexible enough due to the presence of a nonparametric term. In recent years, semiparametric regression models have attracted a growing number of statisticians to study it. Based on the effect of weather on electricity demand, Engle et al. (1986) studied the following semiparametric regression model,

$$\begin{aligned} Y_i=X_i^{'}\beta +g(T_i)+e_i,~~i=1,2,\ldots ,n. \end{aligned}$$
(1.1)

Hong (1991) studied the model (1.1), and gave the estimators \({\hat{\beta }}_n\) and \(g_n^{*}\) of \(\beta \) and g respectively by the methods of least squares and the nearest neighbor weight functions. In addition, he also obtained the asymptotic normality for \({\hat{\beta }}_n\) and the strong consistency for \(g_n^{*}\). Based on the independent and identically distributed (i.i.d.) samples, Gao (1992) proposed the kernel estimator \({\hat{g}}_n (\cdot )\) of \(g(\cdot )\) and the least squares estimator \({\hat{\beta }}_n\) of \(\beta \) in the semiparametric regression model as follows,

$$\begin{aligned} Y_i=X_i\beta +g(t_i)+e_i,~~i=1,2,\ldots ,n, \end{aligned}$$
(1.2)

and obtained some strong and weak consistencies and convergence rates for estimators of \(\beta \) and \(g(\cdot )\). Hu (1999) defined the least squares estimators \({\tilde{\beta }}_{\tau }\) of \(\beta \) and the estimator \({\tilde{g}}_{\tau }(t)\) of g, respectively. In the case of the independent random errors, he established some asymptotic properties for the estimators \({\tilde{\beta }}_{\tau }\) and \({\tilde{g}}_{\tau }(t)\), including the strong consistency, uniform strong consistency, r-th (\(r>2\)) mean consistency and r-th (\(r>2\)) mean uniform consistency. Pan et al. (2003) discussed the semiparametric model (1.2) with \(L^q\) mixingale errors, and obtained r-th (\(r>2\)) mean consistency and complete consistency for estimators of \(\beta \) and g. Hu (2006) studied the model (1.2) with linear time series errors and obtained the r-th (\(r>2\)) mean consistency and complete consistency for the estimators \({\hat{\beta }}_n\) and \({\hat{g}}_n(t)\) of \(\beta \) and g, respectively. Based on model (1.2), Gao et al. (1994) proposed a more general semiparametric regression model,

$$\begin{aligned} y_i=x_i\beta +g(t_i)+\sigma _ie_i,~~i=1,2,\ldots ,n, \end{aligned}$$
(1.3)

where \(\sigma _i^2=f(u_i)\), \((x_i,t_i,u_i)\) are known fixed design points, \(\beta \) is an unknown parameter to be estimated, \(g(\cdot )\) and \(f(\cdot )\) are unknown functions defined on compact set \(A\subset \mathbb {R}\), \(e_i\) are random errors. Additionally, Gao et al. (1994) gave the least squares estimators (LSE) and the weighted least squares estimators (WLSE) of \(\beta \) and g and the estimator of f, and further proved the asymptotic normality for the two estimators of \(\beta \) under i.i.d. random errors. Chen et al. (1998) investigated the strong consistency for the two estimators of \(\beta \) if \(\{e_i,i\ge 1\}\) is i.i.d.. Based on negatively associated random errors, Baek and Liang (2006) studied the strong consistency of estimators of \(\beta \), g and f, and the asymptotic normality of \(\beta \); Zhou and Hu (2010) obtained p-th (\(p>2\)) mean consistency for LSE and WLSE of \(\beta \) and g. For more details about the asymptotic properties of the estimators in semiparametric regression models, one can refer to Chen (1988), Speckman (1988), Hamilton and Truong (1997), Mammen and Van de Geer (1997), Aneiros and Quintela (2001), Zhou and Lin (2013) among others.

Inspired by the above literatures, we will study the p-th (\(p>0\)) mean consistency and strong consistency for LSE and WLSE of \(\beta \) and g in model (1.3) under the random errors \(\{e_i,i\ge 1\}\) being zero mean widely orthant dependent random variables. We will give the details in Sect. 2. Now let us recall the definition of widely orthant dependence structure.

Definition 1.1

For the random variables \(\{X_n,n\ge 1\}\), if there exists a finite real sequence \(\{h_{U}(n),n\ge 1\}\) satisfying for each \(n\ge 1\) and for all \(x_i\in (-\infty ,+\infty )\), \(1\le i\le n\),

$$\begin{aligned} P(X_1>x_1,X_2>x_2,\ldots ,X_n>x_n)\le h_{U}(n)\prod _{i=1}^{n}P(X_i>x_i), \end{aligned}$$

then we say that the \(\{X_n,n\ge 1\}\) are widely upper orthant dependent (WUOD, in short); if there exists a finite real sequence \(\{h_{L}(n),n\ge 1\}\) satisfying for each \(n\ge 1\) and for all \(x_i\in (-\infty ,+\infty )\), \(1\le i\le n\),

$$\begin{aligned} P(X_1\le x_1,X_2\le x_2,\ldots ,X_n\le x_n)\le h_{L}(n)\prod _{i=1}^{n}P(X_i\le x_i), \end{aligned}$$

then we say that the \(\{X_n,n\ge 1\}\) are widely lower orthant dependent (WLOD, in short); if they are both WUOD and WLOD, then we say that the \(\{X_n,n\ge 1\}\) are widely orthant dependent, and \(h_{U}(n),h_{L}(n),n\ge 1\) are called dominating coefficients.

The concept of WOD random variables was firstly introduced by Wang et al. (2013). And they therein gave some examples to show that the class of WOD random variables includes some common negatively dependent random variables, some positively dependent random variables and some others. Subsequently, various properties and applications were obtained. For instance, Liu et al. (2012) gave the asymptotically equivalent formula for the finite-time ruin probability under a dependent risk model with constant interest rate, Shen (2013a) established the Bernstein type inequality for WOD random variables and gave some applications, He et al. (2013) provided the asymptotic lower bounds of precise large deviations with nonnegative and dependent random variables, Wang et al. (2014) further established the complete convergence for arrays of row-wise WOD random variables and gave its applications in nonparametric regression models, Wang and Hu (2015a) studied the consistency of the nearest neighbor estimator of the density function based on WOD samples, Shen et al. (2016) provided some exponential probability inequalities for WOD sequence and gave applications in complete convergence and complete moment convergence, Chen et al. (2016) established a more accurate inequality of WOD random variables, and obtained some limit theorems including the strong law of large numbers, the complete convergence, the almost sure elementary renewal theorem and the weighted elementary renewal theorem, and so on.

Obviously, \(h_U(n)\ge 1\), \(h_L(n)\ge 1\), \(n\ge 1\). If \(h_U(n)=h_L(n)=M\) for any \(n\ge 1\), it is easily seen that the random variables \(\{X_n,n\ge 1\}\) are extended negatively dependent (END, in short), where M is a positive constant. More particularly, if \(M=1\), then the random variables \(\{X_n,n\ge 1\}\) are called negatively orthant dependent (NOD, in short). In other words, NOD is a special case of END. For details about NOD and END sequence, one can refer to Volodin (2002), Asadian et al. (2006), Liu (2009), Wang and Wang (2013), Shen (2013b), Shen et al. (2015), Wang et al. (2015b), and so forth. Furthermore, Joag-Dev and Proschan (1983) pointed out that negatively associated (NA, in short) random variables are NOD. Meanwhile, Hu (2000) introduced the concept of negatively superadditive dependence (NSD, in short) and pointed out that NSD implies NOD [see Property 2 of Hu (2000)]. By the above description, the class of WOD random variables contains END random variables, NOD random variables, NSD random variables, NA random variables and independent random variables as special cases. Thus, it is of practical significance to study the mean consistency and the strong consistency of estimators in the semiparametric model (1.3) with WOD random errors.

The organization of the paper is as follows. In Sect. 2, we first present the LSE and the WLSE of \(\beta \) and \(g(\cdot )\), and some basic assumptions; and then we will establish the main results, including the mean consistency and strong consistency for the LSE and the WLSE of \(\beta \) and \(g(\cdot )\); a numerical simulation to study the consistency of LSE for \(\beta \) and \(g(\cdot )\) is also carried out; finally, some important lemmas to prove the main results are provided. In Sect. 3, we mainly give the proofs of the main results. In “Appendix”, we present the proofs of Lemmas 2.4 and 2.5.

Throughout the paper, denote \(h(n)=\max \{h_U(n),h_L(n)\}\). \(a_n=O(b_n)\) denotes that there exists a positive constant C such that \(a_n\le Cb_n\). Let c, \(c_1\), \(c_2\), C, \(C_1\), \(C_2\), \(\ldots \) denote the positive constants whose values may vary at each occurrence.

2 Main results and lemmas

2.1 Estimators and basic assumptions

The LSE and the WLSE of \(\beta \) and \(g(\cdot )\) given in Gao et al. (1994) are as follows:

$$\begin{aligned}&{\hat{\beta }}_n=S_n^{-2}\sum _{i=1}^n{\tilde{x}}_i{\tilde{y}}_i,~~{\hat{g}}_n(t)=\sum _{i=1}^n W_{ni}(t)(y_i-x_i{\hat{\beta }}_n), \end{aligned}$$
(2.1)
$$\begin{aligned}&{\tilde{\beta }}_n=T_n^{-2}\sum _{i=1}^na_i{\tilde{x}}_i{\tilde{y}}_i,~~{\tilde{g}}_n(t)=\sum _{i=1}^n W_{ni}(t)(y_i-x_i{\tilde{\beta }}_n), \end{aligned}$$
(2.2)

where \(W_{ni}(\cdot )\) are weight functions only depending on the designed points \(t_i~(i=1,2,\ldots ,n)\), \({\tilde{x}}_i=x_i-\sum \nolimits _{j=1}^n W_{nj}(t_i)x_j\), \({\tilde{y}}_i=y_i-\sum \nolimits _{j=1}^n W_{nj}(t_i)y_j\), \(S_n^2=\sum \nolimits _{i=1}^n {\tilde{x}}_i^2\), \(a_i=\frac{1}{f(u_i)}\), \(T_n^2=\sum \nolimits _{i=1}^na_i{\tilde{x}}_i^2\).

In this paper, we will consider the following assumptions:

  • \(\mathbf {H_1}\) \(~(i)~\lim \limits _{n\rightarrow \infty }\frac{1}{n(h(n))^{2r}}\sum \nolimits _{i=1}^n \tilde{x}_i^2=\Gamma \) (\(0<\Gamma <\infty \)), \(\exists ~r\ge 1\);

    \((i)^{'}\) \(~\lim \limits _{n\rightarrow \infty }\frac{1}{n(h(n))^{r}}\sum \nolimits _{i=1}^n \tilde{x}_i^2=\Gamma \) (\(0<\Gamma <\infty \)), \(\exists ~r>0\);

    (ii\(0<m_0\le \min \limits _{1\le i\le n}f(u_i)\le \max \limits _{1\le i\le n}f(u_i)\le M_0<\infty \);

    (iii\(g(\cdot )\) and \(f(\cdot )\) are continuous on compact set A.

  • \(\mathbf {H_2}\) \(\max \limits _{1\le j\le n}\left| \sum \nolimits _{i=1}^nW_{ni}(t_j)-1\right| =o(1)\);

  • \(\mathbf {H_3}\) \(\max \limits _{1\le j\le n}\sum \nolimits _{i=1}^n|W_{ni}(t_j)|I(|t_i-t_j|>a)=o(1),~\forall ~a>0\);

  • \(\mathbf {H_4}\) \(\max \limits _{1\le j\le n}\sum \nolimits _{i=1}^n|W_{ni}(t_j)|=O(1)\);

  • \(\mathbf {H_5}\) \(\max \limits _{1\le i,j\le n}|W_{ni}(t_j)|=O(n^{-s}(h(n))^{-r}),~\exists ~s>0,~r\ge 1\);

  • \(\mathbf {H_5^{'}}\) \(\max \limits _{1\le i,j\le n}|W_{ni}(t_j)|=O(n^{-s}(h(n))^{-r}),~\exists ~s>0,~r>0\).

Remark 2.1

(\(H_1\))(i) (with \(h(n)=1\)) (ii) are some regular conditions, which are assumed in Gao et al. (1994), Chen et al. (1998), Baek and Liang (2006) and so on. Moreover, it can be deduced from (\(H_1\))(i) (or \((i)^{'})\) (ii) here that

$$\begin{aligned} S_n^{-2}\sum _{i=1}^n |{\tilde{x}}_i|\le C,~~T_n^{-2}\sum _{i=1}^n |a_i{\tilde{x}}_i|\le C. \end{aligned}$$
(2.3)

Remark 2.2

Remark 2.3 in Baek and Liang (2006) mentioned that the following two weight functions satisfy assumptions \((H_2)\)\((H_5)\) with \(h(n)=\log n\), \(s=\frac{1}{2}\) and \(r=1\):

$$\begin{aligned} W_{ni}^{(1)}(t)= & {} \frac{1}{h_n}\int _{s_{i-1}}^{s_i}K\left( \frac{t-s}{h_n}\right) d s,\\ W_{ni}^{(2)}(t)= & {} K\left( \frac{t-t_i}{h_n}\right) \left[ \sum _{j=1}^n K\left( \frac{t-t_j}{h_n}\right) \right] ^{-1}, \end{aligned}$$

where \(s_i=(t_i+t_{i+1})/2,~i=1,2,\ldots ,n-1,~s_0=0,~s_n=1\), \(0\le t_1\le t_2\le \cdots \le t_n\le 1\), \(K(\cdot )\) is the Parzen-Rosenblatt kernel function, and \(h_n\) is a bandwidth parameter.

2.2 Consistency

Let \(\{e_i,i\ge 1\}\) be a sequence of mean zero WOD random errors with dominating coefficient h(n), which is stochastically dominated by a random variable e, that is

$$\begin{aligned} P(|e_i|>x)\le CP(|e|>x) \end{aligned}$$

for all \(x\ge 0\), \(n\ge 1\) and some \(C>0\).

Theorem 2.1

(mean consistency) Let \(p>0\). Suppose that conditions \((H_1)\)(i, ii, iii) and \((H_2)\)\((H_5)\) hold. If \(Ee^2<\infty \) for \(0<p\le 2\) or \(E|e|^p<\infty \) for \(p>2\), then

$$\begin{aligned}&\lim _{n\rightarrow \infty }E|{\hat{\beta }}_n-\beta |^p=0, \end{aligned}$$
(2.4)
$$\begin{aligned}&\lim _{n\rightarrow \infty }E|{\tilde{\beta }}_n-\beta |^p=0. \end{aligned}$$
(2.5)

In addition, if \(\max _{1\le j\le n}\left| \sum _{i=1}^n W_{ni}(t_j)x_i\right| =O(1)\), then

$$\begin{aligned}&\lim _{n\rightarrow \infty }\max _{1\le i\le n}E|{\hat{g}}_n(t_i)-g(t_i)|^p=0, \end{aligned}$$
(2.6)
$$\begin{aligned}&\lim _{n\rightarrow \infty }\max _{1\le i\le n}E|{\tilde{g}}_n(t_i)-g(t_i)|^p=0. \end{aligned}$$
(2.7)

In particular, if h(n) is a constant function, we can get the following corollary by Theorem 2.1.

Corollary 2.1

Let \(p>0\). Assume that \(\{e_i,i\ge 1\}\) be a sequence of END random errors with mean zero. Suppose that conditions \((H_1)\)\((H_5)\) hold with \(h(n)=1\). If \(\sup _i Ee_i^2<\infty \) for \(0<p\le 2\) or \(\sup _iE|e_i|^p<\infty \) for \(p>2\), then (2.4) and (2.5) hold. In addition, if \(\max \nolimits _{1\le j\le n}\left| \sum \nolimits _{i=1}^n W_{ni}(t_j)x_i\right| =O(1)\), then (2.6) and (2.7) hold.

Remark 2.3

Under the NA sequence and conditions \((H_1)\)\((H_5)\) with \(h(n)=1\) and \(s=1/2\), Zhou and Hu (2010) obtained results (2.4)–(2.7) under \(\sup _{i}E|e_i|^p<\infty \) for some \(p>2\), while Corollary 2.1 in our paper gives these results for some \(p>0\). Since NA sequence is END, we extend Theorem 2.1 of Zhou and Hu (2010) to the case of END sequence. Furthermore, \(s>0\) in condition \((H_5)\) of Corollary 2.1 is more general than \(s=1/2\) in Theorem 2.1 of Zhou and Hu (2010).

The next theorem gives the strong consistency of estimators under some analogous conditions.

Theorem 2.2

(strong consistency) Suppose that conditions \((H_1)\) (\(i^{'}\), \(\textit{ii}\), \(\textit{iii}\)), \((H_2)\)\((H_4)\) and \((H_5^{'})\) hold. If \(Ee^2<\infty \) and \(\sum \nolimits _{i=1}^{n}i^{-s}(h(i))^{-r}=O(n^{s})\) for some \(s>0\) and \(r>0\), then

$$\begin{aligned}&{\hat{\beta }}_n\rightarrow \beta ~~a.s.,~n\rightarrow \infty , \end{aligned}$$
(2.8)
$$\begin{aligned}&{\tilde{\beta }}_n\rightarrow \beta ~~a.s.,~n\rightarrow \infty . \end{aligned}$$
(2.9)

In addition, if \(\max _{1\le j\le n}\left| \sum _{i=1}^n W_{ni}(t_j)x_i\right| =O(1)\), then

$$\begin{aligned}&\max _{1\le i\le n}|{\hat{g}}_n(t_i)-g(t_i)|\rightarrow 0~~a.s.,~n\rightarrow \infty , \end{aligned}$$
(2.10)
$$\begin{aligned}&\max _{1\le i\le n}|{\tilde{g}}_n(t_i)-g(t_i)|\rightarrow 0~~a.s.,~n\rightarrow \infty . \end{aligned}$$
(2.11)

Remark 2.4

Since \(h(i)\ge 1\), \(\sum \nolimits _{i=1}^{n}i^{-s}(h(i))^{-r}=O(n^{s})\) in Theorem 2.2 always holds as long as \(s\ge 1/2\) and \(r>0\).

Particularly, if h(n) is a constant function and \(s=1/2\) in Theorem 2.2, we can get the following corollary.

Corollary 2.2

Assume that \(\{e_i,i\ge 1\}\) be a sequence of mean zero END random errors, which is stochastically dominated by a random variable e. Suppose that conditions \((H_1)\)\((H_5)\) hold with \(s=1/2\), \(h(n)=1\). If \(Ee^2<\infty \), then (2.8) and (2.9) hold. In addition, if \(\max _{1\le j\le n}\left| \sum \nolimits _{i=1}^n W_{ni}(t_j)x_i\right| =O(1)\), then (2.10) and (2.11) hold.

Remark 2.5

Under mean zero NA random errors, Theorem 2.1 of Baek and Liang (2006) gave the results (2.8)–(2.11) under the conditions \((H_1)\)\((H_5)\) with \(s=1/2\), \(h(n)=1\) and \(\sup _iE|e_i|^p<\infty \) for some \(p>2\). Compared with it, Corollary 2.2 (i) extends the case of NA random variables to END random variables; (ii) lowers the order of the moment from \(p>2\) to 2.

2.3 Simulation

In this section, we will carry out a numerical simulation to study the consistency of LSE for \(\beta \) and \(g(\cdot )\). The data is generated from the model (1.3). Choose \(\sigma _i=1\), \(x_i=(-1)^{i}\frac{i}{n}\), \(i=1,2,\ldots ,n\). We take random error vector \((e_1,e_2,\ldots ,e_n)^{'}\sim N (\mathbf {0},\mathbf {\Sigma })\), where \(\mathbf {0}\) is a zero column vector, and

$$\begin{aligned} \mathbf {\Sigma }{=}\left( \begin{array}{ccccccc} \frac{1}{2}+\nu ^2 &{}\quad -\nu &{}\quad 0 &{}\quad \ldots &{}\quad 0 &{}\quad 0 &{}\quad 0 \\ -\nu &{}\quad \frac{1}{2}+\nu ^2 &{}\quad -\nu &{}\quad \ldots &{}\quad 0 &{}\quad 0 &{} \quad 0 \\ 0 &{}\quad -\nu &{}\quad \frac{1}{2}+\nu ^2 &{}\quad \ldots &{}\quad 0 &{}\quad 0&{}\quad 0 \\ \vdots &{}\quad \vdots &{}\quad \vdots &{}\quad \vdots &{}\quad \vdots &{}\quad \vdots &{} \quad \vdots \\ 0 &{}\quad 0 &{}\quad 0 &{}\quad \ldots &{}\quad \frac{1}{2}+\nu ^2 &{}\quad -\nu &{}\quad 0 \\ 0 &{} \quad 0 &{}\quad 0 &{}\quad \ldots &{}\quad -\nu &{}\quad \frac{1}{2}+\nu ^2 &{}\quad -\nu \\ 0 &{}\quad 0 &{}\quad 0 &{}\quad \ldots &{}\quad 0 &{} \quad -\nu &{}\quad \frac{1}{2}+\nu ^2 \\ \end{array} \right) _{n\times n},~\nu =0.1. \end{aligned}$$

It is obvious that \(e_1,e_2,\ldots ,e_n\) generated as the above method are NA by Joag-Dev and Proschan (1983), which is a special case of WOD (\(h(n)=1\)). Especially, we choose the nearest neighbor weights to be weight functions \(W_{ni}(\cdot )\). Without loss of generality, let \(A = [0,1]\) and \(t_i=\frac{i}{n}\), \(i=1,2,\ldots ,n\). For any \(t\in A\), we rewrite \(|t_1-t|,|t_2-t|,\ldots ,|t_n-t|\) as follows

$$\begin{aligned} |t_{R_1(t)}-t|\le |t_{R_2(t)}-t|\le \cdots \le |t_{R_n(t)}-t|, \end{aligned}$$

if \(|t_i-t|=|t_j-t|\), \(|t_i-t|\) is permuted before \(|t_j-t|\) if \(i<j\). Let \(k_n=\lfloor n^{0.6} \rfloor \) and define the nearest neighbor weight functions as follows

$$\begin{aligned} W_{ni}(t)= {\left\{ \begin{array}{ll} \frac{1}{k_n}, ~~\quad \text {if}~|t_i-t|\le |t_{R_{k_n}(t)}-t|,\\ 0, \qquad \text {otherwise}. \end{array}\right. } \end{aligned}$$

For any \(t=t_i\), \(i=1,2,\ldots ,n\), it is easily checked that

$$\begin{aligned} \sum _{i=1}^nW_{ni}(t)= & {} \sum _{i=1}^{n}W_{nR_{i}(t)}(t)=\sum _{i=1}^{k_n}\frac{1}{k_n}=1,\\ \max _{1\le i\le n}W_{ni}(t)= & {} \frac{1}{k_n}\le Cn^{-0.6},\\ \sum _{i=1}^nW_{ni}(t)I(|t_i-t|>a)\le & {} \sum _{i=1}^nW_{ni}(t)\frac{(t_i-t)^2}{a^2}\\\le & {} \sum _{i=1}^{k_n}\frac{(t_{R_{i}(t)}-t)^2}{k_na^2}\le \sum _{i=1}^{k_n}\frac{(i/n)^2}{k_na^2}\\\le & {} \left( \frac{k_n}{na}\right) ^2\le \frac{C}{n^{0.8}},\\ \left| \sum _{i=1}^nW_{ni}(t)x_i\right|\le & {} \sum _{i=1}^nW_{ni}(t)=1, \end{aligned}$$

which imply that the assumptions in our results are satisfied. Next, we compute \({\hat{\beta }}_n-\beta \) and \({\hat{g}}_n(t)-g(t)\) for 1000 times and obtain the corresponding boxplots by taking \(t=\frac{1}{n},\frac{100}{n}\) and the sample sizes n as 200, 400, 800, 1400, 3400 respectively when \(\beta \) and g(t) are in two different forms.

Case 1 \(\beta =2\), \(g(t)=t^2\).

Case 2 \(\beta =3\), \(g(t)=\sin t\).

In Figs. 1234567 and 8, \({\hat{\beta }}_n-\beta \) and \({\hat{g}}_n(t)-g(t)\), regardless of the values of t, fluctuate to zero and the variation ranges decrease markedly as the sample size n increases. These verify the validity of our results.

Fig. 1
figure 1

Boxplots of \({{\hat{\beta }}}_n-\beta \) with \(\beta =2\), \(\hbox {t}=1/\hbox {n}\), \(\hbox {g}(\hbox {t})={t}^2\)

Fig. 2
figure 2

Boxplots of \({{\hat{g}}}_n(t)-g(t)\) with \(\beta =2\), \(\hbox {t}=1/\hbox {n}\), \(\hbox {g}(\hbox {t})={t}^2\)

Fig. 3
figure 3

Boxplots of \({{\hat{\beta }}}_n-\beta \) with \(\beta =2\), \(\hbox {t}=100/\hbox {n}\), \(\hbox {g}(\hbox {t})={t}^2\)

Fig. 4
figure 4

Boxplots of \({{\hat{g}}}_n(t)-g(t)\) with \(\beta =2\), \(\hbox {t}=100/\hbox {n}\), \(\hbox {g}(\hbox {t})={t}^2\)

Fig. 5
figure 5

Boxplots of \({{\hat{\beta }}}_n-\beta \) with \(\beta =3\), \(\hbox {t}=1/\hbox {n}\), \(\hbox {g}(\hbox {t})=\hbox {sint}\)

Fig. 6
figure 6

Boxplots of \({{\hat{g}}}_n(t)-g(t)\) with \(\beta =3\), \(\hbox {t}=1/\hbox {n}\) , \(\hbox {g}(\hbox {t})=\hbox {sint}\)

Fig. 7
figure 7

Boxplots of \({{\hat{\beta }}}_n-\beta \) with \(\beta =3\), \(\hbox {t}=100/\hbox {n}\), \(\hbox {g}(\hbox {t})=\hbox {sint}\)

Fig. 8
figure 8

Boxplots of \({{\hat{g}}}_n(t)-g(t)\) with \(\beta =3\), \(\hbox {t}=100/\hbox {n}\) , \(\hbox {g}(\hbox {t})=\hbox {sint}\)

3 Proof of main results

It is easy to see that

$$\begin{aligned} {\hat{\beta }}_n-\beta= & {} S_n^{-2}\left[ \sum _{i=1}^n\sigma _i{\tilde{x}}_ie_i-\sum _{i=1}^n{\tilde{x}}_i\left( \sum _{j=1}^n W_{nj}(t_i)\sigma _j e_j\right) +\sum _{i=1}^n{\tilde{x}}_i{\tilde{g}}(t_i)\right] ,\end{aligned}$$
(3.1)
$$\begin{aligned} {\tilde{\beta }}_n-\beta= & {} T_n^{-2}\left[ \sum _{i=1}^na_i\sigma _i{\tilde{x}}_ie_i-\sum _{i=1}^na_i{\tilde{x}}_i\left( \sum _{j=1}^n W_{nj}(t_i)\sigma _j e_j\right) +\sum _{i=1}^na_i{\tilde{x}}_i{\tilde{g}}(t_i)\right] ,\end{aligned}$$
(3.2)
$$\begin{aligned} {\hat{g}}_n(t_i)-g(t_i)= & {} \sum _{j=1}^nW_{nj}(t_i)\sigma _j e_j-({\hat{\beta }}_n-\beta )\sum _{j=1}^nW_{nj}(t_i)x_j-{\tilde{g}}(t_i),\end{aligned}$$
(3.3)
$$\begin{aligned} {\tilde{g}}_n(t_i)-g(t_i)= & {} \sum _{j=1}^nW_{nj}(t_i)\sigma _j e_j-({\tilde{\beta }}_n-\beta )\sum _{j=1}^nW_{nj}(t_i)x_j-{\tilde{g}}(t_i), \end{aligned}$$
(3.4)

where \({\tilde{g}}(t_i)=g(t_i)-\sum \nolimits _{j=1}^nW_{nj}(t_i)g(t_j)\).

Proof of Theorem 2.1

We only prove (2.5) and (2.7), as the proofs of (2.4) and (2.6) are respectively analogous. Denote

$$\begin{aligned} H_{1n}= & {} T_n^{-2}\sum _{i=1}^na_i\sigma _i{\tilde{x}}_ie_i,~H_{2n} =T_n^{-2}\sum _{i=1}^na_i{\tilde{x}}_i\left( \sum _{j=1}^n W_{nj}(t_i)\sigma _j e_j\right) ,\\ H_{3n}= & {} T_n^{-2}\sum _{i=1}^na_i{\tilde{x}}_i{\tilde{g}}(t_i). \end{aligned}$$

From (3.2) and \(C_{r}\) inequality, we have

$$\begin{aligned} E|{\tilde{\beta }}_n-\beta |^p\le C(E|H_{1n}|^p+E|H_{2n}|^p+E|H_{3n}|^p). \end{aligned}$$
(3.5)

Note that \(H_{1n}=\sum \nolimits _{i=1}^n(T_n^{-2}a_i\sigma _i{\tilde{x}}_i)e_i\doteq \sum \nolimits _{i=1}^nb_{ni}e_i\), and

$$\begin{aligned} \max _{1\le i\le n}|b_{ni}|\le & {} \max _{1\le i\le n}\frac{|a_i{\tilde{x}}_i|}{T_n}\cdot \max _{1\le i\le n}\sigma _i\cdot \frac{1}{T_n}=O(n^{-1/2}(h(n))^{-r}), \end{aligned}$$
(3.6)
$$\begin{aligned} \sum _{i=1}^n|b_{ni}|\le & {} \sum _{i=1}^n\frac{|a_i{\tilde{x}}_i|}{T_n^2}\cdot \max _{1\le i\le n}\sigma _i=O(1), \end{aligned}$$
(3.7)

by \(H_1(i, ii)\) and (2.3). Hence, we obtain by Lemma A.1 that

$$\begin{aligned} \lim _{n\rightarrow \infty }E|H_{1n}|^p=0. \end{aligned}$$
(3.8)

Observe that \(H_{2n}=\sum \nolimits _{j=1}^n\left( \sum \nolimits _{i=1}^n T_n^{-2}a_i{\tilde{x}}_i W_{nj}(t_i)\sigma _j\right) e_j\doteq \sum \nolimits _{j=1}^nd_{nj}e_j,\) and

$$\begin{aligned} \max _{1\le j\le n}|d_{nj}|\le & {} \max _{1\le j\le n}\sigma _j\cdot \max _{1\le i,j\le n}|W_{nj}(t_i)|\cdot \sum _{i=1}^n\frac{|a_i{\tilde{x}}_i|}{T_n^2}=O(n^{-s}(h(n)^{-r})), \qquad \end{aligned}$$
(3.9)
$$\begin{aligned} \sum _{j=1}^n|d_{nj}|\le & {} \sum _{j=1}^n\left| \sum _{i=1}^n T_n^{-2}a_i{\tilde{x}}_i W_{nj}(t_i)\sigma _j\right| \le \sum _{i=1}^n\sum _{j=1}^n\sigma _j|W_{nj}(t_i)|\frac{|a_i{\tilde{x}}_i|}{T_n^2}\nonumber \\\le & {} \max _{1\le j\le n}\sigma _j\cdot \max _{1\le i\le n}\sum _{j=1}^n|W_{nj}(t_i)|\cdot \sum _{i=1}^n\frac{|a_i{\tilde{x}}_i|}{T_n^2}=O(1), \end{aligned}$$
(3.10)

by \(H_1(ii)\), \(H_5\) and (2.3). Therefore, we have by Lemma A.1 that

$$\begin{aligned} \lim _{n\rightarrow \infty }E|H_{2n}|^p=0. \end{aligned}$$
(3.11)

We now discuss \(H_{3n}\). It follows from \(H_1(iii)\), \(H_2\), \(H_3\) and \(H_4\) that

$$\begin{aligned} H_{3n}\le \left( \max \limits _{1\le i\le n}|{\tilde{g}}(t_i)|\right) \left( T_n^{-2}\sum \limits _{i=1}^n|a_i{\tilde{x}}_i|\right) \end{aligned}$$

and

$$\begin{aligned} \max _{1\le i\le n}|{\tilde{g}}(t_i)|\le & {} \max _{1\le i\le n}|g(t_i)|\left| \sum _{j=1}^nW_{nj}(t_i)-1\right| \nonumber \\&+\,\max _{1\le i\le n}\sum _{j=1}^n|W_{nj}(t_i)||g(t_i)-g(t_j)|I(|t_i-t_j|>a)\nonumber \\&+\,\max _{1\le i\le n}\sum _{j=1}^n|W_{nj}(t_i)||g(t_i)-g(t_j)|I(|t_i-t_j|\le a)\nonumber \\= & {} o(1). \end{aligned}$$
(3.12)

So, we can obtain by (3.12) and (2.3) that

$$\begin{aligned} \lim _{n\rightarrow \infty }E|H_{3n}|^p=0, \end{aligned}$$

which, together with (3.5), (3.8) and (3.11), yields (2.5).

Now we turn to prove (2.7). It can be seen by (3.4) that

$$\begin{aligned}&\max _{1\le i\le n}E|{\tilde{g}}_n(t_i)-g(t_i)|^p\nonumber \\&\quad \le C\max _{1\le i\le n}E\left| \sum _{j=1}^nW_{nj}(t_i)\sigma _j e_j\right| ^p+C\max _{1\le i\le n}\left| \sum _{j=1}^nW_{nj}(t_i)x_j\right| ^p E|{\tilde{\beta }}_n-\beta |^p\nonumber \\&\qquad +\,C\max _{1\le i\le n}|{\tilde{g}}(t_i)|^p\nonumber \\&\quad \doteq Q_{1n}+Q_{2n}+Q_{3n}. \end{aligned}$$
(3.13)

We can obtain from \(H_1(ii)\), \(H_4\) and \(H_5\) that \(Q_{1n}\rightarrow 0,~n\rightarrow \infty \) by applying Lemma A.1. From (2.5) and the assumption \(\max _{1\le j\le n}\left| \sum \nolimits _{i=1}^n W_{ni}(t_j)x_i\right| =O(1)\), we can get \(Q_{2n}\rightarrow 0,~n\rightarrow \infty \). \(Q_{3n}\rightarrow 0,~n\rightarrow \infty \) follows from (3.12). Therefore, the desired result (2.7) follows from (3.13) and \(Q_{1n}\rightarrow 0\), \(Q_{2n}\rightarrow 0\), \(Q_{3n}\rightarrow 0\), \(n\rightarrow \infty \). This completes the proof of the theorem. \(\square \)

Proof of Theorem 2.2

Using the notations in the proof of Theorem 2.1, we know that

$$\begin{aligned} {\tilde{\beta }}_n-\beta =H_{1n}+H_{2n}+H_{3n}. \end{aligned}$$

Applying Lemma A.2, we have by (3.6) and (3.7) that \(H_{1n}\rightarrow 0~a.s.,~n\rightarrow \infty .\) Likewise, by (3.9) and (3.10), \(H_{2n}\rightarrow 0~a.s.,~n\rightarrow \infty \). From (2.3) and (3.12), we can easily obtain that

$$\begin{aligned} H_{3n}\le \left( \max \limits _{1\le i\le n}|\tilde{g}(t_i)|\right) \left( T_n^{-2}\sum \limits _{i=1}^na_i {\tilde{x}}_i\right) \rightarrow 0,\quad n\rightarrow \infty . \end{aligned}$$

So (2.9) is proved. It follows from (3.4) that

$$\begin{aligned}&\max _{1\le i\le n}|{\tilde{g}}_n(t_i)-g(t_i)|\nonumber \\&\quad \le C\max _{1\le i\le n}\left| \sum _{j=1}^nW_{nj}(t_i)\sigma _j e_j\right| +C\max _{1\le i\le n}\left| \sum _{j=1}^nW_{nj}(t_i)x_j\right| |{\tilde{\beta }}_n-\beta |\nonumber \\&\qquad +\,C\max _{1\le i\le n}|{\tilde{g}}(t_i)|\nonumber \\&\quad \doteq R_{1n}+R_{2n}+R_{3n}. \end{aligned}$$
(3.14)

From \(H_1(ii)\), \(H_4\) and \(H_5\), we obtain that \(R_{1n}\rightarrow 0~a.s.,~n\rightarrow \infty \) by applying Lemma A.2. According to (2.9) and the assumption \(\max _{1\le j\le n}\left| \sum _{i=1}^n W_{ni}(t_j)x_i\right| =O(1)\), we can get \(R_{2n}\rightarrow 0~a.s.,~n\rightarrow \infty \). \(R_{3n}\rightarrow 0,~n\rightarrow \infty \) follows from (3.12). Therefore, the desired result (2.11) follows from (3.14) and \(R_{1n}\rightarrow 0~a.s.\), \(R_{2n}\rightarrow 0~a.s.\), \(R_{3n}\rightarrow 0\), \(n\rightarrow \infty \). The proof is completed. \(\square \)