1 Introduction

The analysis of panel data is the subject of one of the most active and innovative bodies of literature in econometrics. Panel data sets have various advantages over that of pure time-series or cross-sectional data sets, among which the most important one is perhaps that the panel data provide researchers a flexible way to model both heterogeneity among cross-sectional units and possible structural changes over time. Arellano (2003), Baltagi (2005) and Hsiao (2003) provided excellent overviews of statistical inference and econometric analysis of parametric panel data models. However, a misspecified parametric panel data model may result in misleading inference. Therefore, econometricians and statisticians have developed some flexible nonparametric and semi-parametric panel data models. For example, Su and Ullah (2007) proposed a class of two-step estimators for nonparametric panel data with random effects. Cai and Li (2008) studied dynamic nonparametric panel data models. Henderson et al. (2008) considered nonparametric panel data model with fixed effects. Rodriguez-Poo and Soberon (2014) considered varying coefficient fixed effects panel data models, established direct semiparametric estimations. Chen et al. (2013) studied partially linear single-index panel data models with fixed effects, proposed a dummy variable method to remove fixed effects and established a semi-parametric minimum average variance estimation procedure. Baltagi and Li (2002) discussed partially linear panel data models with fixed effects, developed the series estimation procedure and the profile likelihood estimation technique. Hu (2014) proposed the profile likelihood procedure to estimate semi-varying coefficient model for panel data with fixed effects. The partially linear panel data models with fixed effects are widely used in econometric analysis; see, e.g., Henderson et al. (2008), Horowitz and Lee (2004), Hu (2017) and Li et al. (2011).

In this paper, we consider the following partially linear panel data models with fixed effects (e.g. Su and Ullah 2006):

$$\begin{aligned} Y_{it}=X^\tau _{it}\beta +g(Z_{it})+\mu _{i}+\varepsilon _{it},~~i=1,\ldots ,n,~~t=1,\ldots ,T, \end{aligned}$$
(1.1)

where \(Y_{it}\) is the response, \((X_{it},Z_{it})\in R^p\times R\) are strictly exogenous variables, \({\beta }=(\beta _1,\ldots ,\beta _p)^\tau \) is a vector of p-dimensional unknown parameters, and the superscript \(\tau \) denotes the transpose of a vector or matrix. \(g(Z_{it})\) is a unknown functions and \(\mu _{i}\) is the unobserved individual effects, \(\varepsilon _{it}\) is the random model error. Here, we assume \(\varepsilon _{it}\) to be i.i.d. with zero mean and finite variance \(\sigma ^2>0\). We allow \(\mu _i\) to be correlated with \(X_{it}\), and \(Z_{it}\) with an unknown correlation structure. Hence, model (1.1) is a fixed effects model.

It is well known that in many fields, such as engineering, economics, biology, biomedical sciences and epidemiology, observations are measured with error. For example, urinary sodium chloride level (Wang et al. 1996) and serum cholesterol level (Carroll et al. 1995) are often subjects to measurement errors. Simply ignoring measurement errors (errors-in-variables), known as the naive method, would result in biased estimators. Handing the measurement errors in covariates is generally a challenge for statistical analysis. For the past two decades, errors-in-variables can be handled by means of corrected score function (Nakamura 1990), corrected likelihood method (Hanfelt and Liang 1997), the instrumental variables estimation approach (Schennach 2007) and so on.

Specifically, we consider the following partially linear errors-in-variables panel data models with fixed effects

$$\begin{aligned} \left\{ \begin{aligned} Y_{it}&= X_{it}^{\tau }\beta +g(Z_{it})+\mu _{i}+\varepsilon _{it}, \\ W_{it}&=X_{it}+\nu _{it}, \end{aligned} \right. \quad i=1,\ldots ,n;~ t=1, \ldots ,T. \end{aligned}$$
(1.2)

where the covariate variables \(X_{it}\) are measured with additive error and are not directly observable. Instead, \(X_{it}\) are observed \(W_{it}=X_{it}+\nu _{it}\), where the measurement errors \(\nu _{it}\) are independent and identically distributed, independent of \((X_{it},Z_{it},\varepsilon _{it})\), with mean zero and covariance matrix \(\Sigma _{\nu }\). We will assume that \(\Sigma _{\nu }\) is known, as in the papers of Zhu and Cui (2003) and You and Chen (2006) and other. When \(\Sigma _{\nu }\) is unknown, we can estimate it by repeatedly measuring \(W_{it}\); see Liang et al. (1999) and Fan et al. (2013) for details.

It is well-known that high-dimensional data analysis arises frequently in many contemporary statistical studies. The emergence of high-dimensional data, such as the gene expression values in microarray, brings challenges to many traditional statistical methods and theory. One important aspect of the high-dimensional data under the regression setting is that the number of covariates is diverging. When dimensionality diverges, variable selection through regularization has proven to be effective. As argued in Hastie et al. (2009) and Fan and Lv (2008), penalized likelihood can properly adjust the bias-variance trade-off so that the performance improvement can be achieved; Various powerful penalization methods have been developed for variable selection. Fan and Li (2001) proposed a unified approach via nonconcave penalized least squares to automatically and simultaneously select variables. Li and Liang (2008) developed the nonconcave penalized quasilikelihood method for variable selection in semiparametric regression model. Recently, a new and efficient variable selection approach, PEL introduced for the first time by Tang and Leng (2010), was applied to analyze mean vector in multivariate analysis and regression coefficients in linear models with diverging number of parameters. As demonstrated in Tang and Leng (2010), the PEL has merits in both efficiency and adaptivity stemming from a nonparametric likelihood method. Also, the PEL method possesses the same merit of the empirical likelihood (EL) which only uses the data to determine the shape and orientation of confidence regions and without estimating the complex covariance. As far as we know, there are a few papers related to the PEL approach, such as Ren and Zhang (2011) proposed the PEL approach for variable selection in moment restriction models; Leng and Tang (2012) applied the PEL approach to parametric estimation and variable selection for general estimating equations; Wang and Xiang (2017) studied PEL inference for sparse additive hazards regression with a diverging number of covariates.

It is worth pointing out that there is no result available in the literature when the number of covariates is diverging in partially linear errors-in-variables panel data models with fixed effects. In this paper, our aim is to extend the results in Fan et al. (2016) for high-dimensional partially linear varying coefficient model with measurement errors to partially linear error-in-variables panel data models with fixed effects. Our contribution can be summarized as follows. Following the estimation procedure proposed by Fan et al. (2005), we first adapt a local linear dummy variable approach to remove the unknown fixed effects. Moreover, we utilize the EL method to construct confidence regions of unknown parameter and establish asymptotic normality of maximum empirical likelihood (MEL) estimator of the parameter. At last, for building sparse models, we propose an estimating equation-based PEL, a unified framework for variable selection in optimally combining estimating equations. More specifically, this method has the oracle property. Moreover, PEL ratio statistic shows the Wilks’ phenomenon, facilitating hypothesis testing and constructing confidence regions.

The layout of the remainder of this paper is as follows. In Sect. 2, we construct corrected-attenuation EL ratio and test statistic as well as define the MEL and PEL estimators of the parameter and give their asymptotic properties. Moreover, empirical log-likelihood ratio for the nonparametric part is also investigated. Finally, we briefly introduce computational algorithm. The simulated example is provided in Sect. 3. Section 4 summarizes some conclusions and discusses future research. Assumption conditions and the proofs of the asymptotic results are given in Appendix.

2 Methodology and main results

2.1 Modified empirical likelihood

We give vector and matrix notations in the following. Let \(Y=(Y_1^{\tau },\ldots ,Y_n^{\tau })^{\tau }\), \(X=(X_1^{\tau },\ldots , X_n^{\tau })^{\tau }\), \(Z=(Z_1^{\tau }, \ldots ,Z_n^{\tau })^{\tau }\), \(\mu _0=(\mu _1^{\tau },\ldots ,\mu _n^{\tau })^{\tau }\) and \(\varepsilon =(\varepsilon _1^{\tau },\ldots ,\varepsilon _n^{\tau })^{\tau }\) be \(nT\times 1\) vectors, where \(Y_i=(Y_{i1},\ldots ,Y_{iT})^{\tau }\), \(X_i=(X_{i1},\ldots ,X_{iT})^{\tau }\), \(Z_i=(Z_{i1},\ldots ,Z_{iT})^{\tau }\), \(\varepsilon _i=(\varepsilon _{i1},\ldots ,\varepsilon _{iT})^{\tau }\), and \(D_0=I_{n}\otimes i_T \) with \(\otimes \) the Kronecker product, \(I_n\) denotes the \(n\times n\) identity matrix, and \(i_n\) denotes the \(n\times 1\) vector of ones. We rewrite model (1.1) in a matrix format which yields

$$\begin{aligned} Y=X\beta +g(Z)+D_0\mu _0+\varepsilon . \end{aligned}$$
(2.1)

For the identification purpose, we impose the restriction \(\sum ^n_{i=1}\mu _i=0\). Letting \(D=[-i_{n-1}\quad I_{n-1}]\otimes i_T \) and \(\mu =(\mu _2,\ldots ,\mu _n)^\tau \), model (2.1) can then be rewritten as

$$\begin{aligned} Y=X\beta +g(Z)+D\mu +\varepsilon . \end{aligned}$$
(2.2)

Let \(G_{it}(z,h)=(1, (Z_{it}-z)/h)^{\tau }\), \(K_h(z)=K(\cdot /h)/h\) with a kernel function \(K(\cdot )\) and a bandwidth h. The diagonal matrices

$$\begin{aligned} K_h(Z_i,z)= & {} \left[ \begin{array}{ccc} K_h(Z_{i1},z) &{}\quad \cdots &{}\quad 0\\ \vdots &{}\quad \ddots &{}\quad \vdots \\ 0&{}\quad \cdots &{}\quad K_h(Z_{iT},z)\\ \end{array}\right] ,~~\\ W_h(u)= & {} \left[ \begin{array}{ccc} K_h(Z_{1},z) &{}\quad \cdots &{}\quad 0\\ \vdots &{}\quad \ddots &{}\quad \vdots \\ 0&{}\quad \cdots &{}\quad K_h(Z_{n},z)\\ \end{array}\right] , \end{aligned}$$

and \(\zeta =(\mu ^\tau ,\beta ^\tau )^\tau \). Given \(\zeta \), we can estimate the functions \(g(\cdot )=(g(z),\{hg'(z)\}^{\tau })^{\tau }\) by

$$\begin{aligned} \sum _{i=1}^n\sum _{t=1}^T\bigg \{\Big (Y_{it}-X^\tau _{it}\beta -\mu _{i}\Big )-\Big [g(z)+hg'(z)(Z_{it}-z)\Big ]\bigg \}^2K_{h}(Z_{it}-z). \end{aligned}$$
(2.3)

Let \(G(z,h)=[G_1^{\tau }(z,h),\ldots , G_n^{\tau }(z,H)]^{\tau }\), \(G_i(z,h)=(G_{i1}(z,h),\ldots , G_{iT}(z,h))^{\tau },\) \(g'(z)=\partial g(z)/\partial z\) and \(z=z_{it}\) is in a neighborhood of \(Z_{it}\). Then the solution of problem (2.3) is given by

$$\begin{aligned} g(\cdot )= & {} [G^\tau (z,h) W_H(u)G(z,h)]^{-1}G^\tau (z,h) W_H(u)(Y-X\beta -D\mu )\\= & {} S(z,H)(Y-X\beta -D\mu ). \end{aligned}$$

In particular, the estimator for g(z) is given by

$$\begin{aligned} \hat{g}(z)=s(z,h)(Y-X\beta -D\mu ). \end{aligned}$$
(2.4)

where \(s(z,h)=(1~~0)S(z,h)\).

Now we consider a way of removing the unknown fixed effects motivated by a least squares dummy variable model in parametric panel data analysis, for which we solve the following optimization problem:

$$\begin{aligned} \hat{\zeta }=\mathrm{arg}\min _{\zeta }\big [Y-X\beta -D\mu -S(Y-X\beta -D\mu )]^\tau [Y-X\beta -D\mu -S(Y-X\beta -D\mu )], \end{aligned}$$
(2.5)

where the smoothing matrix S is

$$\begin{aligned} S= & {} \left( \begin{array}{ccc} (1~~0)[G^\tau (Z_{11},h)W_{11}(z,h)G({Z_{11},h})]^{-1}G^\tau (Z_{11},h)W_{11}(z,h)\\ \vdots \\ (1~~0)[G^\tau (Z_{1T},h)W_{1T}(z,h)G(Z_{1T},h)]^{-1}G^\tau (Z_{11},h)W_{1T}(z,h)\\ \vdots \\ (1~~0)[G^\tau (Z_{nT},h)W_{nT}(z,h)G(Z_{nT},h)]^{-1}G^\tau (Z_{nT},h)W_{nT}(z,h)\\ \end{array}\right) \\= & {} \left( \begin{array}{ccc} S_{11} \\ \vdots \\ S_{1T} \\ \vdots \\ S_{nT} \end{array}\right) . \end{aligned}$$

Supposing that \(\widetilde{X}=(I_{nT}-S)X\), \(\widetilde{Y}=(I_{nT}-S)Y\), \(\widetilde{D}=(I_{nT}-S)D\), we have \(\widetilde{\mu }=(\widetilde{D}^{\tau }\widetilde{D})^{-1}\widetilde{D}^{\tau }(\widetilde{Y}-\widetilde{X}\beta )\). Let \(H=I_{nT}-\widetilde{D}(\widetilde{D}^{\tau }\widetilde{D})^{-1}\widetilde{D}^{\tau },\) we can obtain \(H\widetilde{D}\mu =0\). Hence, the fixed effects term \(D\mu \) is eliminated in (2.3). Let \(e_{it}\) be the \(nT\times 1\) vector with its \(\{(i-1)T+t\}\)th element being 1 and others 0. We state the approximate residuals as the following:

$$\begin{aligned} \Lambda _i(\beta )=\sum _{t=1}^T\widetilde{X}_{it}^{\tau }H(\widetilde{Y}_{it}-\widetilde{X}_{it}\beta ), ~~~i=1,\ldots ,n. \end{aligned}$$
(2.6)

However, \(X_{it}\)’s can not be observed in our case and we just have \(W_{it}\). If we ignore the measurement error and replace \(X_{it}\) with \(W_{it}\) in (2.4) directly, (2.4) can be used to show that the resulting estimate is inconsistent. It is well known that in linear regression or partially linear regression, inconsistency caused by the measurement error can be overcome by applying the so-called “correction for attenuation”, see Liang et al. (1999) and Hu et al. (2009) for more details.

$$\begin{aligned} \Gamma _i(\beta )=\sum _{t=1}^T\widetilde{W}_{it}^{\tau }H(\widetilde{Y}_{it}-\widetilde{W}_{it}\beta )-(T-1)\Sigma _{\nu }\beta , ~~~i=1,\ldots ,n. \end{aligned}$$
(2.7)

Note that \(E(\Gamma _i(\beta ))=0\), if \(\beta \) is the true parameter. Therefore, similar to Owen (1990), we define a corrected-attenuation empirical likelihood (CAEL) ratio of \(\beta \) as.

$$\begin{aligned} R_{n}(\beta )=-\max \left\{ \sum _{i=1}^n\ln (np_{i})| p_{i}\ge 0,\sum _{i=1}^np_{i}=1,\sum _{i=1}^np_{i} \Gamma _i(\beta )=0\right\} . \end{aligned}$$
(2.8)

With the assumption that 0 is inside the convex hull of the point \((\Gamma _1(\beta ),\ldots ,\Gamma _n(\beta ))\), a unique value for \(R_{n}(\beta )\) exists. By the Lagrange multiplier method, one can obtain that

$$\begin{aligned} R_{n}(\beta )=\sum _{i=1}^n\ln \{1+\gamma ^\tau \Gamma _i(\beta )\}, \end{aligned}$$
(2.9)

where \(\gamma \) is determined by

$$\begin{aligned} \frac{1}{n}\sum _{i=1}^n\frac{\Gamma _i(\beta )}{1+\gamma ^\tau \Gamma _i(\beta )}=0. \end{aligned}$$
(2.10)

Theorem 2.1

Suppose that the conditions of (B1)–(B7) in the Appendix hold. Further assume that \(E(\varepsilon ^3|X,Z)=0\) almost surely or \(k\ge 8\). If \(\beta _0\) is the true value of the parameter vector and \({\mathop {\rightarrow }\limits ^{d}}\) stands for convergence in distribution, \(p^{3+2/(k-2)}/n\rightarrow 0\) as \(n\rightarrow \infty \), then \((2p)^{-1/2}(2R_{n}(\beta _0)-p){\mathop {\rightarrow }\limits ^{d}}N(0,1)\).

Define \(\hat{\beta }_{ME}=\arg \min _{\beta }R_{n}(\beta )\), which is the MEL estimator of the parameter \(\beta \).

Theorem 2.2

Under the conditions of Theorem 2.1, we have

$$\begin{aligned} \sqrt{n}A_n\Omega ^{-1/2}(\widehat{\beta }_{ME}-\beta _0){\mathop {\rightarrow }\limits ^{d}}N(0,\Delta ). \end{aligned}$$

\(A_n\) represents a projection of the diverging dimensional vector to a fixed dimension s, and \(A_n\) is a \(s\times p\) matrix such that \(A_nA_n^{\tau }\rightarrow \Delta \),\(\Delta \) is a \(s\times s\) nonnegative symmetric matrix with fixed s, and \(\Omega =\Sigma _0^{-1}\Sigma _1\Sigma _0^{-1}\), \(\Sigma _1=(T-1)\big \{E(\varepsilon _{11}-\nu _{11}\beta _0)^2\Sigma _2 +\sigma ^2\Sigma _{\nu }+ E[(\nu _{11}\nu _{11}^{\tau }-\Sigma _{\nu })\beta _0]^2\big \}\), \(\Sigma _2=E\{[{X_{11}}-E({X_{11}}|Z_{11})]^{\tau }[{X_{11}}-E({X_{11}}|Z_{11})]\}\) and \(\Sigma _0=(T-1)\Sigma _2\).

Further, \(\widehat{\Sigma }_2^{-1}\widehat{\Sigma }_1\widehat{\Sigma }_2^{-1}\) is a consistent estimator of \(\Sigma _2^{-1}\Sigma _1\Sigma _2^{-1}\) where \(\widehat{\Sigma }_2=\frac{1}{n}\widetilde{W}^{\tau }H\widetilde{W}-(T-1)\Sigma _{\nu }\) and \(\widehat{\Sigma }_1=\{\frac{1}{n}\sum _{i=1}^n\sum _{t=1}^T[\widetilde{W}_{it}^{\tau }H(\widetilde{Y}_{it}-\widetilde{W}_{it}\widehat{\beta })]+(T-1)(\Sigma _{\nu }\widehat{\beta })\}^{\oplus 2}\) and \(\mathbf {A}^{\oplus 2}\) means \(\mathbf {A}\mathbf {A}^{\tau }\). By Theorem 2.2, we obtain that

$$\begin{aligned} (\widehat{\beta }_{ME}-\beta _0)^{\tau }n[\widehat{\Sigma }_2^{-1}\widehat{\Sigma }_1\widehat{\Sigma }_2^{-1}]^{-1}(\widehat{\beta }_{ME}-\beta _0){\mathop {\rightarrow }\limits ^{d}}\chi _s^2. \end{aligned}$$

2.2 Penalized empirical likelihood for variable selection

We use the PEL by combining the profile likelihood method and the smoothly clipped absolute deviation (SCAD) penalized approach. The SCAD penalty is defined in terms of its first derivative.

We define the penalized empirical likelihood (PEL) as follows,

$$\begin{aligned} \mathcal {L}_{n}(\beta )= R_{n}(\beta )+n\sum _{j=1}^pp_{\lambda }(|\beta _j|), \end{aligned}$$
(2.11)

where \(p_{\lambda }(\cdot )\) is a penalty function with tuning parameter \(\lambda \). See Fan and Li (2001) for example of this function. In this paper, we use the smoothly clipped absolute deviation penalty, whose first derivative satisfies

$$\begin{aligned} p_{\lambda }'(\theta )=\theta \Big \{I(\theta \le \lambda )+\frac{(a\lambda -\theta )_{+}}{(a-1)\lambda }I(\theta>\lambda )\Big \},~~\theta>0,~\lambda >0, \end{aligned}$$
(2.12)

for some \(a>2\) and \(p_{\lambda }'(0)=0\). Following Fan and Li (2001), we set \(a=3.7\) in our work.

Maximizing the PEL function (2.8) is equivalent to minimizing

$$\begin{aligned} \mathcal {L}_{n}(\beta )= \sum _{i=1}^n\ln \{1+\gamma ^\tau \Gamma _i(\beta )\}+n\sum _{j=1}^pp_{\lambda }(|\beta _j|), \end{aligned}$$
(2.13)

Let \(\mathcal {B}=\{j:\beta _{0j}\}\) be the set of nonzero components of the true parameter vector \(\beta _0\) and its cardinality \(|\mathcal {B}|=d\) where d is allowed grow as \(n\rightarrow \infty \). Without loss of generality, one can partition the parameter vector as \(\beta =(\beta _1^{\tau },\beta _2^{\tau })^{\tau }\) where \(\beta _1\in \mathbb {R}^d\) and \(\beta _2\in \mathbb {R}^{p-d}\). Hence, the true parameter \(\beta _0=(\beta _{10}^{\tau },0^{\tau })^{\tau }\) and we write \(\widehat{\beta }=(\widehat{\beta }_1^{\tau },\widehat{\beta }_2^{\tau })^{\tau }\) called PEL estimator which is the minimizer of (2.13). The matrix \(\Omega \) can be decomposed as a block matrix according to the arrangement of \(\beta _0\) as \(\Omega =\left( \begin{array}{cc} \Omega _{11}&{} \Omega _{12}\\ \Omega _{21} &{} \Omega _{22} \end{array}\right) .\)

Theorem 2.3

Suppose that Assumptions (B1)–(B9) hold. If \(p^5/n\rightarrow 0\), then with probability tending to 1, the PEL estimator \(\widehat{\beta }\) satisfies

  1. (a)

    (Sparity): \(\widehat{\beta }_2=0\);

  2. (b)

    (Asymptotic normality): \(n^{1/2}W_n\Omega _p^{-1/2}(\widehat{\beta }_1-\beta _{10}){\mathop {\rightarrow }\limits ^{d}}N(0,G)\), where \(W_n\in R^{q \times d}\) such that \( W_n W_n^{\tau }\rightarrow G\) for \(G\in R^{q\times q}\) with fixed q and \(\Omega _p=\Omega _{11}-\Omega _{12}\Omega _{22}^{-1}\Omega _{21}\).

A remarkable advantage of PEL lies in testing hypotheses and constructing confidence regions for \(\beta \). To understand this more clearly, we consider the problem of testing linear hypothesis:

$$\begin{aligned} H_0:L_n\beta _{10}=0~~~vs~~~H_1:L_n\beta _{10}\ne 0 \end{aligned}$$

where \(L_n\) is \(q\times s\) matrix such that \(L_nL_n^{\tau }=I_q\) for a fixed and finite q. A nonparametric profile likelihood ratio statistic is constructed as

$$\begin{aligned} \widetilde{\mathcal {L}}_{n}(\hat{\beta })=-\left\{ \mathcal {L}_{n}(\hat{\beta })-\mathop {\min }\limits _{\beta ,L_n\beta _{10}=0}\mathcal {L}_{n}(\beta )\right\} . \end{aligned}$$

We summarize the property of the test statistic in the following theorem.

Theorem 2.4

Under the conditions of Theorem 2.3. Then under the null hypothesis \(H_0\), we have

$$\begin{aligned} 2\widetilde{\mathcal {L}}_{n}(\hat{\beta }){\mathop {\rightarrow }\limits ^{d}}\chi _q^2, ~~as~ n\rightarrow \infty . \end{aligned}$$

As a consequence of the theorem, confidence regions for the parameter \(\beta \) can be constructed. More precisely, for any \( 0\le \alpha <1\), let \(c_{\alpha }\) be such that \(P(\chi _q^2>c_{\alpha })\le 1-\alpha \). Then \(\ell _{PEL}(\alpha )=\{\beta \in R^p:\widetilde{\mathcal {L}}_{n}({\beta })\le c_{\alpha }\}\) constitutes a confidence region for \(\beta \) with asymptotic coverage \(\alpha \) because the event that \(\beta \) belongs to \(\ell _{PEL}(\alpha )\) is equivalent to the event that \(\widetilde{\mathcal {L}}_{n}({\beta })\le c_{\alpha }\).

2.3 Empirical likelihood for the nonparametric part

For model (2.2), we solve the the following optimization problem:

$$\begin{aligned} \hat{\mu }=\mathrm{arg}\min _{\mu }\big [Y-X\beta -g(Z)-D\mu ]^\tau [Y-X\beta -g(Z)-D\mu ], \end{aligned}$$

we have \(\mu =({D}^{\tau }{D})^{-1}{D}^{\tau }[Y-X\beta -g(Z)]\). For given \(\beta \) and z, an auxiliary random vector for nonparametric part can be stated as

$$\begin{aligned} \Xi _{i}\{g(z)\}=\sum _{t=1}^T(I_{it}-Q)[Y_{it}-X_{it}\beta -g(z)]K_h(Z_{it}-z),~~~i=1,\ldots ,n. \end{aligned}$$

where \(Q=D({D}^{\tau }{D})^{-1}{D}^{\tau }\). Note that \(E[\Xi _{i}\{g(z)\}]=0\) if g(z) is the true parameter. Thus we can define an empirical log-likelihood ratio statistic for g(z) by using the methodology in Owen (1988). We introduce an adjusted auxiliary random vector for g(z) as follows.

$$\begin{aligned}&\hat{\Xi }_{i}\{g(z)\}=\sum _{t=1}^T(I_{it}-Q)[Y_{it}-X_{it}\hat{\beta }-g(z)-\{\hat{g}(Z_{it})-\hat{g}(z)\}]K_h(Z_{it}-z),\\&\quad i=1,\ldots ,n. \end{aligned}$$

By the adjustment in \(\hat{\Xi }_{i}\{g(z)\}\), we not only correct the bias, but also avoid undersmoothing the function g(z), as proved in the Appendix. The adjusted empirical log-likelihood ratio for g(z) can be defined as

$$\begin{aligned} \mathcal {Q}_{n}(g(z))=-\max \left\{ \sum _{i=1}^n\ln (n\tilde{p}_{i})| \tilde{p}_{i}\ge 0,\sum _{i=1}^n\tilde{p}_{i}=1,\sum _{i=1}^n\tilde{p}_{i} \hat{\Xi }_{i}\{g(z)\}=0\right\} . \end{aligned}$$

By the Lagrange multiplier method, one can obtain that

$$\begin{aligned} \mathcal {Q}_{n}(g(z))=\sum _{i=1}^n\ln \{1+\phi ^\tau \hat{\Xi }_{i}\{g(z)\}\}, \end{aligned}$$

where \(\phi \) is determined by

$$\begin{aligned} \frac{1}{n}\sum _{i=1}^n\frac{\hat{\Xi }_{i}\{g(z)\}}{1+\phi ^\tau \hat{\Xi }_{i}\{g(z)\}}=0. \end{aligned}$$

Theorem 2.5

Suppose that the conditions of (B1)–(B9) in the Appendix hold. For a given \(z\in \mathcal {Z}\), if g(z) is the true value of the parameter, then

$$\begin{aligned} 2\mathcal {Q}_{n}(g(z)){\mathop {\rightarrow }\limits ^{d}}\chi _1^2. \end{aligned}$$

2.4 Computational algorithm

This section employs the local quadratic approximation algorithm to obtain the minimizer of PEL ratio defined by (2.13). Specifically, for each \(j=1,\ldots , p, [p_{\lambda }(|\beta _j|)]'\) can be locally approximated by the quadratic function defined as \([p_{\lambda }(|\beta _j|)]'=p'_{\lambda }(|\beta _j|)sgn(\beta _j)\approx \{p'_{\lambda }(|\beta _{j0}|/|\beta _{j0}|)\}\beta _j\) at an initial value \(\beta _{j0}\) of \(\beta _j\) is not close to 0; otherwise, we set \(\hat{\beta }_j=0\). In other words, in a neighborhood of a given nonzero \(\beta _{j0}\), we assume that \(p_{\lambda }(|\beta _{j}|)\approx p_{\lambda }(|\beta _{j0}|)+\frac{1}{2}\{p'_{\lambda }(|\beta _{j0}|/|\beta _{j0}|\}(\beta _j^2-\beta _{j0}^2)\). We then make use of algorithm (see Owen, 2001) to obtain the minimum through nonlinear optimization. The procedure is repeated until convergence.

We apply the following Bayesian information criterion (BIC) to select the tuning parameter \(\lambda \), which is defined by

$$\begin{aligned} BIC(\lambda )=-2\mathcal {L}_{n}(\beta _{\lambda })+\ln (n)df_{\lambda }, \end{aligned}$$

where \(df_{\lambda }\) is the number of nonzero estimated parameters. Then the optimal tuning parameter is the minimizer of the BIC.

3 Simulation studies

In this section, we carry out some simulation to study the finite sample performance of our proposed method. Throughout this section, we choose the Epanechnikov kernel \(K(u)=\frac{3}{4}(1-u^2)I\{|u|\le 1\}\) and use the “leave-one-subject-out” cross-validation bandwidth method to select the optimal handwidth \(h_{opt}\).

Firstly, we consider the following partially linear errors-in-variables panel data models with fixed effects:

$$\begin{aligned} \left\{ \begin{aligned} Y_{it}&= X_{it}^{\tau }\beta +g(Z_{it})+\mu _{i}+\varepsilon _{it}, \\ W_{it}&=X_{it}+\nu _{it} \end{aligned} \right. \quad i=1,\ldots ,n;~ t=1, \ldots ,T. \end{aligned}$$
(3.1)

where \(\beta =(3,1.5,0,0,2,0,\ldots ,0)^{\tau }\), \(g(Z_{it})=\cos (2\pi Z_{it})\), \(Z_{it}{\sim }U(0,1)\), \(\mu _i=\frac{1}{2}\bar{Z_i}+w_i\) and \(w_i{\sim }N(0,0.1^2)\) for \(i=1,2,\ldots ,n\), and \( \bar{Z_i}=\frac{1}{T}\sum _{t=1}^{T}Z_{it}.\) The measurement error \(\nu _{it}{\sim }N(0,\Sigma _{\nu })\) where we take \(\Sigma _{\nu }=0.2^2I_{10}\) and \(0.4^2I_{10}\) to represent different levels of measurement error. The covariate \(X_{it}\) is a p-dimensional normal distribution random vector with mean zero and covariance matrix cov\((X_{it},X_{jt})=0.5^{|i-j|}\).

Table 1 Comparison of coverage probability for MEL, PEL, NMEL and NPEL

In our simulations, we take p as the integer part of \(10(6n)^{1/5.1}-20\) and the sample sizes \((n,T)=(50,4), (50,6)\) and (100, 6), respectively. In order to show the performance of the proposed methods, we compare MEL and PEL estimators with the native maximum empirical likelihood (NMEL) and native penalized empirical likelihood (NPEL) estimators that the neglecting the measurement errors with a direct replacement of X by W in our proposed estimators. In each case the number of simulated realizations is 500.

Seen from Table 1, when the nominal level is 0.9 and 0.95, shows the coverage probability of confidence region for the whole \(\beta \) constructed by MEL and PEL method, respectively. From the results, we can see that the PEL confidence region has slightly higher coverage probability than the NEL confidence region, and the coverage probability tends to the nominal level as the sample size increases.

From Table 2, we can see the average model errors (ME) and the standard deviations (SD) of the \(\beta _1\) that is nonzero components of \(\beta \). based on PEL and MEL estimators decreases as the sample size increases and the PEL estimator gives the smallest ME and SD among the estimators based on PEL, MEL, NPEL and NMEL methods for all settings. The ME is defined as \(ME(\hat{\beta _1})=(\hat{\beta _1}-\beta _1)^{\tau }E(X^{\tau }X)(\hat{\beta _1}-\beta _1).\)

Table 3 summaries the variable selection results, where important variable have large effects. The column labeled “C” gives average number of correct zeros and column labeled “I” gives the average number of incorrect zeros. From Table 3, it can be seen that variable selection method based on the PEL select all three true predictors and the average number of correct zeros are close to \(p-5\) in all settings. Further the smaller measurement errors lead to better performance. It can also be seen that the PEL approach perform better than the NPEL method for all settings. These findings imply that the model selection result based on the PEL approach effectively reduces model complexity and the selected model is very close to the true model in terms of nonzero coefficients.

Table 2 ME and SD of \(\beta _1\) for MEL, PEL, NPEL and NMEL estimators
Table 3 Simulation results for variable selection selection based on the PEL and NPEL methods

From Fig. 1, We see that the method based on the EL performs slighter better than the NA method since the EL method gives shorter confidence intervals than the NA method which is shown in Theorem 4 in Xue and Zhu (2008). Besides, interestingly, seen from Fig. 1, \(\Sigma _v=0.2\) gives shorter confidence intervals and narrower confidence bands than \(\Sigma _v=0.4\) for g(z). This shows the empirical likelihood ratio generally works well.

Fig. 1
figure 1

95% confidence intervals for g(z) for \(\Sigma _v=0.2\) (left panel) and \(\Sigma _v=0.4\) (right panel) based on EL (dotted curve) and NA (dot-dashed curve). The solid curve is the estimated cure of g(z)

4 Conclusion remarks

The partially linear panel data models with fixed effects has received a lot of attention. But there have been few studies about partially linear errors-in-variables panel data models with fixed effects. We apply empirical likelihood both for parameter and nonparametric parts. Moreover, we propose PEL and variable selection procedure for the parameter with diverging numbers of parameters. By using an appropriate penalty function, we show that PEL estimators has the oracle property. Also, we introduce the PEL ratio statistic to test a linear hypothesis of the parameter and prove it follows an asymptotically chi-square distribution under the null hypothesis. We conduct simulation studies to demonstrate the finite sample performance of our proposed method. Still, more work is needed to extend the method to more complex settings, including errors-in-function, cross-sectional dependence and spatial panel data model. The results presented in this paper provide the foundation for additional work in these directions.