- Research
- Open access
- Published:
Empirical likelihood inference of parameters in nonlinear EV models with censored data
Journal of Inequalities and Applications volume 2015, Article number: 379 (2015)
Abstract
In this paper, the authors consider nonlinear error-in-variables regression models with right censored data. Based on the validation data, the authors use the Kaplan-Meier estimate and kernel estimate to repair the censored response variable Y and explanatory X with measurement error. Based on the repaired data, the authors introduce an auxiliary vector suitable to define an estimated empirical log-likelihood function of the unknown parameter which has an asymptotic weighted sum of the \(\chi_{1}^{2}\) variables. Using the result the authors can construct the asymptotic confidence regions of β. But the method of estimating weights will reduce the precision of the confidence regions. Further, the authors adjust the preceding log-likelihood, and it is shown that the adjusted empirical log-likelihood has the asymptotic standard \(\chi_{1}^{2}\) distribution. The result can be used to construct the confidence regions of β.
1 Introduction
Consider the following nonlinear model:
where X is p-variate explanatory variable, Y is a scalar response variable, and where \(\beta=(\beta_{1},\beta_{2},\ldots,\beta_{p})^{\tau }\) is a \(p\times1\) column vector of the unknown regression parameter, \(g(\cdot)\) is a known measurable function, and ε is a random statistical error. In this model, X is an explanatory variable which cannot be observed directly, and X̃ is the observable substitute variable of X, where e is a measurement error, \(\phi (\cdot)\) is an unconditional known function. Usually, there is a complicated relationship between X̃ and X. However, this situation presents serious difficulties toward obtaining a correct statistical analysis. One solution is to use validation data, and the solution has gained much attention by many scholars in recent years. For example: Sepanki and Lee [1] considered error-in-covariable nonlinear models with the help of validation data. Wang [2–4] considered the empirical likelihood inference of the error-in-covariable linear model and partially linear model based on validation data. Xue [5] used the validation data to explore the empirical likelihood inference of nonlinear semiparasitic error-in-variable models. Fang and Hu [6] considered the error-in-response to be nonlinear addressing the dimension reduction estimation of β with validation data.
In practice, the response Y’s may be censored by the censoring variable C. So we cannot observe \((X_{i},Y_{i})\), but we may observe
with \(Z_{i}=\min(Y_{i},C_{i})\), \(\delta_{i}=I(Y_{i}\leq C_{i})\), where \(I(\cdot)\) is the indicator function and \(\delta=0\) if \(Y_{i}\) is censored, otherwise \(\delta=1\). \((\widetilde{X}_{i},Z_{i},\delta_{i})\), \(i=1,\ldots ,n\), are i.i.d. random samples from \((\widetilde{X},Z,\delta)\). Suppose that X̃, Z, are given and δ is independent of Y. In fact, regression models with censored data have been researched in much of the literature. For example, Buckley and James [7] made an unbiased change for data and established a new regression model. Based on [7], Li and Wang [8] supposed the censored variable to be independent of the covariable and applied a new data change to proceed with the empirical likelihood inference in linear models. Cheng et al. [9] developed the empirical likelihood inference in nonlinear models with a right censored response based on the validation data.
In this paper, we consider the model (1.1), where the response Y is randomly censored, and explanatory X with measurement error. We try to construct the confidence region for the parameter β. Similarly, we propose to use the empirical likelihood method to construct the confidence region for β. A typical nonparametric approach to this problem generally includes the following steps: (1) repair the incomplete data with the help of validation data and derive an β̂ to estimate β, (2) define the empirical likelihood function, (3) invert the confidence region by the limiting \(\chi _{1}^{2}\) distribution. To complete these steps, we will carry out the following approach. First, based on the validation data, we use the Kaplan-Meier estimate and the kernel estimate to repair the censored response variable Y and explanatory X, which has been measured erroneously. Second, based on the repaired data, we introduce an auxiliary vector suitable to define an estimated empirical log-likelihood function of the unknown parameter which has an asymptotic weighted sum of the \(\chi_{1}^{2}\) variables. Because the weights are unknown, we can give the corresponding estimators of the weights. Finally, using the result we can construct the asymptotic confidence regions of β. But the method of estimating weights will reduce the precision of the confidence regions. Further, we adjust the preceding log-likelihood, and it is shown that the adjusted empirical log-likelihood has the asymptotic standard \(\chi_{1}^{2}\) distribution.
2 Results and discussion
2.1 Definition of estimated empirical function
Suppose the primary data set \({(\widetilde{X}_{i},Z_{i},\delta _{i})_{i=1}^{n}}\) is independent of the validation data set \({(\widetilde{X}_{j},X_{j})_{j=n+1}^{n+m}}\), \(E[\varepsilon|X]=0\), and \(E[\varepsilon|\widetilde{X}]=0\). Further we suppose \(F(\cdot)\) and \(G(\cdot)\) are the density functions of Y and C, respectively. Because \(\{Y_{i}\}\) is censored, and the completely observed variable \(Z_{i}\) has different expectations to \(Y_{i}\), we cannot directly use the general method to estimate β. When G is known, we define
Denote \(m(\widetilde{X},\beta)=E[g(X,\beta)|\widetilde{X}]\). It is well known that \(E[Y_{iG}|\widetilde{X}_{i}]=E[Y_{i}|\widetilde {X}_{i}]=m(\widetilde{X}_{i},\beta)\). Then based on the complete observed data, from model (1.1) can be switched to the following models:
where \(\eta_{i}=Y_{iG}-m(\widetilde{X}_{i},\beta)\). Write
Suppose G is known, introduce an auxiliary random vector
It is easily shown that \(E[W_{i}(\beta)]=E\{m^{(1)}(\widetilde {X}_{i},\beta)E[\eta_{i}|\widetilde{X}_{i}]\}=0\), if β is the true value of the parameter.
In practice, \(m(\widetilde{X},\beta)\), \(m^{(1)}(\widetilde{X},\beta )\) and G are usually unknown. For establishing the empirical likelihood, their estimators need to be given first. To do this, for G, we employ the Kaplan-Meier estimator
where \(Z_{(1)}\leq Z_{(2)}\leq\cdots\leq Z_{(n)}\) are the order statistics of \(Z_{1},Z_{2},\ldots, Z_{n}\), and \(\delta_{(i)}\) are the indicators associated with \(Z_{(i)}\), \(i=1,\ldots,n\). Suppose
Here \(K(\cdot)\) is a kernel function and \(h=h_{m}\) is a bandwidth tending to 0. Denote \(\hat{f}_{b_{m}}(\tilde{x})=\max\{\hat {f}_{m}(\tilde{x}),b_{m}\}\), where \({b_{m}}\) is bounded zero positive numbers. Using the validation data, we define the blocked estimators of \(m(\widetilde{X},\beta)\) and \(m^{(1)}(\widetilde{X},\beta )\) as follows:
Use their estimators \(\hat{m}(\widetilde{X},\beta)\), \(\hat{m}^{(1)}(\widetilde{X},\beta)\), and \(\widehat{G}(\cdot)\) to replace the unknown functions \(m(\widetilde{X},\beta)\), \(m^{(1)}(\widetilde {X},\beta)\), and G in (2.2), and write
It is easily proved that \(E[\widehat{W}_{i}(\beta)]=o(1)\) when β is the true value. Using this, an estimated empirical log-likelihood-ratio function is defined as
By introducing the Lagrange multipliers, the most fines value of \(p_{i}\) is \(p_{i}=n^{-1}(1+\lambda^{\tau}\widehat{W})^{-1}\), where λ is determined by
So \(\hat{l}(\beta)\) can be represented as
In succession, we define the β’s estimator by minimizing
2.2 Construction of confidence region
Throughout this section, we use \(c>o\) to represent any constant which does not rely on n and m and may take different values for each appearance. Let \(M^{k}\) be a class of all continuous function classes in \(R^{p}\) (\(k>p\)) or subdomains of \(R^{p}\) which make the partial derivatives \(\frac{\partial^{i_{1}}}{\partial x^{i_{1}}_{1}}\cdot \frac{\partial^{i_{2}}}{\partial x^{i_{2}}_{2}}\cdots\frac{\partial ^{i_{p}}}{\partial x^{i_{p}}_{p}}\varphi(x_{1},\ldots,x_{p})\) uniformly bounded for \(o< i_{i}+\cdots+i_{p}\leq k\).
To obtain our result, we need to list the following conditions.
Condition C
- C1.:
-
\(E\|X\|^{2}<\infty\), \(EY^{2}<\infty\).
- C2.:
-
\(g(x,\beta)\) has bounded continuous partial derivatives up to order two in Γ where Γ is the bounded support of x.
- C3.:
-
For some k (>p), there is \(m(\tilde{x},\beta)\in M^{k}\), \(m^{(1)}_{s}(\tilde{x},\beta)\in M^{k}\), \(s=1,2,\ldots,p\).
- C4.:
-
\(K(u)\) is a bounded nonnegative kernel function of order k (\(k>p\)) with bounded support.
- C5.:
-
The density of X̃, say \(f(\tilde{x})\), satisfies
-
(i)
\(f(\tilde{x})\in M^{k}\),
-
(ii)
there exists a positive constant sequence \(b_{m}\), such that \(\sqrt{m}P\{f(\tilde{x})< b_{m}\}\rightarrow0\) as \(m\rightarrow\infty \).
-
(i)
- C6.:
-
\(mh^{2p}b_{m}^{4}\rightarrow\infty\), \(mh^{2k}b_{m}^{-2}\rightarrow 0\), \(h^{k-\frac{1}{2}p}b_{m}\rightarrow0\).
- C7.:
-
\(\sup_{(\tilde{x},x)}E[e^{2}|\widetilde{X}=\tilde{x},X=x]<\infty \).
- C8.:
-
\(\frac{n}{m}\rightarrow\gamma\), where \(\gamma>0\) is a nonnegative constant.
- C9.:
-
For any \(0\leq s<\infty\), there exists \(\Gamma _{1}(s)=E[m^{(1)}(\widetilde{X},\beta)I[s< Y]]\).
- C10.:
-
-
(i)
For any \(s\leq\tau_{Q}=\inf{t:Q(t)=1}\), \(G(s)\), and \(F(s)\) has no common jumps where \(Q(t)=P(Z\leq t)\),
-
(ii)
\(E{\frac{\|g^{(1)}(X,\beta)\|Y}{[(1-G(Y))(1-F(Y))]^{\frac {1}{2}}}}<\infty\),
-
(iii)
\(\int^{\tau_{Q}}_{0}\|H(s)\|\frac {[1-F(s)]}{[1-F(s-)][1-G(s)]}\, dG(s)<\infty\), where \(H(s)=\frac {E[m^{(1)}(\widetilde{X},\beta)Y_{G}I[s< Z]]}{[1-F(s-)][1-G(s)]}\).
-
(i)
- C11.:
-
\(\Sigma_{0}(\beta)=E{[Y_{G}-m(\widetilde{X},\beta )]^{2}m^{(1)}(\widetilde{X},\beta)(m^{(1)}(\widetilde{X},\beta))^{\tau }}=E[W_{1}(\beta)(W_{1}(\beta))^{\tau}]\), \(\Xi=Em^{(1)}(\widetilde{X}, \beta)(m^{(1)}(\widetilde{X},\beta))^{\tau }\), \(\Sigma_{1}(\beta)=\int^{+\infty}_{0}H(s)H^{T}(s)\bar {F}(s-)(1-\Lambda^{G}(s))\, dG(S)\), where \(\Lambda^{G}(s)= \int^{t}_{-\infty }\bar{G}^{-1}(s-)\, dG(s)\), and \(\Sigma_{0}(\beta)\), \(\Sigma_{1}(\beta)\), and Ξ are all positive definite matrices.
Remark 2.1
These conditions are some usual assumptions for studying the semiparametric model and can be satisfied. Here, condition C6 is only explained, h was taken to be \(h=c_{2}m^{-\frac{1}{p+k}}\) if \(b_{m}=c_{1}m^{\frac{p-k}{4(p+k)}}\log m\), where \(c_{1}\) and \(c_{2}\) are positive constants.
Theorem 2.1
Under Condition C, if β is the true value of the parameter, we have
where Ξ is defined in condition C11, \(\Sigma=\Sigma(\beta )=\Sigma_{0}(\beta)-\Sigma_{1}(\beta)+\Sigma_{2}(\beta)\), with \(\Sigma _{0}(\beta)\), \(\Sigma_{1}(\beta)\) being defined in condition C11, and \(\Sigma_{2}(\beta)=\gamma E\{m^{(1)}(\widetilde{X},\beta)(m^{(1)}(\widetilde {X},\beta))^{\tau}[m(\widetilde{X},\beta)-g(X,\beta)]^{2}\}\) where γ is defined in condition C8.
Theorem 2.2
Under Condition C, if β is the true value of the parameter, we have
where \(\chi^{2}_{1,i}\) (\(1\leq i\leq p\)) are independent standard \(\chi ^{2}\) random variables with 1 degree of freedom, and \(\omega_{i}\) (\(1\leq i\leq p\)) are the eigenvalues of \(D(\beta)=\Sigma^{-1}_{0}(\beta)\Sigma (\beta)\).
In order to use the result of Theorem 2.2 to construct the confidence regions of β, the corresponding estimators of the unknown weights \(\omega_{i}\) (\(1\leq i\leq p\)) must be given. Denote by F̂ the Kaplan-Meier estimator of F, and let \(Q_{n}(s)=\frac{1}{n}\sum^{n}_{i=1}I(Z_{i}\leq s)\). Using the plug-in method, we give the following notations:
-
\(\widehat{H}_{n}(s)=\frac{\frac{1}{n}\sum^{n}_{i=1}\hat{m}^{(1)}(\widetilde{X}_{i},\hat{\beta})Y_{i\widehat {G}}I(s< Z_{i})}{(1-\widehat{G}(s))(1-\widehat{F}(s-))}\),
-
\(\Lambda^{\widehat{G}}_{n}(s)=\int^{s}_{-\infty}\frac{1}{1-\widehat {G}(t-)}\, d\widehat{G}(t)=\frac{1}{n}\sum^{n}_{i=1}\frac{(1-\delta _{i})I(Z_{i}\leq t)}{(1-\widehat{G}(t-))(1-Q_{n}(t-))}\),
-
\(\widehat{\Sigma}_{0}(\hat{\beta})=\frac{1}{n}\sum^{n}_{i=1}[Y_{i\widehat{G}}-\hat{m}(\widetilde{X}_{i},\hat{\beta })]^{2}\hat{m}^{(1)}(\widetilde{X}_{i},\hat{\beta})(\hat {m}^{(1)}(\widetilde{X}_{i},\hat{\beta}))^{\tau}=\frac{1}{n}\sum^{n}_{i=1}\widehat{W}_{i}(\beta)\widehat{W}_{i}^{T}(\beta)\),
-
\(\widehat{\Sigma}_{1}(\beta)=\frac{1}{n}\sum^{n}_{i=1}(1-\delta _{i})\widehat{H}_{n}(Z_{i})\widehat{H}_{n}^{\tau}(Z_{i})(1-\Lambda ^{\widehat{G}}_{n}(Z_{i}))\),
-
\(\widehat{\Sigma}_{2}(\hat{\beta})=\frac{\gamma}{m}\sum^{n+m}_{j=n+1}\hat{m}^{(1)}(\widetilde{X}_{j},\hat{\beta})(\hat {m}^{(1)}(\widetilde{X}_{j},\hat{\beta}))^{\tau}[\hat{m}(\widetilde {X}_{j},\hat{\beta})-g(X_{j},\beta)]^{2}\),
-
\(\widehat{\Sigma}(\hat{\beta})=\widehat{\Sigma}_{0}(\hat{\beta })-\widehat{\Sigma}_{1}(\beta)+\widehat{\Sigma}_{2}(\hat{\beta })\).
From this, we can infer that \(\hat{\omega}_{i}\) (\(i=1,2,\ldots,p\)) (the eigenvalues of \(\widehat{D}(\hat{\beta})=\widehat{\Sigma}^{-1}_{0}(\hat{\beta })\widehat{\Sigma}(\hat{\beta})\)) are the corresponding estimators of \(\omega_{i}\) (\(i=1,2,\ldots,p\)). Denote \(\hat{s}=\hat{\omega}_{1}\chi^{2}_{1,1}+\hat{\omega}_{2}\chi ^{2}_{1,2}+\cdots+\hat{\omega}_{p}\chi^{2}_{1,p}\). \(B_{f}(\cdot)\) expresses the conditional distribution of ŝ given \({(\widetilde {X}_{i},Z_{i},\delta_{i})^{n}_{i=1}}\) and \((\widetilde {X}_{j},X_{j})^{n+m}_{j=n+1}\). Let \(\hat{c}_{\alpha}\) be the \(1-\alpha\) fractile of \(B_{f}(\cdot)\), and the \(1-\alpha\) confidence region for β is
Actually, it is convenient to obtain the conditional distribution \(B_{f}(\cdot)\), we can generate independent samples \(\chi^{2}_{1,1},\ldots ,\chi^{2}_{1,p}\) from the \(\chi^{2}_{1}\) distribution and then get it through the Monte Carlo simulation method.
2.3 Definition of adjusted empirical function
Applying the result of Theorem 2.2 to construct the confidence region of β, we need to estimate the weights \(\omega_{i}\), which will reduce the accuracy of the confidence region. So we need to adjust the empirical likelihood function. Let \(r(\beta)=\frac{p}{\operatorname{tr}\{D(\beta)\}}\), according to the result given by Rao and Scott [10], \(r(\beta)\sum^{p}_{i=1}\omega_{i}\chi^{2}_{1,i}\) has an asymptotic standard \(\chi ^{2}\)-distribution with p degrees of freedom. From Theorem 2.1 and the consistency of \(\widehat{\Sigma}(\hat{\beta})\) and \(\widehat{\Sigma}_{0}(\hat {\beta})\), we can find that \(\hat{r}(\hat{\beta})\hat{l}(\beta)\) also has an asymptotic standard \(\chi^{2}\)-distribution with p degrees of freedom, where \(\hat{r}(\hat{\beta})=\frac{p}{\operatorname{tr}\{\widehat{D}(\hat{\beta})\} }=\frac{p}{\operatorname{tr}\{\widehat{\Sigma}_{0}^{-1}(\hat{\beta})\widehat{\Sigma}(\hat{\beta })\}}\). For improving the approximation accuracy, we replace β̂ with β in \(\hat{r}(\hat{\beta})\) and the accuracy will depend on the value of \(\omega_{i}\). We can refine the result given by Rao and Scott [10], and then we give an adjusted empirical log-likelihood ratio. Denote \(\widehat{A}(\beta)=\{\sum^{n}_{i=1}\widehat{W}_{i}(\beta)\}\{\sum^{n}_{i=1}\widehat{W}_{i}(\beta)\}^{\tau}\). If we replace \(\widehat{\Sigma}(\beta)\) with \(\widehat{A}(\beta)\) in \(\hat {r}(\beta)\), we will get a new adjustment factor \(\hat{\rho}(\beta )=\frac{\operatorname{tr}\{\widehat{\Sigma}^{-1}(\beta)\widehat{A}(\beta)\}}{\operatorname{tr}\{\widehat{\Sigma} _{0}^{-1}(\beta)\widehat{A}(\beta)\}}\), then the adjusted empirical log-likelihood ratio function can be defined as
Theorem 2.3
Under Condition C, if β is the true value of the parameter, we have
Based on Theorem 2.3, \(\hat{l}_{\mathrm{ad}}(\beta)\) can be used to construct a confidence region for β,
where \(P(\chi^{2}_{p}\leq c_{\alpha})=1-\alpha\), and \(P\{\beta\in\hat {I}_{\mathrm{ad}}(\tilde{\beta})\}=1-\alpha+o(1)\).
3 Proofs of theorems
Before the proofs of the theorems, we introduce some preliminary results.
Lemma 3.1
Under Condition C, for any \(1\leq i\leq n\), we have
-
(i)
\(E[(\hat{m}(\widetilde{X}_{i},\beta)-m(\widetilde{X}_{i},\beta ))^{2}|\widetilde{X}_{i}]\leq c(mh^{p}b^{2}_{m})^{-1}+ch^{2k}b^{-2}_{m}+cI[F(\widetilde {X}_{i})<2b_{m}]\);
-
(ii)
\(E[(\hat{m}^{(1)}_{s}(\widetilde{X}_{i},\beta )-m^{(1)}_{s}(\widetilde{X}_{i},\beta))^{2}|\widetilde{X}_{i}]\leq c(mh^{p}b^{2}_{m})^{-1}+ch^{2k}b^{-2}_{m}+cI[F(\widetilde {X}_{i})<2b_{m}]\).
This proof is similar to that of Lemma 1 of Xue [5]. Here we omit the details.
Lemma 3.2
Under Condition C, if β is the true value of the parameter, we have
Proof
Employing C9-C11, similar to the proof of Lemma 1 in Lai et al. [11], we have
where \(M_{i}(t)=(1-\delta_{i})I(Z_{i}< t)-\int^{t}_{-\infty}I(c_{i}\geq s,y_{i}>s)\, d\Lambda(s)\), \(Y_{n}(t)=\sum^{n}_{i=1}I(Z_{i}< t)\). The first item \(M_{5}\) is a martingale sequence which has a limiting normal distribution with mean 0 and covariance \(\Sigma_{1}\) according to the Rebolledo central limit theorem of martingales. We have
Write \(m_{b}(\tilde{x},\beta)= m(\tilde{x},\beta)f(\tilde {x})f^{-1}_{b}(\tilde{x})\), \(f_{b}(\tilde{x})=\max\{f(\tilde {x}),b_{m}\}\), then
For \(\forall\varepsilon>0\), we have
Hence, \(M_{312}=o_{p}(1)\). Write
A series simple calculation yields
By conditions C4, C5, and \(\sqrt{m}h^{2k}\rightarrow0\), applying a Taylor expansion, we can prove
Notice that \(f(\tilde{x})-f_{b}(\tilde{x})=0\) when \(f(\tilde {x})\geq b_{m}\). By conditions C2-C4, it is easy to see that \(E(\sqrt {m}K_{m2})\rightarrow0\). Hence, we have \(K_{m2}=o_{p}(m^{-\frac{1}{2}})\).
Using the standard kernel estimation method, we can prove that \(J_{m12}=o_{p}(m^{-\frac{1}{2}})\).
Now, let us prove \(J_{mi}=o_{p}(m^{-\frac{1}{2}})\), \(i=2,\ldots,5\). By conditions C4 and C5, applying a Taylor expansion, we have
According to the condition independence, together with (3.2), it is easily shown that \(E(\sqrt{m}J_{m2})^{2}\rightarrow0\), hence \(J_{m2}=o_{p}(m^{-\frac{1}{2}})\). Now
By condition C4, we easily get
Together with condition C5 and (3.3), and the Markov inequality, it is easily proved that \(L_{m1}=o_{p}(m^{-\frac{1}{2}})\). In addition, for \(\forall\varepsilon>0\), we have
Hence, we have \(J_{m31}=o_{p}(m^{-\frac{1}{2}})\). By an argument similar to \(J_{m31}\), we obtain
Hence, we have \(J_{m3}=o_{p}(m^{-\frac{1}{2}})\). Now
Notice that
Together with \(\sqrt{m}P(f(\widetilde{X}_{i})< b_{m})\rightarrow0\), applying conditional independence, we can obtain \(J_{m4}=o_{p}(m^{-\frac {1}{2}})\).
Similarly, by conditional independence and (3.4), we can prove that \(J_{m5}=o_{p}(m^{-\frac{1}{2}})\). Hence, we have
Consider the sth (\(s=1,\ldots,p\)) component of \(M_{32}\), from the Cauchy-Schwarz inequality and Lemma 3.1, we have
Hence, we have \(M_{32}=o_{p}(1)\),
From Wang [3, 4], we can obtain \(M_{42}=o_{p}(1)\).
Write \(U=\{X_{j},\widetilde{X}_{j}\}^{n+m}_{j=n+1}\), consider the sth (\(s=1,\ldots,p\)) component of \(M_{41}\), from conditional independence and Lemma 3.1 we have
So \(M_{41}=o_{p}(1)\), then \(M_{4}=o_{p}(1)\). Therefore, we have
From the central limit theorem and \(\frac{n}{m}\rightarrow\gamma\), we have
In addition, \(M_{1}\) and \(M_{6}\) are independent of each other, \(M_{5}\) and \(M_{6}\) are independent of each other. Then a simple calculation yields
From the central limit theorem and by (3.6)-(3.8), Lemma 3.2 is proved. □
Lemma 3.3
Under Condition C, if β is the true value of the parameter, we have
Proof
After a complex calculation, we have
By the law of large numbers, we obtain Lemma 3.3. □
Lemma 3.4
Let \(Z\stackrel{L}{\longrightarrow}N(0,I_{p})\), where \(I_{p}\) is the \(p\times p\) identity matrix. Let Q be a \(p\times p\) nonnegative definite matrix with eigenvalues \(\omega _{1},\ldots,\omega_{p}\). Then it follows that
Proof of Theorem 2.1
We have
Let
Using the Lagrange multiplier method, let
that is,
Then we have
Applying the Taylor expansion to \(\hat{m}(\widetilde{X}_{i},\beta)\) and \(\hat{m}^{(1)}(\widetilde{X}_{i},\beta)\) in (3.9), we can obtain
So
where \(\widehat{\Xi}^{-1}(\beta)=\frac{1}{n}\sum^{n}_{i=1}\hat {m}^{(1)}(\widetilde{X}_{i},\beta)\hat{m}^{(1)}(\widetilde{X}_{i},\beta +\theta(\hat{\beta}-\beta))(\hat{\beta}-\beta)\), \(\theta\in(0,1)\) is a constant. This, together with Lemma 3.2, easily proves that \(\widehat{\Xi}^{-1}(\beta )\stackrel{P}{\longrightarrow}\Xi(\beta)\). That is,
From Lemmas 3.2, 3.3, and 3.4, we can obtain
Theorem 2.1 is proved. □
Proof of Theorem 2.2
Applying a Taylor expansion to equation (2.3), we have
By Lemmas 3.2 and 3.3, using the result of Wang [2–4], we can prove
Again,
By (3.11), we can obtain
By (3.13), we can obtain
By (3.10), (3.13), (3.14), we can obtain
By (3.15) and Lemmas 3.2, 3.3, and 3.4, we can obtain
where \(D(\beta)=\Sigma^{\frac{1}{2}}(\beta)\Sigma_{0}^{-1}(\beta)\Sigma ^{\frac{1}{2}}(\beta)\).
Write \(\widetilde{D}=\operatorname{diag}(\omega_{1},\ldots,\omega_{p})\) where \(\omega _{i}\) (\(i=1,\ldots,p\)) are the eigenvalues of D, then there exists an orthogonal matrix Q which makes \(Q^{\tau}\widetilde{D}Q=D(\beta)\). Therefore, we have
From Lemma 3.2, we have
By (3.16) and (3.17), we finish the proof of Theorem 2.2. □
Proof of Theorem 2.3
Review the definition of \(\hat {l}_{\mathrm{ad}}(\beta)\), combined with (3.16), we can infer
Combined with the process of the above proof, similar to the proof of Lemma 3.2, it is easily shown that \(\widehat{\Sigma}(\beta)\stackrel {P}{\longrightarrow}\Sigma(\beta)\).
By Lemmas 3.2, 3.4, and (3.17), we can prove
□
4 Conclusions
In this text, with the help of validation data, the empirical likelihood method is extended to the nonlinear error-in-variables regression models with randomly censored response. We construct an estimated empirical log-likelihood function \(\hat{l}(\beta)\) of the unknown interesting parameter β, and we get the asymptotic distribution of \(\hat{l}(\beta)\). By giving estimators of the weights \(\omega_{i}\) and Monte Carlo simulation method, we construct the confidence region for the parameter β. To avoid estimating \(\omega_{i}\), the adjusted empirical log-likelihood \(\hat {l}_{\mathrm{ad}}(\beta)\) has been defined and the asymptotic distribution of \(\hat{l}_{\mathrm{ad}}(\beta)\) is obtained. Using the result, we may better construct a confidence region for the parameter β. The theoretical analysis shows that the empirical likelihood method can be applied to the nonlinear model with complex incomplete data, and it can be used to obtain a better result.
References
Sepanski, JH, Lee, LF: Semiparametric estimation of nonlinear error-in-variables models with validation study. J. Nonparametr. Stat. 4, 365-394 (1995)
Wang, QH: Estimation of linear error-in-covariables model with validation data under random censorship. J. Multivar. Anal. 74, 245-266 (1999)
Wang, QH, Rao, JNK: Empirical likelihood-based in linear error-in-covariables model with validation data. Biometrika 89(2), 345-358 (2002)
Wang, QH: Survival Data Analysis. Science Press, Beijing (2006)
Xue, L: Empirical likelihood inference in nonlinear semiparametric EV models with validation data. Acta Math. Sin. 49(1), 145-154 (2006)
Fang, L, Hu, F: Empirical likelihood dimension reduction inference in nonlinear EV models with validation data. J. Math. 32(1), 113-120 (2012)
Buckley, J, James, I: Linear regression with censored data. Biometrika 66, 429-436 (1979)
Li, G, Wang, QH: Empirical likelihood confidence regression analysis with right censored data. Stat. Sin. 13(1), 51-68 (2003)
Cheng, F, Li, G, Feng, S, Xue, L: Empirical likelihood inference for nonlinear regression model with right censored data. Acta Math. Appl. Sin. 33(1), 130-140 (2010)
Rao, JNK, Scott, AJ: The analysis of categorical data from complex sample surveys: chi-squared test for goodness of fit and independence in two-way tables. J. Am. Stat. Assoc. 76, 221-230 (1981)
Lai, TL, Ying, ZL, Zheng, ZK: Asymptotic normality of a class of adaptive statistic with applications to synthetic data methods for censored regression. J. Multivar. Anal. 52, 259-279 (1995)
Acknowledgements
The authors are grateful to the referee for carefully reading the manuscript and for offering comments which enabled them to improve the paper. The research of Y Wu was supported by the Humanities and Social Sciences Foundation for the Youth Scholars of Ministry of Education of China (12YJCZH217) and the Key NSF of Anhui Educational Committe (KJ2014A255). The research of L Fang was supported the Anhui provincial natural science research project of university foundation (KJ2013B304) and the natural science research project of the Tongling University foundation (2012tlxy13).
Author information
Authors and Affiliations
Corresponding author
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
All authors contributed equally to the writing of this paper. All authors read and approved the final manuscript.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Fang, L., Wu, Y. Empirical likelihood inference of parameters in nonlinear EV models with censored data. J Inequal Appl 2015, 379 (2015). https://doi.org/10.1186/s13660-015-0903-1
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s13660-015-0903-1