- Research
- Open access
- Published:
Difference-based M-estimator of generalized semiparametric model with NSD errors
Journal of Inequalities and Applications volume 2019, Article number: 61 (2019)
Abstract
In this paper, we consider the generalized semiparametric model (GSPM)
where \(h(\cdot )\) is a known function, \(e_{i}\) are dependent errors. We obtain an estimator of the parametric component β for the model by a difference-based M-estimator. In addition, we prove the asymptotic normality of the proposed estimator and investigate the weak convergence rate of the wavelet estimator of \(f(\cdot )\). Furthermore, we apply these results to a partially linear model with dependent errors.
1 Introduction
Consider the generalized semiparametric model
where \(y_{i}\) are scalar response variables, \(h(\cdot )\) is a continuously differentiable known function, the superscript T denotes the transpose, \(\mathbf{x}_{i}=(x_{i1},\ldots ,x_{id})^{T}\) are explanatory variables, β is a d-dimensional unknown parameter, \(f(\cdot )\) is an unknown function, and \(0\leq t_{1}\leq t_{2}\leq \cdots \leq t_{n} \leq 1\). Some authors commented that the assumption of independence is a serious restriction (see Huber [1] and Hampel [2]); so for the errors \(e_{i}\), we confine ourselves to negatively superadditive dependent (NSD) errors. NSD random variables have been introduced by Hu [3] and are widely used in statistics; see [4,5,6,7,8,9,10,11,12].
The theory of the GSPM is an extension of the classical theory of partially linear models; the component of the generalized parametric \(h (\mathbf{x}_{i}^{T}\beta )\) for GSPM includes the linear parametric component \(\mathbf{x}^{T}_{i}\beta \), exponential parametric component \(e^{\mathbf{x}^{T}_{i}\beta }\), and so on.
As is well known, the generalized partially linear model and partially linear single-index model (\(h(\cdot )\) is an unknown link function) are also derived from the partially linear model. There is a substantial amount of work for generalized partially linear model (see [13,14,15,16,17,18] and, for a partially linear single-index model, [19,20,21,22,23,24]); this research is devoted to presenting various methods to obtain estimators of β and \(f(t_{i})\) and investigating some large-sample properties of these estimators.
In this paper, we consider a difference-based estimator method to estimate the unknown parametric component β. This difference-based estimator is optimal in the sense that the estimator of the unknown parametric component is asymptotically efficient. For example, Tabakan et al. [25] studied a difference-based ridge in a partially linear model. Wang et al. [26] obtained a difference-based approach to the semiparametric partially linear model. Zhao and You [27] used a difference-based estimator method to estimate the parametric component for partially linear regression models with measurement errors. Duran et al. [28] investigated the difference-based ridge and Liu-type estimators in semiparametric regression models. Wu [29] discussed a restricted difference-based Liu estimator in partially linear models. Hu et al. [30] presented a difference-based Huber–Dutter (DHD) estimator to obtain the root variance σ and parameter β for a partially linear model. However, Most of the results rely on the independence errors. Wu [31] studied the difference-based ridge-type estimator of parameters in a restricted partial linear model with correlated errors, but this paper just focuses on estimating the linear component. Zeng and Liu [32] used a difference-based and ordinary least-square method to obtain the estimator of an unknown parametric component, but this paper ignores the fact that a difference-based estimator may cause greater bias in moderately sized samples than other estimators. Inspired by these papers, we propose a difference-based M-estimator (DM) methods for generalized semiparametric model with NSD errors. The M-estimator is a most famous robust estimator, which was introduced by Huber [33]. In addition, once β is estimated, we can estimate \(f(\cdot )\) by a variety of nonparametric techniques. In this paper, the estimator of \(f(\cdot )\) is obtained by the wavelet method.
The paper has the following structure. In Sect. 2, we present the estimation procedure. In Sect. 3, we establish the main results. The proofs of the main results are provided in the Appendix.
2 Estimation method
2.1 Notation
Throughout the paper, Z is the set of integers, N is the set of natural numbers, R is the set of real numbers. A sequence of random variables \(\eta _{n}\) is said to be of smaller order in probability than a sequence \(d_{n}\) (denoted by \(\eta _{n}=o_{P}(d_{n})\)) if \(\eta _{n}/d_{n}\) converges to 0 in probability, and \(\eta _{n}=O_{P}(d_{n})\) if \(\eta _{n}/d_{n}\) is bounded in probability. Convergence in distribution is denoted by \(H_{n}\stackrel{D}{ \rightarrow }H\). For any arbitrary function \(h(\cdot )\), \(h'(\cdot )\), \(h''(\cdot )\), and \(h'''(\cdot )\) are the first, second, and third derivatives of \(h(\cdot )\), respectively. \(\|\mathbf{x}\|\) is the Euclidean norm of x, and \(\lfloor x\rfloor =\max \{k\in \mathbf{Z}:k\leq x\}\). Let \(C_{0}\), \(C_{1}\), \(C_{2}\), \(C_{3}\), \(C_{4}\) be positive constants, and let \(\beta _{0}\) be the true parameter. Let \(\varTheta = \{\beta :|\beta -\beta _{0}|\leq C_{0} \}\).
2.2 Difference-based M-estimation
Let \(\tilde{y}_{i}=\sum_{q=0}^{m}d_{q}y_{i+q}\), \(\tilde{h}_{i}(\beta )=\sum_{q=0}^{m}d_{q}h (\mathbf{x}^{T}_{i+q}\beta )\), \(\tilde{f}(t_{i})=\sum_{q=0}^{m}d_{q}f(t_{i+q})\), and \(\tilde{e}_{i}= \sum_{q=0}^{m}d_{q}e_{i+q}\), where \(d_{0},d_{1},\ldots ,d _{m}\) satisfy the conditions
Then \(\tilde{y}_{i}\), \(\tilde{h}_{i}(\beta )\), \(\tilde{f}(t_{i})\), and \(\tilde{e}_{i}\) can be seen as the mth-order differences of \(y_{i}\), \(h(\mathbf{x}_{i}^{T}\beta )\), \(f(t_{i})\), and \(e_{i}\), respectively. Hence, applying the differencing procedures, model (1) becomes
From Yatchew [34] we find that the application of differencing procedures in model (1) can remove the nonparametric effect in large samples, so we ignore the presence of \(\tilde{f}(\cdot )\). Thus (3) becomes
Let ρ be a convex function. Assume that ρ has a continuous derivative ψ and there is a such that \(\psi (a)=0\). We can propose the difference-based M-estimator given by minimizing
Let a \(d\times 1\) vector \(\hat{\beta }_{n}\) be the minimizer of (5) and \(\hat{\beta }_{n}\in \varTheta \). Write \(\tilde{\mathbf{h}}'_{i}(\beta )= \sum_{q=0}^{m}d_{q}\times h'(\mathbf{x}^{T}_{i+q}\beta )\mathbf{x}_{i+q}\), \(\tilde{h}'_{ik}(\beta )=\sum_{q=0}^{m}d_{q}h'(\mathbf{x}^{T}_{i+q} \beta )x_{(i+q)k}\), with \(1 \leq k \leq d\), \(\tilde{\mathbf{h}}''_{i}( \beta )=\sum_{q=0}^{m}d_{q}h''(\mathbf{x}^{T}_{i+q}\beta )\times\mathbf{x} _{i+q}\mathbf{x}^{T}_{i+q}\), and \(\tilde{\mathbf{h}}'_{i}(\beta ) \tilde{\mathbf{h}}^{\prime \,T}_{j}(\beta )=\sum_{q=0}^{m}d_{q}h'(\mathbf{x} ^{T}_{i+q}\beta )\mathbf{x}_{i+q}\sum_{q=0}^{m}d_{q}h'(\mathbf{x}^{T} _{j+q}\beta )\mathbf{x}^{T}_{j+q}\). Then the estimator satisfies
with \(\hat{\tilde{e_{i}}}=\tilde{y}_{i}-\tilde{\mathbf{h}}_{i}( \hat{\beta }_{n})\). The convexity of ρ guarantees the equivalence of (5) and (6) and the asymptotic uniqueness of the solution; otherwise, it is unimportant.
We estimate the nonparametric function \(f(\cdot )\) by the wavelet method. The formal definition of the wavelet method is the following.
Suppose that there exist a scaling function \(\phi (\cdot )\) in the Schwartz space \(S_{l}\) and a multiresolution analysis \(\{V_{\tilde{m}} \}\) in the concomitant Hilbert space \(L^{2}(\mathbf{R})\) with the reproducing kernel \(E_{\tilde{m}}(t,s)\) given by
Let \(A_{i}=[s_{i-1}, s_{i}]\) denote intervals that partition \([0, 1]\) with \(t_{i} \in A_{i}\) for \(1\leq i\leq n\). Then the estimator of the nonparameter \(f(t)\) is given by
3 Main results
We now list some conditions used to obtain the main results.
-
(C1)
\(\max_{1\leq i \leq n}\|\mathbf{x}_{i}\|=O(1) \), and the eigenvalues of \(n^{-1}\sum^{n}_{i=1}\mathbf{x}_{i}\mathbf{x}^{T}_{i}\) are bounded above and away from zero.
-
(C2)
\(b, bc-d^{2}>0\), where \(b=E\{\psi '(\eta )\}\), \(c=E\{\eta ^{2} \psi '(\eta )\}\), \(d=E\{\eta \psi '(\eta )\}\) with \(\eta =\tilde{e}_{i}+a\).
-
(C3)
\(E\psi (\tilde{e}_{i}+a )=0\).
-
(C4)
The function ρ is assumed to be convex, not monotone, and possessing bounded derivatives of sufficiently high order in a neighborhood of the point \(\mathbf{x}_{i}^{T}\beta _{0}\). In particular, \(\psi (t)\) should be continuous and bounded in a neighborhood of \(\mathbf{x}_{i}^{T}\beta _{0}\).
-
(C5)
\(h(\cdot )\) is assumed to possess bounded derivatives of sufficiently high order in a neighborhood of point \(\mathbf{x}_{i} ^{T}\beta _{0}\).
-
(C6)
\(f(\cdot )\in H^{\alpha }\) (Sobolev space) for some \(\alpha >1/2\).
-
(C7)
\(f(\cdot )\) is a Lipschitz function of order \(\gamma >0\).
-
(C8)
\(\phi (\cdot )\) belongs to \(S_{l}\), which is a Schwartz space for \(l\geq \alpha \), is a Lipschitz function of order 1, and has a compact support, in addition to \(|\hat{\phi }(\xi )-1|=O(\xi )\) as \(\xi \rightarrow 0\), where ϕ̂ denotes the Fourier transform of ϕ.
-
(C9)
\(s_{i}\), \(1\leq i\leq n\), satisfy \(\max_{1\leq i\leq n}(s_{i}-s _{i-1})=O(n^{-1})\), and \(2^{\tilde{m}}=O(n^{1/3})\).
Remark 1
Condition (C1) is often imposed in M-estimation theory of regression models. Condition (C2) is used by Silvapullé [35] for HD estimation. In this paper, this condition is also necessary for M-estimation. Condition (C3) is used by Wu [36] and Zeng and Hu [37] with \(a=0\). We require this in order that the expectation of (5) reaches its minimum at the true value \(\beta _{0}\). For Condition (C4), higher-order derivatives are technically convenient (Taylor expansions), but their existence is hardly essential for the results to hold; see Huber [1]. Condition (C5) is quite mild and can be easily satisfied. Conditions (C6)–(C9) are used by Hu et al. [38].
Remark 2
The assumption of \(\psi (a)=0\) and Condition (C4) are serious restrictions, which shows that the M-estimator in our paper is a particular case of the classical M-estimator. However, in our study, these conditions are necessary.
Theorem 3.1
Let \(\{e_{n}, n\geq 1\}\) be a sequence of NSD random variables with \(Ee_{n}=0\), and let for some \(\delta >0\),
Suppose that
Set \(\tilde{e}_{i}=\sum_{q=0}^{m}d_{q}e_{i+q}\), where \(\{d_{q}, 1 \leq q \leq m\}\) are defined in (2). Let \(\{c_{i}, 1\leq i \leq n-m\}\) be an array of constants satisfying \(\max_{1\leq i\leq n-m}|c_{i}|=O(1)\), and suppose that \(\psi (a)=0\) and Conditions (C3) and (C4) hold. Then
provided that
Theorem 3.2
Let \(\{e_{n}, n\geq 1\}\) be a sequence of NSD random variables with \(Ee_{n}=0\) satisfying conditions (8) and (9). Assume that conditions (C1)–(C5) hold. Then
provided that
is a positive definite matrix, where \(I_{d}\) is the identity matrix of order d.
Corollary 3.1
Let \(h(\mathbf{x}_{i}^{T}\beta )= \mathbf{x}_{i}^{T}\beta \), and let \(\{e_{n}, n\geq 1\}\) be a sequence of NSD random variables with \(Ee_{n}=0\) satisfying conditions (8) and (9). Assume that Conditions (C1)–(C4) hold. Then
provided that
is a positive definite matrix.
Corollary 3.2
Let \(\{e_{n}, n\geq 1\}\) be a sequence of NSD random variables with \(Ee_{n}=0\) satisfying \(\operatorname{Cov}_{|i-j|> \bar{m}}(e_{i},e_{j})=0\) with \(\bar{m}<\infty \). Assume that Condition (C1)–(C5) and (8) hold. Then
provided that
is a positive definite matrix.
By Theorem 3.2 we also easily obtain some corresponding results for \(\rho (t)=t^{2}\). Here we omit their proofs.
Corollary 3.3
(Zeng and Liu [32])
Let \(\rho (t)=t^{2}\), \(h(\mathbf{x}_{i}^{T}\beta )=\mathbf{x}_{i}^{T}\beta \), and let \(\{e_{n}, n\geq 1\}\) be a sequence of NSD random variables with \(Ee_{n}=0\) satisfying conditions (8) and (9). Assume that conditions (C1)–(C2) hold. Then
provided that
is a positive definite matrix.
Corollary 3.4
Let \(\rho (t)=t^{2}\), \(h(\mathbf{x} _{i}^{T}\beta )=e^{\mathbf{x}_{i}^{T}\beta }\), and let \(\{e_{n}, n \geq 1\}\) be a sequence of NSD random variables with \(Ee_{n}=0\) satisfying conditions (8) and (9). Assume that conditions (C1)–(C2) hold. Then
provided that \(\tau ^{2}_{\beta }=\lim_{n\rightarrow \infty }(n-m)^{-1}\operatorname{Var} (\sum_{i=1}^{n-m}\tilde{e}_{i}\sum_{q=0}^{m}d_{q}e^{\mathbf{x} ^{T}_{i+q}\beta _{0}}\mathbf{x}_{i+q} )\) is a positive definite matrix.
Theorem 3.3
Under the conditions of Theorem 3.2, assume that Conditions (C6)–(C9) hold. Then
where \(M_{n}\rightarrow \infty \) in arbitrary slowly rate, and \(\tau _{\tilde{m}}=2^{-\tilde{m}(\alpha -1/2)}\) if \(1/2< \alpha <3/2\), \(\tau _{\tilde{m}}=\sqrt{\tilde{m}}2^{-\tilde{m}}\) if \(\alpha =3/2\), and \(\tau _{\tilde{m}}=2^{-\tilde{m}}\) if \(\alpha >3/2\).
References
Huber, P.J.: Robust regression: asymptotics, conjectures and Monte Carlo. Ann. Stat. 1(5), 799–821 (1973)
Hampel, F.R., Ronchetti, E.M., Rousseeuw, P.J., Stahel, W.A.: Robust Statistics. The Approach Based on Influence Functions. Wiley, New York (1986)
Hu, T.Z.: Negatively superadditive dependence of random variables with applications. Chinese J. Appl. Probab. Statist. 16(2), 133–144 (2000)
Shen, Y., Wang, X.J., Yang, W.Z., Hu, S.H.: Almost sure convergence theorem and strong stability for weighted sums of NSD random variables. Acta Math. Sin. Engl. Ser. 29(4), 743–756 (2013)
Xue, Z., Zhang, L.L., Lei, Y.J., Chen, Z.J.: Complete moment convergence for weighted sums of negatively superadditive dependent random variables. J. Inequal. Appl. 2015, Article ID 117 (2015)
Wang, X.J., Deng, X., Zheng, L.L., Hu, S.H.: Complete convergence for arrays of rowwise negatively superadditive dependent random variables and its applications. Statistics 48(4), 834–850 (2014)
Wang, X.J., Shen, A.T., Chen, Z.Y., Hu, S.H.: Complete convergence for weighted sums of NSD random variables and its application in the EV regression model. Test 24, 166–184 (2015)
Wang, X.J., Wu, Y., Hu, S.H.: Complete moment convergence for double-indexed randomly weighted sums and its applications. Statistics 52(3), 503–518 (2018)
Meng, B., Wang, D., Wu, Q.: On the strong convergence for weighted sums of negatively superadditive dependent random variables. J. Inequal. Appl. 2017, Article ID 269 (2017)
Eghbal, N., Amini, M., Bozorgnia, A.: On the Kolmogorov inequalities for quadratic forms of dependent uniformly bounded random variables. Stat. Probab. Lett. 81, 1112–1120 (2011)
Shen, A.T., Zhang, Y., Volodin, A.: Applications of the Rosenthal-type inequality for negatively superadditive dependent random variables. Metrika 78, 295–311 (2015)
Shen, A.T., Xue, M.X., Volodin, A.: Complete moment convergence for arrays of rowwise NSD random variables. Stochastics 88(4), 606–621 (2016)
Boente, G., He, X., Zhou, J.: Robust estimates in generalized partially linear models. Ann. Stat. 34, 285–2878 (2016)
Cheng, G., Zhou, L., Huang, Z.J.: Efficient semiparametric estimation in generalized partially linear additive models for longitudinal/clustered data. Bernoulli 20(1), 141–163 (2014)
He, X., Fung, W., Zhu, Z.: Robust estimation in generalized partial linear models for clustered data. J. Am. Stat. Assoc. 100, 1176–1184 (2005)
Graciela, B., Daniela, R.: Robust inference in generalized partially linear models. Comput. Stat. Data Anal. 54(12), 2942–2966 (2010)
Qin, G., Zhu, Z., Fung, W.K.: Robust estimation of generalized partially linear model for longitudinal data with dropouts. Ann. Inst. Stat. Math. 68, 977–1000 (2016)
Lin, H., Fu, B., Qin, G., Zhu, Z.: Doubly robust estimation of generalized partial linear models for longitudinal data with dropouts. Biometrics 73(4), 1132–1139 (2017)
Yu, Y., Ruppert, D.: Penalized spline estimation for partially linear single-index models. J. Am. Stat. Assoc. 97, 1042–1054 (2002)
Xia, Y., Hardle, W.: Semi-parametric estimation of partially linear single-index models. J. Multivar. Anal. 97, 1162–1184 (2006)
Wang, J.L., Xue, L.G., Zhu, L.X., Chong, Y.S.: Estimation for a partial-linear single index model. Ann. Stat. 38(1), 246–274 (2010)
Huang, Z.S.: Statistical inferences for partially linear single-index models with error-prone linear covariates. J. Stat. Plan. Inference 141(2), 899–909 (2011)
Lian, H., Liang, H., Carroll, R.: Variance function partially linear single-index models. J. R. Stat. Soc. B 77(1), 171–194 (2015)
Yang, J., Lu, F., Yang, H.: Statistical inference on asymptotic properties of two estimators for the partially linear single-index models. Statistics 52(6), 1193–1211 (2018)
Tabakan, G., Akdeniz, F.: Difference-based ridge estimator of parameters in partial linear model. Stat. Pap. 51, 357–368 (2010)
Wang, L., Brown, L.D., Cai, T.T.: A difference based approach to the semiparametric partial linear model. Electron. J. Stat. 5, 619–641 (2011)
Zhao, H., You, J.: Difference based estimation for partially linear regression models with measurement errors. J. Multivar. Anal. 102, 1321–1338 (2011)
Duran, E.A., Hädle, W.K., Osipenko, M.: Difference based ridge and Liu type estimators in semiparametric regression models. J. Multivar. Anal. 105(1), 164–175 (2012)
Wu, J.: Restricted difference-based Liu estimator in partially linear model. J. Comput. Appl. Math. 300, 97–102 (2016)
Hu, H.C., Yang, Y., Pan, X.: Asymptotic normality of DHD estimators in a partially linear model. Stat. Pap. 57(3), 567–587 (2016)
Wu, J.: Difference based ridge type estimator of parameters in restricted partially linear model with correlated errors. SpringerPlus 5, 178 (2016)
Zeng, Z., Liu, X.D.: Asymptotic normality of difference-based estimator in partially linear model with dependent errors. J. Inequal. Appl. 2018, Article ID 267 (2018)
Huber, P.J.: Robust estimation of a location parameter. Ann. Math. Stat. 35, 73–101 (1964)
Yatchew, A.: An elementary estimator for the partial linear model. Econ. Lett. 5, 135–143 (1997)
Silvapullé, M.J.: Asymptotic behavior of robust estimators of regression and scale parameter with fixed carriers. Ann. Stat. 13(4), 1490–1497 (1985)
Wu, W.B.: M-estimation of linear models with dependent errors. Ann. Stat. 35, 495–521 (2007)
Zeng, Z., Hu, H.C.: Weak linear representation of M-estimation in GLMs with dependent errors. Stoch. Dyn. 17, 1750034 (2017). https://doi.org/10.1142/S0219493717500344
Hu, H.C., Cui, H.J., Li, K.C.: Asymptotic properties of wavelet estimators in partially linear errors-in-variables models with long-memory errors. Acta Math. Appl. Sin. Engl. Ser. 34(1), 77–96 (2018)
Zhou, X., You, J.: Wavelet estimation in varying-coefficient partially linear regression models. Stat. Probab. Lett. 68, 91–104 (2004)
Acknowledgements
This authors would like to thank a referee and an Associate Editor for their comments and suggestions.
Availability of data and materials
Not application.
Funding
The research is supported Support by National Natural Science Foundation of China [grant number 71471075].
Author information
Authors and Affiliations
Contributions
All authors contributed equally to the writing of this paper. All authors read approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
1.1 1.1 Lemmas
In this section, we present the proofs of the main results. We first need some lemmas.
Lemma 1
Under Conditions (C1), (C4), and (C5), suppose that \(e_{i}\) satisfies (8). Then
and
where \(\tilde{\beta }=s\hat{\beta }_{n}+(1-s)\beta _{0}\) for some \(s\in [0,1]\), \(l=1,2,\ldots,d\).
Proof
We have
From (8) we have
Therefore, for a fixed small ε, there exists a positive integer \(\delta =\delta _{\varepsilon }\) such that
for \(1\leq k_{1},k_{2}, l_{1},l_{2} \leq d\), and thus
By Condition (C5) we have that \(\max_{1\leq i \leq n}\{\sum_{q=1}^{m} | d_{q}h'(\mathbf{x}^{T}_{i+q}\tilde{\beta })|\}\), \(\max_{1\leq i \leq n}\{\sum_{q=1}^{m} | d_{q}\times h''(\mathbf{x}^{T}_{i+q}\tilde{\beta })| \}\), and \(\max_{1\leq i \leq n}\{\sum_{q=1}^{m} | d_{q}h'''( \mathbf{x}^{T}_{i+q}\tilde{\beta })|\}\) are bounded by some constant \(C_{1}\). Then by (C4) it follows that, for same constant \(M>0\),
and from (C1) it follows that
By the Chebyshev inequality it suffices to verify that \(I_{1}=O_{P} ((n-m)^{\frac{1}{2}} )\). In the same way, we easily obtain that \(I_{2}=O_{P} ((n-m)^{\frac{1}{2}} )\). Consequently,
Note that, for \(1 \leq l\leq d\),
By Conditions (C1), (C4), and (C5), for \(1\leq k, l, s\leq d\) and some constant \(M>0\), we have
Hence (15) holds, and the proof is completed. □
Lemma 2
If (C1)–(C5) hold, then
Proof
We can prove Lemma 2 by an argument similar to Lemma 4 of Silvapullé [35], so we omit the details. □
Lemma 3
(Zhou and You [39])
If Condition (C8) holds, then
-
(a1)
\(|E_{0}(t,s)|\leq \frac{C_{k}}{(1+|t-s|)^{k}}\), \(|E_{\tilde{m}}(t,s)| \leq \frac{2^{\tilde{m}}C}{(1+2^{\tilde{m}}|t-s|)^{k}}\) (where \(k \in \mathbf{N} \), and \(C=C(k)\) is a constant depending on k only);
-
(a2)
\(\sup_{0\leq s\leq 1}|E_{\tilde{m}}(t,s)|=O(2^{\tilde{m}})\);
-
(a3)
\(\sup_{t}\int ^{1}_{0}|E_{\tilde{m}}(t,s)|\,ds\leq C_{2}\);
-
(a4)
\(\int ^{1}_{0}E_{\tilde{m}}(t,s)\,ds\rightarrow 1, n\rightarrow \infty \).
1.2 1.2 Proof of Theorem 3.1
By Condition (8) we have
from which it follows that
and for all \(\varepsilon >0\),
Then we can find a positive number sequence \(\{\varepsilon _{n}, n \geq 1\}\) with \(\varepsilon _{n}\rightarrow 0\) such that
Now we define the integers: \(m_{0}=0\) and for \(j=0,1,2,\ldots \) ,
Denote
where \(l=l(n)\) is the number of blocks of indices \(I_{j}\). Then
and hence we have \(l\leq C_{3}/\sqrt{\varepsilon _{n}}\). If the remainder term is not zero, then as the construction ends, we put all the remainder terms into a block denoted by \(J_{l}\). Hence, by the Lagrange mean value theorem,
where \(\xi _{i}=t\tilde{e}_{i}+a\) for some \(t\in [0,1]\).
Moreover, setting \(a_{i}=\tau ^{-1}c_{i}\psi '(\xi _{i})\), we have
By the argument in the proof of Theorem 4.1 in Zeng and Liu [32] we have
which implies
The proof is completed.
1.3 1.3 Proof of Theorem 3.2
Now we will use Theorem 3.1 to prove Theorem 3.2. Expanding \(\frac{\partial }{\partial \beta }Q (\hat{\beta }_{n} )\) about \(\beta _{0}\), we have
where \(\tilde{\beta }=s\hat{\beta }_{n}+(1-s)\beta _{0}\) for some \(s\in [0,1]\), and
From (6) we have
We now show that
Let u be a \(1 \times d\) such that \(\|u\|=1\). By the Cramér–Wold theorem it suffices to verify that
where \(\tau ^{2}=\lim_{n\rightarrow \infty }\operatorname{Var} ((n-m)^{-1/2} \sum_{i=1}^{n-m}\psi (\tilde{e}_{i}+a )u \tilde{\mathbf{h}}_{i}'(\beta _{0}) )\); by the definition of \(\tau ^{2}_{\beta }\), \(\tau ^{2}>0\). By Theorem 3.1, (25) follows from \(u\tilde{\mathbf{h}}_{i}'(\beta _{0})=O(1)\). The proof is completed.
1.4 1.4 Proof of Theorem 3.3
By (7) we have
By the argument in the proof of Theorem 3.2 in Hu [30] we have
and
By Lemma 3, (C1), and (C5) we assume that
where \(\xi _{i}=r\mathbf{x}^{T}_{i}\beta +(1-r)\mathbf{x}^{T}_{i} \hat{\beta }_{n}, r\in [0,1]\). It follows that
By Lemma 2 we get
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Fu, F., Zeng, Z. & Liu, X. Difference-based M-estimator of generalized semiparametric model with NSD errors. J Inequal Appl 2019, 61 (2019). https://doi.org/10.1186/s13660-019-2017-7
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s13660-019-2017-7