In this paper, some inequalities for a linearly negative quadrant dependent (LNQD) sequence are obtained. As their application, the asymptotic normality of the weight function estimate for a regression function is established, which extends the results of Roussas et al. (J. Multivar. Anal. 40:162-291, 1992) and Yang (Acta. Math. Sin. Engl. Ser. 23(6):1013-1024, 2007) for the strong mixing case to the LNQD case.
MSC:60E15, 62G08, 62E20.
Hinweise
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
The three authors contributed equally to this work. All authors read and approved the final manuscript.
1 Introduction
We first recall the definitions of some dependent sequences.
Random variables are said to be negatively associated (NA) if for every pair of disjoint subsets and of ,
where and are increasing for every variable (or decreasing for every variable) so that this covariance exists. An infinite sequence of random variables is said to be NA if every finite subfamily is NA.
Remark 1.1 (i) If is a sequence of LNQD random variables, then is still a sequence of LNQD random variables, where a and b are real numbers. (ii) NA implies LNQD from the definitions, but LNQD does not imply NA.
Because of wide applications of LNQD random variables, the concept of LNQD random variables has received more and more attention recently. For example, Newman [2] established the central limit theorem for a strictly stationary LNQD process; Wang and Zhang [4] provided uniform rates of convergence in the central limit theorem for LNQD sequence; Ko et al.[5] obtained the Hoeffding-type inequality for LNQD sequence; Ko et al.[6] studied the strong convergence for weighted sums of LNQD arrays; Wang et al.[7] obtained some exponential inequalities for a linearly negative quadrant dependent sequence; Wu and Guan [8] obtained the mean convergence theorems for weighted sums of dependent random variables. In addition, from Remark 1.1, it is shown that LNQD is much weaker than NA and independent random variables. So, it is interesting to study some inequalities and their applications to a regression function for LNQD sequence.
The main results of this paper depend on the following lemmas.
Suppose thatis a sequence of LNQD random variables with. Then for any, there exists a positive constantDsuch that
2 Main results
Now, we state our main results with their proofs.
Theorem 2.1LetXandYbe NQD random variables with finite second moments. Iffandgare complex-valued functions defined onRwith bounded derivativesand, then
Proof The proof follows easily from the brief outline of the main points of the proof of Theorem 4.1 in Roussas [9], p.773]. □
By Theorem 2.1, we establish an inequality for characteristic function (c.f.) as follows:
Theorem 2.2Ifare LNQD random variables with finite second moments, letandbe c.f.’s ofand, respectively, then for all nonnegative (or nonpositive) real numbers,
Proof Write
(2.1)
Further notice that . Thus,
(2.2)
By the definition of LNQD, it is easy to see that and are NQD for . Then by Theorem 2.1, we can obtain that
(2.3)
Similarly as above, we have
(2.4)
From (2.2) to (2.4), we obtain
(2.5)
Therefore, in view of (2.1) and (2.5), we obtain that
(2.6)
For , using the same decomposition as in (2.1) above, we obtain
Similarly to the calculation of , we get
(2.7)
Thus, from (2.6) and (2.7), constantly repeating the above procedure, we get
(2.8)
Note that for , and are NQD by the definition of LNQD. Similarly as above, we obtain that
This result, along with (2.8), completes the proof of the theorem. □
Theorem 2.3Letbe a sequence of LNQD random variables, and letbe all nonnegative (or nonpositive) real numbers. Then
Remark 2.1 Let , in Theorem 2.3, we can get Lemma 3.1 of Ko et al.[5]; let , , we also get Lemma 1.4 of Wang et al.[7]. Thus, our Theorem 2.3 improves and extends Lemma 3.1 in Ko et al.[5] and Lemma 1.4 in Wang et al.[7].
Proof For , it is easy to see that and are NQD by the definition of LNQD, which implies that and are also NQD for by Lemma 1.1(ii). Then by Lemma 1.1(i) and induction,
(2.9)
For , it is easy to see that and and are NQD by the definition of LNQD, which implies that and are also NQD for by Lemma 1.1(ii). Similar to the proof of (2.9), we obtain
(2.10)
Therefore, the proof is complete by (2.9) and (2.10). □
Theorem 2.4Suppose thatis a LNQD random variable sequence with zero mean anda.s. (). Letand. Then for any,
Proof We obtain the result from the proving process of Theorem 2.3 in Wang et al.[7]. □
Theorem 2.5Letbe a LNQD random variable sequence with zero mean and finite second moment, . Assume thatis a real constant sequence satisfying. Then for any, .
Proof Let , . Notice that
(2.11)
Let . Then is still a sequence of LNQD random variables with by Remark 1.1. Note that . By Lemma 1.2, we obtain
(2.12)
Similarly as above, we have
(2.13)
Combining (2.11)-(2.13), we get the result of the theorem. □
3 Application
To show the application of the inequalities in Section 2, in this section we discuss the asymptotic normality of the general linear estimator for the following regression model:
(3.1)
where the design points , which is a compact set of , g is a bounded real valued function on A, and the are regression errors with zero mean and finite variance . As an estimate of , we consider the following general linear smoother:
(3.2)
where a weight function , , depends on the fixed design points and on the number of observations n.
Here, our purpose is to use the inequalities in Section 2 to establish asymptotic normality for the estimate (3.2) under LNQD condition. The results obtained generalize the results of Roussas et al.[10] and Yang [11] based on strong mixing sequence to LNQD sequence. Adopting the basic assumptions of Yang [11], we assume the following:
Assumption (A1) (i) is a bounded function defined on the compact subset A of ; (ii) is a strictly stationary and LNQD time series with , ; (iii) For each n, the joint distribution of is the same as that of .
Denote
(3.3)
Assumption (A2) (i) for all ; (ii) ; (iii) .
Assumption (A3) for and .
Assumption (A4) There exist positive integers and such that for sufficiently large n and as ,
Here, we will prove the following result.
Theorem 3.1Let Assumptions (A1)∼(A4) be satisfied. Then
Proof We first give some denotations. For convenience of writing, omit everywhere the argument x and set , for , so that . Let . Then may be split as , where
Thus, to prove the theorem, it suffices to show that
(3.4)
(3.5)
By Theorem 2.5, Assumptions (A2)(ii)∼(iii) and (A4)(i)∼(iii), we have
Thus (3.4) holds.
We now proceed with the proof of (3.5). Let and , then . Apply relation (3.4) to obtain . This would also imply that , provided we show that .
Indeed, by Assumption (A3) and , we obtain . Then by stationarity and Assumption (A2), it can be shown that
(3.6)
Next, in order to establish asymptotic normality, we assume that are independent random variables, and the distribution of is the same as that for . Then and . Let , , then are independent random variables with and . Let be the characteristic function of X, then
(3.7)
By Theorem 2.2, relation (3.6) and Assumption (A2), we obtain that
(3.8)
Thus, it suffices to show that which, on account of , will follow from the convergence . By the Lyapunov condition, it suffices to show that for some ,
(3.9)
Using Theorem 2.5 and Assumptions (A2) and (A4)(iv), we have
So, (3.9) holds. Thus, the proof is complete. □
Acknowledgements
This research is supported by the National Natural Science Foundation of China (11061029) and the Science Foundation of Jiangxi Education Department (GJJ12604).
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
The three authors contributed equally to this work. All authors read and approved the final manuscript.