Skip to main content
Erschienen in: Journal of Inequalities and Applications 1/2012

Open Access 01.12.2012 | Research

Some inequalities for a LNQD sequence with applications

verfasst von: Yongming Li, Jianhua Guo, Naiyi Li

Erschienen in: Journal of Inequalities and Applications | Ausgabe 1/2012

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
download
DOWNLOAD
print
DRUCKEN
insite
SUCHEN
loading …

Abstract

In this paper, some inequalities for a linearly negative quadrant dependent (LNQD) sequence are obtained. As their application, the asymptotic normality of the weight function estimate for a regression function is established, which extends the results of Roussas et al. (J. Multivar. Anal. 40:162-291, 1992) and Yang (Acta. Math. Sin. Engl. Ser. 23(6):1013-1024, 2007) for the strong mixing case to the LNQD case.
MSC:60E15, 62G08, 62E20.
Hinweise

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

The three authors contributed equally to this work. All authors read and approved the final manuscript.

1 Introduction

We first recall the definitions of some dependent sequences.
Definition 1.1 (Lehmann [1])
Two random variables X and Y are said to be negative quadrant dependent (NQD) if
P ( X x , Y y ) P ( X x ) P ( Y y ) for any  x , y R .
A sequence of random variables { X n , n 1 } is said to be pairwise negatively quadrant dependent (PNQD) if every pair of random variables in the sequence is NQD.
Definition 1.2 (Newman [2])
A sequence { X n , n 1 } of random variables is said to be linearly negative quadrant dependent (LNQD) if for any disjoint subsets A , B Z + and positive r j ’s, k A r k X k and j B r j X j are NQD.
Definition 1.3 (Joag-Dev and Proschan [3])
Random variables X 1 , X 2 , , X n are said to be negatively associated (NA) if for every pair of disjoint subsets A 1 and A 2 of { 1 , 2 , , n } ,
Cov ( f 1 ( X i ; i A 1 ) , f 2 ( X j ; j A 2 ) ) 0 ,
where f 1 and f 2 are increasing for every variable (or decreasing for every variable) so that this covariance exists. An infinite sequence of random variables { X n ; n 1 } is said to be NA if every finite subfamily is NA.
Remark 1.1 (i) If { X n , n 1 } is a sequence of LNQD random variables, then { a X n + b , n 1 } is still a sequence of LNQD random variables, where a and b are real numbers. (ii) NA implies LNQD from the definitions, but LNQD does not imply NA.
Because of wide applications of LNQD random variables, the concept of LNQD random variables has received more and more attention recently. For example, Newman [2] established the central limit theorem for a strictly stationary LNQD process; Wang and Zhang [4] provided uniform rates of convergence in the central limit theorem for LNQD sequence; Ko et al.[5] obtained the Hoeffding-type inequality for LNQD sequence; Ko et al.[6] studied the strong convergence for weighted sums of LNQD arrays; Wang et al.[7] obtained some exponential inequalities for a linearly negative quadrant dependent sequence; Wu and Guan [8] obtained the mean convergence theorems for weighted sums of dependent random variables. In addition, from Remark 1.1, it is shown that LNQD is much weaker than NA and independent random variables. So, it is interesting to study some inequalities and their applications to a regression function for LNQD sequence.
The main results of this paper depend on the following lemmas.
Lemma 1.1 (Lehmann [1])
Let random variables X and Y be NQD, then
(i)
E X Y E X E Y ;
 
(ii)
If f and g are both nondecreasing (or both nonincreasing) functions, then f ( X ) and g ( Y ) are NQD.
 
Lemma 1.2 (Zhang [4])
Suppose that { X n ; n 1 } is a sequence of LNQD random variables with E X n = 0 . Then for any p > 1 , there exists a positive constant D such that
E | i = 1 n X i | p D E ( i = 1 n X i 2 ) p / 2 .

2 Main results

Now, we state our main results with their proofs.
Theorem 2.1 Let X and Y be NQD random variables with finite second moments. If f and g are complex-valued functions defined on R with bounded derivatives f and g , then
| Cov ( f ( X ) , g ( Y ) ) | f g | Cov ( X , Y ) | .
Proof The proof follows easily from the brief outline of the main points of the proof of Theorem 4.1 in Roussas [9], p.773]. □
By Theorem 2.1, we establish an inequality for characteristic function (c.f.) as follows:
Theorem 2.2 If X 1 , , X m are LNQD random variables with finite second moments, let φ j ( t j ) and φ ( t 1 , , t m ) be c.f.’s of X j and ( X 1 , , X m ) , respectively, then for all nonnegative (or nonpositive) real numbers t 1 , , t m ,
| φ ( t 1 , , t m ) j = 1 m φ j ( t j ) | 4 1 k < l m | t k t l | | Cov ( X k , X l ) | .
Proof Write
| φ ( t 1 , , t m ) j = 1 m φ j ( t j ) | | φ ( t 1 , , t m ) φ ( t 1 , , t m 1 ) φ m ( t m ) | + | φ ( t 1 , , t m 1 ) j = 1 m 1 φ j ( t j ) | = : I 1 + I 2 .
(2.1)
Further notice that e i x = cos ( x ) + i sin ( x ) . Thus,
I 1 = | E exp ( i j = 1 m t j X j ) E exp ( i j = 1 m 1 t j X j ) E exp ( i t m X m ) | | Cov ( cos ( j = 1 m 1 t j X j ) , cos ( t m X m ) ) | + | Cov ( sin ( j = 1 m 1 t j X j ) , sin ( t m X m ) ) | + | Cov ( sin ( j = 1 m 1 t j X j ) , cos ( t m X m ) ) | + | Cov ( cos ( j = 1 m 1 t j X j ) , sin ( t m X m ) ) | = : I 11 + I 12 + I 13 + I 14 .
(2.2)
By the definition of LNQD, it is easy to see that t m X m and j = 1 m 1 t j X j are NQD for t 1 , , t m > 0 . Then by Theorem 2.1, we can obtain that
I 11 | Cov ( j = 1 m 1 t j X j , t m X m ) | j = 1 m 1 t j t m | Cov ( X j , X m ) | .
(2.3)
Similarly as above, we have
I 1 i j = 1 m 1 t j t m | Cov ( X j , X m ) | , i = 2 , 3 , 4 .
(2.4)
From (2.2) to (2.4), we obtain
I 1 4 j = 1 m 1 t j t m | Cov ( X j , X m ) | .
(2.5)
Therefore, in view of (2.1) and (2.5), we obtain that
| φ ( t 1 , , t m ) j = 1 m φ j ( t j ) | 4 j = 1 m 1 t j t m | Cov ( X j , X m ) | + I 2 .
(2.6)
For I 2 , using the same decomposition as in (2.1) above, we obtain
I 2 | φ ( t 1 , , t m 1 ) φ ( t 1 , , t m 1 ) φ m 1 ( t m 1 ) | + | φ ( t 1 , , t m 2 ) j = 1 m 2 φ j ( t j ) | = : I 21 + I 22 .
Similarly to the calculation of I 1 , we get
I 2 4 j = 1 m 2 t j t m 1 | Cov ( X j , X m 1 ) | + I 22 .
(2.7)
Thus, from (2.6) and (2.7), constantly repeating the above procedure, we get
| φ ( t 1 , , t m ) j = 1 m φ j ( t j ) | 4 j = 1 m 1 t j t m | Cov ( X j , X m ) | + 4 j = 1 m 2 t j t m 1 | Cov ( X j , X m 1 ) | + 4 j = 1 m 3 t j t m 2 | Cov ( X j , X m 2 ) | + + 4 | Cov ( X 1 , X 2 ) | = 4 l = 2 m k = 1 l 1 t k t l | Cov ( X k , X l ) | = 4 1 k < l m t k t l | Cov ( X k , X l ) | .
(2.8)
Note that for t 1 , , t m < 0 , t m X m and j = 1 m 1 t j X j are NQD by the definition of LNQD. Similarly as above, we obtain that
| φ ( t 1 , , t m ) j = 1 m φ j ( t j ) | 4 1 k < l m | t k t l | | Cov ( X k , X l ) | .
This result, along with (2.8), completes the proof of the theorem. □
Theorem 2.3 Let X 1 , , X n be a sequence of LNQD random variables, and let t 1 , , t n be all nonnegative (or nonpositive) real numbers. Then
E [ exp ( j = 1 n t j X j ) ] j = 1 n E [ exp ( t j X j ) ] .
Remark 2.1 Let t j = 1 , j 1 in Theorem 2.3, we can get Lemma 3.1 of Ko et al.[5]; let t j = t > 0 , j 1 , we also get Lemma 1.4 of Wang et al.[7]. Thus, our Theorem 2.3 improves and extends Lemma 3.1 in Ko et al.[5] and Lemma 1.4 in Wang et al.[7].
Proof For t 1 , , t n > 0 , it is easy to see that j = 1 i 1 t j X j and t i X i are NQD by the definition of LNQD, which implies that exp ( j = 1 i 1 t j X j ) and exp ( t i X i ) are also NQD for i = 2 , 3 , , n by Lemma 1.1(ii). Then by Lemma 1.1(i) and induction,
E [ exp ( j = 1 n t j X j ) ] E [ exp ( j = 1 n 1 t j X j ) ] E [ exp ( t n X n ) ] = E [ exp ( j = 1 n 2 t j X j ) exp ( t n 1 X n 1 ) ] E [ exp ( t n X n ) ] E [ exp ( j = 1 n 2 t j X j ) ] E [ exp ( t n 1 X n 1 ) ] E [ exp ( t n X n ) ] j = 1 n E [ exp ( t j X j ) ] .
(2.9)
For t 1 , , t n < 0 , it is easy to see that t 1 , , t n > 0 and j = 1 i 1 t j X j and t i X i are NQD by the definition of LNQD, which implies that exp ( j = 1 i 1 t j X j ) and exp ( ( t i X i ) ) are also NQD for i = 2 , 3 , , n by Lemma 1.1(ii). Similar to the proof of (2.9), we obtain
E [ exp ( j = 1 n t j X j ) ] = E [ exp ( j = 1 n t j X j ) ] j = 1 n E { exp [ ( t j X j ) ] } = j = 1 n E [ exp ( t j X j ) ] .
(2.10)
Therefore, the proof is complete by (2.9) and (2.10). □
Theorem 2.4 Suppose that { X j : j 1 } is a LNQD random variable sequence with zero mean and | X j | d j a.s. ( j = 1 , 2 , ). Let t > 0 and t max 1 j n d j 1 . Then for any ε > 0 ,
P ( | j = 1 n X j | ε ) 2 exp { t ε + t 2 i = 1 n E X j 2 } .
Proof We obtain the result from the proving process of Theorem 2.3 in Wang et al.[7]. □
Theorem 2.5 Let { X j : j 1 } be a LNQD random variable sequence with zero mean and finite second moment, sup j 1 E ( X j 2 ) < . Assume that { a j , j 1 } is a real constant sequence satisfying a : = sup j 1 | a j | < . Then for any r > 1 , E | j = 1 n a j X j | r D a r n r / 2 .
Proof Let a i + : = max { a i , 0 } , a i : = max { a i , 0 } . Notice that
E | j = 1 n a j X j | r C { E | j = 1 n a j + X j | r + E | j = 1 n a j X j | r } , E | j = 1 n a j + X j | r = a r E | j = 1 n a j + a 1 X j | r .
(2.11)
Let Y j = a j + a 1 X j . Then { Y n , n 1 } is still a sequence of LNQD random variables with E Y n = 0 by Remark 1.1. Note that 0 < a j + a 1 1 . By Lemma 1.2, we obtain
E | j = 1 n Y j | r D a r n r / 2 , this implies that E | j = 1 n a j + X j | r D a r n r / 2 .
(2.12)
Similarly as above, we have
E | j = 1 n a j X j | r D a r n r / 2 .
(2.13)
Combining (2.11)-(2.13), we get the result of the theorem. □

3 Application

To show the application of the inequalities in Section 2, in this section we discuss the asymptotic normality of the general linear estimator for the following regression model:
Y n i = g ( x n i ) + ε n i , 1 i n ,
(3.1)
where the design points x n i , , x n n A , which is a compact set of R d , g is a bounded real valued function on A, and the { ε n i } are regression errors with zero mean and finite variance σ 2 . As an estimate of g ( ) , we consider the following general linear smoother:
g n ( x ) = i = 1 n w n i ( x ) Y n i ,
(3.2)
where a weight function w n i ( x ) , i = 1 , , n , depends on the fixed design points x n 1 , , x n n and on the number of observations n.
Here, our purpose is to use the inequalities in Section 2 to establish asymptotic normality for the estimate (3.2) under LNQD condition. The results obtained generalize the results of Roussas et al.[10] and Yang [11] based on strong mixing sequence to LNQD sequence. Adopting the basic assumptions of Yang [11], we assume the following:
Assumption (A1) (i) g : A R is a bounded function defined on the compact subset A of R d ; (ii) { ξ t : t = 0 , ± 1 , } is a strictly stationary and LNQD time series with E ξ 1 = 0 , Var ( ξ 1 ) = σ 2 ( 0 , ) ; (iii) For each n, the joint distribution of { ε n i : 1 i n } is the same as that of { ξ 1 , , ξ n } .
Denote
w n ( x ) : = max { | w n i ( x ) | : 1 i n } , σ n 2 ( x ) : = Var ( g n ( x ) ) .
(3.3)
Assumption (A2) (i) i = 1 n | w n i ( x ) | C for all n 1 ; (ii) w n ( x ) = O ( i = 1 n w n i 2 ( x ) ) ; (iii) i = 1 n w n i 2 ( x ) = O ( σ n 2 ( x ) ) .
Assumption (A3) E | ξ 1 | r < for r > 2 and u ( 1 ) = sup j 1 | i j | 1 | Cov ( ξ i , ξ j ) | < .
Assumption (A4) There exist positive integers p : = p ( n ) and q : = q ( n ) such that p + q n for sufficiently large n and as n ,
(i)  q p 1 0 ; (ii)  n q p 1 w n 0 ; (iii)  p w n 0 ; (iv)  n p r 2 1 w n r 2 0 .
Here, we will prove the following result.
Theorem 3.1 Let Assumptions (A1)∼(A4) be satisfied. Then
σ n 1 ( x ) { g n ( x ) E g n ( x ) } d N ( 0 , 1 ) .
Proof We first give some denotations. For convenience of writing, omit everywhere the argument x and set S n = σ n 1 ( g n E g n ) , Z n i = σ n 1 w n i ε n i for i = 1 , , n , so that S n = i = 1 n Z n i . Let k = [ n / ( p + q ) ] . Then S n may be split as S n = S n + S n + S n , where
S n = m = 1 k y n m , S n = m = 1 k y n m , S n = y n k + 1 , y n m = i = k m k m + p 1 Z n i , y n m = i = l m l m + q 1 Z n i , y n k + 1 = i = k ( p + q ) + 1 n Z n i , k m = ( m 1 ) ( p + q ) + 1 , l m = ( m 1 ) ( p + q ) + p + 1 , m = 1 , , k .
Thus, to prove the theorem, it suffices to show that
E ( S n ) 2 0 , E ( S n ) 2 0 ,
(3.4)
S n d N ( 0 , 1 ) .
(3.5)
By Theorem 2.5, Assumptions (A2)(ii)∼(iii) and (A4)(i)∼(iii), we have
E ( S n ) 2 = E ( m = 1 k i = l m l m + q 1 σ 1 w n i ξ i ) 2 D k q σ 2 w n 2 C ( 1 + q p 1 ) 1 n q p 1 w n 0 , E ( S n ) 2 = E ( i = k ( p + q ) + 1 n σ 1 w n i ξ i ) 2 D ( n k ( p + q ) ) σ 2 w n 2 C ( 1 + q p 1 ) p w n 0 .
Thus (3.4) holds.
We now proceed with the proof of (3.5). Let Γ n = 1 i < j k Cov ( y n i , y n j ) and s n 2 = m = 1 k Var ( y n m ) , then s n 2 = E ( S 1 n ) 2 2 Γ n . Apply relation (3.4) to obtain E ( S n ) 2 1 . This would also imply that s n 2 1 , provided we show that Γ n 0 .
Indeed, by Assumption (A3) and u ( 1 ) < , we obtain u ( q ) 0 . Then by stationarity and Assumption (A2), it can be shown that
| Γ n | 1 i < j k μ = k i k i + p 1 ν = k j k j + p 1 | Cov ( Z n μ , Z n ν ) | 1 i < j k μ = k i k i + p 1 ν = k j k j + p 1 σ n 2 | w n μ w n ν | | Cov ( ξ μ , ξ ν ) | C σ n 2 w n i = 1 k 1 μ = k i k i + p 1 | w n ν | sup j 1 t : | t j | q | Cov ( ξ j , ξ t ) | C u ( q ) 0 .
(3.6)
Next, in order to establish asymptotic normality, we assume that { η n m : m = 1 , , k } are independent random variables, and the distribution of η n m is the same as that y n m for m = 1 , , k . Then E η n m = 0 and Var ( η n m ) = Var ( y n m ) . Let T n m = η n m / s n , m = 1 , , k , then { T n m , m = 1 , , k } are independent random variables with E T n m = 0 and Var ( T n m ) = 1 . Let φ X ( t ) be the characteristic function of X, then
| ϕ m = 1 k y n m ( t ) e t 2 2 | | E exp ( i t m = 1 k y n m ) m = 1 k E exp ( i t y n m ) | + | m = 1 k E exp ( i t y n m ) e t 2 2 | | E exp ( i t m = 1 k y n m ) m = 1 k E exp ( i t y n m ) | + | m = 1 k E exp ( i t η n m ) e t 2 2 | = : I 3 + I 4 .
(3.7)
By Theorem 2.2, relation (3.6) and Assumption (A2), we obtain that
I 3 4 t 2 1 i < j k μ = k i k i + p 1 ν = k j k j + p 1 | Cov ( Z n μ , Z n ν ) | C u ( q ) 0 .
(3.8)
Thus, it suffices to show that η n m d N ( 0 , 1 ) which, on account of s n 2 1 , will follow from the convergence m = 1 k T n m d N ( 0 , 1 ) . By the Lyapunov condition, it suffices to show that for some r > 2 ,
1 s n r m = 1 k E | η n m | r 0 .
(3.9)
Using Theorem 2.5 and Assumptions (A2) and (A4)(iv), we have
m = 1 k E | η n m | r = m = 1 k E | y n m | r = m = 1 k E | i = k m k m + p 1 σ n 1 w n i ξ i | r D k σ n r w n r p r 2 C n p r 2 1 w n r 2 0 .
So, (3.9) holds. Thus, the proof is complete. □

Acknowledgements

This research is supported by the National Natural Science Foundation of China (11061029) and the Science Foundation of Jiangxi Education Department (GJJ12604).
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://​creativecommons.​org/​licenses/​by/​2.​0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

The three authors contributed equally to this work. All authors read and approved the final manuscript.
Literatur
1.
Zurück zum Zitat Lehmann EL: Some concepts of dependence. Ann. Math. Stat. 1966, 37: 1137–1153. 10.1214/aoms/1177699260CrossRef Lehmann EL: Some concepts of dependence. Ann. Math. Stat. 1966, 37: 1137–1153. 10.1214/aoms/1177699260CrossRef
2.
Zurück zum Zitat Newman CM: Asymptotic independence and limit theorems for positively and negatively dependent random variables. 5. In Statistics and Probability. Edited by: Tong YL. Inst. Math. Statist., Hayward; 1984:127–140. Newman CM: Asymptotic independence and limit theorems for positively and negatively dependent random variables. 5. In Statistics and Probability. Edited by: Tong YL. Inst. Math. Statist., Hayward; 1984:127–140.
3.
Zurück zum Zitat Joag-Dev K, Proschan F: Negative association of random variables with applications. Ann. Stat. 1983, 11(1):286–295. 10.1214/aos/1176346079MathSciNetCrossRef Joag-Dev K, Proschan F: Negative association of random variables with applications. Ann. Stat. 1983, 11(1):286–295. 10.1214/aos/1176346079MathSciNetCrossRef
4.
Zurück zum Zitat Zhang LX: A functional central limit theorem for asymptotically negatively dependent random fields. Acta Math. Hung. 2000, 86: 237–259. 10.1023/A:1006720512467CrossRef Zhang LX: A functional central limit theorem for asymptotically negatively dependent random fields. Acta Math. Hung. 2000, 86: 237–259. 10.1023/A:1006720512467CrossRef
5.
Zurück zum Zitat Ko MH, Choi YK, Choi YS: Exponential probability inequality for linearly negative quadrant dependent random variables. Commun. Korean Math. Soc. 2007, 22: 137–143. 10.4134/CKMS.2007.22.1.137MathSciNetCrossRef Ko MH, Choi YK, Choi YS: Exponential probability inequality for linearly negative quadrant dependent random variables. Commun. Korean Math. Soc. 2007, 22: 137–143. 10.4134/CKMS.2007.22.1.137MathSciNetCrossRef
6.
Zurück zum Zitat Ko MH, Ryu DH, Kim TS: Limiting behaviors of weighted sums for linearly negative quadrant dependent random variables. Taiwan. J. Math. 2007, 11(2):511–522.MathSciNet Ko MH, Ryu DH, Kim TS: Limiting behaviors of weighted sums for linearly negative quadrant dependent random variables. Taiwan. J. Math. 2007, 11(2):511–522.MathSciNet
7.
Zurück zum Zitat Wang XJ, Hu SH, Yang WZ, Li XQ: Exponential inequalities and complete convergence for a LNQD sequence. J. Korean Stat. Soc. 2010, 39: 555–564. 10.1016/j.jkss.2010.01.002MathSciNetCrossRef Wang XJ, Hu SH, Yang WZ, Li XQ: Exponential inequalities and complete convergence for a LNQD sequence. J. Korean Stat. Soc. 2010, 39: 555–564. 10.1016/j.jkss.2010.01.002MathSciNetCrossRef
8.
Zurück zum Zitat Wu YF, Guan M: Mean convergence theorems and weak laws of large numbers for weighted sums of dependent random variables. J. Math. Anal. Appl. 2011, 377: 613–623. 10.1016/j.jmaa.2010.11.042MathSciNetCrossRef Wu YF, Guan M: Mean convergence theorems and weak laws of large numbers for weighted sums of dependent random variables. J. Math. Anal. Appl. 2011, 377: 613–623. 10.1016/j.jmaa.2010.11.042MathSciNetCrossRef
9.
Zurück zum Zitat Roussas GG: Positive and negative dependence with some statistical applications. In Asymptotics, Nonparametrics, and Time Series. Edited by: Ghosh S, Puri ML. CRC Press, Boca Raton; 1999:757–787. Roussas GG: Positive and negative dependence with some statistical applications. In Asymptotics, Nonparametrics, and Time Series. Edited by: Ghosh S, Puri ML. CRC Press, Boca Raton; 1999:757–787.
10.
Zurück zum Zitat Roussas GG, Tran LT, Ioannides DA: Fixed design regression for time series: asymptiotic normality. J. Multivar. Anal. 1992, 40: 162–291.MathSciNetCrossRef Roussas GG, Tran LT, Ioannides DA: Fixed design regression for time series: asymptiotic normality. J. Multivar. Anal. 1992, 40: 162–291.MathSciNetCrossRef
11.
Zurück zum Zitat Yang SC: Maximal moment inequality for partial sum of strong mixing sequences and application. Acta Math. Sin. Engl. Ser. 2007, 23(6):1013–1024. 10.1007/s10114-005-0841-9MathSciNetCrossRef Yang SC: Maximal moment inequality for partial sum of strong mixing sequences and application. Acta Math. Sin. Engl. Ser. 2007, 23(6):1013–1024. 10.1007/s10114-005-0841-9MathSciNetCrossRef
Metadaten
Titel
Some inequalities for a LNQD sequence with applications
verfasst von
Yongming Li
Jianhua Guo
Naiyi Li
Publikationsdatum
01.12.2012
Verlag
Springer International Publishing
Erschienen in
Journal of Inequalities and Applications / Ausgabe 1/2012
Elektronische ISSN: 1029-242X
DOI
https://doi.org/10.1186/1029-242X-2012-216

Weitere Artikel der Ausgabe 1/2012

Journal of Inequalities and Applications 1/2012 Zur Ausgabe

Premium Partner