Journal
of Statistical
Planning
and Inference
9 (1984) 185-194
185
North-Holland
LAWS OF LARGE U-STATISTICS Krishna
NUMBERS
B. ATHREYA’
Iowa State University,
Malay
Ames,
IA 50011,
USA
GHOSH’
University
Leone
of Florida,
Gainesville,
FL 32611,
USA
Dayton,
OH 45435,
USA
Y. LOW3
Wright State University,
Pranab
K. SEN4
University Received
of North 18 May
Recommended
Abstract:
Carolina,
Chapel Hill, NC 27514,
1982; revised
by N.L.
manuscript
of finiteness
mean,
under the finiteness
U-statistics
under
Key words: tingales;
Classification: Asymptotic
U-statistics;
25 February
1983
law of large
numbers
is obtained
under
the
for some r> 1, and a weak law of large numbers
of the first moment.
conditions.
Stochastic
variance of a bootstrapped U-statistic pivot and the bias of the bootstrapped Subject
a strong
of the rth moment,
is obtained
parallel
received
USA
Johnson
For the bootstrapped
assumption
AMS
FOR BOOTSTRAPPED
The results convergence
are then extended of the jackknifed
is proved. The asymptotic U-statistic are indicated.
normality
to bootstrapped estimator
of the
of the bootstrapped
6OF15, 62699.
theory;
Bootstrap;
Laws of large numbers;
Means;
Reversed
(sub-)mar-
von Mises’ functionals.
1. Introduction Hoeffding’s (1948) U-statistics play a very useful role in problems of estimation and hypothesis testing. The class of one-sample U-statistics includes the sample mean, variance, Gini’s mean difference, signed-rank statistics, Kendall’s tau statist Research ’ Research 3 Research
supported by the National Science Foundation, Contract supported by the Army Research Office, Durham, Grant done while the author was visiting Iowa State University.
4 Research
supported
by the National
Heart,
Lung
and Blood
MCS-8201456. DAAG29-82-K-0136.
Institute,
2243-L.
0378-3758/84/$3.00
0
1984, Elsevier
Science
Publishers
B.V. (North-Holland)
Contract
NIH-NHLBI-71-
K.B. Athreya et al. / Bootstrapped
186
U-statistics
tics and many others. Efron’s (1979) bootstrapping is a resampling scheme in which a statistician attempts to learn the sampling properties of a statistic by recomputing its value on the basis of a new sample realized from the original one (this will be made more precise in the later sections). Bickel and Freedman (1981) have dealt very extensively with the limit laws of many bootstrapped functionals including t-statistics, CT-statistics, von Mises’ functionals and sample quantiles. Singh (1981) has also obtained asymptotic results justifying the validity of bootstrap approximations in certain cases. While studying the laws of large numbers for one-sample U-statistics, in Section 2, we prove first a strong law for the bootstrapped mean under the assumption of finiteness of a moment of order bigger than 1 and a weak law under the finiteness of the first moment. Next, in Section 3, a strong law of large numbers is proved for bootstrapped U-statistics under the finiteness of a moment of order bigger than 1. Also, the stochastic convergence of the jackknifed estimator of the variance of a bootstrapped I/-statistic is proved in this section. This result along with a result of Bickel and Freedman (1981) yield the asymptotic normality of the bootstrapped pivot (defined in Section 3) based on U-statistics.
2. Laws of large numbers for bootstrapped
means
Suppose that X,, . . . , X,, are independent and identically distributed (i.i.d.) random variables (T.v.) with an (unknown) distribution function (d.f.) F, and let xl, . . . ,x, be the corresponding observed realizations. Let X, = (X,, . . . , X,) and x,,). Then, the bootstrapping procedure may be described as follows. x,=(x,,..., (i) Construct the sample (empirical) d.f. F, (from the x,). (ii) With F,, fixed, draw a simple random sample (with replacement) of size m from the set of indices (1, . . . , n). The corresponding sample, called the bootstrapped sample, is denoted by X,*, = (X,*,, . . . , X,*,>, where, conditional on X,=x,,, the X$ are i.i.d. each assuming one of the n possible values x1, . . . ,x, with equal probability n-‘. (iii) Use the bootstrapped sample to estimate the variability or other characteristics of the estimator from the original sample. Our first theorem relates to the weak law of large numbers for the bootstrapped mean under the minimal assumption of finite first moment. 1. For the bootstrapped sample X,*, from X,,, let X$, =m-’ Cy’, Xz andX,,=n-’ En=, Xi. A ssume that EX, = p is finite. Then, X,$, - X, converges in probability to 0 as min(n, m) --t 03. Theorem
Proof.
The method of proof employs the characteristic q&,(t)
= E[exr(it{xz,
function technique.
- -%>)I = E(E[exp(it{%,-%I)
= E{ b#$Wm)l”),
Let
1&I) (2.1)
K.B. Athreya et al. / Bootstrapped U-statistics
187
where @,*(0)=E[exp(iB{X,*, --Xn}) 1X,], f or every real 0. For proving the theorem, it suffices to show that e,,,(t) + 1 as min(n,m) + 03. Since I#,*(t/m)l I 1, using the dominated convergence theorem, it suffices to show that {@,*(t/m)}” 2 1, as min(n, m) -+ 03, for every real t. Writing F,(U) = n-r En=, I(& -X,, I u), where Z(A) stands for the indicator function of the set A, one has the representation @,*(t/m) = 1 + R,(t/m), where
(2.2)
cc R,(t/m)
=
[exp(itu/m)
- 1 -i&/m]
U,(u).
(2.3)
.I’-cc
In view of (2.2), in order to show that {@,*(t/m)}“‘~ 1 as min(n, m) -+ 00, it suffices to show that m IR,(t/m)J 5 0 as min(n, m) --f 00. Now, using standard inequalities for characteristic functions [see e.g. Feller (1966, p. 487)], m IR,(t/m)I
I +t*m-’
u*dF,(u)+2 s ju1 sm”3
5 +t2m-‘/3 +2 ItI n-l
It/
f IXi-znl
lu I fin(u) s Iu/ >m”3
Z((Xj-Xnl>m1’3).
(2.4)
i=l
Thus, it remains only to show that as min(n, m) -+ 00,
=E[JX,-~~IZ(IX1-~~J>m”3)J,0.
(2.5)
This follows directly from the uniform integrability that m1’3 + 00 as min(n, m) + 00. 0
of X1 -2,
along with the fact
The next theorem relates to a strong law of large numbers. Theorem
2. Assume
that EJX, 1’ “<
00 for some BE (0,l).
Then, for every E>O
and d>O, sup sup IX&-XnI>s n>N
-0
asN-roo.
msdn
(2.6)
Remark 1. Note that (2.6) implies in particular
that x$-x,, +O a.s. as n -+ 00. A similar a.s. convergence result is given in Athreya (1983) where conditions relating to the growth rate of m with n and the moments of X, are given to ensure the a.s. convergence of .Y:m -X,, to 0. The method of proof used by Athreya (1983) is entirely different from the present method. Proof of Theorem 2. First, we prove the inequality that for every m 5 dn (d > 0) and
r 11, there exists a positive, finite number C,(d), such that E[(X,*, - Xn)” 1A’,] 5 C,(d)n-‘sir
for every n L 2,
(2.7)
K.B.
188
Athreya
et al. / Bootstrapped
U-statistics
where s,’ = n-r Cl=, (Xi - X,J2. Towards this proof, we note that (XZm_X”)2’ = m-2r i;,***i& m ,fi, (XZ, -%I). *r
(2.8)
Note that if one of the suffixes, say, c, appears exactly once in a summand of (2.8), then, by virtue of the conditional independence of the X2, E(X$ -Xn 1X,) = 0 a-e., so that the conditional expectation of the corresponding summand is also 0 a.e. A general term appearing in the right hand side of (2.8) with all distinct suffixes having possibly non-zero expectation is of the form (X2, - Xn)kl . *. (X2” - XJku where kr + .-. +k,=2r, The conditional
expectation
5
k,z2,
I= l,..., ~(21).
of this typical term, given X,, is equal to
i (Xi -X”)’
‘-“.+
= nrWu$
for every U 1 1.
I
i=l
(2.9)
Also, the number of such terms in (2.8) is O(nP). Hence, from (2.8) through (2.9), we obtain that E&Y& - XJzr 1X,] I K(r)m_2’
( j,
nr-‘mu)
s,” (2.10)
5 C,(d)n-‘sir,
as mend, d >O. Note that K(r) and C,(d) do not depend on n. This proves (2.7). Let us define then W,,, =
n-’ f: IXi-~l'++d, i=l
where 6 is defined as in the theorem.
$5 n-l f (Xi-/4)2S i=l
Note that
max IXi-j.l’-d
I Isicn
5 { i ,xi-a,“d](‘-6)/(i’~)w~,, i=l
Therefore,
W,,, I
= n(1-6)/(1+6)Wnf/bl+6).
(2.,,)
by (2.7) and (2.11), we obtain that E[(X,*, - XJ2’ 1X,] I C,(d)n-2”‘o+6)
W,frb/(‘+‘),
(2.12)
for every m>dn, d>O, nz2. Next, write A N for the event within the braces of (2.6), and let BN = {sup,,,,, W,,, s g], where g (BE IX, -,u I’+‘) is a finite constant. Note that the W,,, form a backward martingale, so that by the Kolmogorov maxi-
K.B. Athreya et al. / Bootstrapped U-statistics
ma1 inequality,
189
we have
P(B,)=P
/X,-pI1+‘>g-E
sup W,,,-E
LIIZN
I
{g-E
-p,ll+a}-lE
IX,
+O
I&-PI’+’
1W,,-E
IX, -p,l’+‘l (2.13)
asN-+=,
where the last step follows from the L,-convergence using the Markov inequality and (2.12), we have P&B/V)5
c P
>&, wn,s rg
I mkdn
fl;rN
s,gN
sup /X;*-XJ
of the mean (Wn,s). Also,
I
E[I(W,,6~g)C,(d)n-26r’(‘+d)W,f’6(1+6$’~2r
~C,*(g,c,r,d)
c n-26r’(1+d)~C*(g,~,r,d)N1-26r’(‘+6),
(2.14)
n>N
where C,* and C* are finite positive constants. Choose r (> 1) such that 2&/( 1 + 6) = 1 + y, for some y>O. Then, noting that P(AN)5P(ANBN) +P@,), (2.6) follows from (2.13), (2.14) and the proper choice of r. q
3. Laws of large numbers for bootstrapped
U-statistics
Let X,, . . . , X, be i.i.d. r.v. with the d.f. F, and for a kernel $Jof degree r, the U-statistic based on X,, . . . , X,, (with n 2 r) is defined as u,= 0 The corresponding denoted by u,*, = For notational
nr -’ F @(X+...,X,); 9’ bootstrapped
C,,,=
U-statistic
{15i,<...
based
on X,*l, . . . , X,*, (m 1 r) is
-1 m 0 r
F @(X& WX,:;). m.r
(3.2)
simplicity, we consider only the case of r = 2. Also, we denote by
IIOIlp= max{E I@(XI,X~)Ip, E I@(X19X2)lp>, The following theorem is a direct generalization Theorem
P>O.
(3.3)
of Theorem 2 to U-statistics.
3. If Ij@lj,+~<00 for some 6 E (0, l), then, for every E> 0, d > 0,
sup sup IU$,nzN
Proof.
(3.1)
Define
madn
U,I>E
+O
as N-t-.
(3.4)
3
V, =nm2 CyS, I,“,, @(Xi, Xi>. Then,
under
the assumption
that
190
K. B. Athreya et al. / Boo&trapped U-statistics
/1@11 t c 00 (implied by (3.3) for p> 1), it follows that P{sup,,, 1f_J,- V, I> E} -+ 0 [see Ghosh and Sen (1970)]. Hence, in (3.4), it suffices to replace U,, by V,,. Next, we make use of the Hoeffding (1961) decomposition for U,*, under the conditional setup, given X,,. Then, we have i-J,*, -
v n = 2u”’“In + u(2) nmr
(3.5)
where
U$=m-'
i
,
n-'i @(X,:.,Xj)-v,
i=l
(3.6)
j=l
(3.7) Note that if we let
Z~~=#(Xn*iX$)-n-’ 2 @(X~*Xj)-n-'
i
@(Xi,X$)+Vn,
i=l
j=l
for i,j= I,..., m, then, given X,,, Zn:j and Zz& are conditionally orthogonal whenever (i,j)#(r,s) (i.e. at least one of the two indices is different) (given X,,), the same is true for more than two Zzj’s. As such, we may repeat the steps in (2.8) through (2.12) and claim that for every r-2 1, mzdn (d>O) and nz2, E[( U$)”
1X,,] 5 Cr(d)n-2r’(’
+‘I( W;J’(’
+ @,
(3.8)
where WnT6= n-* i i /@(Xi,Xj)l’+’ i=[ j=l =n
-I
I
n-1
$,
= n-‘w;‘j+.-‘(,_])w(*f
Since W,$,
n,61
P
we proceed as in (2.13)-(2.15) sup sup IU(*)l>c nm
medn
nzN
-+O
;)-’
F
I@(Xi,Xj)[‘+d n.2
say.
j = 1,2, are both reverse martingales
by ll@ll,+~<~)~ and d>O,
Therefore,
+y(
l@(xi,X,)ll+‘]
(3.9)
with finite expectations (ensured and conclude that for every E> 0
as N-00.
(3.10)
3
to prove (3.4), it suffices to show that for every e>O and d >O, SUP
supIU,$I>e
madn
nzN
-0
asN--+aJ.
(3.11)
V,, i= Now, given X,, Un(fn)=m-’ C!” , r=l ZniO where Znio=n-’ C!‘= I 1 4(X? nrt X,)1, -**, m, are conditionally independent r.v. and ZniO can only assume n values w,, with equal (conditional) probability n-l. (=n-’ C,“,, @(Xa,Xj)V,), 1 casn,
K. B. Athreya et al. / Bootstrapped
U-statistics
191
Note that C”, , w,; = 0 and
= 2,wvcl+a){~n
3)241+6),
(3.12)
say,
where U,,, is a von Mises functional with finite expectation and hence converges a.s. to a finite constant as n -+ 00. As such, we proceed as in (2.8) through (2.15) and conclude that (3.11) holds. This proves (3.4). 0 The final result of this section concerns the convergence in probability of the jackknifed estimator of the variance of a bootstrapped U-statistic under the minimal assumption of the finiteness of the second moment of the kernel. Towards this, we introduce the following. Let
and, for every i (= 1, . . . , m), let
b(‘) nm = (m
f
- 1)-t
@(X,*i,X$),
(3.14)
j=l(j+i)
s2nm = (m - 2)-24(m - 1)2 (m - l)-’ f (@A - U,*,)’ i=l [ =4(m-1)2(m-2)-2
(m-1)-$,
{b~~}2-_(111--1)-1,{U,*,}2
4.
If 1/@/j2<00, then, as min(m, n) + 03, s,‘, --t 46,
Proof.
1.
(3.15)
Then, we have the following. Theorem
1
in probability.
(3.16)
First, note that E[U,*, 1X,] = V,, and H(LSr, - VII21= Eb%vL 5 3m-‘E 5
3m-’
- v,12 1&II} ne2
ll@l12~
(3.17)
192
K. B. Athreya et al. / Bootstrapped U-statistics
where the penultimate step follows by using (3.5)-(3.7). Chebyshev inequality, under 11 @I/*< 03, for every E < 0, P{IU,*,-V,J>s}+O
Hence, by (3.17) and the
as m-m.
(3.18)
Again, under the assumption that /I@J II1< w (implied by 11#[12
in probability
as min(m, n) + 03.
(3.19)
Hence, it suffices to show that
m-l f, {bg}2-
4-I+
e2 in probability as min(m, n) + 00.
(3.20)
To this end, we write a,; = ZniO+ V,, where the ZniOare defined after (3.11). Then, noting that ani =E[@(X$,X,$) 1X,,Xz], we have
E[(b$h-Qa,i)2] = E{E[(btJj-C7a,i)2 1Xn]} 5(m-1)-‘~~@~~2+0
as m+o3.
(3.21)
Note that by (3.21), E
m-’
i
i=l
(b$,-aa,i)2
as m-t-,
=E[bA$-a,l]2-*0 I
so that m-’ ir, (b:i - Q,;)~ + 0
in probability
as m + co.
(3.22)
Let us now define
a,*i= @(X~,y)dF(y)=&(X~),
i=l,...,
m.
(3.23)
s
In view of (3.22) and Lemma 1 of Sen (1960), to prove (3.20), it suffices to show that in probability
as min(m, n) -+ 00,
(3.24)
and m-’ f,
a,*i?+ [, + ~9~ in probability
as min(m, n) + 03.
Note that (3.25) follows directly from Theorem 1 by noting that E&(X,)s To prove (3.24), it suffices to show that as min(m, n) + 00,
me’i=l i (a,i-a,*,)2 =E[u,~-u,*~]~-+~. I
(3.25) II@112.
(3.26)
K.B. Athreya et al. / Bootstrapped U-statistics
Now,
193
by definition, E(a,, -
a,*J2= E (n - l)(n - 2)
#l)
n2 n +2(n-l)Uc2)+n-1 7 n2 ”
1,
Ui3’ + ne2UJ4)
where
(3.27)
U(l) n = [n(n - l)(n - 2)]-’ x {wG,&) UA2’= [n(n - l)]-’
C
- @*(X;)},
(3.28)
{@Cxi,xi>- @lwi>>
I cifjcn
Ui3) = [n(n - 1)1-l
C
x ~@-cPl(xi)~9
(3.29)
{0(xi~xj)-4~(xi))2~
(3.30)
lri#jrn
and
VA’)= n-l J, {O(xi~xi)-OI(xi))2~
(3.31)
Symmetric versions of these kernels may be obtained by averaging over permutations. Since U$), r= 1, . . . , 4, are all U-statistics with EU,(‘)= 0 and E 1U$‘j < 00, for r= l,..., 4, using the backward martingale property of U-statistics, it follows that E IUn(‘)/‘0, as n-+ 03 [see Chow and Teicher (1978, p. 377)], so that by (3.27), 0 E(a,, - azI)2 + 0 as n + 00. This completes the proof of the theorem. < 03 for some 6> 0, then, expressing s& as a linear combinaRemark 2. If 1/@1(2+6 tion of bootstrapped U-statistics and using Theorem 2, one can show that for every e>O and d>O, sup sup js&--4~,l>c naN
madn
+O
asN-,=.
(3.32)
1
Remark 3. Although the unbootstrapped U-statistics U-statistics are typically biased estimators of 8 = EU,. vation that E[U,*, 1X, = V,] is a von Mises functional, unbiased estimator of 8. The bias is however of the
are unbiased, bootstrapped This follows from the obserand, in general, V, is not an order O(n-‘).
Remark 4. Although we have not spelled explicitly, by virtue of the backward martingale property of U,*, and the components of V,, the weak law of large numbers for U$, holds under the condition that (I@11,< 03. Remark 5. The bootstrappedpivot for U-statistics can be taken as m”2(U:m - U,)/ nm. From Theorem 3.1 of Bickel and Freedman (1981) [which can easily be S generalized to the case where m # n], it follows that 9 m”2(U,*m - U,)/(4[,) m, N(0, 1) as min(m,n)-
K.B. Athreya
194
when 114112< 03. Combining /1#1/2<~,
this
result
with
U-statistics
out
Theorem
4, we have
under
as min(m,n)+m,
z
rrP2(U,*m - UJ/s,, A similar
et al. / Bootstrapped
conclusion
also holds
N(O,1).
(3.33)
for bootstrapped
von Mises functionals.
References Athreya, Bickel,
K.B. (1983). P.J.
and
D.A.
A strong
law for the bootstrap.
Freedman
(1981).
Statist.
Some asymptotic
Probability
theory
Letters.
1, 147-150.
for the bootstrap.
Ann.
Statist.
9,
1196-1217. Chow,
Y.S. and H. Teicher
Springer,
(1978). Probability
Theory,
methods:
look at the jackknife.
Efron,
B. (1979).
Feller,
W. (1966). An Introduction
Bootstrap
Another fo Probability
York. Ghosh, M. and P.K. Sen (1970). On the almost functions. Hoeffding,
Calcutta
Statist.
Assoc.
Bull.
W. (1948). A class of statistics
19, 293-325. Hoeffding, W. (1961). Carolina,
Independence,
Interchangeability,
Martingale.
New York.
Mimeo
The strong
Ann.
Theory and Applications,
sure convergence
Sfatist.
7, l-26.
Vol. 2. John
of von Mises’ differentiable
Wiley,
New
statistical
19, 41-44. with asymptotically
law of large
numbers
normal for
distribution.
u-statistics.
Inst.
Ann. Math. Statist.
Univ.
Statist. North
Ser. No. 302.
Sen, P.K. (1960). On some convergence properties of u-statistics. Calcutta Statisf. Assoc. Bull. 10, 1-18. Singh, K. (1981). On the asymptotic accuracy of Efron’s bootstrap. Ann. Sfatist. 9, 1187-1195.