The central limit theorem for Tukey's 3R smoother

The central limit theorem for Tukey's 3R smoother

Statistics & Probability North-Holland 2 January Letters 13 (1991) 29-37 The central limit theorem smoother Wlodzimierz Department Bryc and Magda ...

697KB Sizes 0 Downloads 66 Views

Statistics & Probability North-Holland

2 January

Letters 13 (1991) 29-37

The central limit theorem smoother Wlodzimierz Department

Bryc and Magda Peligrad

of Mathematical

Sciences,

University

1992

for Tukey’s 3R

*

of Cincinnati, Cincmnati, OH 45221-0025, USA

Received May 1990 Revised May 1991

Abstract: Asymptotic properties of a non-linear smoothing algorithm are analyzed. It is shown that under appropriate output of the algorithm is absolutely regular and that partial sums are asymptotically normal. AMS

1980 Subject Ckzssifications:

Keywords:

3R smoother,

Primary

60F05;

weak dependence,

Central

Secondary

conditions

the

60GlO.

Limit Theorem.

1. Introduction and notation In his discussion of exploratory analysis of time series Tukey (1977) defined and advocated a non-linear smoothing algorithm in which the running median of three consecutive quantities is applied repeatedly until no further change occurs. The algorithm, called the 3R smoother, has the properties of being resistant to occasional outliers and at the same time responsive to changes in trend. Mallows (1979, 1980) discussed some theoretical aspects of the 3R smoother for strictly stationary sequences of random variables and he also raised the question as to under what conditions the output of the algorithm is either strongly (Rosenblatt) mixing, or +mixing; see Mallows (1979, p. 83). Bradley (1984) answered both questions, showing that: (i) the ouput of the 3R smoother is strongly mixing, provided the input is; he gave also explicit bounds for the convergence rate of the strong mixing coefficients. (ii) the output of the 3R smoother may fail to be +mixing even for an i.i.d. input. In this paper we prove that the absolute regularity condition is preserved by the 3R smoother, complementing (i) above. We also have something to say about case (ii): we use the ‘two part mixing’ conditions, see Bradley and Peligrad (1986), to analyze the dependence structure of the output of the 3R smoother if the input is a +mixing strictly stationary random sequence. As a consequence we show that if the input sequence is +mixing then even though the output of the 3R smoother may fail to be $-mixing, under appropriate moment conditions we have enough information to establish the Central Limit Theorem. Furthermore, if the input sequence is i.i.d., then we show that the output of the 3R smoother satisfies the Central Limit Theorem with the rate of convergence close to the best possible, even though the input sequence is not assumed to have second moments. None of those Central Limit Theorems could be obtained from the previously known fact that the strong mixing condition is preserved by the 3R * Supported

in part by NSF grants

0167-7152/92/$05.00

DMS-8905615

and DMS-9007986.

0 1992 - Elsevier Science Publishers

B.V. All rights reserved

29

STATISTICS & PROBABILITY

Volume 13, Number 1

LETTERS

2 January 1992

smoother. Our Lemma 5.2 below is also of independent interest, see condition A5 in Mallows (1980, p. 701). To give the formal definition of the 3R smoother, consider the running median mapping R : RE ---)Iw‘, defined for x = (x~)~ E E by R(x) ByRN=Ro

= (med(-+,,

xk?

Xk+l))k~E~

0 R we denote the n th iteration X0 is defined as

. . .

with the input

X” = lim.,,

(1)

(composition)

of R. The output

Xm of the 3R smoother

R"(X')

(2)

provided the limit exists (on each coordinate). From discussion in Tukey (1977), Mallows (1979) and from Theorem B in Bradley (1984) it follows that the limit (2) exists a.s. and is stationary and mixing, provided the input X0 = (X,“) is strictly stationary and mixing in the ergodic theoretic sense. Notice that when applied to finite sequences of data, the algorithm is usally modified by discarding end-point observations in each iteration, see e.g. Tukey (1977, pp. 210-213). This does not lead to significant loss of information, as the algorithm usually reaches fixed point in several iterations; the last claim is supported both by the numerical evidence on several sets of real data in Tukey (1977) and by theoretical estimates (3.1) and (3.2) below. It might be also worth pointing out that the 3R smoother algorithm with an i.i.d. input is an example of the ‘cellular automaton’, see e.g. Schonmann (1990) for another ‘transition rule’ and references. Except in the simplest case d = 1, little seems to be know about asymptotic behavior of variants of the 3R smoother defined for general Zd-indexed (even O-l valued) i.i.d. input sequences. We shall be interested in the strong mixing properties of the output X”. Those will be expressed in terms of the strong mixing coefficients defined as follows: be a strictly stationary sequence of random variables. Define past and future u-fields Let x=(xk>kGz 9,(X) = u( X1 : j < k) and pn( X) = a( X, : j > n) (the dependence of u-fields on X will be omitted, if no misunderstanding results). Put CX,(X)=SU~{(P(A~B)-P(A)P(B)):AEP~(X), (P,(x)=sup{IP(BIA)-P(B)I:AE~o(x),

4??(X)= sup

P(A n B)

il

P(A)P(B)

- 1 : A Ego(X),

BEF~(X)}, BE%(X),

BEE(X),

(3)

%+O},

P(A)P(B)

(4)

> 0 ) 1

i,J where the supremum in (6) is taken over all finite PO(X)-measurable partitions {A;} and over all finite sn (X)-measurable partitions { B, }. A stationary sequence X = (Xk) is called: - strongly mixing, if a,(X) -0 as n -+ cc; - +-mixing, if +n(X) + 0 as n ---f co; - +mixing, if q,(X) + 0 as n --j co; - absolutely regular, if p,(X) + 0 as n + 00. It is known that 0 G a,(X) G /I,(X) G I&(X) G 4,(X). Dependence of left hand sides of (3)-(6) on X will be omitted, if no misunderstanding results. ‘ e as m + 00’ instead of O(m); w denotes the standard Brownian In this paper we use notation motion on [O,l], [x] denotes the integer part of x.

30

Volume

13, Number

STATISTICS

1

& PROBABILITY

LETTERS

2 January

1992

2. Results Theorem 1. Suppose that X is a stationary absolutely regular randon sequence. Furthermore there is 0 < q < 1 and C > 0 such that for all n, k >, 1, P&k+4(Xa))

G P,,,(X)

+ CW

Then X” = Rm(X)

exists.

” %1(X)).

(7)

(Our proof gives C = 4(1 - q)-‘,) The following corollary lists some of the consequences of Theorem 1; in particular, it shows that the absolute regularity is preserved by the 3R smoother and explicit bounds for the rate of convergence to 0 are found. Corollary (i) Zf (ii) Zf (iii) If n -+ co. (iv) If

1. If X p,,(X) p,(X) p,,(X)

is + < <

a strictly stationary random sequence then the following implications hold: 0 as n -+ co, then X” = R”(X) is defined a.s. and &(Xm) + 0 as n ---) 00. ne8 as n + 00 for some 8 > 0, then /3,(X”) -=z nPej2 asn-+co. pn as n + co for some 0 < p < 1 then there is 0 < q < 1 such that /3,(X”) < qfi

X is m-dependent,

then j3,(Xcc)

Since it is known that aperiodic Harris (1986), we get also the following result.

as

+z p” for some 0 -=zp < 1. recurrent

Corollary 2. If X = (X,), rZ is an aperiodic X” = R”(X) exists and is absolutely regular.

Markov

Harris

chains

recurrent

are absolutely

stationary

Markov

regular,

see e.g. Bradley

chain,

then the output

q

From Corollary 1, using limit theorems known in the literature, one can deduce limit theorems for partial sums of X” under appropriate assumptions on a rate of convergence to 0 of b,(X) and under some additional moment assumptions. However, if the input sequence is $-mixing, we do not need to impose any conditions on the rate of convergence of Q,,(X) to 0. This is achieved by using the approach developed in Bradley and Peligrad (1986), together with the following result. Theorem 2. Suppose that X is a strictly stationary a.s. and there are C < co, 0 6 p < 1 such that IP(A for allA

f-l B) - P(AP(B)

ELP,,(X~),

IG

G-mixing

~[.,2-4](X)f$4)

random sequence.

+

CP”

is defined

(8)

BEAM.

The following result establishes the Central Limit Theorem sequence; no mixing rate assumption is required. Let ST=

Then X* = R*(X)

i:

xr,

under

G-mixing

dependence

of the input

u,’ = Var( Snm).

k=O

Theorem 3. Let X be a strictly stationary $-mixing random sequence. Suppose E{ 1X,, ) 2ts} -C co for some 6 > 0. Then X” = R”(X) exists, and E{ 1X,M ) ‘+‘} < co. If in addition for each E > 0, u,‘/n’-’

+ 00

(9) 31

Volume 13, Number 1

STATISTICS & PROBABILITY LE’M’ERS

as n + CO. then (Snw - nE{ Xr })/a,, + N(0, 1) in distribution

2 January 1992

as n + 00. Moreover,

(10) weakly in D[O,l] as n + co. Furthermore, if X is symmetric then E{ XF} = 0.

(i.e., ( XP)P E z and ( - XP)P E z have same finite dimensional

distributions),

The following theorem shows that if the input is an i.i.d. sequence, the moment assumptions for the Central Limit Theorem can be significantly relaxed. In particular the CLT holds even when the input sequence does not have finite second moments. Furthermore, the rate of convergence in the Central Limit Theorem can be analyzed, and (9) can be dropped. Theorem 4. Suppose X is an i.i.d. non-degenerate random sequence such that E{ 1X0 )log’ +&12(1 + 1X0 I)} < cc for some 6 > 0. Then X” = R”(X) exists and E{ ] X7 I2 log2+“(1 + ( X,M I)} < co. Furthermore o,‘/n --$ a2 > 0 as n + M and (S,m - nE { Xr })/an’12 -+ N(0, 1) in distribution as n + 00. Suppose in addition that E{ 1X,, ( 3/2 } < 00. Then there is a constant B = B(X) such that < Bn-7’2

-CX Moreover,

if X,-, is symmetric,

then XF IS symmetric

log’ n.

and hence E { XT}

Remark. If for some 0 the distribution of X,, - 8 is symmetric, symmetric; in particular Theorem 4 allows to write the normal has no second moments. Theorems 1, 2 and the corollaries and 5 respectively.

are proved

in Section

= 0.

then the same proof shows that XF - 13is confidence interval for 8 even when X0 - 8

3. Theorems

3 and 4 are proved

in Sections

4

3. Proofs of Theorems 1 and 2 Our proof monotonic

follows the general idea of Bradley (1984). An ordered triplet of numbers if either a 6 b < c or c < b < a holds. Define the ‘random times’: T+=inf{t>,O:(X,_,, T-=inf{s>O:(X_,_,,

X, , X, + , ) is monotonic} X_,,

X_,+,)

(a,

b, c) is called

,

ismonotonic)

We shall show that that for every n, m, k >, 1, P(T”

>, n + m + k) < P(T”

>, n)P(TO

> m) + CY~(X),

P(T”

>, n + m + k) < P(T”

>, n)P(T”

> m) + $,(X)P(T”

(11) >, n),

(12)

where To is either T+ or T-. To prove (11) and (12) we shall consider the case To = T+. Since T- is the same as Tt for the time-reversed sequence, the proof for the case To = TP is similar and is omitted.

magnitude. In particular Observe that Tf > n means that the numbers X0, X,, . . . , X, _ 1 are of alternating and T+ > n + m + k implies the occurrence of both events A = { T+ > n } and T+>, n is .%$_, -measureable 32

STATISTICS & PROBABILITY

Volume 13, Number 1

2 January 1992

LETTERS

X,, X,,,) for any n + k G r Q n + m + k}. Hence P(T” B = {there is no monotonic triple (X,-i, using (3) or (4) respectively we get + k) G P(A n B). Since A Ed,_, and B E.S$+~-,(X),

>, n + m

P(AnB)~P(A)P(B)+cu,(X), P(A

n B) G P(A)P(B)

+G,(X)P(A).

By stationarity P(B) = P(T o > m). Therefore (11) and (12) follow. We shall also need the following observation: If X is strictly stationary and mixing (in the ergodic theoretic sense), then P(( X_,, X,,, X,) is monotonic) > 0, see Bradley (1984, p. 254). In particular P(T” an) G P(T” 2 1) = 1 - P((X_,, X0, X,) is monotonic) -C 1 for every n 2 1. Put q = 1 - P(( X_,, The following

X0, X,) is monotonic).

two lemmas

(13)

make use of (11) and (12) to estimate

tails of random

variables

T+ and T-.

Lemma 3.1. For every n, k >, i, P(T”

> nk) < 2(1-

q)-‘(a,_,

V

(14)

q”),

where 0 < q < 1 is defined by (13) and cxk = CQ(X)

is defined by (3).

Proof. Since P(T o 2 k) < q, from (11) we get P(T”

>, (a + 1)k)


> nk) + CQ_,.

Therefore P(T”&nk)
...

<2(1-q)-1(q”vcu,_,).

+q(Cy,_,+q2)...)=qn+cx_l

‘;g;’

0

Lemma 3.2. If $” + 0, then there are 0 Q fi < 1 and d < cc such that P(TO>n)<@

foreueryn>l.

Proof. Let k 2 1 be such that 0 = c@~_,+ q < 1, where q is defined P(T o 2 (n + 1)k) G 0P(T o > nk). Hence P(T o > nk) 6 8” and the

C=eP.

by (13). Inequality (12) implies conclusion holds with F = 61’k,

0

We shall also use the following Lemma A. ZfrfA~a(R~(X),:O
B

lE+,v+i

of Bradley

(1984) Lemma

j< -n), thenAn{T-
The proof of both Theorems A EEn-,?-,

consequence

1 and 2 is now concluded

(X”)ca(Rk(X),:j<

as follows:

2.4.

0 Fix N, k, n > 1 and let

-n-N-l),

(X”)~cr(R~(X),:j>n+N+l). 33

L 01~uX 13, Number

STATISTICS

1

& PROBABILITY

LETTERS

2 January

1992

Define TN =inf{t>O:(X,+,_,,

XrfN,

T; =inf{t>O:(X_,_._,,

X,,,,,)

Xmr-N,

ismonotonic},

XPr-N+,

) is monotonic}.

Clearly TN+ and T; satisfy the conclusions of Lemma 3.1 and Lemma 3.2. Indeed, and T- respectively for shifted stationary sequences (Y,), t L = (X, * N)k E E. We have

they are equal to T+

{T;~n}nBn{T,i~n})

P(An

~P(AnB)~P(An{T~~n}f3BB{T~fn})+P(max{T~,TT,-}>n).

(15)

By Lemma A applied to the shifted sequences (Y,), E z = (X, * N)k E =, we have P_,+,(X) and B n {T,,, G rz}E.F&~(X).Therefore

n {T;GH} nsn

P(A

A

n {T; G n } E

{TN+
G'(A~{T,-Q~})P(B~{T,+~})+~,,~,(X)P(A~{T,-~~}) ~fyA)~(B)+(P,,-,~(A). Lemma

3.2. together

with (15), (16) gives therefore

nB)-P(A)P(B)<+,,_,P(A)+

P(A

(16)

2&n+‘.

(17)

Since for a complement B' of B we have P(A nB')- P(A)P(B')=P(A)P(B)- P(A nB), inequality (8) follows for n = 4k + 2 from (17) by picking N = n = k in (17). The general case then follows by the standard argument, yielding C = 2@j2 and p = $/4. 0 Similarly we prove Theorem 1. Let { Ai) be a R n _ v_ 1(X*)-measurable partition. Then F n+ N+, (X”)-measurable ;C/P(AinB,) i,j

partition

and let { B, > be an

-%4)+,)/

<:CIP(A,n

{T;
nB,n {T~~n})-P(A,n{T',
i.i ++xP(A,nB,n{T,f i,i

>n})+:~P(A,nB,n{T,->n}) i,i

+iCP(A,)f'(B,n(T,t >n})+;~P(A,n{T,-w})P(B,) i,i i,i
nk for n and putting P 4&+4(Xm)

GP,,,(X)

N = nk + 1 we get

+40 -~)-l(%-lVcf)~

q

which is a Proof of Corollary 1. All cases except (iv) follow from /34n~+4(X”) G (C + 1)q” V /3,-l(X), trivial consequence of (7) applied with k = n. To obtain (iv) we choose k > 1 such that cy,_,(X) = 0. Then for n 2 1, Dznk = 0 and by (7) we have P 4n/c+4 34

G

cc

+

lh”.

0

Volume

13. Number

STATISTICS

1

& PROBABILITY

2 January

LETTERS

1992

4. Proof of Theorem 3 The following

simple observation

med( -a,

-b,

-c)

is a consequence = -med(a,

Lemma 4.1. Zf R”( X) exists, then R”( -X) 0 is a symmetric random variable. The next observation

of the fact that

b, c). = -R”(X).

shows that the 3R smoother

Lemma 4.2. Zf X is a stationary

In particular,

preserves

4.1, inequality

P((X,“12t)63P((XO(&t)

(18)

(18) implies

XF > t implies

that at least one of the variables

also that P( Xr < t) < 3P( X0 < t) for every t E R;

foreveryt20.

(19)

Proof of Theorem 3. We shall apply Theorem 1 in Bradley and Peligrad (1986) By (8), condition (1.5) of Bradley and Peligrad (1986) is satisfied (X?)kEZ. geometrically. Indeed for every A E .FY,,(Xco), B E F2,+ s( Xm), n@

-P(A)P@)J

G /p(B)/\&4 < /P(B)

then XT

exists then

foreveryttE.

Proof. Indeed, it is enough to observe that the inequality X _,, X0, Xi cannot be smaller than t. 0

IP(A

in distribution

the moments.

sequence of random variables such that R”(X)

P(X,“>t)<3P(X,>t)

Remark. In view of Lemma in particular

if X G -X

n@

to stationary sequence with {a,} decreasing

-P(A)P@)\

+,,J’( A) + Cp2n G &&GQ~~

+ GP”.

By (19) we have

E{ 1x7 12+6} = (2 + fi)/atl+aP(

1X,m ) > t) dt < 3E( ) X,, )*+a} < CO.

0

Since by assumption 0,:/n' -’ + co as n ---*CO for every E > 0, (10) follows from the conclusion and Peligrad (1986). If in addition X0 is symmetric, then E{ XT} = 0 by Lemma 4.1. q

of Bradley

5. Proof of Theorem 4 Recall that the sequence (X,), E z is associated, see Esary, Proschan and Walkup (1967) if for each m 2 1 and every bounded continuous coordinatewise nondecreasing functions f, g : R*“‘+’ + Iw we have Cov(f(X-, ,..., X,), 8(X_, ,..., X,)) >, 0. Lemma 5.1. Zf X is i.i.d.,

then R”(X)

exists and is associated.

Proof. Indeed, X (being i.i.d.) is associated and since x, y, z + med(x, y, z) is the nondecreasing function of each coordinate, therefore R”(X) is associated for each n 2 1. The limit R”(X) exists because 35

Volume

13, Number

1

STATISTICS

of the i.i.d. assumption 0

& PROBABILITY

and hence is also associated,

2 January

1992

(1967) property

P5.

LETTERS

see Esary, Proschan

and Walkup

Remark. The proof of the lemma shows that if X is associated, then so is X”. In particular 2 0 for every k 2 1 and lim inf, _ ,n-’ Var(C, d n XT) > 0 in the nondegenerate case.

Cov( X7,

XF)

Lemma 5.2. Suppose f ( .) 2 0 is a nondecreasing and differentiable function, f(x) Z 0 for x f 0 and X is an i.i.d. sequence such that E { f ( l X0 1)) < co. Then the output X” = R”(X) satisfies E{ f2( I Xl

I>} < CQ.

(20)

Proof. Let (XL),,, = R(X). Since XA > t implies that at least two of the variables therefore by the i.i.d. assumption P(Xi 2 t) < 3P2(X0 > t). Lemma 4.1 gives

X_,,

X0, Xi exceed t,

P( I x; (2 t) < 3p2( I x, I 2 1) for each t 2 0. This, together

(21)

with (19) implies

P( 1x,m I > t) f 9P2( I x, ) > t). Since by Chebyshev’s known tail integration

inequality formula:

(22)

P( l X,, l > t) G E{ f( l X0 ])}/f(t),

(20) follows

from (22) and the well

sc+18E{f(lX,l)}~mf~(t)~(lXol~t)dt~~.

Remark. Inequality

(22) can be complemented

P( 1x; 1 > t) ,, Indeed

XT

:P’( 1x,

•I

with the lower bound

1 > t).

> t holds on { X0 > t } f~ { X, > t } and Lemma

(23) 4.1 ends the proof.

Remark. The constant in (22) can be improved to 9 by using Mallows (1979) formula for the distribution of X$‘. Lemma 5.2 holds also for +mixing input sequences X which satisfy q,(X) < cc (with the constant in (21) replaced by 3(1 + $i(X))). The condition #,(X) < cc is satisfied in particular for Markov chains on ([E, .G@)with one step transition probabilities P(x, A) such that there is C < cc satisfying P(x, A) G CP( y, A) Vx, y E E, VA E .?8 (such Markov chains were studied e.g. in large deviation theory). We shall also need the following Lemma 5.3. Suppose c%(Y)

Y = (Y,)

= P”

E{ q2 log’+‘(l

lemma.

is strictly stationary,

forsome

O
+ I r, I)} < co.

Assume also E{ Y,} = 0 and put S,, = Ck <“Yk, u,’ = Var(S,). Then n-’ Moreover, if CJf 0, then S[nt]/u, + w weakly in D[O,l] as n + 00. 36

(24) (25) a~*+u2forsome

a’>,0

asn+oo.

Volume

13, Number

1

STATISTICS

& PROBABILITY

LETTERS

2 January

Proof. First by Lemma 1 in Herrndorf (1985) specialized to the function G(x) = x*(log(x + l))‘+‘, S > 0, and with f,(x) = 2i+’ llog x 1-u+‘) as x + 0, x > 0, we can find a constant C such that IE{Y,~})I~Clloga,I-(‘+“)(E{Yo’ Therefore

log’+s~l+Y,l})2

the sum u2 = E{ Yi} + 2 CyZ1 E{ Y,?} Var

=nE{Y;}

+2

5

converges.

(n-j)E{Y,Y,}

1992

where

foreveryial.

Now by stationarity =no*(l

+0(l))

asn+cc,

j=l

which proves the first part of the lemma. The second part is now a particular after Theorem 2 in Herrndorf (1985). 0

case of the corollary,

stated

Proof of Theorem 4. In order to prove the first part of Theorem 4, we have only to verify the assumptions of Lemma 5.3. As follows from Theorem 1 (or Bradley, 1984, Corollary l(iv)), (XF), =z is stationary and strongly mixing with (Y, < p” for some 0 < p < 1, which proves (24). By Lemma 5.2 applied with f(x) = x log’ +&‘*(1 + x), condition (25) is satisfied. To verify that u f 0, notice that by Lemma 5.1 (XF) is an associated sequence and in particular lim inf, _ m Var(E.jl= i X:)/n > Var( XF). From the lower bound (23) we have Var( XT) # 0, provided X0 is not a constant. Therefore u Z 0 and the first part of Theorem 4 follows now from Lemma 5.3. The second part of Theorem 4 follows from Theorem 2.1 in Birkel (1988) applied to stationary sequence ( Xp)k E z. Indeed, (Xp) is associated by Lemma 5.1. Moreover, by Lemma 5.2 we have E{ 1XT 13} < cc; therefore assumption (2.3) in Birkel (1988) is satisfied. Since it is well known that Cov(X,“, X;) < [a,JX”)J”3(E{ I X,m 13}>*I3 by Theorem 1 (or Corollary l(iv) in Bradley, 1984), condition (2.1) in Birkel (1988) is satisfied. Therefore the conclusion of Theorem 2.1 in Birkel (1988) ends the proof of Theorem 4. q

References Birkel, T. (1988) On the convergence rate in the Central Limit Theorem for associated processes, Ann. Probab. 16, 16851698. Bradley, R.C. (1984), Some mixing properties of Tukey’s 3R smoother, Stochastics 11, 249-264. Bradley, R.C. (1986) Basic properties of strong mixing conditions, in: E. Eberlein and M.S. Taqqu, eds., Dependence in Probability and Statistics (Birkhauser, Boston, MA). Bradley, R.C. and M. Peligrad (1986) Invariance principle under a two part mixing assumption, Stochastic Process Appl. 22, 211-289. Esary, J.D., F. Proschan and D.W. Walkup (1967), Association of random variables with applications, Ann. Math. Statist. 38, 1466-1474.

Hermdorf, N. (1985), A functional central limit theorem for strongly mixing sequences of random variables, Z. Wahrsch. Verw. Gebiete 69, 541-550. Mallows, C.L. (1979), Some theoretical results on Tukey’s 3R smoother, in: T. Gasser and M. Rosenblatt, eds., Smoothing Techniques for Curve Estimation. Lecture Notes in Math. No. 757 (Springer, Berlin). Mallows, CL. (1980) Some theory of non-linear smoothers, Ann. Statist. 8, 695-715. Schonman, R. (1990), Critical points of two dimensional bootstrap percolation like cellular automata, J. Statist. Phys. 58, 1239-1244. Tukey, J.W. (1977) Exploratory Data Analysis (Addison-Wesley, Reading, MA).

31