A problem of minimax estimation with directional information

A problem of minimax estimation with directional information

STATISTICS & PROBABILITY LETTERS ELSEVIER Statistics & Probability Letters 26 (1996) 205-211 A problem of minimax estimation with directional inform...

282KB Sizes 0 Downloads 51 Views

STATISTICS & PROBABILITY LETTERS ELSEVIER

Statistics & Probability Letters 26 (1996) 205-211

A problem of minimax estimation with directional information Thomas S. Ferguson Mathematics Department, 405 Hilgard Ave., Los Angeles, CA 90024, USA

Received October 1994; revised December 1994 Dedicated with great respect to Paul G. Hoel

Abstract This problem is in the area of minimax selection of experiments. Nature chooses a number 0 in the closed interval [ - 1 , 1]. The statistician chooses a number y in the same interval (an experiment) and is informed whether 0 < y, 0 = y, or 0 > y. Based on this information, the statistician then estimates 0 with squared error loss. The minimax solution of this problem is found. The minimax value is 1/(2e). The least favorable distribution involves the truncated t-distribution with two degrees of freedom. The minimax choice of experiment involves the truncated t-distribution with zero degrees of freedom. Keywords: Selection of experiments; Optimal design; Search games; Minimum variance partitions

1. Introduction Paul Hoel has done important work on many problems. There is, for example, his early work on approximation, factor analysis, and indices o f dispersion. Or the more mathematical work on unbiased estimation, similar regions, optimum Bayesian classification, confidence bounds for polynomial regression, and the sequential F-test. But it is probably his mature work on optimal design in regression problems that has had the most impact in the statistical world. This work appeared in a series o f 6 or 7 papers in the Annals o f Mathematical Statistics starting in 1958. In this work he solves some important and difficult problems in minimax design. In this paper I am going to emphasize how difficult these problems are by looking at one o f the simplest nontrivial problems in minimax selection o f experiment. This little result is not in any way comparable to the really fine work Hoel has done on experimental design and is only offered as a tidbit that m a y be amusing. The problem treated is the estimation o f a parameter based on information about whether the parameter is greater or less than selected points (design points). When the minimax criterion is used, these problems are called search games with directional information. The discrete version, generalizing "twenty questions", Delivered August 1994 at the Joint Statistical Meetings, Toronto, in the Session in Honor of Paul G. Hoel's 90th Birthday. 0167-7152/96/$12.00 (~) 1996 Elsevier Science B.V. All rights reserved SSDI 01 67-71 52(95)0001 1-9

T.S. Feryuson/Statistics & Probability Letters 26 (1996) 205-211

206

has been studied by Gilbert (1962), Johnson (1964) and Gal (1974). A generalization to sequential selection of an infinite sequence of design points in the unit interval, with loss equal to the sum of the distances to the true point, has been studied by Baston and Bostock (1985) and Alpem (1985). Problems using a Bayes criterion with respect to a uniform prior have been considered in Cameron and Narayanamurthy (1964), Murakami (1971, 1976) and Hinderer (1990). If lying is allowed, this has been called Ulam's game and treated in Spencer (1984, 1992) and Czyzowicz et al. (1989). Berry and Mensch (1986) consider a related Bayes problem allowing false directional information to be given with known probabilities. A search game with a continuous parameter and with possibly false directional information has been treated in an interesting paper of Gal (1978). Here we treat a continuous parameter version of a search game with directional information, but restrict the searcher to one question, after which he must estimate the parameter with squared error loss.

2. Statement of the problem The rules of the game are as follows. There are two players. Player I, the hider, chooses a number 0 in the interval [ - 1 , 1]. Player II, the searcher, not knowing 0, chooses a number y in the interval [ - 1 , 1]. Then player II is informed whether 0 < y, 0 = y, or 0 > y and estimates 0 by a number z in the interval [ - 1 , 1]. The payoff is ( 0 - z ) 2 given by player II to player I. When player II is told that 0 = y, it is clearly optimal for him to choose z = y. With this understanding, a pure strategy for II is a triplet, (y, z0, Zl ), where z0 denotes his estimate of 0 when he is told 0 < y and zl is his estimate of 0 when he is told 0 > y. We may assume that z0 is in the interval [ - 1 , y ] and Zl is in the interval [y, 1]. Player II's loss is then

L(O, (y, z0, zl )) = (0 - z0)2I(0 < y) + (0 -

zl

)2I(0 > y),

(1)

where I(A) represents the indicator function of the set A. A mixed strategy for player I is a probability distribution, F(O), on the interval [ - 1 , 1]. A mixed strategy for player II is a probability distribution on the three quantities (y, z0, zl ), but since the loss to player II is a convex function of z0 and Zl, he may restrict attention to strategies that are nonrandomized in the choice of z0 and Zl. With this restriction, II's mixed strategies consist of a distribution G(y) on the interval [ - 1 , 1], and two functions, zo(y) and zl(y), with the understanding that if y is the value chosen from G and II is told that 0 < y (resp. 0 > y), he will estimate y to be zo(y) (resp. zl(y)).

3. Solution The minimax value of the game is v = (2e) - l = 0.184... The unique optimal strategy for I is to choose 0 in [ - 1 , 1] according to the distribution F(O) that has mass:

mass:

1 ----0.269... ( e + 1)

at 0 = - 1 ,

1 ---0.269...

at 0 = 1,

(e+ 1)

and density: f(O) = 2e(e_l + 02)3/2

for 10[ < 0.5(1 - e - l ) -~ 0.316...

(2)

207

T.S. Ferguson I Statistics & Probability Letters 26 (1996) 205-211

~

.269 -i

-.316

0

.824

824 731

.5~4 .25

.2691

25 |

.316

1

-I

-.316

0

.316

1

g(~)

f(O) 1.0

.368

J

-.316

0

.316

Zl(y) Fig. 1.

The unique optimal strategy for II is to choose y in [ - 1 , 1] according to the distribution G ( y ) that has mass : 0.25

at y = 0 . 5 ( 1 - e - l ) ,

mass : 0.25

at ? = - 0.5(1 - e - 1 )

and

1 density : O(Y) - 2(e_ 1+ y2)1/2

for

I f 0 < y, choose z -- zo(y) = y - X / ~

lyl

< 0.5(1 - e -1).

(3)

+ y2. I f 0 > y, choose z = zl(y) = y + x/e -l + y2.

IfO = y, choose z = y (see Fig. 1).

4. P r o o f It is convenient to introduce a generalization o f the distribution (2). The scaled version o f the t-distribution with 2 degrees o f freedom has density 1 ( f(0[cr)=~

02) -3/2 1+~-~

(The t2-distribution has a = x/-2.) For 0 < ~ < 1, let F(0[cr) denote the distribution o f a random variable O that has density: f(O]o)

for [0[ < 7

T.S. FergusonlStatisties & ProbabiliO, Letters 26 (1996) 205-211

208 and mass: p(0-) =

f(O[a) dO at 0 ----+ l ,

f7 ~

(4)

where 1

-

7=

0-2

-

2

The distribution (2) has a = e -1/2. W e use the following indefinite integral formulas:

0

f f(oI0-)dO =

(5)

/ Of(Ol0-)dO = - - ~°

(

1 + -~ j

,

and

j 'O2f(Ol0-)dO

=

log

+

1+

-

1+

The mass p ( a ) m a y be found from (5) as

p(a)=~

1--

1+

0-

0-~j

j

=

~

1

~i7~

.

Against the strategy where I chooses O according to the distribution (4), player II chooses y c [ - 7 , 7 ] and estimates 0 to be E(6)~9 < y) i f he learns O < y, and E(O[6) > y) if he learns O > y. The expected payoff is

V ( y ) = P ( O < y)Var(O[O < y) + P(O > y)Var(O[O > y) = E ( O 2) - P(O < y ) E ( O l O < y)2 _ P(O > y)E(O[O > y)2

(E(69I(O > y)))2 .

=E(O2)_

(6)

P(O > y)P(O < y) Here we have used the formula E ( O I ( O > y)) = - E ( 0 1 ( O < y ) ) , which follows from the symmetry o f the distribution o f O. For lYl < Y,

E ( O I ( O > y)) =

Of(Ola)dO + p(a) 1

=o2

1_

(

1+

,+

1+

T.S. Fergusonl Statistics & ProbabilityLetters 26 (1996) 205-211

209

The first term of (6) is independent of y:

~ l "-~0"2 log

q-\l+ ( 7'~I/'~--(a2+I)~ 2j (I+ .j 72~

(~

= O'2(1 -- 1og(tY)). Hence,

V(y) = 0 - 2 ( 1 - l o g ( a ) ) - 0-2 = _ t~2 log(O') is also independent of y. This is maximized (by player I) at a = e -1/2, giving a value V(e -1/2) = 1/(2e). Now, we show that the strategy (3) for player II guarantees a value of at most 1/2e. Again we generalize the strategy (3) for purposes of analysis. For e -1 < a < 1, let Y be chosen from the distribution G(y]~) that has

y2)-1/2

for lyl < y,

9(yla)dy

at y = 4- 7,

density:

g ( y l a ) = ½(a2 +

mass:

q(e) = ½ -

choose

z=z0(y)=y-v'~+y2

choose

z=zl(y)=y+v/a

choose

z=y

2+y2

if

0
if

0>y,

(7)

if0=y,

where, again, 7 = ( 1 - ~2)/2. For a = e -1/2, this reduces to (3). Here are some useful indefinite integral formulas 1 S 9(Yl')dy= ~log (y+ ~ ) ,

f zl(y)g(yl~r)dy =

' (y+

/ zo(y )o(yla ) d y = --~ f

y

l(y)29(ylcr ) dy = -~ zo(y)2 g(ylcr) dy = -

The expected payoff to I using 0 against (7) is

W(O) = E(O--zI(Y))21(y < O) + E(O-zo(y))2I(y > O) = 02-20[EzI(Y)I(Y < O) + Ezo(Y)I(Y > 0)] + Ez1(Y)2I(Y < O)+ Ezo(y)2I(y > 0).

210

T.S. Ferguson/Statistics & ProbabiBty Letters 26 (1996) 205-211

We consider two cases. Case 1: 10[ < 7- Then Ezt ( Y ) I ( Y < O) = [ o z~ (Y)g(Yl0.) dy + q(~)zj ( - 7 ) .I-; = ½[(0+ V / ~ +02) - ( -

7 + V/--~+ 72)] + q(0.)Zl(-y).

Similarly, E z o ( r ) I ( r > O) = ½[(0 - v ~ so that using z l ( - 7 ) =

+ 02) - (7 - V ~ + 72)] + q(0.)zo(7),

- z o(7), one obtains

E z , ( Y ) I ( Y < O) + Ezo(Y)I(Y > O) = O. Also, EzI(Y)2I(Y < O) =

zl(y)29(yl~)dy + q ( a ) z l ( - 7 ) 2 ,/

=

o

,

( 0 + V~a2+O 2) + ~ ( - 7 +

X / ~ + 7 2 ) + q ( a ) z . ( - 7 ) 2,

and similarly, Ezo(y)2I(y > O) = ~0 ( 0 + V ~ + 0 2 Now, using - 7 + ~

+ 72 = 0.2

) + 57 ( - 7 + V / ~ + 72)] + q(a)z0(-7) 2.

and z0(7) 2 = z l ( - 7 ) 2 = 0.4, we have

EzI(Y)zI(Y < O) + Ezo(Y)2I(Y > O) = 02 + Try2 + 20.4q(0.). Combining these, we have W(O) = 02 - 202 + 02 + ~0.2 + 2cr4q(a)

__- ]20.2 +

20.4q(cr)

independent of 0. Case 2 : 7 < [0[ < 1. Assume, without loss of generality, that 7 < 0 < 1. Then Y < 0 with probability one, and W(O) = 02 - 20Ezl(Y) + Ezl(Y) 2. But

Ezl(Y) = q(0.)Zl(--7) +

zl(y)g(yl0.)dy + q(0.)zl (7)

= q(0.)(1 + 0.2) + (1 - 0"2)/2

T.S. FergusonlStatistics & Probability Letters 26 (1996) 205-211

211

and f-. E z , ( Y ) 2 = q ( o ' ) t z l ( - y ) 2 + z,(y)21 + J _ , , z l ( y ) 2 o ( y l a ) d Y /

= q(a)(o'4+ 1) + (1 -- a4)/4. At cr = e -Vz, we have E z 1 ( Y ) = (3 - e - l ) / 4

and E z j ( Y ) 2 = ½, so

1 W(O) = 0 z - 20(3 - e - 1 ) / 4 + 3"

This is a quadratic in 0 with m i n i m u m at the point (3 - e -1 )/4, which is the midpoint of the interval (% 1). Hence, on this interval, W(O) is less than the value at its endpoints, that is,

W(O)<.W(1)=l-Z(3-e-l)/4+½=e-~/2

for7 < 0~<1.

The third case is 0 = 7, but this allows II to estimate 0 exactly when y = 7, which happens with probabilty 1 so W(y) < l i m 0 ~ , W(O) = e-1/2. Combining this with the above, we have g,

W(O) = e - l / 2
for 101 < 7

and

0 = 5=1,

for7~<[01 < 1.

5. R e m a r k

Jim MacQueen points out that this result has an interesting implication in the area of m i n i m u m variance partitions. We are given the distribution of a random variable X, and we desire to choose a partition of the sample space into k subsets for fixed k, say disjoint sets A1 . . . . . Ak whose union is the whole space, such that the conditional variances averaged over the partitions is as small as possible; that is we desire to have P(X E A i ) V a r ( X I X ~ Ai) as small as possible. Such a m i n i m u m variance partition may be considered as an efficient mapping of the observation into a k-valued random variable. When k = 2 and when the distribution of X is F ( x ) , where F(O) is the optimal strategy o f the hider given above, then it does not matter how you partition the interval [ - 1 , + 1] into two intervals. All such partitions are m i n i m u m variance partitions.

References

Alpern, S. (1985), Search for a point in interval, with high-low feedback, Math. Proc. Camb. Phil Soc. 98, 569-578. Baston, V.J. and F.A. Bostock (1985), A high-low search game on the unit interval, Math. Proc. Camb. Phil Soc. 97, 345-348. Berry, D.A. and R.F. Mensch (1986), Discrete search with directional information,Oper. Res. 34, 470-477. Cameron, S.H. and S.G. Narayanamurthy(1964), A search problem, Oper. Res. 12, 623-629. Czyzowicz, J., D. Mundici and A. Pelc (1989), Ulam's game with lies, J. Combin. Theory A 52, 62-76. Gal, S. (1974), A discrete search game, SIAM J. AppL Math. 27, 641-648. Gal, S. (1978), A stochastic search game, SlAM J. Appl. Math. 34, 205-210. Gilbert, E.N. (1962), Games of identificationor convergence, SIAM Rev. 4, 16-24. Hinderer, K. (1990), On dichotomous search with direction-dependentcosts for a uniformly hidden object, Optimization 21, 215-229. Johnson, S.M. (1964), A search game, Advances in Game Theory, AMS Study #52, pp. 39-48. Murakami, S. (1971), A dichotomous search, J. Oper. Res. Soc. Japan 14, 127-142. Murakami, S. (1976), A dichotomous search with travel cost, J. Oper. Res. Soc. Japan 19, 245-254. Spencer, J. (1984), Guess a number - with lying, Math. Mag. 57, 105-108. Spencer, J. (1992), Ulam's searching game with a fixed number of lies, Theoret. Comput. Sci. 95, 307-321.