A Streamlined Approach for Calculating Expected Utility and Expected Value of Perfect Information Bruce R. H A R T S O U G H Agricultural Engineering Department, Unwersity of California, Davis, CA 95616, USA J o h n L. T U R N E R Department of Mechanical Engineering, University of South Carolina. Columbia. SC 29208, USA An approach based on the criteria of maximizing expected utility was developed for comparing alternatives, and the expected value of perfect information was incorporated to indicate the relative importance of uncertainty. Multi-dimensional Gauss quadrature was used to integrate over the uncertain variables. Problem characteristics were taken into account to subdivide utility functions, reducing the amount of computation when integrating utility. Additional factors were included to streamline the calculation of expected value of perfect information. Problems were taken from the literature to test the convergence rate of the method. The quadrature results converged quickly, although the rate depended on the type of probability distributions assigned to the variables. Keywords: Expected utility algorithm; Perfect information value; Gauss quadrature; Convergence.
Bruce Hartsough graduated in Forest Engineering from the University of California at Davis 1976. He joined Weyerhaeuser Company, evaluating and developing new systems for harvesting small trees, in Oklahoma, Washington and Oregon. In 1980, he returned to U.C. Davis as a development engineer, studying logging with helicopters and modeling cable yarders and systems. He received his Ph.D. in Agricultural Engineering at Auburn University in 1986. His current research involves systems analysis of eucalyptus plantations, methods for harvesting timber on steep terrain and collection of forest residues for energy.
1. Introduction This p a p e r deals with some c o m p u t a t i o n a l aspects of the decision analysis p r o b l e m of selecting from alternative choices when u n c e r t a i n t i e s in the available i n f o r m a t i o n are to be considered. This s i t u a t i o n arises frequently in the early stages of system design where the objectives of the system p e r f o r m a n c e are specified b u t the influence a n d values of m a n y variables are n o t k n o w n with certainty. F o r example, in d e s i g n i n g a n engineering system, say a piece of e q u i p m e n t , several alternative concepts might be initially considered. Each concept can be represented in terms of relev a n t system attributes (e.g., cost, productivity, reliability, expected life, user acceptability, etc.). The attributes might be r a n k e d in relative i m p o r t a n c e to develop a characterizing " o b j e c t i v e " or " u t i l i t y " f u n c t i o n to be m a x i m i z e d for e v a l u a t i n g a n d comp a r i n g the concepts. In the initial stages, the available i n f o r m a t i o n for r a n k i n g a t t r i b u t e s is usually limited. F u r t h e r , the a t t r i b u t e s themselves are freq u e n t l y expressible as f u n c t i o n s of more basic variables of the p r o b l e m , whose values m a y be k n o w n only a p p r o x i m a t e l y or within ranges. In the previous e x a m p l e for i n s t a n c e the basic variables might i n c l u d e geometric d i m e n s i o n s , material properties (strength, fatigue properties, mass density), materials cost per u n i t volume, etc. The p r o b l e m of m a k i n g a r a t i o n a l decision as to which alternative is best u n d e r such c o n d i t i o n s has received relatively little a t t e n t i o n . Also, the q u e s t i o n
North-Holland Decision Support Systems 6 (1990) 1-11 016%9236/90/$3.50 © 1990, Elsevier Science Publishers B.V. (North-Holland)
John Turner received his Ph.D. in Theoretical and Applied Mechanics. from the University of Illinois in 1975. He has worked in numerous areas of engineering mechanics including composite materials, biomechanics, optical methods of stress analysis, finite element analysis and vehicle mobility modeling. His current research involves thermal strain analysis by digital imaging techniques.
2
B.R, Hartsough, ,I.L. Turner / Expected utility and value of perfect information
of assessing the value of missing or uncertain information in reaching the right decision (selecting the best alternative) is an important one that needs study. For example, determining the merit of expending additional effort and dollars to collect more information about certain problem variables to assure that an optimum solution is not overlooked would be valuable information to the design process. Decision support systems which incorporate information uncertainty at the most basic levels of analysis and deal with the questions previously posed, are needed to make contemporary design processes fully effective. Clearly, such procedures would have applications in numerous other areas as well as design. An attractive approach to selecting from alternative choices under uncertainty is the expected utility (EU) method [10]. This approach is viable when a utility function can be described in terms of attributes whose uncertain values can be characterized with probability distributions. In m a n y cases, the decisionmaker can express the attributes as explicit functions of basic problem variables whose values are uncertain. Traditional decision support systems do not explicitly evaluate EU in such circumstances. Instead, other approaches such as dominance [19] and mean-variance methods [18] have been used. This has been largely due to difficulties in the required integration of the utility function over the ranges of the variables. Explicit evaluation of EU has several advantages, among them the ability to calculate the expected value of perfect information (EVPI) for each of the uncertain variables [25]. EVPI is an upper bound on the amount that may be spent in order to reduce uncertainty thus it can be used to decide whether to choose an alternative now or to collect more information. This research effort addressed the problem of explicitly evaluating EU and EVPI.
2. Literature Review In the decision analysis approach to selection, the utility function is determined, then the expected utility is explicitly calculated for each alternative [9]. The calculation of EU and EVPI involve multi-dimensional integration over the ranges of the variables. In most cases, the utility functions cannot be integrated analytically.
Several alternative approaches have been considered. Expanding a part of the integrand in a Taylor series has been widely used. The value of the function at the expected value of the variable is often taken as the first term in the Taylor expansion. The utility function has also been expanded about the means of state variables [1,5,13]. Pulley [18] expanded a logarithmic utility function to develop several mean-variance approximations. He noted criticism of the approach because the Taylor series does not converge for many situations to which it has been applied. Pearson and Tukey [16] developed an empirical formula for the mean of a distribution, based on fractiles. Perry and Greig [17] suggested transforming to the utility function and evaluating the utility of the fractiles of an attribute. Then EU can be approximated with the Pearson and Tukey formula. More work with this formula has shown it to give good results in some situations [7,8], however, the formula requires knowing the fractiles of an attribute, which are not available if the attribute is modeled as a function of other uncertain quantities. Numerical techniques involve evaluating (i.e., sampling) the integrand at some discrete number of m points over the integration range. Each evaluation is multiplied by an appropriate weighting factor and the weighted evaluations are summed to arrive at an approximation to the exact integral. Within a class of techniques, the error decreases as m increases [12]. The trapezoidal rule is a simple numerical integration technique and has been commonly used in decision analysis for integration over a single variable [15,21]. Unfortunately, the technique is not very efficient. Gauss quadrature rules optimize the locations of the sampling points and thus are more efficient than any others which use sums of weighted function evaluations [2]. Product rules for numerically integrating over multiple-dimensioned regions are developed from single-dimensioned rules. A one-dimension rule requiring m points forms the basis for an m'-point rule for the n-dimensional region. A limitation is the exponential growth of the number of evaluations that are required for multi-dimensional integration. Monte Carlo sampling techniques have been advocated in the decision analysis field [10] due in part to their simple structure [22]. However, they
B.R. Hartsough, J.L Turner / Expected utility and value of perfect information
converge very slowly compared to systematic methods. The average quasi-Monte Carlo method has been shown to given poorer results than the 2~-point product Gauss quadrature rule for any number of variables n less than 108 [23]. The systematic techniques are generally preferable to Monte Carlo methods for many practical decision analysis problems. Although several decision support systems have been developed for comparing discrete alternatives via the explicit EU approach [e.g., 11,20,21,26], they ignore several factors which are important in many situations. All assume that attributes are probabilistically independent, and none of them treat attributes as functions of more basic problem variables. None of the multiattribute packages consider EVPI, nor do they consider uncertainty about the utility function (except in after-the-fact sensitivity analysis).
3. Approach 3.1. Attributes and Utifity Functions Any selection problem can be represented with an objective function to be maximized. The attributes included in the objective function may be, for each alternative, functions of many uncertain quantities or variables. This work was limited to utility functions of the linear additive form: U= E CiA,(x),
where
3
has been drawback because a decisionmaker may be as uncertain about the trade-offs between attributes as about the other variables; in the problem. In the current analysis the decisionmaker is permitted to assign a range to the possible values of the scaling coefficients by stipulating a range on the r,. The importance of reducing these uncertainties can be studied through the approach described herein.
3.2. Calculating Probabifity Distributions To characterize the uncertainty of the basic problem variables each variable must be assigned a probability distribution. Jaynes [6] proposed maximizing entropy, subject to the constraints imposed by known information, as a criteria for choosing a distribution. His approach was shown to be correct when the distribution of prior information is uniform [4]. Three forms of maximum entropy distributions were employed to test convergence of the proposed numerical integration scheme: If only finite upper and lower bounds are known, the distribution is uniform; for any finite range and a specified mean of a problem variable the form is a truncated exponential; when a finite range, mean and variance are specified the distribution takes on a truncated normal form. A previously-developed method [24] to calculate the distribution coefficients was modified for this work.
3.3. Calculating Expected Utifity
(1)
i
U = utility A i = the value of the ith. attribute, which is a function of the basic problem variables _x = the array of basic problem variables C, = the value of the scaling coefficient for the ith attribute It was assumed that a decisionmaker could supply, for each alternative, the functions describing the attributes A, in terms of the basic variables. Each of these variables is assumed independent of all others. In this work the coefficients C~ are defined in terms of their ratio r, to the first coefficient C 1. The coefficients are scaled such that they sum to unity. In previous work, the scaling coefficients have been assumed to have constant values. This
In theory, the utility function is integrated over the multidimensional region of uncertain variables, but this is frequently not feasible: the integrals usually can't be evaluated analytically and a large number of variables can make precise numerical integration computationally impractical. Two steps were taken to make the problem tractable: (1) an approach was developed to divide the integrand into small blocks, each of which could be integrated separately and (2) a computationally efficient multi-dimensional Gauss quadrature procedure was employed for the integrations. For the ith alternative, expected utility is
EU,-EU
d x dr,
Where
(2)
= the expected utility of the ith alternative
4
B.IL Hartsough, £ I - Turner / Expected utility and value of perfect information
Cj(r)
= the value of the scaling coefficient for the j t h attribute, which is a function of the vector r of scaling coefficient ratios A,j(x) = the value of the j t h attribute for the ith alternative x = the array of problem variables which influence the attributes. The summation and integration operators can be exchanged and, since the scaling coefficient ratios r are assumed independent of the other variables x, the expected values of the scaling coefficients can be calculated separately. The expected utility can then be expressed as
where E[@] is the expected value of the j th scaling coefficient. The decisionmaker specifies the ranges of ratios for the coefficients. In addition, the coefficients must sum to one. Since the j t h coefficient can be written as its ratio times the first coefficient C~, the unity constraint in terms of the ratios becomes C, + ~ ~)C, -- 1,
where
(4)
j=2
@ -- the scaling coefficient for the j t h attribute §. = the ratio of Cj to C, a = the number of attributes considered in the problem For specific values of the scaling coefficient ratios, the scaling coefficients are
j--2
~
j=2
]
To find the expected values of the coefficients, equations (5) and (6) must be integrated over the ranges of the ratios. It is efficient to integrate all of the coefficients simultaneously. A product Gauss quadrature approach was also used to evaluate the expected values of the attributes. An existing algorithm [2] was used to generate the quadrature points and weights. The expected value of attribute Ajj is:
E[A,j]=fa,j(x)p(x)dx,
where
(7)
E[Aij] -- the expected value of the j t h attribute p(_x)
for the ith alternative = the probability density for the variables x
One approach to evaluating E[A,j] is direct numerical integration of the function f ( x ) where
(8)
f(s) = A,j(x)p(s).
This approach converged rapidly for cases where the probability distributions were uniform or rather broad. For peaked distributions where variables had high probabilities of occurring within very small segments of the ranges, the approximations fluctuated and converged very slowly. In theory, the slow convergence is due to the large values of higher derivatives of p(_x) for very peaked distributions. Practically, the problem arises because few of the quadrature sampling points are located within the high probability region of the variable space. Another choice of integrand was made to improve the rate of convergence. The cumulative probability distribution G (x) is defined for a single variable as
G(x) =
Ep ( x ) dx,
(9)
where a is the lower bound for variable x. The differential dG is p ( x ) d x , and substituting into (8), gives, for the single-dimensional case
E[ A,j] =
f ,o, A,j(x(G)) dG
= fo
G(b)
da.
(10)
The variable of integration is now the cumulative probability G, and the integrand includes only the attribute function. Gauss quadrature can be applied to this integrand as long as x(G) can be evaluated. This function can be found analytically for the uniform and truncated exponential distributions, but a numerical approach was required for the truncated normal distribution. Integrating over the cumulative probability rather than over the variable x significandy improved the convergence for peaked distributions. Fig. 1 compares results for a single problem solved by each of the two approaches. When the utility function is described as a linear combination of attributes, the EU of the ith alternative can be written as
EU~=
EE[Cj]E[Aij(x_)]. J
(11)
B.R. Hartsough, J.L. Turner / Expected utility and value of perfect information
attribute function. The algorithm, flowcharted in fig. 2, was carried out manually for the test problems, however, it could be done symbolically within a decision support system. Each attribute is now described as a sum of products of independent blocks, so the expected value of an attribute is
o
Lu
5
0
"~ -2rr
v
P
nsllies
E[Aij] = E
q
I-I E[bij,,,]
(14)
I=1 m = l -6
and the expected value of each block is evaluated by numerically integrating
i
0
2
4 Quadrature
6
8
Points,
10 rn
Fig. 1. Comparison of Convergence for Integration over the Cumulative Densities Versus Integration over the Variable Ranges.
=
dx,
where
(15) E[bijt,,, ] = the expected value of block b,jt, ~
Each of the attributes for each alternative may be a function of many variables. The integration to find the expected value of an attribute is not tractable unless the function includes approximately ten or fewer variables. However, in many cases, the integration can be subdivided into two or more smaller problems by taking advantage of two properties of the expectation operator
E [ a + b] = E[a] + E[b]
and
E [ c * d] = E[c] * E [ d ]
if c and d are independent,
(12)
(13)
where a, b, c, and d are general quantities. Since the expected value of each attribute is required in order to calculate expected utility, equations (12) and (13) are applied to the functions which describe each attribute for each alternative. An algorithm was developed which insures that all possible subdivisions are made. First, operations are carried out algebraically as far as possible, eliminating as many parentheses as possible from the attribute function. For example, the function fragment ((a + b) * (c + d)) is written as (ac + ad + bc + bd). The function is separated into its additive components and each of these is separated into multiplicative fragments. Since each element of the variable vector x is assumed independent of every other, a fragment is independent of the others if the variables included in the fragment do not occur in any of the others. The fragments are grouped into blocks which have common variables occurring in two or more fragments. These blocks represent the smallest possible subdivisions of the
xB
= the subset of _x included in the block Subdivision can make many large problems tractable. An ideal division would break a problem with n variables into q blocks, each with n / q variables. Using m sampling points per variable in a product Gauss quadrature scheme, only q * m ( n / q ) integrand evaluations are required for the subdivided problem, versus m" for the complete case. For example, if a problem with 20 variables and 10 integration points per variable is divided into 4 blocks with 5 variables each, the number of evaluations is reduced from 10 20 to 4 x lO s. For a single alternative, the blocks which make up its attributes are identified. For each block, the associated variables are next identified and arrays of quadrature points and weights are generated. Then the product Gauss routine is used to calculate the expected value of the block. After all blocks have been considered, the expected values of the attributes are calculated and combined with the scaling coefficient values, returning EU for the alternative. The framework is flowcharted in fig. 3.
3.4. Calculating E VPI In theory, the best alternative is identified at each possible value of the variable. This alternative is termed the conditional best. The difference between the utility of the conditional best and that of the unconditional best is the loss if that specific value of the uncertain variable were actu-
B.R. Hartsough, J.L Turner / Expected utility and value of perfect information
6
A : f(x) ]
Describe attribute in terms of uncertain variables:
+ Carry out
all multiplication/diviaion
operations
algebraically I
,[
+ Divide attribute function Into additive components:
A=
?
:~ C I I=1
:For I = 1 t o pJ
T
q
Divide Ith component into multlpllcetlve blocks:
CI = m'T~r'tbl = m
? IFor m = I to ql
l'or.- m'c.l t
Does l u b l a l of x In b Im Intersect subset In b l r ? . ~
/
Yea '~-
[ Absorb b Im into b I r Drop b I m
,ll
91
I
I=
ms1
Final attribute function Is: A =
b Im
Fig. 2. Algorithm to Subdivide Attribute Functions into Smaller Blocks.
ally to occur. This loss function is integrated over the probability distribution of the variable to give the EVPI. The EVPI concerning variable k can be written as
EVPIk = falkp( Xk )
X Max{~ i
$x k
(U~(_x)
- U . ( x ) ) v ( x ) dx) dXk,
(16)
where a~, b k -- the lower and upper bounds for variable Xk
U/(_x) = the utility function for the ith alternative The star subscript denotes the unconditional best alternative, i.e., the one with the largest EU. The
integrand for the outer integration loop is the loss function. Two problems hinder the calculations. Since the best alternative must be found at each point over the range of the variable, the number of function evaluations to calculate EVPI for a single variable is roughly the same as are required to find the EU's of all the alternatives, if no special characteristics of the problem are taken into account. Most problems deal with many more variables than alternatives, so the time requirements for EVPI calculations typically exceed those for EU by at least an order of magnitude. Secondly, the loss function is piecewise continuous, with a separate segment for each alternative that is best over some part of the range of the variable (fig. 4 displays a hypothetical example). The first and higher derivatives are not continuous at the segment junctions, so a high-order integration
B.R. Hartsough, J.IL Turner / Expected utility and value of perfect information I Identify blocks associated with alternative ]
.
tl
For each block
I ,dent,ty
I
associated variables
I
t
Generate vector of quadrature points for each variable
÷
I n,..boo,..o,.o..u.=ol t For each comblnsllon of quadrature
points
~ 1 ~
f
Call user-supplied BLOCK routine to calculate value of block[
f
I Multiply by corresponding quadrature weight and add to block expected value
Csll user-supplied ALTFUNC routine to calculate expected values of attributes
+ Call UTIL routine to calculate expected utility
Fig. 3. Algorithm to Calculate EU of an Alternative.
scheme is required in the outer loop to achieve precise results. To reduce the time requirements and eliminate the discontinuity problem, an initial screening is performed to identify the endpoints of and condi-
tional best alternatives on each segment. Then the actual integration is carried out for each segment, considering only the conditional best alternative and the unconditional best. Rewriting in terms of the segments bkh
:'10 ----
~
.IU1
U3 U2
(17)
0
0
where the Bh subscript indicates the alternative which is best on segment h, akh and bkh are the lower and upper endpoints of the segment. As with the E U calculations, the results for Gauss quadrature are more precise when the cumulative probabifity is substituted for the actual variable in the integrand Variable Value, x
Fig. 4. Utifity Functions for Three Alternatives, and the Piece-
wise Loss Function Assuming Alternative 2 is the Unconditional Best Alternative.
EVPI k
- E/. h
Gtb~h) G(akh) G_¢G,
- U,(x(c))) d~ d6~.
(18)
8
B.R. Ha.sough, J.L Turner / Expected utility and value of perfect information pertinent alternatives on the segment. This results in
On each segment, the integral can be multiplied and divided by the integral of the cumulative probability differential over the segment
EVPIk
=
EVPI k = y ' ( ( G ( b k h ) - G(akh)) h
[G(bkh) dG fG(b~h) Y'~"G(akh) dG h "G(akh)
where (20)
×
G(bkh), G(akh ) - U , ( x ( G ) ) ) dG dG k.
(19)
Eh[Ush ]
But the first integral is just the probability of the variable occurring within the segment, and the remaining integral can be carried out in two parts, giving the difference in EU between the two
i lnltlelize
EVPI : oJ
?
; For ..ch block I
--
[--~s
- Yes block dependent Tupon EVPI varlsbls?')~'~"-I
[Stere expected value of in working vector block]] ~ ~, < [ 4
I ,..l..x.
No r-~ls
[ Put block Into vector I i of dependent variables
bess Eu : - = 1
I For each alternative [ , J . Yes alternative dependent upon EVPI vsrlable?)---I~ 1
] l If alternative EU ~. base EU ~
[i..t be.
Eh[U.]
Put alternative into vector
..,.,.,..,,v.
of d.pend.n, .',.rn.t'v., I
and Search to find segments of loss function conditional best alternative on each segment
?
I
Identify dependent blocks which affect unconditional best alternative i
?
;For each segment of loss function]
T IFInd probability Ps of being within segment[
?
Identify dependent blocks which affect conditional best alternative]
?
[ For each pertinent block ]
?
i
Calculate expected value of block on segment ] snd store in working vector I J Calculate EU's, on segment, of conditional and unconditional best a ternatives, subtraCtmultipiytObyfindpsexpectedand add tol°SSEvPI°n the segment,
Fig. 5. Algorithm to Calculate
EVPI
for a Variable.
I
the cumulative probability values for variable x k at the upper and lower bounds of loss segment h = t h e EU, on segment h, of the conditional best alternative = t h e EU, on segment h, of the unconditional best alternative
=
B.R. Hartsough, J.L. Turner / Expected utility and value of perfect information
This shows that the integration on a segment develops the expected loss on the segment, weighted by the probability of the variable occurring within the segment. A bisection routine is used to identify the segments of the loss function for the EVP1 variable, continuing until the remaining distance between the search bounds for a segment transition is less than a prespecified tolerance, stated as a fraction of the range of the EVP1 variable. Initially, the search bounds are set at the lower and upper bounds of the range of the variable. The best alternatives at these endpoints are identified. If both are the same as the unconditional best alternative, it is assumed that no other alternative has greater conditional utility at any point over the range of the variable, i.e. the E V P I is zero. This assumption has been found to be valid in the test problems analyzed thus far. Theoretically, there could be exceptions requiring the loss function to be evaluated at more points. Characteristics of individual problems are considered in order to reduce the number of calculations. The alternatives are checked to identify those which do not depend on the E V P I variable. The alternative from this group with the greatest E U is considered the " b a s e " alternative: its E U is a lower bound on conditional best EU. The rest of the unaffected alternatives can then be ignored. Some of the blocks making up the attribute functions of the dependent alternatives may not include the E V P I variable. Since their conditional expected values are constant, the blocks do not have to be re-integrated at each value of the EVP1 variable. These blocks are identified and their expected values placed in a working array of conditional block values. At any specified value of the E V P I variable, the calculation of conditional E U ' s for the dependent alternatives reduces to integrating the dependent blocks over the ranges of the rest of the variables. The conditional block values are stored in the working array, then the conditional values of the attributes are calculated using the problemspecific attribute function. Finally, the utility routine is called to sum the attributes, premultiplied by the expected values of the scaling coefficients. The unconditional best alternative is the conditional best on one or more of the segments of the range of the E V P I variable. N o integration is
9
required on these segments, as the loss and contribution to E V P I is zero. The overall procedure for calculating E V P I of a variable is flowcharted in fig. 5. During the final integration over a segment, the product Gauss routine is used to determine the expected values of the pertinent blocks on the segment. The E U ' s on the segment are evaluated, then combined using equation (20) to arrive at E V P I . A computer program incorporating the algorithms described above was written in Microsoft Fortran Version 3.0 [14], for an IBM Personal Computer with an 8087 coprocessor.
4. Convergence Testing The algorithms were evaluated on a n u m b e r of problems to determine the convergence characteristics of the E U and E V P I solutions as the number of quadrature points was increased. Each problem was run with one through nine integration points over the range of each variable. Factors which were considered as having possible impacts on the convergence of the quadrature algorithms included (a) the size of the problem, indicated by the number of variables, (b) the problem structure, indicated by the block size (variables per block), (c) the level of uncertainty, as measured by the ranges of the variables, and (d) the type of distribution (uniform, exponential or normal) assumed for the variables. Test problems were taken from the published literature on the comparison of forest harvesting systems or components. To facilitate algorithm testing, supplemental data were generated for the problems, as necessary, in order to vary the levels of the factors listed above. The test problems are described in detail in reference 3.
4.1. Expected Utility Convergence Thirty-two alternatives were chosen for the analysis, half from small problems (10 variables) with small blocks (one variable per block). The rest were from large problems (70 variables), with five variables per block. All combinations of range and distribution type were represented. Regression analysis was used to quantify the
B.R. Hartsough, J.L Turner / Expected utility and value of perfect information
10
rates of convergence. The convergence response variable was LOGeR
.. = log o t( E u o - -
EUQ,)/EUQ,1,
0-
°2-
E
•
~
•
,
Other
Distributions
(21) where L O G E R ev,, = the logarithm of the magnitude of the relative error in expected utility, evaluated with m quadrature points EUQm = the expected utility result for m quadrature points EUQ9 = the results for 9 quadrature points The number of variables and size of the blocks had no significant effect on the rate of convergence, nor did the relative range of the variables. The only important factor was the form of probability distribution: no significant difference was found between the convergence rates for the exponential and truncated normal situations, but the cases with uniform distributions converged at a higher rate: F L O G E R ev,, =
-
2.048
- 0.362m - 0.676m U N I F
N--222,
106.0 124.0 (22)
R 2=0.49,
where 1 UNIF =
0
if variable distributions were uniform, otherwise.
This relationship, along with the data points, is plotted in fig. 6. The poorer performance with the non-uniform distributions may be attributed to the less-smooth utility functions when described in terms of the cumulative variables _G. Since the precision of the Gauss quadrature approach depends on the higher derivatives of the utility function, situations with exponential and truncated normal distributions have inherent disadvantages over the uniform cases. Even though the quadrature method did not converge as quickly for the non-uniform situations, it still gave good results in most cases (compared with the results at nine quadrature points).
Distri-,,o: \ 0
2
7-,. 6
4
8
•
10
Quadrature Points, m
Fig. 6. Regression Relationships for Convergence of EU, Plotted with the Means of the Data for the U n i f o r m and Other Distributions.
When eight points were used, the maximum relative error for any of the test problems was only 0.8 percent. Much of the variation in convergence was not explained by any of the factors considered in the analysis. This variation is due to differences in mathematical form of the attribute functions, which are not readily quantifiable. 4.2. E V P I Convergence
Data were collected on eight variables, four with uniform distributions and four with truncated normal distributions. When searching for segments of the loss function, a tolerance of 10-4 gave excellent results. Tighter tolerances changed the E V P I figures by less than 0.01 percent. Convergence over the number of quadrature points was evaluated with search tolerance set at 10 -4 . The dependent variable was L O G E R e v m m = logt01( EVPIQm - EVPIQ9 ) / E V P I Q 9 [,
where
(23)
L O G E R E v e j m = the logarithm of the magnitude of the relative error in E V P I , evaluated with m quadrature points EVPIQ,. = the E V P I result for m quadrature points EVPIQ9 = the result for 9 quadrature points Trends were similar to those for EU. Cases with uniform distributions converged significantly
B.R. Hartsough, J.1- Turner/ Expected utility and value of perfect information faster than those with truncated n o r m a l distributions: F L O G E R EVm,. =
- 0.760
- 0.254m - 1.868m U N I F
n=45,
24.6 391.0 (24)
r2=0.90.
I n the w o r s t cases, E V P 1 results d i d n o t c o n verge as q u i c k l y as those for EU. T h i s m a y b e d u e to the piecewise n a t u r e o f the loss f u n c t i o n s a n d the errors a s s o c i a t e d w i t h f i n d i n g the e n d s o f the loss f u n c t i o n s e g m e n t s . H o w e v e r , the m a x i m u m e r r o r i n E V P 1 w i t h eight q u a d r a t u r e p o i n t s was o n l y three p e r c e n t w h e n c o m p a r e d w i t h the r e s u l t using nine points.
5. C o n c l u s i o n s
A p r a c t i c a l m e t h o d for c a l c u l a t i n g e x p e c t e d u t i l i t y ( E U ) of a l t e r n a t i v e choices a n d e x p e c t e d v a l u e of perfect i n f o r m a t i o n ( E V P I ) was develo p e d a n d tested. T h e a l g o r i t h m is c o m p u t a t i o n a l l y efficient and can make an otherwise intractable c a l c u l a t i o n feasible. E x a m p l e p r o b l e m s with u p to 70 v a r i a b l e s a n d u p to five u n c e r t a i n v a r i a b l e s p e r b l o c k were effectively a n a l y z e d o n a m a t h c o p r o c e s s o r - e q u i p p e d I B M PC. T h e m e t h o d c o n v e r g e d q u i c k l y for all of the test p r o b l e m s . T h e d e v e l o p e d m e t h o d o l o g y p e r m i t s c o n s i d e r a t i o n of u n c e r t a i n i n f o r m a t i o n i n d e c i s i o n a n a l y s i s in a n o v e l fashion.
References
[1] S.E. Bodily, A Multiattribute Decision Analysis for the Level of Frozen Blood Utilization, IEEE Transactions SMC-7, Nr. 10 (1977) 683-694. [2] P.J. Davis and P. Rabinowitz, Methods of Numerical Integration (Academic Press, New York, 1985). [3] B.R. Hartsough, Evaluation of Alternatives Considering Uncertainty with Examples in Forest Harvesting, Auburn University Doctoral Dissertation (Univ. Microfilms, Ann Arbor, 1986). [4] A. Hobson and B.K. Cheung, A Comparison of the Shannon and Kullback Information Measures, Journal of Statistical Physics 7, Nr. 4 (1973) 301-310. [5] R.A. Howard, Proximal Decision Analysis, Management Science 17, Nr. 9 (1971) 507-541. [6] E.T. Jaynes, Information Theory and Statistical Mechanics I, Physics Revue 106 (1957) 620-630.
11
[7] D.U Keefer, Allocation Planning for R and D with Uncertainty and Multiple Objectives, IEEE Transactions M25, Nr. 1 (1978) 8-14. [8] D.U Keefer and S.M. Pollock, Approximations and Sensitivity in Multi-objective Resource Allocation, Operations Research 28, Nr. 1 (1980) 114-128. [9] R.L. Keeney, Decision Analysis: An Overview, Operations Research 30, Nr. 5 (1982) 803-838. [10] R.L. Keeney and H. Ralffa, Decision With Multiple Objectives: Preferences and Value Tradeoffs (Wiley, New York, 1976). [11] R.L. Keeney and A. Sicherman, An Interactive Computer Program for Assessing and Analyzing Preferences Conceming Multiple Objectives, Behavioral Science 21, Nr. 3 (1976) 173-182. [12] J.N. Lyness, Quid, Qou, Quadrature.'?, In: D. Jacobs, Ed., The State of the Art in Numerical Analysis (Academic Press, London, 1977). [13] M.W. Merkhofer, The Value of Information Given Decision Flexibility, Management Science 23, Nr. 7 (1977) 716-727. [14] Microsoft Corporation, FORTRAN Compiler for the MSDOS Operating System, Version 3.0 (Microsoft Corporation, Bellevue, 1984). [15] A.C. Miller, M.W. Merkhofer, R.A. Howard, J.E. Matheson and T.R. Rice, Development of Automated Aids for Decision Analysis, Technical Report 3309 (SRI International, Menlo Park, 1976). [16] E.S. Pearson and J. Tukey, Approximating Means and Standard Deviations Based on Distances Between Percentage Points of Frequency Curves, Biometrika 52 (1965) 533-546. [17] C. Perry and I.D. Greig, Estimating the Mean and Variance of Subjective Distributions in PERT and Decision Analysis, Management Science 21, Nr. 12 (1975) 1477-1480. I18] UB. Pulley, Mean-Variance Approximation to Expected Logarithmic Utility, Operations Research 31, Nr. 4 (1983) 685-696. [19] A.P. Sage and C.C. White, ARIADNE: A KnowledgeBased Interactive System for Planning and Decision Support, IEEE Transactions SMC-14, Nr. 1 (1984) 35-47. [20] M. Sakawa and F. Sco, Integrated Methodology for Computer-Aided Decision Analysis, In: R.T. Trappl, F. De P. Hanika and R. Tomlinson, Eds., Progress in Cybernetics and Systems Research, Volume X (Henusphere, Washington, 1982). [21] R. Schlaifer, Computer Programs for Elementary Decision Analysis (Graduate School of Business, Harvard University, 1971). [221 l.M. Sobol, The Monte Carlo Method (University of Chicago Press, Chicago, 1974). [23] A.H. Stroud, Approximate Calculation of Multiple Integrals (Prentice-Hall, Englewood Cliffs, 1971). [24] M. Tribus, Rational Descriptions, Decisions and Designs (Pergamon Press, New York, 1969). [25] R.E. Trueman, An Introduction to Quantitative Methods for Decision-Making (Holt Rinehart and Winston, New York, 1974). [26] Woodward-Clyde Consultants, Decision Framework for Technology Choice - Volume 2: Decision Analysis Users' Manual, Report EA-2153 (Electric Power Research Institute, Palo Alto, 1982).