J. ltal. Statist. Soc. (1992) 3, pp. 325-334
A BAYESIAN JUSTIFICATION FOR THE L I N E A R P O O L I N G OF OPINIONS* Ma...
4 downloads
347 Views
372KB Size
Report
This content was uploaded by our users and we assume good faith they have the permission to share this book. If you own the copyright to this book and it is wrongfully on our website, we offer a simple DMCA procedure to remove your content from our site. Start by pressing the button below!
Report copyright / DMCA form
J. ltal. Statist. Soc. (1992) 3, pp. 325-334
A BAYESIAN JUSTIFICATION FOR THE L I N E A R P O O L I N G OF OPINIONS* Mario Di Bacco Universit~ degli Studi di Bologna
Vergilius Mocellin Universiut degli Studi di Venezia Summary It is known that the problem of combining a number of expert probability evaluations is frequently solved with additive or multiplicative rules. In this paper we try to show, with the help of a behavioural model, that the additive rule (or linear pooling) derives from the application of Bayesian reasoning. In another occasion we will discuss the multiplicative rule. Keywords: Bayesian inference, Expert resolution, linear pool. 1. Introduction
Let X be an h-dimensional (h i> 1) random vector (r. v.) and suppose the Decision-maker, DM, has assigned the distribution function, D.F., oF to r.v. X. Now DM consults k (k I> 1) Experts to learn their opinions on X; Expert iE answers this question by assigning the D.F. iF (i = 1 .... , k). The issue then is: how would DM combine (or aggregate) F Ck) = (1F, ..., kF) and oF to update his opinion on X or, equivalently, what would his or D.F. EF be? This is the > (Morris, 1983) which has been carefully studied by a number of statisticians (for a thorough, critical survey see Genest and Zidek, 1989). It seems that the most popular solutions are the additive rule (or linear pool), i.e.: k
k
eF = ~ wi ~F + Wo oF ; (wi , Wo) >~ 0 ;i=
Wi + w 0 = 1
(1.1)
i=1
* Research supported by C.N.R. and Ministry of University and Technological and Scientific Research. 325
M . D I B A C C O 9 V. M O C E L L I N
or the multiplicative rule (logarithmic p o o l ) , i.e. - denoting by f the density function (d.f.) associated to F -
(1.2)
or, for wi = 1, i = 1 . . . . , k, and Wo = 1 - X wi, Bordley's slight variant (Bordley, 1982)
It is usual enough to validate rules (1.1)-(1.3) in an axiomatic way, that is, to state them as ~optimal rules>) because they have some properties which are ~obviously>> or ~maturally>) typical for ~rationab~ pooling. We will not discuss this approach, but turn our attention to the Bayesian one or, following Genest and Zidek's terminology, the supra Bayesian approach, i.e. what was proposed by Morris's pioneer work. Denote L(x;d) as the likelihood that data (or information) d grives to x; then Morris proposes ef(x) o: of(X) L(x;F~k))
(1.4)
The resolution (1.4) has been receiving a lot of attention and there is a large number of contributions to its theoretic foundation as well as its practical usefulness. For our part we think everyone would agree that (1.4) is an effective tool only if DM is able to make L ( x ; F (k)) explicit. This is precisely the crux of the issue: it was Morris himself who wrote (Morris, 1986) ~Problems with the direct Bayesian approach of assessing likelihood functions over expert opinions are not conceptual; they are pratical. Most would agree that it is virtually impossible to directly assess a joint likelihood distribution over all possible expert densities>>. To overcome the impasse of ~virtual impossibility>>, he proposed a system of axioms which he considered selfevident and that would guide DM to a compelled choice of the likelihood version. Then (following this) he puts: L ( x ; F (k)) oc 17 if(x) and, obviously, arrives at rule (1.2) with wi = Wo = 1. This simplification undoubtely makes the application of (1.1) straightforward - indeed, almost automatic. Nevertheless this cannot be a good motive for distorting Bayesian inference, thrusting it - as de Finetti liked to say into a Procrustean bed. In fact, by giving an immutable version of likelihood, 326
LINEAR POOLING OF OPINIONS
this is, obviously, doing violence to the essence of inductive reasoning in the Bayesian sense. In our opinion, the desire to find a ~rational>~ (or reasonable, at least) rule able to resolve any problem of aggregation of opinions has been, and is, the main cause of genuine Bayesian disagreement with Morris's axiomatic proposal and of the criticisms made by - for instance, Lindley (Lindley, 1986). It is clear that, from the Bayesian viewpoint no single rule can be given to solve the expert problem. For DM, F (k) is an empirical datum that he now uses to update his present opinion on r.v.X. How he makes this explicit is his own business, subject only to the need to be coherent. Thus to guarantee his own coherence and, at the same time, facilitate interpersonal communication, we can ask him to use the empirical datum F ~k) in accordance with the rules of inductive reasoning. But DM wilt be able to do that only if, before consulting the experts, he holds precise opinions about the r.v. X and the expert evaluations, i.e. about the choice iF of expert iE (i = 1 . . . . . k). In spite of all this, there is no doubt that going in search of an ~optimal>> (or at least, a ~good~0 rule to aggregate probabilities is appealing when, for instance, the problem of resolving different probability judgements arises in the context of a decision-making involving more than one person. So appealing, indeed, as to silence the foundational scruples we discussed above. Then, it is natural to restrict the attention to the most popular rules, the additive or the multiplicative. This conflict between pragmatic requirements and strict Bayesian observance can be weakened, if there are situations of practical significance in which the explicitness of the likelihood function leads to the additive or the multiplicative rule. If both classes of situations are of common occurrence in practice, then it is right to say that from a pragmatic viewpoint there exist coherent rules for combining probability judgements, and DM may be able to find among them one that mirrors his own present beliefs. In this paper we try to describe some typical situations in which it appears appropriate to use the additive rule. We will give a Bayesian interpretation of a class of multiplicative rules elsewhere.
2. The additive rule We now present our paradigm, which will lead us to combine F ~k) and oF according to (1.1), i.e. according to the additive rule. Suppose that in DM's opinion the phenomenon q~- whose random realization is described by r.v. X - is placed in one, and only one, of the alternative 327
M . D I B A C C O 9 V. M O C E L L I N
environments, or state of the world, belonging to set 6) = {0}. T h e term ~environment>> and ~state of the world>> can be characterized in the following way: if D M and all the Experts of group E knew that the environment of 9 is O, then he and the whole E group would assign the D.F. F to X but, if it were known to all of them that the environment is O', then they would agree to assign the D.F. F' to X. For D M the true ~world's statue>> - i.e. the ~environment>> in which p h e n o m e n o n 9 is actually placed - is a random entity. On the contrary, he believes that each of the experts of group E is sure about what is the true world's state: in other words, D M believes that if iE assigns the D.F. iF to X, this happens because he is sure that Oi is the true environment. Then, if the answers of the experts do not coincide - i.e. if their probability assignments are 1F, .... iF, ..... kF, with iF ~ iF, i 4=j - D M can, at most, give the whole credit to only one expert. But if he did this, he would be incoherent, since he has just deemed experts all those who belong to group E, and then decided to consult all of them. It is natural, therefore, to think that D M believes n o n e of his experts to be totally credible thus admitting that each of them can be wrong about his diagnosis of the true world's state. As an example, let us suppose D M and his friend V. are about to play the following game. V. has got two coins, a fair coin C - i.e. P r { h e a d / C } = 1/2 and a biased coin C, so that Pr{head/C} = 1. V. will make two throws, each time choosing the coin to be thrown and D M who knows the rule but is in the dark about V.'s choice, will forecast the result. H e then consults 1E and 2E obtaining - if ~head>> is coded with 1 and ~tail>> with 0 - lp(i,j) = 1/4 (i = 0,1; j = 0,1) and, respectively, 2p(0,1) = 2p(1,1) = 1/2. Neither 1E or 2E want, nor is the opportunity offered them, to justify their probability assessments. It is D M himself who interprets them by saying that in 1E's [2E's] opinion it is 0 = (C,C) [ 0 = (C,C)]; there seems to be a paradox, of course, which immediately disappears if D M believes both Experts can be wrong! In the above setup how does the D M use data bxk) to update his own opinion about X?. H e r e we present our soultion. In giving it we assume O is a r a n d o m number (r.n.) with support T= {O:_O~i 1
(2.13)
k
T(e,O) = f3 iT c (e) i=1
where (.)c denotes complementation and {]l,...,ju} denote the whole set of choices of u (u ~> 1) suffixes - all different - among 1 ..... k. Let us then assume
[
~--~, O e iT(e) , O < e < l
a) floilo) = 0 < a(e)