29
On Bernoulli experiments with imprecise prior probabilities Citation for published version (APA): Coolen, F. P. A. (1992). On Bernoulli experiments with imprecise prior probabilities. (Memorandum COSOR; Vol. 9219). Eindhoven: Technische Universiteit Eindhoven. Document status and date: Published: 01/01/1992 Document Version: Publisher’s PDF, also known as Version of Record (includes final page, issue and volume numbers) Please check the document version of this publication: • A submitted manuscript is the version of the article upon submission and before peer-review. There can be important differences between the submitted version and the official published version of record. People interested in the research are advised to contact the author for the final version of the publication, or visit the DOI to the publisher's website. • The final author version and the galley proof are versions of the publication after peer review. • The final published version features the final layout of the paper including the volume, issue and page numbers. Link to publication General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal. If the publication is distributed under the terms of Article 25fa of the Dutch Copyright Act, indicated by the “Taverne” license above, please follow below link for the End User Agreement: www.tue.nl/taverne Take down policy If you believe that this document breaches copyright please contact us at: [email protected] providing details and we will investigate your claim. Download date: 24. Jan. 2020

On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

  • Upload
    others

  • View
    5

  • Download
    0

Embed Size (px)

Citation preview

Page 1: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

On Bernoulli experiments with imprecise priorprobabilitiesCitation for published version (APA):Coolen, F. P. A. (1992). On Bernoulli experiments with imprecise prior probabilities. (Memorandum COSOR; Vol.9219). Eindhoven: Technische Universiteit Eindhoven.

Document status and date:Published: 01/01/1992

Document Version:Publisher’s PDF, also known as Version of Record (includes final page, issue and volume numbers)

Please check the document version of this publication:

• A submitted manuscript is the version of the article upon submission and before peer-review. There can beimportant differences between the submitted version and the official published version of record. Peopleinterested in the research are advised to contact the author for the final version of the publication, or visit theDOI to the publisher's website.• The final author version and the galley proof are versions of the publication after peer review.• The final published version features the final layout of the paper including the volume, issue and pagenumbers.Link to publication

General rightsCopyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright ownersand it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights.

• Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal.

If the publication is distributed under the terms of Article 25fa of the Dutch Copyright Act, indicated by the “Taverne” license above, pleasefollow below link for the End User Agreement:

www.tue.nl/taverne

Take down policyIf you believe that this document breaches copyright please contact us at:

[email protected]

providing details and we will investigate your claim.

Download date: 24. Jan. 2020

Page 2: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

;

EINDHOVEN UNIVERSITY OF TECHNOLOGYDepartment of Mathematics and Computing Science

Memorandum COSOR 92-19

On Bernoulli Experiments withImprecise Prior Probabilities

F.P.A. Coolen

Eindhoven, June 1992The Netherlands

Page 3: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

Eindhoven University of TechnologyDepartment of Mathematics and Computing ScienceProbability theory, statistics, operations research and systems theoryP.O. Box 5135600 MB Eindhoven - The Netherlands

Secretariate: Dommelbuilding 0.03Telephone: 040-473130

ISSN 0926 4493

Page 4: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

On Bernoulli Experiments with ~recise Prior Probabilities

F.P.A. Coolen

Eindhoven University of Technology

Department of Mathematics and Computing Science

P.o. Box 513, 5600 MB Eindhoven, The Netherlands.

Abstract

This paper deals with a situation where the prior distribution in

a Bayesian treatment of a Bernoulli experiment is not known

precisely. Imprecision for Bernoulli experiments is discussed for

the case where the prior densities are defined in the form of

intervals of measures, and a simple model with conjugate imprecise

prior densities is used for ease of calculation in case of

updating. Attention is focussed on imprecision with regard to

predictive probabilities.

The main aim of the paper is an analysis of imprecision,

emphasizing the important relation between information and

imprecision. This paper contributes to the theory advocated by

Walley (1991). Imprecision within the chosen model is compared to

intuition, and it is shown that complete lack of prior information

can be reasonably well represented.

Key Words

Bayesian analysis, Bernoulli experiments, conjugate prior

densities, imprecision, information, intervals of measures,

predictive probabilities.

Paper presented at the Third International Conference on Practical

Bayesian Statistics of The Institute of Statisticians, Nottingham

(U.K.), 8-11 july 1992.

Page 5: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

1. Introduction

We consider a standard Bernoulli situation. Independent random

variables X. (i=l, 2, .. ) with identical binomial probability mass~

functions, p(xi=xln) = nX (l-n) 1-x, are sampled (x=O or 1 and

O~n~l) . We refer to the observation of a number, say k, of the

independent random variables X, as a Bernoulli experiment (if k=l,~

the experiment is also called a trial). An observation of X, with~

value 1 is called a success and an observation with value 0 a

failure. Let X be the number of successes in an experiment of size

k (ke~ predetermined), then the distribution of X is the binomial:

px(x=xln) = (~)nx(l_n)k-X, for xe{O,l, .. ,k} and O~n~l, denoted

by X-Bin(k,n) .

We are interested in the probability of a success in a future

trial, or, more generally, in the probability of x successes in a

future experiment of size k. Inferences must be based on the

available information, which can consist of historical data

(results of past experiments) and opinions (subjective data) .

Walley (1991) in an extensive discussion shows that the amount of

information available should be reported in a model by means of a

concept of imprecision (or precision). The more information one

has, the more precise statements can be made, even about events

under uncertainty (unless new information is contradictory to old

information) .

We use the concept of imprecise probabilities (Walley (1991» with

lower and upper probabilities denoted by P and P respectively,

following axioms derived by Smith (1961) {see also Wolfenson and

Fine (1982)}.

2

Page 6: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

In this paper, inferences for the parameter n of a Bernoulli

process are discussed within the Bayesian framework using

imprecise prior densities for n in the form of intervals of

measures (DeRobertis and Hartigan (1981». We do not discuss

elicitation here, but remark that assessment of such intervals of

measures is more suitable for practical application than

assessment of a single prior distribution (Walley (1991, ch.4».

In section 2 of this paper a short literature survey is given,

discussing both precise and imprecise models. In section 3 a

statistical model based on imprecise probabilities is proposed, in

which conjugate imprecise prior densities are used to simplify

calculation, and imprecision for a Bernoulli trial is discussed.

In section 4 the behaviour of the model is discussed for

experiments of size k, again with an emphasis on imprecision.

Finally, in section 5 we discuss lack of prior knowledge and give

an example of possible practical application.

2. Survey of the Literature

2.1 Precise Models

If the parameter n is regarded as a random variable, the density

of a conjugate prior distribution is a beta (Colombo and

Constantini (1980» a-I ~-1g (p) ~ P (l-p) for O~p~l (a,~>O). Wen

write n-Be(a,~). If an experiment of size n leads to s successes

this prior can be updated giving a posterior n-Be(a+s,~+n-s). We

summarise the data as (n,s). Lee (1989, ch.3) shows that any

3

Page 7: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

reasonably smooth unimodal distribution on [0,1] can be well

approximated by some beta distribution.

The predictive distribution for X, corresponding to the conjugate

prior, is a Polya distribution, X-pol(k,a,~), also called a

beta-binomial distribution (Raiffa and Schlaifer (1961)), with

probability mass function Px(X=x) = (~)B(a+x,~+k-X)JB(a,~) for

xelO,1, .. ,kl, with B the beta function. On updating (a,~) is

replaced by (a+s,~+n-s).

Many priors have been proposed to represent total lack of

knowledge (Walley (1991, section 5.5.2)). Perhaps the most obvious

proposition is n-Be(1,1), the uniform distribution on [0,1]. Bayes

(1763) and Laplace (1812) used uniform priors. Also proposed are

Be (1/2, 1/2) (Jeffreys (1983) and Perks (1947)), and an improper

(not integrable over [0,1]) prior that can be represented by

Be(O,O) (Haldane (1945), Novick and Hall (1965), Jaynes (1968) and

Villegas (1977)). Priors that do not belong to the beta family

have also been proposed (Zellner (1977)). Other interesting

contributions are Lindley and Phillips (1976) and Geisser (1984).

Press (1989) extensively discusses Bayesian analysis for the

binomial distribution, with n both discrete and continuous.

Bayesian analysis with discrete n is disregarded here, as it is of

little practical interest.

The fact that authors do not agree about a 'noninformative' prior

distribution indicates that none of the priors has all properties

required to model ignorance (there is even no agreement about a

criterion for ignorance within the usual concept of precise

probability). Further, each prior distribution leads to a single

4

Page 8: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

predictive distribution, this does not represent the lack of

knowledge. For example, all prior distributions that are symmetric

around 0.5 (e.g. n-Be(a,a) for every a) lead, for k=l, to

predictive probability mass function P (X=O)=P (X=1)=1/2. It isX X

remarkable that, within the standard Bayesian framework an optimal

decision is almost always obtained, even if one has tried to use

no information at all.

2.2 Imprecise Models

A full overview is given by Walley (1991), so we restrict

ourselves to a short survey. The idea of imprecise probabilities

goes back to Boole (1854). For more than a hundred years little

attention was paid to the subject until the early sixties when

Smith (1961) and Good (1962) provided important contributions to

the theory. Dempster (1968) proposed a theory that has been shown

to fail in some cases (Walley (1991, section 5.13)). The method of

DeRobertis and Hartigan (1981), 'intervals of measures', is

especially useful when imprecise probabilities are used for

parameters, as in the Bayesian framework (see Pericchi and Walley

(1991)). The method of DeRobertis and Hartigan is adopted here. We

regard imprecision to be the essential concept in reporting an

amount of information, it is in this way that our method differs

from the sensitivity approach and robust Bayesian analysis (Berger

(1985), Berger and Berliner (1986), Lavine (1991a,b), Moreno and

Cano (1991)).

The problem of inference for Bernoulli trials was discussed by

Walley (1991, chapter 5), but his suggested method is less general

5

Page 9: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

than here, as the set of prior densities consists only of beta

densities. Our method, following DeRobertis. and Hartigan (1981),

defines bounds for prior densities, and depends less on an assumed

parametric model. Walley (1991) restricts himself to a set of beta

densities, which is out of tune with his overall approach.

We adopt Walley's interpretation of imprecise probabilities in

terms of betting behaviour (see also Smith (1961) and de Finetti

(1974) ) .

For an event of interest, say A, with lower probability P(A) and

upper probability P(A), imprecision is defined as

~(A) := P(A) - P(A), and the amount of information corresponding

to these imprecise probabilities is I(A) := ~(A)-l - 1 (Walley

(1991». Such a relation between imprecision and information (with

zero imprecision if and only if one has an infinite amount of

information, and maximum imprecision if and only if one has no

information) is essential for a theory of statistics, but widely

disregarded.

3. Modelling Imprecision

3.1 The Model

Following the method of intervals of measures, proposed by

DeRobertis and Hartigan (1981), imprecise prior densities can be

chosen such that a conjugacy property holds. To achieve this let

a-1 fj-1jBtn(p) = p (l-p) (a,fj) be the lower prior density for n

(O~p~l), with a>O and fj>O, and Un(P) = tn(p) + cOan(p) the upper

6

Page 10: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

prior density, where;\-1 /-l-1

a (p) = p (l-p) jB(;\,/-l),l[with ;\>0 and

/-l>0, and CO~O. Here we emphasize imprecision and updating.

Lower and upper prior predictive probability mass functions are

(x=O,l, .. ,k):

ex(x) (~)B(<<+X,~+k-X)jB(<<,~) and

Px(x) = (~)B(<<+X,~+k-X)jB(<<,~) + Co(~)B(;\+X'/-l+k-X)jB(A'/-l)'

respectively.

Corresponding imprecise probabilities for the event X=x are

p (X=x)-xex(x)

ex (x) + ) px (i)i~X

and PX(X=x)Px(x)

If data (n,s) become available, the imprecise prior densities are

updated to imprecise posterior densities

ul[(pln,s)

al[(pln,s)

t (pin,s) + c a (pin,s), wherel[ n l[

pA+S-1(1_p)/-l+n-s-1jB(A,/-l). Note that these updated

densities t and a are not normalized, and that Co is replaced byl[ l[

c (this will be explained later on). The corresponding updatedn

imprecise predictive probabilities are again derived from these

densities as above, using the lower and upper posterior predictive

probability mass functions

ex(xln,s) (~)B(<<+s+x,~+n-s+k-X)jB(<<,~) and

px(xln,s) (~)B(<<+s+x,~+n-s+k-X)jB(<<,~)

+ Cn(~)B(A+s+x'/-l+n-s+k-X)jB(A'/-l)'

Corresponding to the prior densities t and U (and analogously tol[ l[

posterior densities) for l[ are, respectively, lower and upper

cumulative distribution functions (cdf, see Walley (1991, ch. 4»

7

Page 11: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

p

A(p) = I an(v)dv ando

F (p)-n

F (p)n

pI in (v) dvo L(p)

L + Co [A-A(p)]

L(p) + COA(P)

L + COA(P)

and

with

00

A = I an(v)dv.o

For the proposed prior i and a we have L=A=l,n n

while this does not hold after updating.

The imprecision corresponding to these imprecise cdf's for n~p is

!J. (n~p)n

F (p)-F (p) =n -n

2cO/LIp) [A-A(p)]+A(p) [L-L(p)]} + COA(p) [A-A(p)]

2 2L +COLA+COA(p) [A-A(p)]

a non-decreasing function of cO. Further !J.n(n~p) is increasing in

Co unless L(p)=A(p)=O or both L(p)=L and A(p)=A hold (then

imprecision is zero for all cO) .

After updating the imprecision has a similar form, and is

increasing as function of c , except in the trivial situations.n

The amount of information, according to Walley's measure, is a

decreasing function of cO' except in trivial situations with

infinite information.

Next we discuss briefly some earlier results on the imprecision

for n~p, and then the imprecision for X corresponding to imprecise

predictive probabilities.

8

Page 12: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

3.2 Imprecision

Coolen (1992b) shows that (if L and A are known)COA

U\r (ll:Sp) IO:Sp:Sl ICOA

and if is such that:S max :S+ C A' Pm2L + COA L

°fJ. (ll:SP ) = max {fJ. (ll:Sp) 10:sp:Sll, then p E [m , mtl , with m and m

t1l m 1l m a a

the medians of the probability distributions with normalized

densities a IA and t IL respectively. If density a is identical1l 1l 1l

to density t (as in the following discussion), then the maximum1l

Coimprecision for ll:SP is (Coolen (1992a)).

2 + Co

For Bernoulli experiments it is more useful to analyze the

imprecision corresponding to the predictive probabilities, as this

relates to the variable of interest, x, itself. We discuss the

above model for k=l, with A=a, ~=~ and c =c for all n (a, ~ and cn

positive). Given data (n,s), the imprecision for X=O becomes

{remark that fJ.x(x=Oln,s)=fJ.x(x=lln,s), as Px(X=O)=l-~x(X=l) and

~x(X=O)=l-Px(X=l)l;

fJ.x(x=Oln,s) = px(x=Oln,s) - ~x(x=Oln,s)

2(c +2c) (a+s) (~+n-s)

2 2C (a+s) (~+n-s) + (c+l) (~+a+n)

The prior imprecision for X=O is2

I (c +2c) a~fJ.x(x=O 0,0) = 2 2

C a~ + (c+l) (~+a)

As a function of a, for a constant ~, this prior imprecision is a

maximum if a=~, strictly increasing if a<~ and strictly decreasing

if a>~. The following analysis is based on this simple form for

The effect of data on the imprecision for X=O is of interest, so

we analyze the sign of Dx(n,s);= fJ.x(x=OIO,O)

it is assumed that n>O. For simplicity we use a continuous

9

Page 13: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

variable sE[O,n] in place of the discrete SE{O,l, .. ,n}, thus

making Dx a continuous function. We write

with a positive denominator and numerator

DX(n,s) as a quotient,

2(c +2c) (c+1) gIn, s),

with 2 2 2 2 2gIn,s) = (<<.+(3) (s-n)s + (<<. -(3 ) (<<.+(3)s + (Xn«(3n+(3 -«.). Since

the sign of Dx(n,s) is determined by the sign of gIn,s), only the

sign of gIn,s) needs to be analyzed. The results are given in

table 3.1. Moreover g(n,s»O means that imprecision decreases

(precision increases), so a '+' in table 3.1 indicates that

precision increases. The sign of gIn,s) also indicates the change

in Walley's measure of information ('+' indicates more

information) .

gIn,s) = 0 <=* s = (<<.:(3)n =: sl or s = (3-«'+(<<.~(3)n =: s2

{s2E[0,n] does not always hold}.

This analysis shows some interesting facts about imprecision in

this simplified model. First of all, if data become available with

S=sl' sosn

(the relative number of successes fully

agrees with our prior ideas), then D(n,s)=O for all n, and neither

the imprecision nor Walley's measure of information change. This

situation would surely suggest decreasing imprecision (increasing

information), and is a strong argument in favour of choosing c ton

be a decreasing function of n instead of being constant (according

to an analogous argument, it seems to be logical that c ~O ifn

10

Page 14: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

Table 3.1

a, ~, n sign of g sl' s2

g: + 0 +n/2s1 =

I. a = ~ I I I s2 = n/2s: 0 n/2 n

II. a < ~ 0<sl<n/2<s2

g: + 0 -< ~-a(a+~) s1n < ----ex (~-a)

I I I s2 > ns: 0 s1 n

g: + 0 - 0~-a(a+~) s1 =

n = ----ex (~-a) I I I s2 = ns: 0 sl n=s2

g: + 0 - 0 + > ~-an > (a+~) sl----ex (~-a) I I I I s2 < n

s: 0 sl s2 n

III. a > ~ s2<n/2<sl<n

g: - 0 +(a-~)a/~(a+~) s1 <n < T (a-~)

I I I s2 < 0s: 0 s1 n

g: 0 - 0 +(a-~)a/~(a+~) sl =n = T (a-m I I I

I s2 = 0s: O=s s1 n

2

(a+~)g: + 0 - 0 +

(a-~)a/~n > T (a-~)

sl >I I I I

s2 > 0s: 0 s2 sl n

Secondly, it is worth remarking that data contradictory to the

prior ideas increase the imprecision if n is small, but a large n

can also make the imprecision decrease (as shown by the role of

S2). Such behaviour seems to be natural, and is called prior-data

conflict (Walley (1991, section 5.4».

11

Page 15: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

The behaviour of imprecision around sl is also remarkable, but

easily explained within the model by considering the behaviour of

~x as function of ~ and ~. It is more interesting to discuss this

behaviour from an intuitive point of view. Here the idea is that,

if the updated densities are further away from n=1/2 than the

priors are, imprecision decreases. If we feel, after we have got

new data, that n will probably be closer to the nearest extreme

value (O or 1) than we thought before, imprecision for the

probability of success at the next trial decreases.

The concept of imprecision is important as it corresponds to the

amount of information available on which to base subjective

statements. An even better interpretation is as the value of the

information available to us, with regard to the way in which it

changes our ideas about the event of interest. Thus imprecision is

the correct tool to report the amount of information. This

generalization of the standard theory of probability is necessary

to accommodate this point of view.

The effect of new data can be divided in two parts. First, some

important aspects of the data, e.g. mean value (these aspects

are often summarized by sufficient statistics), have influence on

our thoughts. Secondly, the amount of data is important, and this

aspect was not taken into consideration by the simple model for

k=l discussed above, but it has been remarked that this could be

solved by choosing c a decreasing function of n. A very simplen

1 + n7~' For the modelCn

by Coolen {1992b, 1992a (slightly

Coform for c is proposed

n

different definition of the model»:

12

Page 16: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

discussed above, with a equal to t , ~ is such that the minimum1f 1f

posterior amount of information (corresponding to the posterior

densities for 1f, according to Walley's measure), if n=~, is twice

the minimum prior amount of information. This form of c isn

adopted here, and in the following chapter we discuss imprecision

according to the complete model for Bernoulli experiments.

4. Imprecision for the Binomial Model

For experiments of size k, the above model leads to imprecise

prior predictive cdf's (x=O,I, .. ,k):

F (x)-x p (x~x)-x

Px(x~x)

x k

Lex (i) + L Px (i )i=O i=x+l

and

in section 3.1, depending on hyperparameters tt, ~, A, M and cO'

The corresponding prior imprecision for x~x is

bX(X~x) = FX(X) - ~x(x), for all x=O,l, .. ,k. After updating, the

hyperparameters are changed as before, and the imprecise posterior

predictive cdf's and posterior imprecision are again easily

derived. However, the resulting form for the difference between

posterior and prior imprecision is awkward for theoretical

analysis, except for its behaviour related to ~. If we define, for

13

Page 17: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

n and Co positive, thendDX(X~xln,s)

d~ < 0 for x~k-1. This follows

easily fromctax(x~xln,s)

dcn

> 0 for these values of x.

can be interpreted as the loss of imprecision, or the gain of

information about the event X~x. The imprecision of the event X=x,

~X(x=x), depends analogously on ~. So for larger ~ the gain of

information from data is less than for smaller ~ (this gain of

information might be negative) .

To show the behaviour of imprecision within this model an example

is used.

Example 4.1

In table 4.1 the results are given for the following situation:

«=2, ~=B, A=4, ~=6 and cO=l. First the lower and upper

probabilities for the events X~x (edf's) and X=x are given for the

prior situation, and thereafter posteriors for five different

cases.

14

Page 18: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

Tab1e 4.1

Prior Post. (10,8) ~=10

x F FX

P Px F FX

P Px-x -x -x -x0 .1070 .2354 .1070 .2354 .0007 .0098 .0007 .00981 .2436 .5159 .1294 .3077 .0046 .0679 .0038 .05882 .3831 .7355 .1123 .3057 .0155 .2309 .0107 .17703 .5196 .8712 .0813 .2732 .0379 .4841 .0209 .34434 .6514 .9431 .0509 .2280 .0767 .7189 .0312 .49475 .7729 .9773 .0278 .1771 .1399 .8692 .0369 .58426 .8736 .9921 .0133 .1252 .2434 .9467 .0340 .60717 .9437 .9977 .0054 .0775 .4150 .9815 .0236 .56708 .9819 .9995 .0018 .0395 .6684 .9951 .0119 .45949 .9968 .9999 .0004 .0149 .9097 .9993 .0041 .2819

10 1 1 .0001 .0032 1 1 .0007 .0903

Post. (5,4) ~=10 Post. (10,3) ~=10

x F FX

P P F F P Px-x -x X -X X -X

0 .0050 .0361 .0050 .0361 .0628 .1144 .0628 .11441 .0219 .1680 .0168 .1361 .2014 .3524 .1344 .24852 .0555 .4013 .0320 .2772 .3874 .6112 .1623 .30963 .1085 .6466 .0445 .4031 .5837 .8037 .1396 .28984 .1847 .8225 .0488 .4773 .7565 .9151 .0925 .22275 .2923 .9214 .0430 .4951 .8818 .9687 .0489 .14376 .4435 .9693 .0302 .4624 .9543 .9904 .0210 .07687 .6402 .9899 .0167 .3837 .9866 .9977 .0072 .03308 .8401 .9975 .0071 .2657 .9973 .9996 .0019 .01089 .9650 .9997 .0021 .1331 .9997 1.0000 .0003 .0024

10 1 1 .0003 .0350 1 1 .0000 .0003

Post. (10,3) ~=5 Post. (10,3) ~=20

x F FX

P Px F FX

P Px-X -X -X -X

0 .0714 .1090 .0714 .1090 .0561 .1198 .0561 .11981 .2265 .3364 .1517 .2342 .1813 .3676 .1207 .26232 .4280 .5904 .1818 .2877 .3539 .6299 .1466 .33023 .6290 .7873 .1559 .2631 .5445 .8178 .1265 .31464 .7936 .9060 .1034 .1957 .7227 .9227 .0836 .24805 .9044 .9648 .0550 .1214 .8604 .9718 .0441 .16486 .9644 .9892 .0237 .0624 .9444 .9915 .0188 .09087 .9899 .9974 .0082 .0258 .9834 .9980 .0064 .04008 .9980 .9996 .0021 .0082 .9966 .9997 .0017 .01339 .9998 1. 0000 .0004 .0018 .9996 1.0000 .0003 .0030

10 1 1 .0000 .0002 1 1 .0000 .0004

15

Page 19: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

In this discussion we consider only imprecision for X=x, but it is

also interesting to study imprecision for XSx from table 4.1. For

the prior, the maximum imprecision is ~x(X=2) = 0.1934. First,

with ~=10, we compare data (10,8) and (5,4). Here, the maximum

imprecision is ~x(X=6110,8) = 0.5731 and ~x(X=515,4) = 0.4521

respectively. For these examples, the data obviously conflict with

the prior thoughts, and this leads to increasing imprecision for

almost all X=x, except for small values of x, for which the

probabilities become very small. The other three cases represented

in table 4.1 all have data (10,3), that do not conflict with prior

opinion, and from the numbers in table 4.1 the behaviour of the

imprecise probabilities depending on the value of ~ can be

studied. For ~=5 the information of the data has more value,

compared to our prior thoughts, than for ~=10 and ~=20, and leads

to less posterior imprecision. The maximum imprecision for X=x

confirms this, as ~x(X=31~=5)

and ~x(X=31~=20) = 0.1881.

0.1072, ~x(X=31~=10) = 0.1502

5. Lack of Prior Knowledge and Possible Practical Application

The situation of no information on event A can theoretically be

covered by the so-called vacuous probabilities P(A)=O and

P(A)=l, implying ~(A)=l and I(A)=O. However, Walley (1991,

section 2.9.1) shows that these probabilities are not useful in

practical problems to model prior beliefs about a statistical

parameter, as they give rise to vacuous posterior probabilities.

It is also unlikely in practice that one has no information at

16

Page 20: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

all, and the problem of vacuous posterior probabilities disappears

for all non-vacuous prior probabilities, so total lack of prior

information can be approximated by choosing the hyperparameters

such that the lower probabilities for all events are arbitrarily

close to zero and the upper probabilities close to one. Within the

model proposed in this paper this can be reached by choosing Co

large. For Co increasing to infinity the limiting values of the

predictive probabilities for X, ~x(x=x) and Px(x=x), are zero and

one respectively (these predictive probabilities relate to X, and

are therefore of primary interest to us). For the model of section

3.2 (where c=cO

) this easily follows from2

(c +2c)~~ I2 2' so lim x~(X=O 0,0) = 1.

c ~~ + (c+1) (~+~) c~

Example 5.1 shows how this idea of 'almost no prior information'

can be used in practice (elicitation is not considered). Again the

model of section 3.2 is used.

It is important to remark that, as example 5.1 will also show,

although this method seems to be promising for modelling lack of

prior information, further research is necessary, for example on

the form of c . The c suggested above performs well if priorn n

information is regarded as informative, but its relation to Co is

doubtful if Co becomes very large, as is necessary when

approximating prior lack of information. With the form of cn

proposed above, small changes in the prior imprecision, if this is

close to one, can lead to large changes in the posterior

imprecision. However, the relative change (by updating) of

Walley's measure of information for this discrete predictive

distribution is almost equal for different cO.

17

Page 21: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

Example 5.1

Let k=10 and a=A=~=~=l. Then ex(x) = 1/11

for all x=O,l, .. ,10. This leads to p (X=x)-x

and Px(X) = (CO+1)/11

= 1/(11+10CO

) and

Px(x=x) (CO+1)/(C

O+11) for all x. The corresponding imprecise

cdf's for X are F (x)-xx + 1

11 + Co (10-x)and F X(x)

(Co+1) (x+1)

11 + Co (x+1) .

It is easy to see that Co can be chosen such that P (X=x) and-xPx(x=X) are arbitrarily close to 0 or 1, although 0 and 1 cannot

be adopted for finite cO. This can also be done for all F (x) and-xFX(X)' where it should be remarked that F (10) = F (10) 1.-x X

As a practical example of where this can be used, suppose that

there are 10 identical machines and the object of interest is the

number X of these machines that will operate for one month without

breakdown, while nothing at all is known about these machines. A

possible choice to model this prior lack of information is to

define Co such that ~x(x=x)~O.Ol and Px(x=x)~0.99. These

constraints hold for cO

=1000, so this value is adopted.

Suppose that the first month 8 of these 10 machines operated

successfully (and there is no wearout). Then your prior opinions

can be updated. The parameters ~ and Co should already be chosen,

to indicate the value you assign to new data, compared to your

prior knowledge. Remember that ~ can be interpreted as the number

of data that gives you an equal amount of information as you had

before, so it is logical to choose ~ very small. In table 5.1 the

results of updating are given for ~=1 and ~=0.01. This last value

means you regard the value of one observation as 100 times the

value of your prior knowledge. Also results are given if the data

were (60,48) instead of (10,8), indicating for example a situation

18

Page 22: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

wherein after six months you know that on 48 occasions a machine

has operated for a month successfully, again with 10 machines in

the sample. Finally, the results are given for a situation with

CO

=10000, which implies P (X=x)SO.OOl and P (X=x)~0.999. Note that~ X

the results seem to be very sensitive, but remember the above

remark that Walley's measure of information is not (for example,

if imprecision 6x

(X=x) increases from 0.98 to 0.998, then

information Ix(X=X) decreases from 2/98 to 2/998) .

The prior for cO=10000 is not given in table 5.1. For this Co we

have P (X=x)=O.OOOO and P (X=x)=0.9990 for all x. So if we regard~ X

the prior imprecision for X=x, then

6x

(X=xlco=1000) = 0.9901-0.0001 = 0.9900, and

6x

(x=xlco=10000) = 0.9990-0.0000 = 0.9990, and for Walley's

measure of information we get Ix

(x=xlco=1000) = 0.010 and

IX

(X=X!C O=10000) = 0.001. For the corresponding posterior

probabilities, after data (60,48) with ~=0.01, the maximum values

for 6x (X=x) are 6x(X=8Ico=1000, (60,48» = 0.0620 and

6x(X=8Ico=10000, (60,48» = 0.3817. The corresponding values of

Walley's information measure are Ix

(x=xlco=1000, (60,48» = 15.1

and Ix

(x=xlco

=10000, (60,48» = 1.62.

The relative difference in the amount of information has changed

very little after updating.

19

Page 23: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

Table 5.1

Prior cO

=1000 Post. (10,8) ~=1 c =1000-- 0x F F

XP Px F F

XP P-x -x -x -x X

0 .0001 .9901 .0001 .9901 .0000 .0169 .0000 .01691 .0002 .9955 .0001 .9901 .0000 .1277 .0000 .11442 .0004 .9973 .0001 .9901 .0001 .4043 .0001 .34673 .0006 .9983 .0001 .9901 .0003 .6948 .0002 .61154 .0008 .9988 .0001 .9901 .0007 .8617 .0004 .78995 .0012 .9992 .0001 .9901 .0018 .9374 .0009 .88416 .0017 .9994 .0001 .9901 .0040 .9711 .0016 .93087 .0027 .9996 .0001 .9901 .0088 .9869 .0024 .95358 .0045 .9998 .0001 .9901 .0215 .9946 .0030 .96269 .0099 .9999 .0001 .9901 .0713 .9985 .0028 .9599

10 1 1 .0001 .9901 1 1 .0015 .9287

Post. (60,48) ~=1 c =1000 Post. (10,8) ~=.01 cO

=10000

x F FX

P P F FX

P Px-X -x X -X -X

0 .0000 .0000 .0000 .0000 .0001 .0004 .0001 .00041 .0000 .0006 .0000 .0005 .0008 .0032 .0007 .00282 .0000 .0063 .0000 .0058 .0037 .0145 .0029 .01143 .0002 .0440 .0001 .0381 .0122 .0472 .0085 .03314 .0008 .1925 .0006 .1607 .0328 .1193 .0201 .07565 .0031 .4882 .0023 .4103 .0754 .2457 .0399 .14236 .0103 .7597 .0065 .6635 .1547 .4223 .0682 .22647 .0305 .9050 .0142 .8130 .2906 .6208 .1004 .30858 .0905 .9678 .0219 .8712 .5029 .8017 .1230 .35919 .3109 .9927 .0191 .8551 .7794 .9338 .1154 .3426

10 1 1 .0073 .6891 1 1 .0662 .2206

Post. (60,48) ~=.01 c O=1000 Post. (60,48) ~=.01 C O=10000- -

x F FX

P Px F F P Px-x -x -x X -X

0 .0000 .0000 .0000 .0000 .0000 .0000 .0000 .00001 .0000 .0000 .0000 .0000 .0000 .0001 .0000 .00012 .0003 .0004 .0003 .0004 .0001 .0010 .0001 .00093 .0023 .0031 .0019 .0027 .0010 .0070 .0009 .00604 .0116 .0157 .0093 .0127 .0051 .0353 .0041 .02855 .0449 .0601 .0331 .0446 .0201 .1276 .0148 .09646 .1348 .1749 .0885 .1168 .0638 .3264 .0408 .23217 .3194 .3897 .1764 .2258 .1703 .5934 .0857 .39998 .5972 .6687 .2500 .3120 .3935 .8218 .1273 .50909 .8706 .9015 .2253 .2835 .7464 .9544 .1129 .4749

10 1 1 .0985 .1294 1 1 .0456 .2536

20

Page 24: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

References

Bayes, T. (1763) An essay towards solving a problem in the

doctrine of chances, Philosophical Transactions, Royal

Society of London, 45, pp. 370-418. Reprinted in Press

(1989) .

Berger, J.O. (1985) Statistical decision theory and Bayesian

analysis (Springer-Verlag, New York) .

Berger, J. and Berliner, L.M. (1986) Robust Bayes and empirical

Bayes analysis with c-contaminated priors, The Annals of

Statistics, 14 no.2, pp. 461-486.

Boole, G. (1854) An investigation of the laws of thought on which

are founded the mathematical theories of logic and

probabilities (Macmillan, London; reprinted by Dover (1958».

Colombo, A.G. and Costantini, D. (1980) Ground-hypotheses for beta

distribution as Bayesian prior, IEEE Transactions of

Reliability, R-29 no.l, pp. 17-20.

Coolen, F.P.A. (1992a) Imprecise conjugate prior densities for the

one-parameter exponential family of distributions, submitted

to Statistics & Probability Letters.

Coolen, F.P.A. (1992b) Maximum imprecision related to intervals of

measures and Bayesian inference with imprecise conjugate

prior densities, Cosor-Memorandum 92-07, Eindhoven University

of Technology.

Dempster, A.P. (1968) A generalization of Bayesian inference (with

discussion), Journal of the Royal Statistical Society, series

B, 30, pp. 205-247.

21

Page 25: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

DeRobertis, L. and Hartigan, J.A. (1981) Bayesian inference using

intervals of measures, Annals of Statistics 9, pp. 235-244.

de Finetti, B. (1974) Theory of probability (2 volumes) (Wiley,

London) .

Geisser, S. (1984) On prior distributions for binary trials (with

discussion), The American Statistician, 38 no.4, pp. 244-251.

Good, I.J. (1962) Subjective probability as the measure of a

non-measurable set, in: E. Nagel, P. Suppes and A. Tarski

(eds.) Logic, Methodology and Philosophy of Science, pp.

319-329 (Stanford University Press) .

Haldane, J.B.S. (1945) On a method of estimating frequencies,

Biometrika 33, pp. 222-225.

Jaynes, E.T. (1968) Prior probabilities, IEEE Transactions on

Systems Science and Cybernetics SSC-4, pp. 227-241.

Jeffreys, H. (1983) Theory of probability, 3rd edition (1st

edition 1939) (Clarendon Press, Oxford).

Laplace, P.S. de (1812) Theorie analytique des probabilites

(Courcier, Paris).

Lavine, M. (1991a) Sensitivity in Bayesian statistics: The prior

and the likelihood, Journal of the American Statistical

Association, 86 no.414, pp. 396-399.

Lavine, M. (1991b) An approach to robust Bayesian analysis for

multidimensional parameter spaces, Journal of the American

Statistical Association, 86 no.414, pp. 400-403.

Lindley, D.V. and Phillips, L.D. (1976) Inference for a Bernoulli

process (a Bayesian view), The American Statistician, 30

no.3, pp. 112-119.

22

Page 26: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

Lee, P.M. (1989) Bayesian statistics: An introduction (Oxford

University Press, New York) .

Moreno, E. and Cano, J.A. (1991) Robust Bayesian analysis with

e-contaminations partially known, Journal of the Royal

Statistical Society, series B, 53 no.1, pp. 143-155.

Novick, M.R. and Hall, W.J. (1965) A Bayesian indifference

procedure, Journal of the American Statistical Association

60, pp. 1104-1117.

Pericchi, L.R. and Walley, P. (1991) Robust Bayesian credible

intervals and prior ignorance, International Statistical

Review 58, no.1, pp. 1-23.

Perks, W. (1947) Some observations on inverse probability

including a new indifference rule, Journal of the Institute

of Actuarians, 73, pp.285-312.

Press, S.J. (1989) Bayesian statistics: Principles, models, and

applications (Wiley, New York) .

Raiffa, H. and Schlaifer, R. (1961) Applied statistical decision

theory (MIT Press, Cambridge, USA).

Smith, C.A.B. (1961) Consistency in statistical inference and

decision (with discussion), Journal of the Royal Statistical

Society, series B 23, pp. 1-25.

Villegas, C. (1977) Inner statistical inference, Journal of the

American Statistical Association 72, pp. 453-458.

Walley, P. (1991) Statistical reasoning with imprecise

probabilities (Chapman and Hall, London).

Wolfenson, M. and Fine, T.L. (1982) Bayes-like decision making

with upper and lower probabilities, Journal of the American

Statistical Association 77, no.377, pp. 80-88.

23

Page 27: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

Zellner, A. (1977) Maximal data information prior distributions,

in: A. Aykac and C. Brumat (eds.) New Developments in the

Applications of Bayesian Methods, pp. 211-232 (North-Holland,

Amsterdam) .

24

Page 28: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

,

List of CaSaR-memoranda - 1992

Number Month Author Title92-01 January F.W. Steutel On the addition of log-convex functions and sequences

92-02 January P. v.d. Laan Selection constants for Uniform populations

92-03 February E.E.M. v. Berkum Data reduction in statistical inferenceH.N. LinssenD.A. Overdijk

92-04 February H.J.C. Huijberts Strong dynamic input-output decoupling:H. Nijmeijer from linearity to nonlinearity

92-05 March S.J.L. v. Eijndhoven Introduction to a behavioral approachJ .M. Soethoudt of continuous-time systems

92-06 April P.J. Zwietering The minimal number of layers of a perceptron that sortsE.H.L. AartsJ. Wessels

92-07 April F .P.A. Coolen Ma.x.imum Imprecision Related to Intervals of Measuresand Bayesian Inference with Conjugate Imprecise PriorDensities

92-08 May I.J.B.F. Adan A Note on "The effect of varying routing probability inJ. Wessels two parallel queues with dynamic routing under aW.H.M. Zijm threshold-type scheduling"

92-09 May I.J .n.F. Adan Upper and lower bounds for the waiting time in theG.J.J.A.N. v. Houtum symmetric shortest queue systemJ. v.d. Wal

92-10 May P. v.d. Laan Subset Selection: Robustness and Imprecise Selection

92-11 May R.J .M. Vaessens A Local Search TemplateE.H.L. Aarts (Extended Abstract)J .K. Lenstra

92-12 May F.P.A. Coolen Elicitation of Expert Knowledge and Assessment of Im-precise Prior Densities for Lifetime Distributions

92-13 May M.A. Peters Mixed H 2 / H oo Control in a Stochastic FrameworkA.A. Stoorvogel

Page 29: On Bernoulli experiments with imprecise prior probabilities · independent random variables X, as a Bernoulli experiment (if k=l, ~ the experiment is also called a trial). An observation

Number92-14

92-15

92-16

92-17

92-18

92-19

MonthJune

June

June

June

June

June

AuthorP.J. ZwieteringE.H.L. AartsJ. Wessels

P. van der Laan

J.J .A.M. BrandsF.W. SteutelR.J.G. Wilms

S.J.L. v. EijndhovenJ.M. Soethoudt

J .A. HoogeveenH. OosterhoutS.L. van der Velde

F .P.A. Coolen

-2-

TitleThe construction of minimal multi-layered perceptrons:a case study for sorting

Experiments: Design, Parametric and NonparametricAnalysis, and Selection

On the number of maxima in a discrete sample

Introduction to a behavioral approach of continuous-timesystems part II

New lower and upper bounds for scheduling around asmall common due date

On Bernoulli Experiments with Imprecise PriorProbabilities