Professional Documents
Culture Documents
Bayesian Statistics
Lecture 1
Motivation
Alvaro Mauricio Montenegro Daz
August, 2011
Table of contents
Brief History
Bayesian-frequentist controversy
Motivating Examples.
Table of contents
1 Brief History
2 Bayesian-frequentist controversy
3 Some basic Bayesian models
Classical Statistics
Classical Age
1 Probability is old. Several hundreds of years ago, most mathematicians
Francis Galton in the late 1800s. Karl Pearson added correlation and
goodness-of-fit measures around the turn of the last century.
3 Statistics blossomed. Between 1920 and 1930 R.A. Fisher developed the
Bayesian Statistics
Bayesian Age
1 Bayesian methods are much older. The original paper dating in 1763
century.
4 Several non-statisticians as Jeffreys (a physicist) and Bolye (an
Table of contents
1 Brief History
2 Bayesian-frequentist controversy
3 Some basic Bayesian models
(x) = x 1.96s/ n
Interpretation. Before any data are collected the probability that the interval
contains the true value is 0.95, for any value of and 2 .
After collecting the data and computing (x), the interval contains the true
or it does not; its coverage probability is not 0.95, but either 0 or 1.
It is not correct to say that the true has a 95% chance of falling in (x).
Alvaro Mauricio Montenegro Daz
12
j=9
Using the negative binomial likelihood we have
!
X
2+j j
(1 )3 = 0.0325.
2 = P= 1 (X 9) =
2
j
j=9
Uppss!!
Using the usual type I error level = 0.05, the two model assumptions lead to
two different decisions. The problem is that there is not sufficient information
in the problem setting to help us in the selection of the correct model.
Alvaro Mauricio Montenegro Daz
Table of contents
1 Brief History
2 Bayesian-frequentist controversy
3 Some basic Bayesian models
Bayes theorem
The most basic Bayesian model has three stages:
1 Likelihood specification Y f (y |).
2 Prior specification p(), where Y or can be vectors. p() can be
(|y ) =
where
f (y |)p()
,
m(y )
Z
m(y ) =
f (y |)p()d
2 known.
In this case both the prior and the likelihood (the observational model) are
Gaussian distributions, namely,
N(, 2 )
Y | N(, 2 ),
thus, the marginal distribution of Y is N(, 2 + 2 ), and the posterior
distribution of is Gaussian with mean and variance given by
E [|Y ] = B + (1 B)Y
Var [|Y ] = (1 B) 2 ,
where
B = 2 /( 2 + 2 ).
Alvaro Mauricio Montenegro Daz
A beta/binomial model
Y Bin(n, )
This is a model which could be used in the problem of test of hypothesis
presented above. After observe Y , the number of success in n independent
trials, the sampling distribution (observational model) is given by
!
n y
P(Y = y |) = f (y |) =
(1 )1y
y
To obtain a closed form for the marginal distribution, we use the Beta(a, b)
prior distribution. That is
P() =
1
a1 (1 )b1 ,
B(a, b)
A beta/binomial model II
Y Bin(n, )
The prior variance given by (1 )/(M + 1) is a decreasing function of of M.
The marginal distribution of Y is called beta-binomial, and can be shown that
Y
E
=
n
(1 )
Y
n1
Var
=
1+
.
n
n
M +1
The posterior distribution of is again a Beta distribution with
M
n
Y
+
M +n
M +n n
h
i
)/(M
Var [|Y ] = (1
+ n) .
= E [|Y ] =
Poisson distribution
Yi Po(),
i = 1...,n
x 1 e x
,
()
n
Y
#
P
e yi 1 e = e (n+)
yi +1
i=1
Yi Po(),
i = 1...,n
Then,
p(|y) = G
n
X
!
yi + , n +
i=1
i=1 Yi +
i=1 yi
b=
+
.
=
n+
n+
n
n+