Professional Documents
Culture Documents
Abstract
The exponential-logarithmic is a new lifetime distribution with decreasing fail-
ure rate and interesting applications in the biological and engineering sciences.
Assuming different noninformative prior distributions for the parameters of the
distribution, we introduce a Bayesian analysis using MCMC (Markov Chain
Monte Carlo) methods. In this paper distinct noniformative priors, such as Jef-
freys, reference, MDIP (maximal data information prior) and independent gam-
mas, are derived and compared for the Bayesian inference of the two-parameters
exponential-logarithmic distribution.
then consider noninformative and weakly informative priors in this family.
We use an example to illustrate serious problems with the inverse-gamma family
of “noninformative” prior distributions. We suggest instead to use a uniform
prior on the hierarchical standard deviation, using the half-t family when the
number of groups is small and in other settings where a weakly informative prior
is desired. We also illustrate the use of the half-t family for hierarchical modeling
of multiple variance parameters such as arise in the analysis of variance. Various
noninformative prior distributions have been suggested for scale parameters in
hierarchical models.
A maximum cross-entropy method for determining the reference prior distri-
bution, which represents partial information or complete ignorance, is proposed
and studied in this paper, several properties and interpretations are gi
Much criticism of Bayesian analysis concerns the fact that the result of the
analysis depends on the choice of prior, and that the assignment of this prior
seems rather subjective. Is there some objective way of assigning a prior in the
case that we know little about its possible distribution?
The fundamental limitation of Bayesian statistics lies in the selection of a
suitable expression for the prior probability; in our example the prior probability
of various maps of the radio sky before are given any dat
∗ Department of Statistics, State University of Sao Paulo, Brazil, Email:
femoala@fct.com.br
† Department of Statistics, State University of Sao Paulo, Brazil, Email: ???
1
The problem of determining appropriate prior distributions corresponding
to such ignorance has led to considerable controversy since the time of Laplace
and there is, as yet, no entirely satisfactory way of assigning priors, even in
the simplest situations. Therefore, at this stage, we suggest that it is most
useful to understand the properties of the prior distributions that arise from
the various entropy expressions and to collect arguments that lead to different
priors, without attempting to decide which, if any, are correct
A simulation study is also carried out to compare the performance of the
prior distributions presented in this paper.
Bayesian phylogenetic methods require the selection of prior probability dis-
tributions for all parameters of the model of evolution. These distributions
allow one to incorporate prior information into a Bayesian analysis, but even
in the absence of meaningful prior information, a prior distribution must be
chosen. In such situations, researchers typically seek to choose a prior that
will have little effect on the posterior estimates produced by an analysis, al-
lowing the data to dominate. Sometimes a prior that is uniform (assigning
equal prior probability density to all points within some range) is chosen for
this purpose. In reality, the appropriate prior depends on the parameterization
chosen for the model of evolution, a choice that is largely arbitrary. There is
an extensive Bayesian literature on appropriate prior choice, and it has long
been appreciated that there are parameterizations for which uniform priors can
have a strong influence on posterior estimates. We here discuss the relationship
between model parameterization and prior specification, using the general time-
reversible model of nucleotide evolution as an example. We present Bayesian
analyses of 10 simulated data sets obtained using a variety of prior distributions
and parameterizations of the general time-reversible model. Uniform priors can
produce biased parameter estimates under realistic conditions, and a variety of
alternative priors avoid this bias.
For a Bayesian analysis of the EL(p, β) distribution, we assume different
prior distributions for p and β. In a situation which we have no prior informa-
tion, we want a prior with minimal influence on the inference. There is in the
Bayesian literature several forms of formulating noninformative priors, for ex-
ample, Jeffreys prior (1967), reference prior (Bernardo (1979), Berger&Bernardo
(1991)), MDIP prior (Zellner, 1977, 1984, 1990) and others.
In this work, we first computethe Bayes estimates and construct the credible
intervals of á and ë with respect to different priors, and compare these with
the classical maximum likelihood estimators (MLEs) and with the confidence
intervals based on the asymptotic distributions of the MLEs.
The problem of assigning probability distributions which reflect the prior in-
formation available about experiments is one of the major stumbling blocks in
the use of Bayesian methods of data analysis. In this paper the method of max-
imum (relative) entropy (ME) is used to translate the information contained in
the known form of the likelihood into a prior distribution for Bayesian inference.
2
1 Introduction
p=0.3, β=1
p=0.5, β=2
p=0.9, β=3
2.0
1.5
f(x)
1.0
0.5
0.0
3
ln 1 − (1 − p)e−βt
F (t; p, β) = 1 − (1)
ln p
and hence the survival and hazard function associated to (1) are given, respec-
tively, by,
ln 1 − (1 − p)e−βt
S(t; p, β) = P {T > t} = (2)
ln p
and
−β(1 − p)e−βt
h(t; p, β) = , (3)
1 − (1 − p)e−βt ln 1 − (1 − p)e−βt
In Figure 2, we have the plots of the survival function (3) and hazard function
(4) assuming different values of p and β.
1.0
p=0.3, β=1
0.8
p=0.5, β=2
p=0.8, β=3
0.6
S(x)
0.4
0.2
0.0
x
1.4
p=0.2, β=1
p=0.5, β=1
p=0.8, β=1
1.3
h(x)
1.2
1.1
1.0
0 1 2 3 4 5
4
has the EL(p, β) distribution with parameters p and β.
Several other properties, such as mean lifetime, moments generating mo-
ments, mode, mean and variance, etc, for the EL(p, β) distribution are derived
in Tahmasbi and Rezaei (2008).
The estimation of the parameters attained by the EM algorithm and their
asymptotic variances and covariances of the proposed distribution are discussed
in Tahmasbi and Rezaei (2008).COPIEI ENTAO MUDAR
Now, suppose we have a complete random sample T1 , . . ., Tn from EL(p, β).
The likelihood function in the parameters p and β, based on T , is then:
Pn
1 n β n (1 − p)n e−β i=1 ti
L(p, β| t) ∝ − Qn −βti
. (5)
i=1 1 − (1 − p)e
ln p
The log-likelihood function based on the observed sample is given by
n
X n
X
log L(p, β| t) = −nlog(−log p)+nlog β+nlog(1−p)−β ti −− log[1−(1−p)e−βti ].
i=1 i=1
(6)
The maximum likelihood estimators (M LE) for the parameters p and β of
EL(p, β) are obtained by the first derivatives of L = log L(p, β | t) given by
n
∂ n n X e−βti
ln L = − − − (7)
∂p p ln p 1−p i=1
1 − (1 − p)e−βti
and
n n
∂ n X X (1 − p)ti e−βti
ln L = − ti − . (8)
∂β β i=1 i=1
1 − (1 − p)e−βti
∂
Because it is not easy to solve the likelihood equations ∂p log L = 0 and
∂
∂β log L = 0, directly, we need to use an iterative approach to find the M LE for
p and β.
To achieve estimations via MLE approach, Tahmasbi and Rezaei (2008)
derive the conditions for the existence and uniqueness of the MLE for p and β
when the other parameter is given or known.
Tahmasbi and Rezaei (2008) also shown in detail that for the EL(p, β)
distribution the Fisher information matrix is given by,
(1−4p)/p2 −2 ln p+3
" #
− (plnlnp+1
p)2 +
1
(1−p)2 + 2(1−p)2 ln p
1−p+p ln p
2βp(1−p)ln p
I(p, β) = n , (9)
1−p+p ln p
2βp(1−p)ln p − dilog(p)
β 2 ln p
R x ln x
where dilog(·) is the dilogarithm function definided as dilog(x)= 1 1−x dx.
Thus, the maximum likelihood estimators pb and β have asymptotic joint
b
normal distribution given by
5
2 Bayesian inference and Jeffreys prior
6
ln p + 1 1 (1 − 4p)/p2 − 2 ln p + 3 dilog(p) 1 − p + p ln p 2
h(p) = − − − .(16)
(p ln p) (1 − p)2
2 2(1 − p)2 ln p ln p 2p(1 − p)ln p
(13)
It is interesting note that this prior provides an independence prior although
its dependence structure.
Other common specification of noninformative prior considered in the liter-
ature is given by the product of independent Gamma distributions,
7
by E and (b) has the maximum entropy, relative to a reference measure in the
continuous case:
It is of interesting that the data gives more information about the param-
eter than the information on the prior density, otherwise, there would not be
justification for the realization of the experiment. Thus, we wish a prior dis-
tribution π(φ) that provides the gain in the information supplied by the data
the largest as possible relative to the prior information of the parameter, that
is, maximizes the information on the data. With this idea Zellner (1977, 1984,
1991) and Zellner and Min (1992) derived a prior which maximize the average
information in the data density relative to that one in the prior.
Shannon defines entropy as
Z b
H(φ) = f (x | φ)lnf (X | φ)dx(11) (16)
a
8
Theorem 1: The MDIP for the parameters p and β of exponential-logarithmic
distribution (1) is given by:
n o
1 β(1 − p)exp − polylog(2,1−p)
ln p
πZ (p, β) ∝ − √ .(19) (19)
ln p p
P∞ k
where polylog(z, α) = k=1 kzα .
proof: Firstly, we have to evaluate the measure information,
∞
1 β(1 − p)e−βt 1 β(1 − p)e−βt
Z
H(p, β) = − ln − dt(20) (20)
0 ln p 1 − (1 − p)e−βt ln p 1 − (1 − p)e−βt
1 h i
H(p, β) = ln(− ) + ln β + ln(1 − p) − βE(T ) − E ln 1 − (1 − p)e−βT .(21)
ln p
(21)
polylog(2,1−p)
Since the mean of the EL(p, β) distribution is E(T ) = β ln p and
h i
−βT 1
E ln 1 − (1 − p)e = 2 ln p , we have,
1 polylog(2, 1 − p) 1
H(p, β) = ln(− ) + ln β + ln(1 − p) − − lnp.(22) (22)
ln p ln p 2
Now, using H(p, β) above in equation (18), the resulting MDIP for the
parameters p and β will be given by:
n o
n o 1 β(1 − p)exp − polylog(2,1−p)
ln p
πZ (p, β) ∝ exp H(p, β) ∝ − √ (23)
ln p p
Corollary: On combining the likelihood (6) and prior (19), the joint posterior
distribution for the parameters p and β is
n Pn o
n+1 n+1 polylog(2,1−p)
1 n+1 β (1 − p) exp − ln p + β i=1 t i
p(p, β| t) ∝ − √ Qn .(23)
ln p p i=1 1 − (1 − p)e−βti
(24)
4 Refence prior
9
(1992). Reference prior requires the knowledgment of which parameter is nui-
sance or interest.
The idea is to derive a prior π(φ) that maximizes the expected posterior
information about the parameters provided by independent replications of an
experiment relative to the information in the prior. A natural measure of the
expected information about φ provided by data x is given by
p(φ | x)
Z
K(p(φ|x), π(φ)) = p(φ|x)log dφ(37) (26)
Φ π(φ)
is the Kullback-Leibler distance. So, the reference prior is defined as the prior
π(φ) that maximizes the expected Kullback-Leibler distance between the pos-
terior distribution and the prior distribution π(φ), taken over the experimental
data.
The prior density π(φ) which maximizes the functional (36) is found through
calculus of variation and, the solution is not explicit. However, when the pos-
terior p(φ | x) is asymptotically normal, this approach leads to Jeffreys prior
for a single parameter situation. If on the other hand, we are interested in one
of the parameters, being the remaining parameters nuisances, the situation is
quite different, and the appropriated reference prior is not a multivariate Jef-
freys prior. Bernardo argues that when nuisance parameters are present the
reference prior should depend on which parameter(s) are considered to be of
primary interest. The reference prior in this case is derived as follows. We will
present here the two-parameters case in details. For the multiparameter case,
see Berger and Bernardo (1992).
Let θ = (θ1 , θ2 ) be the whole parameter, θ1 being the parameter of interest
and θ2 the nuisance parameter. The algorithm is as follows:
Step 1: Determine π2 (θ2 | θ1 ), the conditional reference prior for θ2 assuming
that θ1 is given,
p
π2 (θ2 | θ1 ) = I22 (θ1 , θ2 )(38) (27)
where I22 (θ1 , θ2 ) is the (2,2)-entry of the Fisher Information Matrix.
Step 2: Normalize π2 (θ2 | θ1 ).
Case π2 (θ2 | θ1 ) is improper, choose a sequence of subsets Ω1 ⊆ Ω2 ⊆ . . . Ω
on which π2 (θ2 | θ1 ) is proper. Define
1
cm (θ1 ) = R (39) (28)
Ωm
π 2 (θ2 | θ1 )dθ2
10
n 1 Z
det I(θ1 , θ2 )
o
πm (θ1 ) ∝ exp pm (θ2 | θ1 )log
dθ2 (41) (30)
2 Ωm I22 (θ1 , θ2 )
Step 4: Compute the reference prior for (θ1 , θ2 ) when θ2 is a nuisance pa-
rameter:
cm (θ1 )πm (θ1 )
π(θ1 , θ2 ) = li m π(θ2 | θ1 )(42) (31)
m→∞ cm (θ ∗ ) πm (θ ∗ )
1 1
where θ1∗ is any fixed point in with positive density for all πm .
*************FAZER UM LINK AQUI**********************
Let f (x | θ1 , θ2 ), (θ1 , θ2 )∈ Φ × Λ ⊆ <2 be a probability model with two
real-valued parameters θ1 and θ2 , where θ1 is the quantity of interest and sup-
pose that the joint posterior distribution of (θ1 , θ2 ) is asymptotically normal
with covariance matrix S(θ1 , θ2 ) with S(θ1 , θ2 )= I −1 (θ1 , θ2 ) where S11 (θ1 , θ2 )=
I22 (θ1 , θ2 )
det I(θ1 , θ2 ) .
If the nuisance parameter space Λ( θ1 )= Λ is independent of θ1 , and the
−1/2 1/2
functions S11 (θ1 , θ2 ) and I22 (θ1 , θ2 ) factorize in the form
−1/2 1/2
S11 (θ1 , θ2 ) = f1 (θ1 )g1 (θ2 ) , I22 (θ1 , θ2 ) = f2 (θ1 )g2 (θ2 )(24) (32)
then,
11
where Ipp is given from (12) by
ln p + 1 1 (1 − 4p)/p2 − 2 ln p − 3
Ipp = − 2
+ 2
+ .(30) (37)
(p ln p) (1 − p) 2(1 − p)2 ln p
proof: a) The inverse of the Fisher matrix given in (12) is,
" dilog(p) #
− h(p) ln p Spβ
Spp Spβ
S(p, β) = = β 2 Ipp , (31) (38)
Spβ Sββ Spβ h(p)
−1
As Spp (p, β) and Iββ (p, β)= −dilog(p)/(β 2 ln p) factorize, we find the refer-
ence prior for the nuisance parameter conditionally on the parameter of interest
given by
1
π(β|p) ∝ , (32) (39)
β
and the marginal reference prior of the parameter of interest p as
1
π(p) ∝ p , (33) (40)
Spp
The joint reference prior needed to obtain a reference posterior for the pa-
rameter of interest p and β is given by
1
πp (p, β) = π(β|p)π(p) = p . (34) (41)
β Spp
b) In a similar way, we obtain the reference prior considering β as the pa-
rameter of interest and p the nuisance parameter.
12
projetista a respeito das caractersticas dos componentes pode ser representado,
ento este conhecimento adicional (a priori) pode ser utilizado dentro do sistema
Bayesiano de inferncia. Contudo, o conhecimento a priori no frequentemente
considerado nas anlises Bayesianas devido dificuldade em represent-lo ou ale-
gando que pela grande quantidade de dados o conhecimento a priori tende a
ter pouco efeito nas inferncias finais ou ainda devido s vrias tcnicas disponveis
para utilizar prioris no-informativas. Porm, ao se investigar fenmenos novos
ou raros, os poucos dados podem no ser suficientes para se tomar uma deciso
estatstica. Como exemplos tm-se os modelos dos danos causados por terremo-
tos, o estudo de doenas raras, etc. Nem sempre haver dados suficientes para
que se possa ignorar o conhecimento a priori, e consequentemente, a opinio do
perito pode ser uma das poucas fontes de informao. Alm disso, h casos em
que torna-se difcil obter concluses sobre certos parmetros em modelos estat-
sticos, mesmo com uma quantia razovel de dados. Quando se trabalha com
dados reais, alm de poucos dados, pode-se encontrar tambm dados censurados,
o que justifica ainda mais a utilizao da Inferncia Bayesiana. Tal contexto en-
contrado principalmente na indstria, onde no possvel obter muitas amostras
devido ao encarecimento da pesquisa ou pela demora em ocorrer uma falha.
Quando isso acontece, possvel acelerar o processo, o que pode levar a dados
no condizentes com a realidade. Neste caso, todas as fontes de conhecimento
como a opinio de especialistas devem ser levadas em conta. Vrios autores tm
buscado o estudo da elicitao da priori (), que formaliza o conhecimento do es-
pecialista. Na prtica, numa tomada de deciso, a escolha de uma priori deve-se
considerar o seu prtico manuseio (recursos explcitos e fcil amostragem), que
de grande importncia para estudos de sensibilidade. Definies estatsticas como
matriz de Fisher inversa so, na maioria das vezes, demasiadamente tcnicas e
inacessveis. A principal ideia de elicitao vem de uma viso simples da opinio de
um especialista. possvel considerar que a opinio de um especialista perfeito
deve ser, a grosso modo, semelhante a um levantamento de dados reais, e deve
proporcionar amostras independentes e identicamente distribudas (i.i.d.) dos
dados. Porm, impraticvel esperar que um especialista seja capaz de estabele-
cer valores da densidade a priori. Sabe-se que no fcil responder a perguntas
como: ”qual a probabilidade da varivel de interesse assumir tal valor?”. O
mais razovel seria pedir valores vinculados probabilidades, como P(X ¿ x), ou
seja, o mais conveniente perguntar ”qual o valor que a varivel poderia assumir
para uma certa probabilidade?”. Alm disso, estudos indicam que os momentos
de segunda ordem no so recomendados para serem elicitados. A recomendao
sobrevm da dificuldade de especialistas com conhecimento em conceitos estat-
sticos (ou at mesmo estatsticos) fornecerem medidas de variabilidade precisas
(como varincia e desvio-padro). Em geral, a varincia subestimada. A escolha
mais comum obter quantis, como a mediana ou os quartis, pois so mais fceis
de explicar aos especialistas. Da mesma forma, geralmente intervalos de credi-
bilidade so bem entendidos e avaliados com bastante preciso, desde que no seja
exigido do especialista o fornecimento de intervalos com alta probabilidade de
cobertura (por exemplo 95a elicitao de quantis permite a avaliao adequada da
varincia da distribuio a priori do especialista, em vez de elicit-la diretamente.
13
OHagan (1998) d um exemplo do uso de percentis elicitados para estimao da
varincia. Portanto, o melhor estabelecer probabilidades de fcil entendimento
para leigos em estatstica e pedir a eles que proponham valores correspondentes
essas porcentagens. Desta maneira a informao fornecida pelos especialistas ser
obtida de forma mais fiel realidade.
onde σ = h(φ̂)−1/2
[?] argumentam que:
I = Iˆ · 1 + O(n−1 ) (44)
2 1/2
onde Iˆ = 2πσ
n f (φ̂)exp{−nh(φ̂)}.
Para o caso m-dimensional,
Z ∞ m/2
2π
f (φ)exp{−nh(φ)}dφ ∼
= det(Σ)1/2 f (φ̂)exp{−nh(φ̂)} (45)
0 n
14
6 Elicita??o da confiabilidade
No caso da Elicita??o da confiabilidade R(t), o especialista indica um valor
condicionado a uma probabilidade especificada. Ou seja, dada uma probabil-
idade p, pergunta-se ao especialista qual o valor que ele acredita que p% dos
produtos est?o acima desse valor.
Portanto, k valores t1 , t2 , ..., tk s?o escolhidos e o especialista fornece suas
probabilidades Rt1 , Rt2 , ..., Rtk com
√
2πΓ(a + b)dc β̂ c (1 − e−û )b
1 1
=− p exp − a − û + dβ̂ ln[1−(1−e−û )e−β̂t ]
Γ(a)Γ(b)Γ(c) (b − 1)(c − 1) 2 û
(47)
Γ(a + b) a−1 d
π(p|a, b) = p (1 − p)b−1 e π(β|c, d) = (dβ)c−1 e−dβ .
Γ(a)Γ(b) π(c)
Sabe-se que:
Z 1 Z ∞
S(t) = P (T > t|p, β)π(p, β)dpdβ. (48)
0 0
Supondo p e β independentes, tem-se que π(p, β) = π(p)π(β), ou seja,
15
Γ(a + b)dc a−1
π(p, β) = p (1 − p)b−1 β c−1 e−dβ . (49)
Γ(a)Γ(b)Γ(c)
Portanto, de (??) tem-se:
1 ∞
Γ(a + b)dc ln[1 − (1 − p)e−βt ] a−1
Z Z
S(t) = p (1 − p)b−1 β c−1 e−dβ dpdβ.
Γ(a)Γ(b)Γ(c) 0 0 ln p
(50)
? f?cil notar que a integral em (??) n?o possui solu??o anal?tica. Neste
caso, a aproxima??o de Laplace torna-se muito ?til pois ? um m?todo eficaz
para aproxima??o de integrais complicadas.
Para que se possa utilizar o M?todo de Laplace em (??), ? preciso fazer uma
transforma??o em p para se obter limites 0 a ∞. Seja ent?o a transforma??o
u = − ln p tal que p → 0 implica em u → ∞, p → 1 implica em u → 0, p = e−u
e dp = −e−u du.
Assim,
1
ln[1 − (1 − p)e−βt ] a−1
Z
p (1 − p)b−1 dp
0 ln p
∞
ln[1 − (1 − e−u )e−βt ] −(a−1)u
Z
=− e (1 − e−u )b−1 e−u du.
0 u
Portanto,
1 ∞
ln[1 − (1 − p)e−βt ] a−1
Z Z
p (1 − p)b−1 β c−1 e−dβ dpdβ
0 0 ln p
∞ ∞
ln[1 − (1 − e−u )e−βt ]
Z Z
=− (1 − e−u )b−1 e−au β c−1 e−dβ dudβ.
0 0 u
−u
)e−βt ]
Sejam f (u, β) = − ln[1−(1−eu e n = 1 pois n?o h? dados em X. Ent?o,
16
Observa-se que β̂ ? a moda da distribui??o Gama.
∂2h ∂2h ∂2h
Agora ? preciso calcular as derivadas segundas ∂u 2, ∂β 2 e ∂u∂β para obter
−1
a matriz Σ , que ? a Matriz Hessiana dada a seguir.
2 2
" #
∂ h ∂ h
Σ−1 = ∂u2
∂2h
∂u∂β
∂2h
. (51)
∂u∂β ∂β 2
1
Nota-se que det(Σ) = det(Σ −1 )
Al?m disso,
−1
(b − 1)(c − 1)e−u β 2 (1 − e−u )2
1
det(Σ) = = = .
det(Σ−1 ) β 2 (1 − e−u )2 (b − 1)(c − 1)e−u
√
Γ(a + b)dc 2π c−1 −dβ̂ −aû β̂(1 − e−û ) û 1
S(t|θ) = − β̂ e e (1−e−û )b−1 p e 2 ln[1−(1−e−û )e−β̂t ]
Γ(a)Γ(b)Γ(c) (b − 1)(c − 1) û
(53)
√
2πΓ(a + b)dc β̂ c (1 − e−û )b
1 1
S(t|θ) = − p exp − a − û + dβ̂ ln[1−(1−e−û )e−β̂t ]
Γ(a)Γ(b)Γ(c) (b − 1)(c − 1) 2 û
(54)
17
• Se forem considerados desconhecidos os quatro hiperpar?metros a, b, c, d,
ser? necess?rio quatro valores S(t) dado quatro valores t.
• Pode-se considerar desconhecidos apenas ”a” e ”b”, sendo c e d conhecidos.
Assim, ser? preciso apenas dois diferentes valores para S(t) dados t1 e t2 .
√
2πΓ(a + b)dc β̂ c (1 − e−û )b
1
S(t|θ) = − p exp − a − û + dβ̂ ·
Γ(a)Γ(b)Γ(c) (b − 1)(c − 1) 2
1
· ln[1 − (1 − e−û )e−β̂t ],
û
a+b−1 c−1
onde û = ln 2 e β̂ = d , de modo que a + b > 1 e c > 1.
7 Simulation study
?????????????
In this section, let us consider the data set related to the lifetime of a type
of electrical insulator subject to a constant voltage stress, introduced in Lawless
(1982). The data are not censored and represent the lifetime (in minutes) to
fail: 0.96, 4.15, 0.19, 0.78, 8.01, 31.75, 7.35, 6.50, 8.27, 33.91, 32.52, 3.16, 4.85,
2.78, 4.67, 1.31, 12.06, 36.71, 72.89. Let us denote this data as “Lawless data”.
Lawless used this data set to illustrate the fit of the model Weibull lifetime
data and we assume a exponential-logarithmic distribution with density (1) to
analyse the data.
18
To determine the appropriate distribution for fitting this data, the Weibull
proposed by Lawless or Exponential-Logarithmic, various selection criteria were
examined. These include information-based criteria (AIC, BIC, and DIC) given
in the Table 5 for each distribution. From this table, we conclude that Exponential-
Logarithmic gives better fit (smaller value for AIC and BIC) while by considering
the DIC criterion it seems that both Exponential-Logarithmic and Weibull dis-
tributions are appropriate for the data. However, if one has to choose preferably
one of them based on DIC, the chosen will also be the Exponential-Logarithmic
(smaller value for DIC).
Table 1: Information-based Model Selection Criteria (AIC, BIC and DIC) for
Lawless data
19
Table 3: Posterior summaries for β (dados de Lawless).
Prioris M?dia Desvio-Padr?o Intervalo 95%
Beta/Gama 0.0483 0.0187 (0.0165, 0.0873)
Uniforme/Gama 0.0513 0.0177 (0.0195, 0.0879)
Jeffreys 0.0396 0.0182 ( 0.0259, 0.0796)
MDIP/Gama 0.0514 0.0182 (0.0389,0.0916)
Refp 0.0366 0.0171 (0.0242,0.0749)
20
Figure 3: Marginal posterior densities of parameters p and β (Lawless data)
9 Conclusions
????????????? In this paper, noninformative priors are derived for the pa-
rameters α and λ of the generalized exponential distribution. A fully Bayesian
analysis was developed and a simulation study was performed to determinate
which prior provides less information, mainly for small sample size.
We shown that maximal data information prior causes a improper posterior
distribution for parameters α and λ. We also shown that reference prior pro-
cedure for generalized exponencial distribution provides nonuniqueness of prior
due to the choice of the parameter of interest.
The results obtained through Bayesian summaries and credible intervals are
similars for all the priors studied here. There is a tendency of the measures of
performance of the estimators and intervals get better as the sample size in-
creases. In addition, a comparison of frequentist coverage probabilities among
the priors shows that they are very good but the reference prior slightly domi-
nates the others when n is small.
21
Thus, these results are of great interest in applications of this distribution.
It is observed by Tahmasbi and Rezaei (2008) that the EL(p, β) distribution
can be used quite erectively in analyzing several lifetime data, particularly in
place of two-parameter gamma or two-parameter Weibull distribution
10 References
?????????????
Berger, J. O. ; Bernardo, J. M. (1992). On the development of reference
priors. Bayesian Statistics 4 (J. M. Bernardo, J. O. Berger, A. P. Dawid and A.
F. M. Smith, eds.) Oxford: University Press, 35–60 (with discussion).
Bernardo, J. M. (1979). Reference Posterior Distributions for Bayesian In-
ference. Journal Royal Statistical Society, v.41, n.2, p. 113-147.
Box, G.E.P., and Tiao, G.C. (1973). Bayesian inference in statistical analy-
sis. Addison Weiley.
Chib, S. ; Greenberg, E. (1995). Understanding The Metropolis-Hasting
Algorithm. The American Statistician, Vol. 49, No 4.
Gelfand, A. E.; Smith, F. M. (1990). Sampling-based approaches to calcu-
lating marginal densities. Journal of the American Statistical Association, n.
85, p. 398-409.
Jeffreys, Sir Harold (1967). Theory of probability. 3rd rev. ed., London:
Oxford U. Press.
Sinha, B. and Zellner, A. (1990), A note on the prior distributions of Weibull
parameters, SCIMA, 19, 5-13.
Tahmasbi, R., Rezaei, S., (2008). “A two-parameter lifetime distribution
with decreasing failure rate”, Computational Statistics and Data Analysis, Vol.
52, pp. 3889-3901.
Zellner, A. (1977). Maximal Data Information Prior Distributions, In New
Methods in the applications of Bayesian Methods, A. Aykac e C. Brumat (eds).
North-Holland, Amsterdam.
Zellner, A. (1984). Maximal Data Information Prior Distributions, Basic
Issues in Econometrics, U. of Chicago Press.
Zellner, A. (1990). Bayesian Methods and Entropy in Economics and Econo-
metrics. In W.T. Grandy, Jr. and L.H. Schick (eds.), Maximum Entropy and
Bayesian Methods, Dordrecht, Netherlands: Kluwer Academic Publishers, p.
17-31.
22