Professional Documents
Culture Documents
Statistical Inference
Goal: The goal in statistical inference is to make conclusions about a
phenomenon based on observed data.
Frequentist: Observations made in the past are analyzed with a specified
model. Result is regarded as confidence about state of real world.
Probabilities defined as frequencies with which an event occurs if
experiment is repeated several times.
Parameter Estimation:
o Relies on estimators derived from different data sets and a specific
sampling distribution.
o Parameters may be unknown but are fixed
Bayesian: Interpretation of probability is subjective and can be updated with
new data.
Parameter Estimation: Parameters described as density
Bayesian Inference
Framework:
Prior Distribution: Quantifies prior knowledge of parameter values.
Likelihood: Probability of observing a data if we have a certain set of
parameter values.
Posterior Distribution: Conditional probability distribution of unknown
parameters given observed data.
Joint PDF: Quantifies all combination of data and observations
Bayesian Inference
Uninformative Prior: No a priori information parameters
Informative Prior: Use conjugate priors; prior and posterior from same
distribution
Evaluation Strategies:
Analytic integration --- Rare
Classical quadrature; e.g., p = 2
Monte Carlo quadrature Techniques
Markov Chains
Bayesian Inference
Example:
Bayesian Inference
Example:
1 Head, 0 Tails
Note:
5 Heads, 9 Tails
49 Heads, 51 Tails
Bayesian Inference
Example: Now consider
5 Heads, 5 Tails
50 Heads, 50 Tails
Note: Poor informative prior incorrectly influences results for a long time.
Assumption:
Note:
Sample Distribution
Posterior Distribution:
Markov Chains
Definition:
Markov Chains
Example:
Chapman-Kolmogorov Equations:
rain
sun
Question:
rain
sun
p1
p2
Periodicity: A Markov chain is periodic if parts of the state space are visited at
regular intervals. The period k is defined as
Stationary Distribution
Theorem: A finite, homogeneous Markov chain that is irreducible and aperiodic
has a unique stationary distribution
and the .chain will converge in the sense of
distributions from any initial distribution .
Recurrence (Persistence):
Matrix Theory
Definition:
Lemma:
Example:
Matrix Theory
Theorem (Perron-Frobenius):
Corollary 1:
Proposition:
Stationary Distribution
Corollary:
Proof:
Convergence: Express
Stationary Distribution
Note:
In Markov chain theory, we are given a Markov chain, P, and we
construct its equilibrium distribution.
In MCMC theory, we are given a distribution and we want to construct
a Markov chain that is reversible with respect to it.
Metropolis Algorithm
Metropolis Algorithm: [Metropolis and Ulam, 1949]
Metropolis-Hastings Algorithm
Metropolis-Hastings Algorithm:
Examples:
Proposal Distribution
Proposal Distribution: Significantly affects mixing
Too wide: Too many points rejected and chain stays still for long periods;
Too narrow: Acceptance ratio is high but algorithm is slow to explore
parameter space
Ideally, it should have similar shape to posterior (target) distribution.
Problem:
Anisotropic posterior,
isotropic proposal;
Efficiency nonuniform for
different parameters
Result:
Recovers efficiency of
univariate case
Acceptance Probability:
Case i:
Note:
Model:
Estimated Parameters:
Note:
as parameter to be estimated.
Strategy:
Choose prior so that posterior is from same family --- termed conjugate prior.
For normal distribution with unknown variance, conjugate prior is inverse
Gamma distribution
which is equivalent to inverse
Strategy:
Note:
Note:
Related Topics
Note: This is an active research area and there are a number of related topics
Burn in and convergence
Adaptive algorithms
Population Monte Carlo methods
Sequential Monte Carlo methods and particle filters
Gaussian mixture models
Development of metamodels, surrogates and emulators to improve
implementation speeds
References:
A. Solonen, Monte Carlo Methods in Parameter Estimation of Nonlinear Models,
Masters Thesis, 2006.
H. Haario, E. Saksman, J. Tamminen, An adaptive Metropolis algorithm, Bernoulli,
7(2), pp. 223-242, 2001.
C. Andrieu and J. Thomas, A tutorial on adaptive MCMC, Statistics and
Computing, 18, pp. 343-373, 2008.
M. Vihola, Robust adaptive Metropolis algorithm with coerced acceptance rate,
arXiv:1011.4381v2.