Professional Documents
Culture Documents
Senaka Samarasekera
SSP algorithm
development process
Estimators
Estimator
: A function of DATA (a.k.a a STATISTIC) that will
Cost of an estimator
Bias
: Systemic error (try to avoid this if possible but not at all costs)
Unbiased estimators B = 0
Variance
Minimum variance estimator arg
Mean-square error
Minimum man square estimator MMSE = arg
Find the bias, variance and the that will minimize the
mean square error.
Likelihood function
A pdf parameterized by an unknown parameter
Eg : If noise was Gaussian in the DC measurement case
What would be the case for multiple observations
assuming i.i.d Gaussian noise
A stock of a growing company can be modelled as
where is i.i.d AWGN. What is the likelihood function? Is
this stationary or non stationary?
Assignment 3 question 2
Find the likelihood function of the autoregressive
moving average model
where is a i.i.d zero mean Gaussian noise.
Deterministic vs Random
parameters
Deterministic Parameters : Only one value possible but
we dont know their value. (this lecture)
Random Parameters: An a priori pdf can be defined on
the parameter. This accounts for the prior uncertainty
about parameters. Which means parameters are
assumed to be random.
In the next lecture we will look at Bayesian methods
which will estimate random parameters.
Estimator development
Deterministic parameter case
Least squares estimator
Maximize the Likelihood
Method of moment estimator
No explicit guarantee on the estimator cost optimality. So will have to derive
conditions of optimality for these estimators.
Maximum likelihood
estimators
MLE
Advantages
Most versatile estimator of the lot
Has a nice physical intuition to it
Has a direct way of finding it
Is consistent asymptotically unbiased, and asymptotically optimal (w.r.t to mse)), and
invariant for functional transformations
Therefore for large number of observations, MLE-> MVUE
DC level
DC level example
Let
for be observations from a noisy voltmeter measuring a DC
level
Assignment 3 question 3
A noisy oscillator at a modulator is modeled as
Where is a i.i.d standard Gaussian noise. If the phase
offset of the oscillator is a constant for N samples, find an
estimator using this N samples.
For each iteration k, with the input and data x and the likelihood
Expectation step :F
Compute
Maximization step :
For real
Invariance properties
Looks at the MLE of a transformed parameter
If a measurement equation is given by the likelihood
function , and if is one-to-one, then
Method of moments
method
Let vector of moments of the pdf (likelihood) and be the underlying parameter
vector. Then using the pdf definition we can find the function s.t
If is invertible
Now if we plugin the estimators of the respective moments we get the MOM estimators
The multiple moments needed can be easily found via the moment
generation function
Example
Let the likelihood function be a 2 components Gaussian
mixture with unknown variance and weights. Let both
means be 0.
Where
Find the MOM estimators
Example
Let
Let , and
Then the parameters (after some algebra) can be shown
to be
Minimum variance
unbiased estimators
(MVUE)
MVUE
two observations with the likelihood functions
Assume
MVUE
Some time the variance and the bias equations become
so complicated that direct optimization methods fail.
Some times no function of data will give a minimum
variance for all possible possible parameter values.
Sufficient statistic
All the information about the parameter in the likelihood
function comes through the sufficient statistic.
Note : Raw data it self is a sufficient statistic.
An MVUE should be a function of the sufficient statistic
Minimal sufficient statistic : The smallest of them all ( in
dimensions)
Minimal sufficient statistic is always a function of other
sufficient statistics.
Complete statistics
If the whole parameter set is identifiable using the sufficient statistic
these are called a COMPLETE statistic.
A sufficient statistic is complete iff
How does this condition relate to parameter identifiability?
Note that
Since the only function in the null space of is the function , the space of all
possible functions is spanned by the parametric family .
RaoBlackwell theorem
applied to estimators
Let
be anestimatorof a parameter , then theconditional
expectationof given a sufficient statistic
is always a better estimator of , and is never worse.
From a mean square error perspective this means
LehmannScheff theorem
If a statistic that is
UNBIASED,COMPLETEandSUFFICIENTfor some
parameter , then this statistic has the minimum
expected loss for ANYCONVEXLOSS FUNCTION.
In many practical applications with the squared lossfunction, it has a smaller mean squared error among
any estimators with the sameexpected value.
Hence a unbiased, complete, sufficient statistic is a
MVUE.
Exponential family
Not to be confused with exponential distribution ( it is
also a member in this family when the parameter of
interest is the mean).
This is concerned with how the pdfs are parameterized.
A good number of common pdfs belong to this family
when some of there parameters are known
Eg:
Poisson distribution with unknown mean
Exponential distribution with unknown mean
Gaussian distribution with unknown mean/unknown variance/
both mean and variance unknown
Exponential family
Definition
Example
Univariate Poisson distribution with unknown mean
Recall that the pmf (since is
This can be rearranged to
So
This is a exponential family member of degree 1.
Example
Find the sufficient statistic, natural parameters, log
normalizer, and the carrier measure for the Univariate
Gaussian distribution with unknown mean and variance.
Since
Therefore
Exponential Family
Assignment question
Prove that (a properly normalized) product of arbitrary
exponential family members is also a member of the
exponential family.
Is it the same for mixtures of exponential family
members?
Log-normalizer
Exponential families are characterized by their strictly
convex and differentiable functions F, called lognormalizer or the partition function
Since we have
Log normalizer
It
is also related to the moment generating and cumulant
generating functions of the sufficient statistics.
The moment generating function of the sufficient statistic is
Log- normalizer
Therefore in exponential family we can easily find the
mean and the variance of the test statistic as
Assignment questions
Show that the MLE and MVUE both are efficient in for
natural parameters of the exponential family.
Show that Expectation Maximization method finds the
local maximum of an exponential family joint
distribution (of observed and unobserved data)
statistic.
Proof
We can write this as
Let and
Which is a scaled Laplace transformation of the function.. Therefore
From the uniqueness of the Laplace transform this implies Since , this implies
Weiner Filter
.
In our compact notation this can be written as
If the number of samples are equal to number of coefficients we
can find a that make .
If N > m (which is normally the case) we will try to find that
minimize
Weiner Filter
Weiner filter
Where
Applications of
Weiner filter
Noise cancellation
System identification
Channel
equalization
BLUE
Is
an extension of the Weiner filtering process.
Now we assume a linear estimator
To this to be unbiased
For this to be the case for some matrix .If we assume and then
If the covariance matrix of is , then
and
BLUE
Taking
the gradient of each variance and equating it to zero we
Weighted LSE
What
if we change the error criterion
When would we use this? All data points are not same
Now
and
If we add a probabilistic description to the noise which will enable
us to formulate a likelihood function then W characterize the
spectral characteristic of the noise
Geometrical interpretation
Geometrical interpretation
What happens if Then the projection on the signal
space becomes
Geometrical interpretation
Lets
look at the signal estimate with LSE
where