- This is part of probstat.
Previously we tried to estimate the population means and variances using the sample means and variances. In this section, we shall see the justification why what we did makes sense.
There are many ways to estimate parameters.
Method of moments estimators
- See also wikipedia article.
This is probably the simplest estimators. However, they are often biased (as we shall show in the example).
Definition: For a random variable
,
is called the k-th moment of
. Note that the first moment is the mean
. The variance of a random variable depends on the first and the second moments.
If we want to estimate a parameter
, using the method of moments, we start by writing the parameter as a function of the moments, i.e.,
We then estimate the sample moments
for
. Our estimate
is thus
EX1: We show how to estimate the variance with the method of moments. Recall that the variance
We first estimate the first moment
and the second moment
. The estimator is
Note that the estimate
is biased, because
.
As this example shows, other estimation techniques are usually preferred over the method of moments.
Maximum likelihood estimators
Suppose that we want to estimate parameter
based on observations (or sample)
. If we can find a joint-distribution function
,
that gives the probability that we observe
given a particular value
. The maximum likelihood estimator is
such that
is maximum, i.e.,
.
EX1: maximum likelihood estimator for a Bernoulli parameter
EX2: maximum likelihood estimator for the Poisson mean
Bayes estimators