STAT 330 Lecture 14
Reading for Today's Lecture: 6.2, 8.1, 8.2
Goals of Today's Lecture:
Today's notes
Two general methods to develop estimators:
This is easier to understand for discrete data.
General framework for MLEs:
If are independent and is the probability mass function for an individual and if we actually observe then the likelihood function is
Example: independent Poisson rvs. So
The likelihood is
It is easier to maximize the logarithm of this:
To do the maximization set the derivative with respect to equal to 0. Get
from which we get
so that is the mle (maximum likelihood estimate) of .
Extension to Continuous Distributions: If are independent and is the probability density of an individual (so that the X's are continuous rvs) then we can interpret the density as follows. Let denote a small positive number. Then
(The equation is the definition of density and the approximation is the meaning of integration over very small intervals -- it expresses the idea that integration is the opposite of differentiation.)
This prompts us to define
and
The MLE of maximizes the Likelihood or equivalently the log likelihood.
Example: a sample from population.
We need to set and to find estimates of and of .
We find:
and
Set to get
so that . Then put this solution in the second equation to get
which produces the solution
Notice that in the denominator there is an n and not n-1. The MLE is not quite the usual estimate of .
Property of MLE
Suppose a different statistician used where as the parameters? What would and be? The log likelihood is now
When you take the derivatives and set them equal to 0 the equation is unchanged and . The derivative with respect to gives the estimate (I leave you to do the algebra)
so that . This is a general principle. If
is a transformation (for some function like say or any other function) of the parameters then the mles satisfy
Power, , Sample Size
Definition: the power function of a hypothesis test procedure is
Recall: Type I error is incorrect rejection. Type 2 error is incorrect acceptance.
Definition: The level of a test is the largest probability of rejection for in the null hypothesis:
Typically if the null hypothesis is like the value of is actually , that is, the edge of the null hypothesis gives the highest risk of incorrect rejection.
Definition: The probability of a type two error is a function defined for by
Thus for not in the Null: