Title: Maximum Likelihood Estimation
1Maximum Likelihood Estimation
- Multivariate Normal distribution
2The Method of Maximum Likelihood
- Suppose that the data x1, , xn has joint
density function - f(x1, , xn q1, , qp)
- where q (q1, , qp) are unknown parameters
assumed to lie in W (a subset of p-dimensional
space). - We want to estimate the parametersq1, , qp
3Definition The Likelihood function
- Suppose that the data x1, , xn has joint
density function - f(x1, , xn q1, , qp)
- Then given the data the Likelihood function is
defined to be - L(q1, , qp)
- f(x1, , xn q1, , qp)
- Note the domain of L(q1, , qp) is the set W.
4Definition Maximum Likelihood Estimators
- Suppose that the data x1, , xn has joint
density function - f(x1, , xn q1, , qp)
- Then the Likelihood function is defined to be
- L(q1, , qp)
- f(x1, , xn q1, , qp)
- and the Maximum Likelihood estimators of the
parameters q1, , qp are the values that
maximize - L(q1, , qp)
5- i.e. the Maximum Likelihood estimators of the
parameters q1, , qp are the values
Such that
Note
is equivalent to maximizing
the log-likelihood function
6The Multivariate Normal Distribution
- Maximum Likelihood Estiamtion
7denote a sample (independent)
Let
from the p-variate normal distribution
with mean vector
and covariance matrix
Note
8The matrix
is called the data matrix.
9The vector
is called the data vector.
10The mean vector
11The vector
is called the sample mean vector
note
12also
13In terms of the data vector
where
14Graphical representation of sample mean vector
The sample mean vector is the centroid of the
data vectors.
15The Sample Covariance matrix
16The sample covariance matrix
where
17There are different ways of representing sample
covariance matrix
18(No Transcript)
19(No Transcript)
20(No Transcript)
21Maximum Likelihood Estimation
- Multivariate Normal distribution
22denote a sample (independent)
Let
from the p-variate normal distribution
with mean vector
and covariance matrix
Then the joint density function of
is
23The Likelihood function is
and the Log-likelihood function is
24To find the Maximum Likelihood estimators of
we need to find
to maximize
or equivalently maximize
25Note
thus
hence
26Now
27Now
28Summary the Maximum Likelihood estimators of
are
and
29Sampling distribution of the MLEs
30Note
is
The joint density function of
31This distribution is np-variate normal with mean
vector
32Thus the distribution of
is p-variate normal with mean vector
33(No Transcript)
34Summary
- The sampling distribution of
- is p-variate normal with
35The sampling distribution of the sample
covariance matrix S and
36The Wishart distribution
- A multivariate generalization of the c2
distribution
37Definition the p-variate Wishart distribution
be k independent random p-vectors
Each having a p-variate normal distribution with
Then U is said to have the p-variate Wishart
distribution with k degrees of freedom
38The density ot the p-variate Wishart distribution
Then the joint density of U is
where Gp() is the multivariate gamma function.
It can be easily checked that when p 1 and S
1 then the Wishart distribution becomes the c2
distribution with k degrees of freedom.
39Theorem
then
Corollary 1
Corollary 2
Proof
40Theorem
are independent, then
Theorem
are independent and
Suppose
then
41Theorem
Let
be a sample from
then
Theorem
Let
be a sample from
then
42Theorem
Proof
etc
43Theorem
Let
be a sample from
then
is independent of
Proof
be orthogonal
Then
44Note H is also orthogonal
45Properties of Kronecker-product
46(No Transcript)
47(No Transcript)
48This the distribution of
is np-variate normal with mean vector
49Thus the joint distribution of
is np-variate normal with mean vector
50Thus the joint distribution of
is np-variate normal with mean vector
51Summary Sampling distribution of MLEs for
multivatiate Normal distribution
Let
be a sample from
then
and