Postscript version of this file
STAT 801 Lecture 3
Goals of Today's Lecture:
- Learn properties of independence
- Define Multivariate Normal Dist.
- Do 1 sample normal distribution theory.
Review of end of last time
I defined independent events and then independent random variables:
Def'n: Ai,
independent:
any distinct indices
.
Def'n: Rvs
independent:
for any
.
Theorem:
- 1.
- If X and Y are independent then
FX,Y(x,y) = FX(x)FY(y)
for all x,y
- 2.
- If X and Y are independent and have joint density
fX,Y(x,y) then X and Y have densities, say fX and fY,
and
- 3.
- If X and Y are independent and have marginal densities
fX and fY then (X,Y) has joint density
fX,Y(x,y) given by
- 4.
- If
FX,Y(x,y) = FX(x)FY(y)
for all x,y then X and Y are independent.
- 5.
- If (X,Y) has density f(x,y) and there are functions
g(x) and h(y) such that
f(x,y) = g(x) h(y)
for all (well technically almost all) (x,y) then
X and Y are independent and they each have a density
given by
and
Proof:
- 1.
- Since X and Y are independent so are the events
and ;
hence
- 2.
- Suppose X and Y real valued.
Asst 2: existence of fX,Y implies that of fX and fY(marginal density formula).
Then for any sets A and B
Since
It follows (measure theory) that the quantity in [] is 0
(almost every pair (x,y)).
- 3.
- For any A and B we have
If we define
g(x,y) = fX(x)fY(y) then we have proved that
for
To prove that g is the joint density of (X,Y) we need only
prove that this integral formula is valid for an arbitrary Borel set
C, not just a rectangle
.
This is proved via a
monotone class argument. The collection of sets Cfor which identity holds has closure properties which guarantee that
this collection includes the Borel sets.
- 4.
- Another monotone class argument.
- 5.
-
Take B=R1 to see that
where
.
So c1 g is the density of X.
Since
we see that
so that
.
Similar argument for Y.
Theorem: If
are independent and
Yi =gi(Xi) then
are independent.
Moreover,
and
are
independent.
Conditional probability
Def'n:
P(A|B) = P(AB)/P(B) if
.
Def'n: For discrete X and Y the conditional
probability mass function of Y given X is
For absolutely continuous X
P(X=x) = 0 for all
x. What is P(A| X=x) or
fY|X(y|x)?
Solution: use limit
If, e.g., X,Y have joint density fX,Y then with
we have
Divide top, bottom by ;
let
.
Denom converges to fX(x); numerator converges to
Define conditional cdf of Y given X=x:
Differentiate wrt y to get def'n of
conditional density of Y given X=x:
in words ``conditional = joint/marginal''.
The Multivariate Normal Distribution
Def'n:
iff
Def'n:
iff
(a column vector for later use) with the
Zi independent and each
.
In this case according to our theorem
superscript t denotes matrix transpose.
Def'n:
has a multivariate normal distribution if it
has the same distribution as
for some
,
some
matrix of constants A and
.
If the matrix A is singular then X will not have a density. If A is
invertible then we can derive the multivariate normal density
by the change of variables formula:
So
Now define
and notice that
and
Thus fX is
the
density. Note density is
the same for all A such that
.
This justifies the notation
.
For which vectors
and matrices
is this a density?
Any
but if
then
where y=At x. Inequality strict except for y=0 which
is equivalent to x=0. Thus
is a positive definite symmetric
matrix. Conversely, if
is a positive definite symmetric matrix
then there is a square invertible matrix A such that
so that
there is a
distribution. (A can be found
via the Cholesky decomposition, e.g.)
More generally X has multivariate normal distribution
if it has the same distribution as
(no
restriction that A be non-singular). When A is singular X will not
have a density:
such that
;
X is confined to a hyperplane. Still
true that the distribution of X depends only on the matrix
:
if
AAt = BBt then
and
have the same distribution.
Properties of the MVN distribution
1: All margins are multivariate normal: if
and
then
implies
.
2: All conditionals are normal: the conditional distribution of X1given X2=x2 is
3:
:
affine
transformation of MVN is normal.
Normal samples: Distribution Theory
Theorem:
Suppose
are independent
random
variables.
Then
- 1.
-
(sample mean)and s2 (sample variance) independent.
- 2.
-
- 3.
-
- 4.
-
Proof: Let
.
Then
are
independent N(0,1) so
is multivariate
standard normal. Note that
and
Thus
and
where
.
So: reduced to
and .
Step 1: Define
(So Y has same dimension as Z.) Now
or letting M denote the matrix
It follows that
so we need
to compute
MMt:
Solve for Z from Y:
Zi = n-1/2Y1+Yi+1for
.
Use the identity
to get
.
So M
invertible:
Now we use the change of variables formula to compute the density
of Y. Let
denote vector whose
entries are
.
Note that
Then
Note: ftn of y1 times a
ftn of
.
Thus
is independent of
.
Since sZ2 is a function of
we see that
and
sZ2 are independent.
Also, density of Y1 is a multiple of the function of y1 in
the factorization above. But factor is standard normal
density so
.
First 2 parts done. Third part is
a homework exercise. Derivation of the density:
Suppose
are independent N(0,1). Define
distribution to be that of
.
Define angles
by
(Spherical co-ordinates in n dimensions. The
values run
from 0 to
except last
from 0 to .)
Derivative formulas:
and
Fix n=3 to clarify the formulas.
Matrix of partial derivatives is
Find determinant by adding
times col 1 to col
j+1 (no change in determinant). Resulting
matrix lower triangular; diagonal entries
,
and
.
We multiply these together to get
(non-negative for all U and ).
For general n we see that every term in the first column contains a factor
U-1/2/2 while
every other entry has a factor U1/2. Multiplying a column in a matrix by c multiplies
the determinant by c so the Jacobian of the transformation is
u(n-1)/2u-1/2/2 times
some function, say h, which depends only on the angles.
Thus the joint density
of
is
To compute the density of U we must do an n-1 dimensional
multiple integral
.
We see that the answer has
the form
for some c which we can evaluate by making
Substitute y=u/2, du=2dy to see that
so that the
density is
Fourth part is consequence of
first 3 parts and def'n of distribution, namely,
if it has same distribution
as
where
,
and Z and U are independent.
Derive density of T in this definition:
Differentiate wrt t by
differentiating inner integral:
by fundamental thm of calculus. Hence
Now I plug in
to get
Substitute
,
,
and
:
or
Richard Lockhart
2000-01-19