next up previous

STAT 330 Lecture 21

Reading for Today's Lecture: 10.1.

Goals of Today's Lecture:

Today's notes

Geometry

Write the data, tex2html_wrap_inline226 out as a big vector

displaymath228

Now we write

eqnarray26

Writing the equation for a particular component of this vector we have

displaymath230

Here are the Geometric facts:

  1. Y, A, T, and R are vectors in Euclidean space of tex2html_wrap_inline240 dimensions.
  2. A, T and R are all perpendicular.
  3. The ANOVA identity is

    displaymath248

    The squared lengths of these 4 vectors are called the Total Sum of squares ( tex2html_wrap_inline250 ), the Sum of Squares due to the Grand Mean ( tex2html_wrap_inline252 ), the Sum of Squares due to Treatment ( tex2html_wrap_inline254 ) and the Sum of Squares due to Error ( tex2html_wrap_inline256 ).

  4. Usually we move the grand mean to the other side and write

    displaymath258

    and

    displaymath260

    The quantity tex2html_wrap_inline262 is called the Corrected Total Sum of Squares.

Here is a proof that A and T are perpendicular. The proofs are similar for tex2html_wrap_inline268 and tex2html_wrap_inline270 .

Recall that if

displaymath272

and

displaymath274

then x and y are perpendicular ( tex2html_wrap_inline280 ) if

displaymath282

Then

eqnarray48

eqnarray67

because tex2html_wrap_inline284 . The sum of deviations from average in any list is 0. This shows tex2html_wrap_inline270 .

Model equations

For this section we take all tex2html_wrap_inline288 -- all sample sizes equal.

Our model is

displaymath290

which we can rewrite as

displaymath292

If we define tex2html_wrap_inline294 which we call the underlying, or true, residuals and we define the population Grand Mean by

displaymath296

and we can write

displaymath298

We call this equation a Model Equation. The three pieces of the right hand side of this equation correspond to three pieces on the right hand side of our decomposition of the data

displaymath230

and we have

displaymath302

displaymath304

This identities for the equal sample size case motivate the general definition for unequal sample sizes

displaymath306

and

displaymath308

It is automatically true that

displaymath310

Least squares

Write our model as

displaymath312

where the errors tex2html_wrap_inline314 are iid tex2html_wrap_inline316 . If we stack up our data in the vector Y as

displaymath228

then the method of least squares consists of estimating tex2html_wrap_inline322 , tex2html_wrap_inline324 by finding the vector of the form

displaymath326

which is closest to the data vector Y. We measure distance in the usual Euclidean distance way. The solution is to find the orthogonal projection of Y onto the space of vectors of the form

displaymath332

It is then automatic that the projection of Y is perpendicular to Y minus the projection of Y.

How do we compute this projection? We find the vector

displaymath326

closest to Y by minimizing the squared distance from this vector to Y. This squared distance is just

displaymath346

To minimize this we set the partial derivatives with respect to each tex2html_wrap_inline348 equal to 0 for tex2html_wrap_inline350 . We get

eqnarray148

which is equal to 0 if and only if

displaymath352

This method is called least squares.

Remark: This method shows that A+T (which is the projection of Y) is perpendicular to R (which is Y-(A+T) by definition).

Null hypothesis Case

If tex2html_wrap_inline362 is true then the model equation is

displaymath364

If the errors were 0 then the vector Y would simply be

displaymath368

We find the vector of the form

displaymath370

closest to Y. Again we project Y onto the subspace of vectors spanned by

displaymath376

by minimizing the squared distance which is

displaymath378

This is minimized by

displaymath380

Least squares and maximum likelihood

The likelihood function is

displaymath382

If we assume that the null hypothesis is true then the likelihood simplifies to

displaymath384

The log likelihood is

displaymath386

or in the null hypothesis case

displaymath388

To find the MLE of the parameters tex2html_wrap_inline390 and tex2html_wrap_inline392 we must maximize these log likelihood functions. But notice that the tex2html_wrap_inline322 's occur only in the sums of squares which are multiplied by a negative sign. For any value of tex2html_wrap_inline392 we can choose the tex2html_wrap_inline322 's to maximize the log likelihood by minimizing the sum of squares. This means:

Least squares is the same as maximum likelihood for normally distributed errors -- at least in terms of estimating means.

What do you do after testing tex2html_wrap_inline362 ?

In our coagulation example we concluded that all the mean coagulation times were not the same. What next?

1: Model diagnostics

1: Confidence Intervals


next up previous



Richard Lockhart
Tue Feb 10 10:35:49 PST 1998