STAT 350: Lecture 17

The function

is a quadratic form. The coefficient of a cross product term like is so the function is unchanged if each of and is replaced by their average. In other words we might as well assume that the matrix Q is symmetric. Consider for example the function . The matrix Q is

What I did in class is the n-dimensional version of the following: Find new variables and and constants and such that . Put in the expressions for in terms of the and you get

Comparing coefficients we can check that

where A is the matrix with entries and is a diagonal matrix with and on the diagonal. In other words we have to diagonalize Q.

To find the eigenvalues of Q we can solve The characteristic polynomial is whose two roots are 2 and 7. To find the corresponding eigenvectors you ``solve'' . For you get the equations

These equations are linearly dependent (otherwise the only solution would be v=0 and would not be an eigenvalue). Solving either one gives so that is an eigenvector as is any non-zero multiple of that vector. To get a normalized eigenvector you divide through by the length of the vector, that is, by . The second eigenvector may be found similarly. We get the equation so that is an eigenvector for the eigenvalue 2. After normalizing we stick these two eigenvectors in the matrix I called P obtaining

Now check that

This makes the matrix A above be and and . You can check that as desired.

As a second example consider a sample of size 3 from the standard normal distribution, say, , and . Then you know that is supposed to have a distribution on n-1 degrees of freedom where now n=2. Expanding out

for which the matrix Q is

The determinant of may be found to be . This factors as so that the eigenvalues are 1, 1, and 0. An eigenvector corresponding to 0 is . Corresponding to the other two eigenvalues there are actually many possibilities. The equations are which is 1 equation in 3 unknowns so has a two dimensional solution space. For instance the vector is a solution. The third solution would then be perpendicular to this, making the first two entries equal. Thus is a third eigenvector.

The key point in the lecture, however, is that the distribution of the quadratic form depends only on the eigenvalues of Q and not on the eigenvectors. We can rewrite in the form . To find and we fill up a matrix P with columns which are our eigenvectors, scaled to have length 1. This makes

and we find to have components

and

You should check that these new variables all have variance 1 and all covariances equal to 0. In other words they are standard normals. Also check that . Since we have written as a sum of square of two of these independent normals we can conclude that has a distribution.

Richard Lockhart
Fri Feb 28 10:43:31 PST 1997