MAT225 Section Summary: 6.1

Inner Product, Length, and Orthogonality

Summary

Our objective here is to solve the least squares problem: there are times when we would like to the equation tex2html_wrap_inline420 exactly, but when the solution does not, in fact exist. The question then is, what's the best non-solution? We need to do something, so what should we do when the exact solution isn't a possibility? Do the next best thing....

What do we mean by ``next best thing''? We mean that we want to make the distance between tex2html_wrap_inline422 and tex2html_wrap_inline424 as small as possible; that will have to do with definitions of distance, which will fall out of something called an inner product.

The classic example of this is the standard least-squares line, which students of any science are familiar with:

  figure88

In terms of matrix operations, we're trying to find coefficients tex2html_wrap_inline426 and tex2html_wrap_inline428 such that

displaymath404

for all points. Unfortunately, we have more than two points, so the system becomes over-determined:

displaymath405

We can't (generally) find an actual solution vector tex2html_wrap_inline430 that makes this true, so we make due with an approximate solution tex2html_wrap_inline432 that gives us a ``best fit'': that minimizes the distance between the two vectors tex2html_wrap_inline434 and tex2html_wrap_inline436 .

inner product: The inner product between vectors u and v in tex2html_wrap_inline438 , or their dot product, is defined as

displaymath406

Example: #1, p. 382.

Properties of inner products (Theorem 1): Let u, v, and w be vectors in tex2html_wrap_inline438 , and c be any scalar. Then

  1. tex2html_wrap_inline444
  2. tex2html_wrap_inline446
  3. tex2html_wrap_inline448
  4. tex2html_wrap_inline450 , and tex2html_wrap_inline452 if and only if tex2html_wrap_inline454 .

norm: The length or norm of vector tex2html_wrap_inline456 is the non-negative scalar tex2html_wrap_inline458 .

Example: #7, p. 382.

unit vector: a vector whose length is 1 is called a unit vector, and one can ``normalize'' a vector (that is, give it unit length) by dividing the vector by its norm:

displaymath407

Example: #9, p. 382.

distance: For u and v in tex2html_wrap_inline438 , the distance between u and v, denoted dist(u,v), is the length of the vector u-v. That is,

displaymath408

Example: #13, p. 382.

orthogonal: two vectors u and v in tex2html_wrap_inline438 are orthogonal (to each other) if and only if tex2html_wrap_inline468 .

Example: #15, p. 382.

Theorem 2 (the Pythagorean Theorem): Two vectors u and v are orthogonal if and only if

displaymath409

orthogonal complement: If a vector z is orthogonal to every vector in a subspace W of tex2html_wrap_inline438 , then z is said to be orthogonal to W. The set of all such vectors is called the orthogonal complement of W, and denoted tex2html_wrap_inline478 .

Example: #26, p. 383.

It is easy to deduce the following facts concerning the orthogonal complement of W:

  1. A vector x is in tex2html_wrap_inline478 if and only if x is orthogonal to every vector in a spanning set of W.
  2. tex2html_wrap_inline478 is a subspace of tex2html_wrap_inline438 .

Demonstration: #29 and 30, p. 383

Theorem 3: Let A be an tex2html_wrap_inline492 matrix. The orthogonal complement of the row space of A is the nullspace of A, and the orthogonal complement of the column space of A is the nullspace of tex2html_wrap_inline500 :

displaymath410

The angle between two vectors in tex2html_wrap_inline438 can be defined using the familiar formula from calculus:

displaymath411

One interpretation of the cosine of this angle in higher dimensional space is as a correlation coefficient.

Example:

  1. If one vector is defined as the difference between heights of individuals and the mean height, and
  2. the other as the difference between weights of individuals and the mean weight, then
  3. the correlation between the two variables is given by the cosine of this angle.
The correlation coefficient in the case of the figure shown in the linear regression above is .86: the variables are strongly positively correlated.

LONG ANDREW E
Sat Jan 29 21:03:49 EST 2011