Remark: Given a nonzero vector in , consider the problem of decomposing a vector in into the sum of two vectors, one a multiple of and the other orthogonal to . We wish to write for some scalar and is some vector orthogonal to and . is orthogonal to if an only if if and only if if and only if which is the wight of with respect to of the linear combination of an orthogonal set if is part of . The vector is called the orthogonal projection of onto , and the vector is called the component of orthogonal to . Notice when . Hence is determined by the subspace spanned by ( the line passing and ). is denoted by proj and is called the orthogonal projection of onto .
Example 1: Let and .
Find the orthogonal projection of onto .
Then write as the sum of two orthogonal vectors, one in Span and one orthogonal to .
Exercise 1: Let and . Find the orthogonal projection of onto . Then write as the sum of two orthogonal vectors,
one in Span and one orthogonal to .
The orthogonal projection of a point in onto a line through the origin has an important analogue in . Given a vector and a subspace in , there is a vector in such that is the unique vector in for which is orthogonal to , and is the unique vector in closest to . These two properties of provide the key to finding the least-squares solutions of linear systems.
Theorem: Let be a subspace of . Then each in can be written uniquely in the form where is in and is orthogonal to . In fact, if is any orthogonal basis of , then and .
Definition: The vector is called the orthogonal projection of onto and often is written as .
Proof: Use }
Example 2: Let . Notice that
is an orthogonal basis of . Let . Write as the sum of a vector in and a vector orthogonal to .
Exercise 2: Let . Notice that
is an orthogonal basis of . Let . Write as the sum of a vector in and a
vector orthogonal to .
Remark: If is an orthogonal basis for and if happens to be in , then the formula for is exactly the same as the representation of given in above theorem, i.e. .
Theorem: Let be a subspaceof , let be any vector in ,and let be the orthogonal projection of onto . Then is the closest point in to , in the sense that for all in distinct from .
Remark: The vector in is called the best approximation to by elements of . The distance from to , given by , can be regarded as the \textquotedblleft error\textquotedblright{} of using in place of . The theorem says that this error is minimized when . This theorem also shows that does not depend on the particular orthogonal basis used to compute it.
Example 3: The distance from a point in to a subspace is defined as the distance from to the nearest point in . Find the
distance from to , where and .
Exercise 3: The distance from a point in to a subspace is defined as the distance from to the nearest point in . Find the distance from to , where and .
Example 4: Find the best approximation to by vectors of the form where and .
Exercise 4: Find the best approximation to by vectors of the form
where , and .
Definition: If isand is in , a least-squares solution of is an in such that for all in .
Remark: 1. Given and , apply the Best Approximation Theorem to the subspace and let . Because is in the column space , the equation is consistent, and there is an in such that . Since is the
closest point in to , a vector is a least-squares solution of if and only if satisfies .
2. has the property that is orthogonal to, so is orthogonal to each column of , i.e. for any column of or which is equivalent to .
Definition: is called the normal equations for . A solution of is often denoted by .}
Theorem: The set of least-squares solutions of coincides with the nonempty set of solutions of the normal equation .
Example 5: Find a least-squares solution of the inconsistent system for
,
Exercise 5: Find a least-squares solution of the inconsistent system for ,
Theorem: Let be an matrix. The following statements are logically equivalent:
(a) The equation has a unique least-squares solution for each in .
(b) The columns of are linearly independent.
(c) The matrix is invertible.
When these statements are true, the least-squares solution is given by .
Remark: When a least-squares solution is used to produce as an approximation to , the distance from to is called the least-squares error of this approximation.
Example 6: Find a least-squares solution of the inconsistent system for
. ,
Exercise 6: Find a least-squares solution of the inconsistent system for
,
GroupWork 1: True or False. is an matrix and is in .
a. The general least squares problem is to find an that makes as close as possible to .
b. A least squares solution of is a vector that satisfies where is the orthogonal projection of onto .
c. A least squares solution of is a vector such that for all in .
d. Any solution of is a least squares solution of .
e. If the columns of are linearly independent, then the equation has exactly one least squares solution.
f. For each and each subspace , the vector is orthogonal to .
g. The orthogonal projection of onto a subspace can sometimes depend on the orthogonal basis for used to compute .
GroupWork 2: Find a formula for the least squares solution of when the columns of are orthonormal.
GroupWork 3: True or False. is an matrix and is in .
a. If is in the column space of , then every solution of is a least squares solution.
b. The least squares solution of is the point in the column space of closest to .
c. A least squares solution of is a list of weights that, when applied to the columns of , produces the orthogonal projection of onto .
d. If is a least squares solution of , then .
e. If is in a subspace , then the orthogonal projection of onto is itself.
f. The best approximation to by elements of a subspace is given by the vector .
GroupWork 4: Describe all least squares solutions of the system
x+y = 2
x+y = 4