Orthogonality and Least Squares

Slides:



Advertisements
Similar presentations
Vector Spaces A set V is called a vector space over a set K denoted V(K) if is an Abelian group, is a field, and For every element vV and K there exists.
Advertisements

10.4 Complex Vector Spaces.
5.4 Basis And Dimension.
Linear Equations in Linear Algebra
Eigenvalues and Eigenvectors
1 1.8 © 2012 Pearson Education, Inc. Linear Equations in Linear Algebra INTRODUCTION TO LINEAR TRANSFORMATIONS.
4 4.3 © 2012 Pearson Education, Inc. Vector Spaces LINEARLY INDEPENDENT SETS; BASES.
Eigenvalues and Eigenvectors
6 6.1 © 2012 Pearson Education, Inc. Orthogonality and Least Squares INNER PRODUCT, LENGTH, AND ORTHOGONALITY.
THE DIMENSION OF A VECTOR SPACE
Symmetric Matrices and Quadratic Forms
Chapter 5 Orthogonality
4 4.6 © 2012 Pearson Education, Inc. Vector Spaces RANK.
Linear Equations in Linear Algebra
6 6.1 © 2012 Pearson Education, Inc. Orthogonality and Least Squares INNER PRODUCT, LENGTH, AND ORTHOGONALITY.
6 6.3 © 2012 Pearson Education, Inc. Orthogonality and Least Squares ORTHOGONAL PROJECTIONS.
5 5.1 © 2012 Pearson Education, Inc. Eigenvalues and Eigenvectors EIGENVECTORS AND EIGENVALUES.
Orthogonal Sets (12/2/05) Recall that “orthogonal” matches the geometric idea of “perpendicular”. Definition. A set of vectors u 1,u 2,…,u p in R n is.
5.1 Orthogonality.
Length and Dot Product in R n Notes: is called a unit vector. Notes: The length of a vector is also called its norm. Chapter 5 Inner Product Spaces.
Linear Equations in Linear Algebra
Inner Product Spaces Euclidean n-space: Euclidean n-space: vector lengthdot productEuclidean n-space R n was defined to be the set of all ordered.
Chapter 5: The Orthogonality and Least Squares
CHAPTER FIVE Orthogonality Why orthogonal? Least square problem Accuracy of Numerical computation.
Chapter 5 Orthogonality.
4 4.2 © 2012 Pearson Education, Inc. Vector Spaces NULL SPACES, COLUMN SPACES, AND LINEAR TRANSFORMATIONS.
CPSC 491 Xin Liu Nov 17, Introduction Xin Liu PhD student of Dr. Rokne Contact Slides downloadable at pages.cpsc.ucalgary.ca/~liuxin.
Eigenvalues and Eigenvectors
4 4.4 © 2012 Pearson Education, Inc. Vector Spaces COORDINATE SYSTEMS.
1 1.4 Linear Equations in Linear Algebra THE MATRIX EQUATION © 2016 Pearson Education, Inc.
Linear Equations in Linear Algebra
AN ORTHOGONAL PROJECTION
1 1.7 © 2016 Pearson Education, Inc. Linear Equations in Linear Algebra LINEAR INDEPENDENCE.
1 1.3 © 2012 Pearson Education, Inc. Linear Equations in Linear Algebra VECTOR EQUATIONS.
Orthogonality and Least Squares
6 6.1 © 2016 Pearson Education, Inc. Orthogonality and Least Squares INNER PRODUCT, LENGTH, AND ORTHOGONALITY.
Elementary Linear Algebra Anton & Rorres, 9th Edition
Chapter 10 Real Inner Products and Least-Square
4.8 Rank Rank enables one to relate matrices to vectors, and vice versa. Definition Let A be an m  n matrix. The rows of A may be viewed as row vectors.
Section 5.1 Length and Dot Product in ℝ n. Let v = ‹v 1­­, v 2, v 3,..., v n › and w = ‹w 1­­, w 2, w 3,..., w n › be vectors in ℝ n. The dot product.
Chap. 5 Inner Product Spaces 5.1 Length and Dot Product in R n 5.2 Inner Product Spaces 5.3 Orthonormal Bases: Gram-Schmidt Process 5.4 Mathematical Models.
4 © 2012 Pearson Education, Inc. Vector Spaces 4.4 COORDINATE SYSTEMS.
Class 24: Question 1 Which of the following set of vectors is not an orthogonal set?
Advanced Computer Graphics Spring 2014 K. H. Ko School of Mechatronics Gwangju Institute of Science and Technology.
4.8 Rank Rank enables one to relate matrices to vectors, and vice versa. Definition Let A be an m  n matrix. The rows of A may be viewed as row vectors.
6 6.3 © 2016 Pearson Education, Inc. Orthogonality and Least Squares ORTHOGONAL PROJECTIONS.
5 5.1 © 2016 Pearson Education, Ltd. Eigenvalues and Eigenvectors EIGENVECTORS AND EIGENVALUES.
4 4.1 © 2016 Pearson Education, Ltd. Vector Spaces VECTOR SPACES AND SUBSPACES.
4 4.5 © 2016 Pearson Education, Inc. Vector Spaces THE DIMENSION OF A VECTOR SPACE.
4 4.2 © 2016 Pearson Education, Inc. Vector Spaces NULL SPACES, COLUMN SPACES, AND LINEAR TRANSFORMATIONS.
1 1.4 Linear Equations in Linear Algebra THE MATRIX EQUATION © 2016 Pearson Education, Ltd.
6 6.5 © 2016 Pearson Education, Ltd. Orthogonality and Least Squares LEAST-SQUARES PROBLEMS.
Inner Product Spaces Euclidean n-space: Euclidean n-space: vector lengthdot productEuclidean n-space R n was defined to be the set of all ordered.
Eigenvalues and Eigenvectors
VECTOR SPACES AND SUBSPACES
Orthogonality and Least Squares
VECTOR SPACES AND SUBSPACES
Orthogonality and Least Squares
Linear Algebra Lecture 39.
Linear Algebra Lecture 38.
Symmetric Matrices and Quadratic Forms
Linear Algebra Lecture 41.
THE DIMENSION OF A VECTOR SPACE
NULL SPACES, COLUMN SPACES, AND LINEAR TRANSFORMATIONS
Vector Spaces COORDINATE SYSTEMS © 2012 Pearson Education, Inc.
Linear Equations in Linear Algebra
Orthogonality and Least Squares
Symmetric Matrices and Quadratic Forms
Linear Equations in Linear Algebra
VECTOR SPACES AND SUBSPACES
Presentation transcript:

Orthogonality and Least Squares ORTHOGONAL SETS © 2012 Pearson Education, Inc.

ORTHOGONAL SETS A set of vectors {u1,…,up} in is said to be an orthogonal set if each pair of distinct vectors from the set is orthogonal, that is, if whenever . Theorem 4: If is an orthogonal set of nonzero vectors in , then S is linearly independent and hence is a basis for the subspace spanned by S. © 2012 Pearson Education, Inc.

ORTHOGONAL SETS Proof: If for some scalars c1,…,cp, then because u1 is orthogonal to u2,…,up. Since u1 is nonzero, is not zero and so . Similarly, c2,…,cp must be zero. © 2012 Pearson Education, Inc.

ORTHOGONAL SETS Thus S is linearly independent. Definition: An orthogonal basis for a subspace W of is a basis for W that is also an orthogonal set. Theorem 5: Let {u1,…,up} be an orthogonal basis for a subspace W of . For each y in W, the weights in the linear combination are given by © 2012 Pearson Education, Inc.

ORTHOGONAL SETS Proof: The orthogonality of {u1,…,up} shows that Since is not zero, the equation above can be solved for c1. To find cj for , compute and solve for cj. © 2012 Pearson Education, Inc.

AN ORTHOGONAL PROJECTION Given a nonzero vector u in , consider the problem of decomposing a vector y in into the sum of two vectors, one a multiple of u and the other orthogonal to u. We wish to write ----(1) where for some scalar α and z is some vector orthogonal to u. See the following figure. © 2012 Pearson Education, Inc.

AN ORTHOGONAL PROJECTION Given any scalar α, let , so that (1) is satisfied. Then is orthogonal to u if an only if That is, (1) is satisfied with z orthogonal to u if and only if and . The vector is called the orthogonal projection of y onto u, and the vector z is called the component of y orthogonal to u. © 2012 Pearson Education, Inc.

AN ORTHOGONAL PROJECTION If c is any nonzero scalar and if u is replaced by cu in the definition of , then the orthogonal projection of y onto cu is exactly the same as the orthogonal projection of y onto u. Hence this projection is determined by the subspace L spanned by u (the line through u and 0). Sometimes is denoted by projLy and is called the orthogonal projection of y onto L. That is, ----(2) © 2012 Pearson Education, Inc.

AN ORTHOGONAL PROJECTION Example 1: Let and . Find the orthogonal projection of y onto u. Then write y as the sum of two orthogonal vectors, one in Span {u} and one orthogonal to u. Solution: Compute © 2012 Pearson Education, Inc.

AN ORTHOGONAL PROJECTION The orthogonal projection of y onto u is and the component of y orthogonal to u is The sum of these two vectors is y. © 2012 Pearson Education, Inc.

AN ORTHOGONAL PROJECTION That is, The decomposition of y is illustrated in the following figure. © 2012 Pearson Education, Inc.

AN ORTHOGONAL PROJECTION Note: If the calculations above are correct, then will be an orthogonal set. As a check, compute Since the line segment in the figure on the previous slide between y and is perpendicular to L, by construction of , the point identified with is the closest point of L to y. © 2012 Pearson Education, Inc.

ORTHONORMAL SETS A set {u1,…,up} is an orthonormal set if it is an orthogonal set of unit vectors. If W is the subspace spanned by such a set, then {u1,…,up} is an orthonormal basis for W, since the set is automatically linearly independent, by Theorem 4. The simplest example of an orthonormal set is the standard basis {e1,…,en} for . Any nonempty subset of {e1,…,en} is orthonormal, too. © 2012 Pearson Education, Inc.

ORTHONORMAL SETS Example 2: Show that {v1, v2, v3} is an orthonormal basis of , where , , Solution: Compute © 2012 Pearson Education, Inc.

ORTHONORMAL SETS Thus {v1, v2, v3} is an orthogonal set. Also, which shows that v1, v2, and v3 are unit vectors. Thus {v1, v2, v3} is an orthonormal set. Since the set is linearly independent, its three vectors form a basis for . See the figure on the next slide. © 2012 Pearson Education, Inc.

ORTHONORMAL SETS When the vectors in an orthogonal set of nonzero vectors are normalized to have unit length, the new vectors will still be orthogonal, and hence the new set will be an orthonormal set. © 2012 Pearson Education, Inc.

ORTHONORMAL SETS Theorem 6: An matrix U has orthonormal columns if and only if . Proof: To simplify notation, we suppose that U has only three columns, each a vector in . Let and compute ----(3) © 2012 Pearson Education, Inc.

ORTHONORMAL SETS The entries in the matrix at the right are inner products, using transpose notation. The columns of U are orthogonal if and only if , , ----(4) The columns of U all have unit length if and only if , , ----(5) The theorem follows immediately from (3)–(5). © 2012 Pearson Education, Inc.

ORTHONORMAL SETS Theorem 7: Let U be an matrix with orthonormal columns, and let x and y be in . Then if and only if Properties (a) and (c) say that the linear mapping preserves lengths and orthogonality. © 2012 Pearson Education, Inc.