Presentation is loading. Please wait.

Presentation is loading. Please wait.

Lecture 11 Inner Product Spaces Last Time Change of Basis (Cont.) Length and Dot Product in R n Inner Product Spaces Elementary Linear Algebra R. Larsen.

Similar presentations


Presentation on theme: "Lecture 11 Inner Product Spaces Last Time Change of Basis (Cont.) Length and Dot Product in R n Inner Product Spaces Elementary Linear Algebra R. Larsen."— Presentation transcript:

1

2 Lecture 11 Inner Product Spaces Last Time Change of Basis (Cont.) Length and Dot Product in R n Inner Product Spaces Elementary Linear Algebra R. Larsen et al. (6th Edition) TKUEE 翁慶昌 -NTUEE SCC_12_2008

3 11 - 2 Lecture 11: Inner Product Spaces Today Inner Product Spaces (Cont.) Orthonormal Bases:Gram-Schmidt Process Mathematical Models and Least Squares Analysis Applications & Term Project Reading Assignment: Secs 5.3-5.5

4 5.2 Inner Product Spaces (1) (2) (3) (4) and if and only if Inner product: Let u, v, and w be vectors in a vector space V, and let c be any scalar. An inner product on V is a function that associates a real number with each pair of vectors u and v and satisfies the following axioms. 11 - 3

5 11 - 4 Note: A vector space V with an inner product is called an inner product space. Vector space: Inner product space:

6 11 - 5 Thm 5.7: (Properties of inner products) Let u, v, and w be vectors in an inner product space V, and let c be any real number. (1) (2) (3) Norm (length) of u: Note:

7 11 - 6 u and v are orthogonal if. Distance between u and v: Angle between two nonzero vectors u and v: Orthogonal:

8 11 - 7 Notes: (1) If, then v is called a unit vector. (2) (the unit vector in the direction of v) not a unit vector

9 11 - 8 Ex 6: (Finding inner product) is an inner product Sol:

10 11 - 9 Properties of norm: (1) (2) if and only if (3) Properties of distance: (1) (2) if and only if (3)

11 11 - 10 Thm 5.8 : Let u and v be vectors in an inner product space V. (1) Cauchy-Schwarz inequality: (2) Triangle inequality: (3) Pythagorean theorem : u and v are orthogonal if and only if Theorem 5.5 Theorem 5.6 Theorem 5.4

12 11 - 11 Orthogonal projections in inner product spaces: Let u and v be two vectors in an inner product space V, such that. Then the orthogonal projection of u onto v is given by Note: If v is a init vector, then. The formula for the orthogonal projection of u onto v takes the following simpler form. Q: Why called orthogonal projection?

13 11 - 12 Ex 10: (Finding an orthogonal projection in R 3 ) Use the Euclidean inner product in R 3 to find the orthogonal projection of u=(6, 2, 4) onto v=(1, 2, 0). Sol:

14 11 - 13 Thm 5.9: (Orthogonal projection and distance) Let u and v be two vectors in an inner product space V, such that. Then

15 11 - 14 Keywords in Section 5.2: inner product: 內積 inner product space: 內積空間 norm: 範數 distance: 距離 angle: 夾角 orthogonal: 正交 unit vector: 單位向量 normalizing: 單範化 Cauchy – Schwarz inequality: 科西 - 舒瓦茲不等式 triangle inequality: 三角不等式 Pythagorean theorem: 畢氏定理 orthogonal projection: 正交投影

16 11 - 15 Lecture 11: Inner Product Spaces Today Inner Product Spaces (Cont.) Orthonormal Bases:Gram-Schmidt Process Mathematical Models and Least Squares Analysis Applications & Term Project

17 11 - 16 5.3 Orthonormal Bases: Gram-Schmidt Process Orthogonal: A set S of vectors in an inner product space V is called an orthogonal set if every pair of vectors in the set is orthogonal. Orthonormal: An orthogonal set in which each vector is a unit vector is called orthonormal. Note: If S is a basis, then it is called an orthogonal basis or an orthonormal basis.

18 11 - 175 - 17 Ex 1: (A nonstandard orthonormal basis for R 3 ) Show that the following set is an orthonormal basis. Sol: Show that the three vectors are mutually orthogonal.

19 11 - 18 Show that each vector is of length 1. Thus S is an orthonormal set.

20 11 - 19 The standard basis is orthonormal. Ex 2: (An orthonormal basis for ) In, with the inner product Sol: Then

21 11 - 20

22 11 - 21 Thm 5.10: (Orthogonal sets are linearly independent) If is an orthogonal set of nonzero vectors in an inner product space V, then S is linearly independent. Pf: S is an orthogonal set of nonzero vectors

23 11 - 22 Corollary to Thm 5.10: If V is an inner product space of dimension n, then any orthogonal set of n nonzero vectors is a basis for V.

24 11 - 23 Ex 4: (Using orthogonality to test for a basis) Show that the following set is a basis for. Sol: : nonzero vectors (by Corollary to Theorem 5.10)

25 11 - 24 Thm 5.11: (Coordinates relative to an orthonormal basis) If is an orthonormal basis for an inner product space V, then the coordinate representation of a vector w with respect to B is is orthonormal ( unique representation ) Pf: is a basis for V

26 11 - 25 Note: If is an orthonormal basis for V and, Then the corresponding coordinate matrix of w relative to B is

27 11 - 26 Ex 5: (Representing vectors relative to an orthonormal basis) Find the coordinates of w = (5, -5, 2) relative to the following orthonormal basis for. Sol:

28 11 - 27 Gram-Schmidt orthonormalization process: is a basis for an inner product space V is an orthogonal basis. is an orthonormal basis.

29 11 - 28 Sol: Ex 7: (Applying the Gram-Schmidt orthonormalization process) Apply the Gram-Schmidt process to the following basis.

30 11 - 29 Orthogonal basis Orthonormal basis

31 11 - 30 Ex 10: (Alternative form of Gram-Schmidt orthonormalization process) Find an orthonormal basis for the solution space of the homogeneous system of linear equations. Sol:

32 11 - 31 Thus one basis for the solution space is ( orthogonal basis) (orthonormal basis )

33 11 - 32 Keywords in Section 5.3: orthogonal set: 正交集合 orthonormal set: 單範正交集合 orthogonal basis: 正交基底 orthonormal basis: 單範正交基底 linear independent: 線性獨立 Gram-Schmidt Process: Gram-Schmidt 過程

34 11 - 33 Lecture 11: Inner Product Spaces Today Inner Product Spaces (Cont.) Orthonormal Bases:Gram-Schmidt Process Mathematical Models and Least Squares Analysis Applications & Term Project

35 11 - 34 http://road.uww.edu/road/zhaoy/Marketing%20770%20Summer%202008/bass.ppt Diffusion of Innovations: Bass Model S(t)=m[p+qF(t)][(1-F(t)] m=ultimate market potential p=coefficient of innovation q=coefficient of imitation

36 11 - 35 (c) Frank M. Bass (1999) An Empirical Generalization

37 11 - 36 (c) Frank M. Bass (1999) Another Example 35 mm Projectors

38 11 - 37 (c) Frank M. Bass (1999) Another Example: Overhead Projectors

39 11 - 38 5.4 Mathematical Models and Least Squares Analysis Let W be a subspace of an inner product space V. (a) A vector u in V is said to orthogonal to W, if u is orthogonal to every vector in W. (b) The set of all vectors in V that are orthogonal to W is called the orthogonal complement of W. (read “ perp”) Orthogonal complement of W:  Notes:

40 11 - 39 Notes:  Ex:

41 11 - 40 Direct sum: Let and be two subspaces of. If each vector can be uniquely written as a sum of a vector from and a vector from,, then is the direct sum of and, and you can write. Thm 5.13: (Properties of orthogonal subspaces) Let W be a subspace of R n. Then the following properties are true. (1) (2) (3)

42 11 - 41 Thm 5.14: (Projection onto a subspace) If is an orthonormal basis for the subspace W of V, and, then

43 11 - 42 Ex 5: (Projection onto a subspace) Find the projection of the vector v onto the subspace W. Sol: an orthogonal basis for W an orthonormal basis for W

44 11 - 43 Find by the other method:

45 11 - 44 Thm 5.15: (Orthogonal projection and distance) Let W be a subspace of an inner product space V, and. Then for all, ( is the best approximation to v from W)

46 11 - 45 Pf: By the Pythagorean theorem

47 11 - 46 Notes: (1) Among all the scalar multiples of a vector u, the orthogonal projection of v onto u is the one that is closest to v. (2) Among all the vectors in the subspace W, the vector is the closest vector to v.

48 11 - 47 Thm 5.16: (Fundamental subspaces of a matrix) If A is an m×n matrix, then (1) (2) (3) (4)

49 11 - 48 Ex 6: (Fundamental subspaces) Find the four fundamental subspaces of the matrix. (reduced row-echelon form) Sol:

50 11 - 49 Check:

51 11 - 50 Ex 3: Let W is a subspace of R 4 and. (a) Find a basis for W (b) Find a basis for the orthogonal complement of W. Sol: (reduced row-echelon form)

52 11 - 51 is a basis for W Notes:

53 11 - 52 Least squares problem: (A system of linear equations) (1) When the system is consistent, we can use the Gaussian elimination with back-substitution to solve for x (2) When the system is consistent, how to find the “best possible” solution of the system. That is, the value of x for which the difference between Ax and b is small.

54 11 - 53 Least squares solution: Given a system Ax = b of m linear equations in n unknowns, the least squares problem is to find a vector x in R n that minimizes with respect to the Euclidean inner product on R n. Such a vector is called a least squares solution of Ax = b.

55 11 - 54 (the normal system associated with Ax = b)

56 11 - 55 Note: The problem of finding the least squares solution of is equal to he problem of finding an exact solution of the associated normal system. Thm: For any linear system, the associated normal system is consistent, and all solutions of the normal system are least squares solution of Ax = b. Moreover, if W is the column space of A, and x is any least squares solution of Ax = b, then the orthogonal projection of b on W is

57 11 - 56 Thm: If A is an m×n matrix with linearly independent column vectors, then for every m×1 matrix b, the linear system Ax = b has a unique least squares solution. This solution is given by Moreover, if W is the column space of A, then the orthogonal projection of b on W is

58 11 - 57 Ex 7: (Solving the normal equations) Find the least squares solution of the following system and find the orthogonal projection of b on the column space of A.

59 11 - 58 Sol: the associated normal system

60 11 - 59 the least squares solution of Ax = b the orthogonal projection of b on the column space of A

61 11 - 60 Keywords in Section 5.4: orthogonal to W: 正交於 W orthogonal complement: 正交補集 direct sum: 直和 projection onto a subspace: 在子空間的投影 fundamental subspaces: 基本子空間 least squares problem: 最小平方問題 normal equations: 一般方程式

62 11 - 61 Lecture 11: Inner Product Spaces Today Inner Product Spaces (Cont.) Orthonormal Bases:Gram-Schmidt Process Mathematical Models and Least Squares Analysis Applications & Term Project

63 11 - 62 62 2016/7/9 Application: Cross Product Cross product (vector product) of two vectors  向量 (vector) 方向 : use right-hand rule The cross product is not commutative: The cross product is distributive:

64 11 - 63 63 2016/7/9 Parallelogram representation of the vector product x y θ Bsinθ Area Application: Cross Product

65 11 - 64 64 2016/7/9 向量之三重純量積 Triple Scalar product The dot and the cross may be interchanged : 純量 (scalar)

66 11 - 65 65 2016/7/9 向量之三重純量積 Parallelepiped representation of triple scalar product x y z Volume of parallelepiped defined by,, and

67 11 - 66 Fourier Approximation 12 - 66

68 11 - 67 Fourier Approximation The Fourier series transforms a given periodic function into a superposition of sine and cosine waves The following equations are used

69 11 - 68 Lecture 11: Inner Product Spaces Today Inner Product Spaces (Cont.) Orthonormal Bases:Gram-Schmidt Process Mathematical Models and Least Squares Analysis Applications & Term Project Reading Assignment: Secs 5.3-5.5

70 11 - 69 Lecture 11: Inner Product Space Next Time Introduction to Linear Transformations The Kernel and Range of a Linear Transformation Matrices for Linear Transformations Reading Assignment: Secs 6.1-6.3


Download ppt "Lecture 11 Inner Product Spaces Last Time Change of Basis (Cont.) Length and Dot Product in R n Inner Product Spaces Elementary Linear Algebra R. Larsen."

Similar presentations


Ads by Google