Presentation is loading. Please wait.

Presentation is loading. Please wait.

2.III. Basis and Dimension 1.Basis 2.Dimension 3.Vector Spaces and Linear Systems 4.Combining Subspaces.

Similar presentations


Presentation on theme: "2.III. Basis and Dimension 1.Basis 2.Dimension 3.Vector Spaces and Linear Systems 4.Combining Subspaces."— Presentation transcript:

1 2.III. Basis and Dimension 1.Basis 2.Dimension 3.Vector Spaces and Linear Systems 4.Combining Subspaces

2 2.III.1. Basis Definition 1.1: Basis A basis of a vector space V is an ordered set of linearly independent (non-zero) vectors that spans V. Notation: Example 1.2: is a basis for R 2 B is L.I. : → → B spans R 2 : → → L.I. → Minimal Span → Complete

3 Example 1.3: is a basis for R 2 that differs from B only in order. Definition 1.5: Standard / Natural Basis for R n k th component of e i =

4 Example 1.6: For the function space a natural basis is Another basis is Proof is straightforward. Example 1.7: For the function space of cubic polynomials P 3, a natural basis is Other choices can be Proof is again straightforward. Rule: Set of L.C.’s of a L.I. set is L.I. if each L.C. contains a different vector.

5 Example 1.8: The trivial space { 0 } has only one basis, the empty one  . Note: By convention, 0 does not count as a basis vector. ( Any set of vectors containing 0 is linearly dependent. ) Example 1.9: The space of all finite degree polynomials has a basis with infinitely many elements  1, x, x 2, … . Example 1.10: Solution Set of Homogeneous Systems The solution set of is ( Proof of L.I. is left as exercise )

6 Example 1.11: Matrices Find a basis for this subspace of M 2  2 : Solution: ∴ Basis is ( Proof of L.I. is left as exercise ) Theorem 1.12: In any vector space, a subset is a basis if and only if each vector in the space can be expressed as a linear combination of elements of the subset in a unique way. Proof: A basis is by definition spanning → every vector can be expressed as a linear combination of the basis vectors. Letthen ∴ L.I.  uniqueness

7 Definition 1.13: Representation wrt a Basis Let B =  β 1, …, β n  be a basis of vector space V and Then the representation of v wrt B is c j are called the coordinates (components) of v wrt B. Subscript B is often omitted Example 1.14: P 3 Let Then

8 Exercises 2.III.1 2. A square matrix is symmetric if for all indices i and j, entry i, j equals entry j, i. (a) Find a basis for the vector space of symmetric 2  2 matrices. (b) Find a basis for the space of symmetric 3  3 matrices. (c) Find a basis for the space of symmetric n  n matrices. 1. Find a basis for each. (a) The subspace { a 2 x 2 + a 1 x + a 0 | a 2  2a 1 = a 0 } of P 2. (b) The space of three-wide row vectors whose 1st and 2nd components add to zero. (c) This subspace of the 2  2 matrices

9 is a vector space under these operations. 3. One of the exercises in the Subspaces subsection shows that the set Find a basis.

10 2.III.2. Dimension Definition 2.1 A vector space is finite-dimensional if it has a basis with only finitely many vectors. Lemma 2.2:Exchange Lemma Assume that B =  β 1, …, β n  is a basis for a vector space, and that for the vector v the relationship holds: To be proved: All bases for a vector space have the same number of elements. →Dimension  Number of vectors in basis. →Basis = Minimal spanning = L.I. set = Smallest set. Then exchanging β j for v yields another basis for the space. where c j  0. Proof: See Hefferon p.120.

11 Theorem 2.3: In any finite-dimensional vector space, all of the bases have the same number of elements. Proof: Let B =  β 1, …, β n  be a basis of n elements. Any other basis D =  δ 1, …, δ m  must have m  n. Let with c k  0. By lemma 2.2, D 1 =  β 1, …, β k  1,δ 1,β k + 1, …, β n  is a basis. Next, replacing β j in D 1 begets D 2 =  β 1, …, β k  1,δ 1,β k + 1, …, β j  1,δ 2,β j + 1, …, β n  Repeating the process n times results in a basis D n =  δ 1, …, δ n  that spans V. Which contradicts with the assumption that D is L.I. with at least one c k  0. If m > n, then we can write Hence m = n.

12 Definition 2.4: Dimension The dimension of a vector space is the number of vectors in any of its bases. Example 2.5: R n Any basis for R n has n vectors since the standard basis E n has n vectors. → R n is n-D. Example 2.6: P n dim P n = n+1. since its natural basis,  1, x, x 2, …, x n , has n+1 elements. Example 2.7: A trivial space is 0-D since its basis is empty. Comments: All results in this book are applicable to finite-D vectors spaces. Most of them are also applicable to countably infinite-D vectors spaces. For uncountably infinite-D vectors spaces, e.g., Hilbert spaces, convergence most be taken into account.

13 Corollary 2.8: No L.I. set can have a size greater than the dimension of the enclosing space. Example 2.9 : Only subspaces in R 3. 2-D: Planes thru 0 1-D: Lines thru 0 0-D: {0}

14 Corollary 2.10: Any L.I. set can be expanded to make a basis. Corollary 2.11: Any spanning set can be shrunk to a basis. Corollary 2.12: In an n-D space, a set of n vectors is L.I. iff it spans the space. Remark 2.13: The statement ‘any infinite-dimensional vector space has a basis’ is known to be equivalent to a statement called the Axiom of Choice. Mathematicians differ philosophically on whether to accept or reject this statement as an axiom on which to base mathematics (although, the great majority seem to accept it).

15 Exercises 2.III.2. 2. Observe that, where S is a set, the functions f : S → R form a vector space under the natural operations: ( f + g ) (s) = f(s) + g(s) and (r f )(s) = r f(s). What is the dimension of the space resulting for each domain? (a) S = {1} (b) S = { 1, 2 } (c) S = { 1, 2, …, n } 1. What is the dimension of the span of the set { cos 2 θ, sin 2 θ, cos 2θ, sin 2θ} This span is a subspace of the space of all real-valued functions of one real variable. 3. Prove that if U and W are both three-dimensional subspaces of R 5 then U  W is non-trivial. Generalize.

16 2.III.3. Vector Spaces and Linear Systems Definition 3.1: Row Space & Row Rank The row space of a matrix is the span of the set of its rows. The row rank is the dimension of the row space, the number of L.I. rows. Example 3.2: → Lemma 3.3: Row-equivalent matrices have the same row space & hence the same row rank. Proof:Let A & B be row-equivalent matrices.  Each row of B is a lin.comb. of rows of A → RowSpace(B)  RowSpace(A) Each row of A is a lin.comb. of rows of B → RowSpace(A)  RowSpace(B) Hence, RowSpace(A) = RowSpace(B)

17 Lemma 3.4: The nonzero rows of an echelon form matrix make up a L.I. set. Proof: This is just a re-statement of Lemma III.2.5, which states that, in an echelon form matrix, no nonzero row is a linear combination of the other rows. Gaussian reduction ~ Finding a basis for the row space. Example 3.5: →→ Basis for the row space is { (1 3 1), (0 1 0), (0 0 3) }.

18 Definition 3.6: Column Space & Column Rank The column space of a matrix is the span of the set of its columns. The column rank is the dimension of the column space, the number of L.I. columns. A linear system is equivalent to the linear combination of the columns of the corresponding coefficient matrix. ~ Basis for the column space can be found by applying the Gaussian reduction to the transpose of the corresponding coefficient matrix. Definition 3.8: Transpose The transpose of a matrix is the result of interchanging the rows and columns of that matrix, i.e.,

19 Example 3.7: Find a basis for the column space of ~ → Basis is

20 Example 3.9: Get a basis for Solution: Vector spaces T & S are isomorphic. Let ~ → Basis is

21 Lemma 3.10: Row operations do not change the column rank. Proof:Rationale for Gaussian reduction: Row operations do not affect linear relationships among columns. Example:Reading basis from reduced echelon form ~ Basis for row space (rows with leading entries): Basis for column space (columns containing leading entries):

22 Theorem 3.11: The row rank and column rank of a matrix are equal. Proof: Lemmas 3.3 & 3.10 → row rank of matrix = row rank of its echelon form. In reduced echelon form: Row rank = Number of leading entries = Column rank. Definition 3.12: Rank The rank of a matrix is its row rank or column rank. → dim(RowSpace) = dim(ColumnSpace) Theorem 3.13: For linear systems with n unknowns and matrix of coefficients A, the following statements are equivalent. (1) rank of A is r (2) space of solutions of the associated homogeneous system has dim n  r Proof:rank A = r  reduced echelon form has r non-zero rows (L.I. eqs)  there are n  r free variables (L.D.eqs)

23 Corollary 3.15: Where the matrix A is n  n, the following statements are equivalent. (1) the rank of A is n (2) A is nonsingular (3) the rows of A form a linearly independent set (4) the columns of A form a linearly independent set (5) any linear system with matrix of coefficients A has one and only one solution Proof:Trivial (see Hefferon p.129)

24 Exercises 2.III.3. 1. Find a basis for the span of each set. (a) (b) (c) (d) 2. (a) Show that the following set of column vectors is a subspace of R 3. (b) Find a basis.

25 2.III.4. Combining Subspaces Definition 4.1:Sum of Vector Subspaces Where W 1, …, W k are subspaces of a vector space, their sum is the span of their union: Example 4.2: R 3 → R 3 = x-axis + y-axis + z-axis Example 4.3: A sum of subspaces can be less than the entire space. Let L = { a + b x | a, b  R } and C = { c x 3 | c  R } be subspaces of P 4. Then L + C = { a + b x + c x 3 | a, b, c  R }  P 4.

26 Example 4.4: R 3 A space can be described as a combination of subspaces in more than one way. → R 3 = xy-plane + yz-plane This decomposition of v is obviously not unique. Definition 4.7: The concatenation of the sequences … is their adjoinment

27 Lemma 4.8: Let V be a vector space that is the sum of some of its subspaces V = W 1 + … + W k. Let B 1,..., B k be any bases for these subspaces. Then the following are equivalent. (1) For every v  V, the expression v = w 1 + …+ w k (with w i  W i ) is unique. (2) The concatenation B 1  …  B k is a basis for V. (3) The nonzero members of {w 1, …, w k } (with w i  W i ) form a linearly independent set — among nonzero vectors from different W i ’s, every linear relationship is trivial. Proof: See Hefferon, p.134.

28 Definition 4.9: Independent Set of Subspaces A collection of subspaces { W 1, …, W k } is independent if no nonzero vector from any W i is a linear combination of vectors from the other subspaces W 1, …, W i  1 W i + 1, …, W k. Definition 4.10: ( Internal ) Direct Sum A vector space V is the direct sum of its subspaces W 1, …, W k if V = W 1 +…+ W k and { W 1, …, W k } is independent. We write V = W 1  …  W k Example 4.11: R 3 = x-axis  y-axis  z-axis. Example 4.12:

29 Corollary 4.13: The dimension of a direct sum is the sum of the dimensions of its summands. Proof. Follows directly from Lemma 4.8 Definition 4.14: Complement Subspaces When a vector space is the direct sum of two of its subspaces, then they are said to be complements. Lemma 4.15: A vector space V is the direct sum of two of its subspaces W 1 and W 2 iff V = W 1 +W 2 and W 1  W 2 = { 0 } Caution:  fails if there are more than 2 subspaces. See Example 4.19 below. Proof  : V = W 1  W 2 → W 1 and W 2 are L.I. → W 1  W 2 = { 0 } Proof  : W 1  W 2 = { 0 } → W 1 and W 2 are L.I.

30 Example 4.16: Direct Sum Decomposition is not Unique In R 2, the x-axis and the y-axis are complements, i.e., R 2 = x-axis  y-axis. So are lines y = x and y = 2 x. Example 4.17:Complement to a Subspace is not Unique In the space F = { a cosθ + b sinθ | a,b  R }, the subspaces W 1 = { a cosθ | a  R } and W 2 = { b sinθ | b  R } are complements. Another complement of W 1 is W 3 = { b cosθ + b sinθ | b  R }. Example 4.18: In R 3, the xy-plane and the yz-planes are not complements.

31 Example 4.19: R 3 Let W 1 = x-axis, W 2 = y-axis, If there are more than two subspaces, then having a trivial intersection is not enough to guarantee unique decomposition. → Decomposition is not unique: Reason: vector from W 3 can be a L.C. of those from W 1 & W 2 →

32 Exercises 2.III.4. 1. Let W 1,W 2 be subspaces of a vector space. (a) Assume that the set S 1 spans W 1, and that the set S 2 spans W 2. Can S 1  S 2 span W 1 +W 2 ? Must it? (b) Assume that S 1 is a linearly independent subset of W 1 and that S 2 is a linearly independent subset of W 2. Can S 1  S 2 be a linearly independent subset of W 1 +W 2 ? Must it? 3. Let W 1, W 2, W 3 be subspaces of a vector space. Prove that (W 1  W 2 ) + (W 1  W 3 )  W 1  (W 2 +W 3 ). Does the inclusion reverse? 2. The example of the x-axis and the y-axis in R 2 shows that W 1  W 2 = V does not imply that W 1  W 2 = V. Can W 1  W 2 = V and W 1  W 2 = V happen?

33 Let V and W be vector spaces. Use wikipedia to find out the meanings of their Direct sum, V  W. Direct product, V  W. Tensor product, V  W.


Download ppt "2.III. Basis and Dimension 1.Basis 2.Dimension 3.Vector Spaces and Linear Systems 4.Combining Subspaces."

Similar presentations


Ads by Google