Download presentation
Presentation is loading. Please wait.
Published byBarbara Russell Modified over 9 years ago
1
Solving Systems of Equations
2
Rule of Thumb: More equations than unknowns system is unlikely to have a solution. Same number of equations as unknowns system is likely to have a unique solution. Fewer equations than unknowns the system is likely to have infinitely many solutions. Solving a System with Several Equations in Several Unknowns We are interested in this last situation!
3
Systems of Equations Linear system. Matrix methods. Note: We expect to have as many “free variables” as the difference between the number of unknowns and the number of equations.
4
Systems of Equations Linear system. Matrix methods. Non-linear system. Can be solved analytically More difficult!
5
Linear Systems of Equations We can write this as a matrix/vector equation as follows: We can solve this system using “elementary row operations”
6
Solving the System Row-reduced coefficient matrix: Now we can reinterpret this reduced matrix as a solution to the system of equations.
7
The Solution Function We started with three equations in five unknowns. As expected, our first three unknowns depend on two “free variables,” two being the difference between the number of unknowns and the number of equations.
8
Where are we? Solving a system of m equations in n unknowns is equivalent to finding the “zeros” of a vector valued function from m→n.m→n. When m > n, such a system will “typically” have infinitely many solutions. In “nice” cases, the solution will be a function from m-n → n.
9
What about finding solutions? most are impossible to solve analytically! Though we completely understand how to solve systems of linear equations... and we “luck out” with a small number of non-linear systems...
10
The Implicit Function Theorem... Now we Introduce a useful theoretical framework... gives us conditions under which a system of m equations in n unknowns will have a solution.... and it tells us that (under reasonable hypotheses) the solution function will be well behaved.
11
F: 5 → 3 g : 2 → 3
12
So How do we Proceed? This equation is already too hard…. We’ll start with something a bit simpler. What would you do with this one?
13
So How do we Proceed? This equation is already too hard…. We’ll start with something a bit simpler. What would you do with this one?
14
So How do we Proceed? This equation is already too hard…. We’ll start with something a bit simpler. What would you do with this one? Iteration!
15
Recall: x0x0 x0x0 x0x0 R R we are looking for a fixed point
16
Solving equations Suppose we want to find a root for the equation We can add x to both sides of the equation and reinterpret the problem as a fixed point problem.
17
Root finding via iteration When we iterate the function starting at x 0 =0, we get The limit is, indeed, a solution to the equation
18
Problem... This works nicely when the iterated map converges, but it may not converge. For instance, if we try the same trick with the iterated map goes to infinity. What’s the difference? What’s going on?
19
Recall: Attractors and repellors x0x0 The key difference is the derivative of the function at the fixed point. M
20
The Derivative at the Fixed Point Theorem: Suppose that R is continuously differentiable on [a,b] and that p is a fixed point of R lying in [a,b]. If then there exists some subinterval I of [a,b] containing p such that if x 0 is in I, then Conversely, if then there is some interval I about p such that every iterated map based a point of I (other than p, of course) will eventually leave I and will thus not converge to p.
21
Let x be a real number and K a non-zero real number. Note that x is a solution to the equation if and only if it is also a solution to the equation Rethinking the scheme So our fixed point root-finding scheme can be applied to any equation of the form Choose the value of K so that the derivative is “favorable”
22
The best of all possible worlds It is not hard to show that the rate at which the iterated maps on R converge to the fixed point p is governed by the size of | R '(p) |. The closer | R '(p) | is to zero, the faster the convergence rate. While we are making choices, we might as well be ambitious and choose K so that we not only get convergence to the root, but we get really fast convergence! For we want Works, provided that f '(p) 0.
23
So in the best of all possible worlds, we want to iterate the function where p is a root of f. Notice that this looks remarkably like the iteration function for Newton’s method: Constant (given p!), and ideally chosen for convergence near p. Evaluated each step, but not far from f’(p) when x is close to p.
24
Quasi-Newton methods: Iterative ways to find p with f (p) = 0 Must evaluate new f’(x) each step. Requires no “magic knowledge” and converges in a neighborhood of p. Newton’s method “Pretty good” method “Leibnitz” method If we choose D near enough to f’(p) so that |Q’(p)| < 1/2, iterated maps will converge in a neighborhood of p. “Best of all possible worlds”: Since R’(p) = 0, iterated maps converge very quickly in a neighborhood of p.
25
One Equation in Two Unknowns Can we use this set of ideas to tackle Yes! First we recall that the solutions to this equation are the zeros of the 2-variable function
26
0-Level Curves The pairs (x,y) that satisfy the equation f(x,y)=0 lie on the intersection of the graph of f and the horizontal plane z = 0. That is, they lie on the 0-level curves of the function f.
27
Taking a “piece” The 0-level curves of f. Though the points on the 0-level curves of f do not form a function, portions of them do.
28
x y Consider the contour line f (x,y) = 0 in the xy-plane. Idea: At least in small regions, this curve might be described by a function y = g(x). Our goal: find such a function!
29
x y (a,b)(a,b)(a,b)(a,b) (a,b)(a,b)(a,b)(a,b) y = g(x) Start with a point (a,b) on the contour line, where partial with respect to y is not 0: In a small box around (a,b), we can hope to find g(x). Make sure all of the y-partials in this box are close to D.
30
(a,b)(a,b)(a,b)(a,b) How to construct g(x): For this x, construct function Iterate x (y) to find its fixed point, where f (x,y) = 0. Let fixed point be g(x). x yz f (x,y) (x fixed) What’s going on here? For our fixed value of x, we find g(x) so that f (x,g(x)) = 0. We do this by iterating x. g(x)g(x) Fix x.
31
(a,b)(a,b)(a,b)(a,b) How to construct g(x) Given x, construct function Iterate x (y) to find its fixed point, where f (x,y) = 0. Let fixed point be g(x). x How do we know this works? For x = a, this is just the ideal “Leibnitz” method for finding g(a) = b. For small enough boxes, this is a “pretty good” method elsewhere. Non-trivial issues: Make sure the iterated map converges. Make sure you get the right fixed point. (Don’t leave the box!) b
32
x y (a,b)(a,b)(a,b)(a,b) (a,b)(a,b)(a,b)(a,b) y = g(x) Staying in the Box?
33
Staying in the box? Not so obvious... we have to work a bit to make sure we get the right fixed point. (We don’t leave the box!) (a,b)(a,b)(a,b)(a,b) x y x
34
Looking ahead We are interested in solving systems of equations. The Implicit Function Theorem will tell us when there are solutions to a system and will describe the nature of those solutions. We will prove the implicit function theorem by employing quasi-Newton’s methods as a theoretical tool.
Similar presentations
© 2024 SlidePlayer.com Inc.
All rights reserved.