Analysis of Algorithms CS 477/677

Slides:



Advertisements
Similar presentations
CS Section 600 CS Section 002 Dr. Angela Guercio Spring 2010.
Advertisements

Greedy Algorithms.
Comp 122, Spring 2004 Greedy Algorithms. greedy - 2 Lin / Devi Comp 122, Fall 2003 Overview  Like dynamic programming, used to solve optimization problems.
1.1 Data Structure and Algorithm Lecture 6 Greedy Algorithm Topics Reference: Introduction to Algorithm by Cormen Chapter 17: Greedy Algorithm.
Greedy Algorithms Greed is good. (Some of the time)
Analysis of Algorithms
Greedy Algorithms Be greedy! always make the choice that looks best at the moment. Local optimization. Not always yielding a globally optimal solution.
Greedy Algorithms Basic idea Connection to dynamic programming Proof Techniques.
Lecture 7: Greedy Algorithms II Shang-Hua Teng. Greedy algorithms A greedy algorithm always makes the choice that looks best at the moment –My everyday.
Greedy Algorithms Reading Material: –Alsuwaiyel’s Book: Section 8.1 –CLR Book (2 nd Edition): Section 16.1.
Greedy Algorithms CIS 606 Spring Greedy Algorithms Similar to dynamic programming. Used for optimization problems. Idea – When we have a choice.
Lecture 7: Greedy Algorithms II
16.Greedy algorithms Hsu, Lih-Hsing. Computer Theory Lab. Chapter 16P An activity-selection problem Suppose we have a set S = {a 1, a 2,..., a.
Dynamic Programming Adapted from Introduction and Algorithms by Kleinberg and Tardos.
CSCI 256 Data Structures and Algorithm Analysis Lecture 14 Some slides by Kevin Wayne copyright 2005, Pearson Addison Wesley all rights reserved, and some.
Greedy Algorithms and Matroids Andreas Klappenecker.
Prof. Swarat Chaudhuri COMP 482: Design and Analysis of Algorithms Spring 2012 Lecture 16.
CSCI 256 Data Structures and Algorithm Analysis Lecture 6 Some slides by Kevin Wayne copyright 2005, Pearson Addison Wesley all rights reserved, and some.
COSC 3101A - Design and Analysis of Algorithms 8 Elements of DP Memoization Longest Common Subsequence Greedy Algorithms Many of these slides are taken.
Analysis of Algorithms CS 477/677 Instructor: Monica Nicolescu Lecture 14.
1 Chapter 6-1 Dynamic Programming Slides by Kevin Wayne. Copyright © 2005 Pearson-Addison Wesley. All rights reserved.
CS 3343: Analysis of Algorithms Lecture 18: More Examples on Dynamic Programming.
CSC5101 Advanced Algorithms Analysis
Dynamic Programming.  Decomposes a problem into a series of sub- problems  Builds up correct solutions to larger and larger sub- problems  Examples.
Greedy Algorithms BIL741: Advanced Analysis of Algorithms I (İleri Algoritma Çözümleme I)1.
CS 3343: Analysis of Algorithms Lecture 19: Introduction to Greedy Algorithms.
CSCI-256 Data Structures & Algorithm Analysis Lecture Note: Some slides by Kevin Wayne. Copyright © 2005 Pearson-Addison Wesley. All rights reserved. 18.
1 Algorithms CSCI 235, Fall 2015 Lecture 29 Greedy Algorithms.
Analysis of Algorithms CS 477/677 Instructor: Monica Nicolescu Lecture 18.
6/13/20161 Greedy A Comparison. 6/13/20162 Greedy Solves an optimization problem: the solution is “best” in some sense. Greedy Strategy: –At each decision.
Analysis of Algorithms CS 477/677 Instructor: Monica Nicolescu Lecture 17.
PREPARED BY: Qurat Ul Ain SUBMITTED TO: Ma’am Samreen.
Greedy Algorithms Many of the slides are from Prof. Plaisted’s resources at University of North Carolina at Chapel Hill.
Greedy Algorithms Many of the slides are from Prof. Plaisted’s resources at University of North Carolina at Chapel Hill.
Analysis of Algorithms CS 477/677
Greedy Algorithms General principle of greedy algorithm
Greedy algorithms: CSC317
Dynamic Programming Typically applied to optimization problems
Algorithm Design and Analysis
CS 3343: Analysis of Algorithms
Greedy Algorithms (Chap. 16)
Greedy Algorithms Basic idea Connection to dynamic programming
ICS 353: Design and Analysis of Algorithms
CS 3343: Analysis of Algorithms
CS6045: Advanced Algorithms
Greedy Algorithms Many optimization problems can be solved more quickly using a greedy approach The basic principle is that local optimal decisions may.
Chapter 16: Greedy algorithms Ming-Te Chi
Richard Anderson Lecture 16a Dynamic Programming
Advanced Algorithms Analysis and Design
Greedy Algorithms.
Advanced Algorithms Analysis and Design
Lecture 6 Topics Greedy Algorithm
Chapter 15-1 : Dynamic Programming I
Data Structures and Algorithms (AT70. 02) Comp. Sc. and Inf. Mgmt
Chapter 16: Greedy algorithms Ming-Te Chi
Analysis of Algorithms CS 477/677
Data Structure and Algorithms
Chapter 6 Dynamic Programming
ICS 353: Design and Analysis of Algorithms
Algorithms CSCI 235, Spring 2019 Lecture 29 Greedy Algorithms
DYNAMIC PROGRAMMING.
Greedy Algorithms Comp 122, Spring 2004.
Algorithm Design Techniques Greedy Approach vs Dynamic Programming
Analysis of Algorithms CS 477/677
Greedy algorithms.
Asst. Prof. Dr. İlker Kocabaş
Advance Algorithm Dynamic Programming
Dynamic Programming.
Analysis of Algorithms CS 477/677
Analysis of Algorithms CS 477/677
Presentation transcript:

Analysis of Algorithms CS 477/677 Instructor: Monica Nicolescu Lecture 17

Segmented Least Squares Foundational problem in statistic and numerical analysis Given n points in the plane: (x1, y1), (x2, y2) , . . . , (xn, yn) Find a line y = ax + b that minimizes the sum of the squared error: Solution – closed form Minimum error is achieved when y the term "regression" stems from the work of Sir Francis Galton (1822-1911), a famous geneticist, who studied the heights of fathers and their sons. He found that offspring of a parent of larger than average height tended to be smaller than their parents, and offspring of parents of smaller than average size tended to be larger than their parents. "regression toward mediocrity". Gauss (1809-1811) used Gaussian elimination to solve multiple linear regression Gauss-Markov theorem (least squares estimates are uniformly minimum variance estimates (UMVUE) among all linear unbiased estimates) x CS 477/677 - Lecture 17

Segmented Least Squares Points lie roughly on a sequence of several line segments Given n points in the plane (x1, y1), (x2, y2) , . . . , (xn, yn) with x1 < x2 < ... < xn, find a sequence of lines that minimizes f(x) What is a reasonable choice for f(x) to balance accuracy and parsimony? y goodness of fit number of lines x CS 477/677 - Lecture 17

Segmented Least Squares Points lie roughly on a sequence of several line segments Given n points in the plane (x1, y1), (x2, y2) , . . . , (xn, yn) with x1 < x2 < ... < xn, find a sequence of lines that minimizes: the sum of the sums of the squared errors E in each segment the number of lines L Tradeoff function: E + c L, for some constant c > 0 y CS 477/677 - Lecture 17 x

(1,2) Making the Choice and Recursive Solution Notation OPT(j) = minimum cost for points p1, pi+1 , . . . , pj e(i, j) = minimum sum of squares for points pi, pi+1, .. , pj To compute OPT(j) Last segment uses points pi, pi+1 , . . . , pj for some i Cost = e(i, j) + c + OPT(i-1) CS 477/677 - Lecture 17

3. Compute the Optimal Value INPUT: n, p1,…,pN , c Segmented-Least-Squares() { M[0] = 0 for j = 1 to n for i = 1 to j compute the least square error eij for the segment pi,…, pj M[j] = min 1 ≤ i ≤ j (eij + c + M[i-1]) return M[n] } Running time: O(n3) Bottleneck = computing e(i, j) for O(n2) pairs, O(n) per pair using previous formula CS 477/677 - Lecture 17

Greedy Algorithms Similar to dynamic programming, but simpler approach Also used for optimization problems Idea: When we have a choice to make, make the one that looks best right now Make a locally optimal choice in the hope of getting a globally optimal solution Greedy algorithms don’t always yield an optimal solution When the problem has certain general characteristics, greedy algorithms give optimal solutions CS 477/677 - Lecture 17 7

Activity Selection Problem Schedule the largest possible set of non-overlapping activities for a given room Start End Activity 1 8:00am 9:15am Numerical methods class 2 8:30am 10:30am Movie presentation (refreshments served) 3 9:20am 11:00am Data structures class 4 10:00am noon Programming club mtg. (Pizza provided) 5 11:30am 1:00pm Computer graphics class 6 1:05pm 2:15pm Analysis of algorithms class 7 2:30pm 3:00pm Computer security class 8 4:00pm Computer games contest (refreshments served) 9 5:30pm Operating systems class CS 477/677 - Lecture 17 8

Activity Selection Schedule n activities that require exclusive use of a common resource S = {a1, . . . , an} – set of activities ai needs resource during period [si , fi) si = start time and fi = finish time of activity ai 0 ≤ si < fi < ∞ Activities ai and aj are compatible if the intervals [si , fi) and [sj, fj) do not overlap fi ≤ sj fj ≤ si i j j i CS 477/677 - Lecture 17

Activity Selection Problem Select the largest possible set of non-overlapping (compatible) activities. E.g.: i 1 2 3 4 5 6 7 8 9 10 11 si 12 fi 13 14 Activities are sorted in increasing order of finish times A subset of mutually compatible activities: {a3, a9, a11} Maximal set of mutually compatible activities: {a1, a4, a8, a11} and {a2, a4, a9, a11} CS 477/677 - Lecture 17

Optimal Substructure Define the space of subproblems: Sij = { ak ∈ S : fi ≤ sk < fk ≤ sj } activities that start after ai finishes and finish before aj starts Add fictitious activities a0 = [-∞, 0) an+1 = [∞, “∞ + 1”) Range for Sij is 0 ≤ i, j ≤ n + 1 S = S0,n+1 entire space of activities CS 477/677 - Lecture 17

Representing the Problem We assume that activities are sorted in increasing order of finish times: f0 ≤ f1 ≤ f2 ≤ … ≤ fn < fn+1 What happens to set Sij for i ≥ j ? For an activity ak ∈ Sij: fi ≤ sk < fk ≤ sj < fj contradiction with fi ≥ fj! ⇒ Sij = ∅ (the set Sij must be empty!) We only need to consider sets Sij with 0 ≤ i < j ≤ n + 1 CS 477/677 - Lecture 17

Optimal Substructure Subproblem: Select a maximum-size subset of mutually compatible activities from set Sij Assume that a solution to the above subproblem includes activity ak (Sij is non-empty) Solution to Sij = (Solution to Sik) ⋃ {ak} ⋃ (Solution to Skj) ⎥Solution to Sij⎥= Sij Sik Skj ⎥Solution to Sik⎥+ 1 + ⎥Solution to Skj⎥ CS 477/677 - Lecture 17

Optimal Substructure Aij = Optimal solution to Sij Aik Akj Aij = Optimal solution to Sij Claim: Sets Aik and Akj must be optimal solutions Assume ∃ Aik’ that includes more activities than Aik Size[Aij’] = Size[Aik’] + 1 + Size[Akj] > Size[Aij] ⇒ Contradiction: we assumed that Aij has the maximum # of activities taken from Sij CS 477/677 - Lecture 17

Recursive Solution Any optimal solution (associated with a set Sij) contains within it optimal solutions to subproblems Sik and Skj c[i, j] = size of maximum-size subset of mutually compatible activities in Sij If Sij = ∅ ⇒ c[i, j] = 0 CS 477/677 - Lecture 17

Recursive Solution Sij Sik Skj If Sij ≠ ∅ and if we consider that ak is used in an optimal solution (maximum-size subset of mutually compatible activities of Sij), then: c[i, j] = c[i,k] + c[k, j] + 1 CS 477/677 - Lecture 17

Recursive Solution 0 if Sij = ∅ c[i, j] = max {c[i,k] + c[k, j] + 1} if Sij ≠ ∅ There are j – i – 1 possible values for k k = i+1, …, j – 1 ak cannot be ai or aj (from the definition of Sij) Sij = { ak ∈ S : fi ≤ sk < fk ≤ sj } We check all the values and take the best one We could now write a dynamic programming algorithm i < k < j ak ∈ Sij CS 477/677 - Lecture 17

Theorem Let Sij ≠ ∅ and am the activity in Sij with the earliest finish time: fm = min { fk: ak ∈ Sij } Then: am is used in some maximum-size subset of mutually compatible activities of Sij There exists some optimal solution that contains am Sim = ∅ Choosing am leaves Smj the only nonempty subproblem CS 477/677 - Lecture 17

Proof Assume ∃ ak ∈ Sim fi ≤ sk < fk ≤ sm < fm ⇒ fk < fm contradiction ! am must have the earliest finish time ⇒ There is no ak ∈ Sim ⇒ Sim = ∅ Sij sm fm ak am Sim Smj CS 477/677 - Lecture 17

: Greedy Choice Property Proof : Greedy Choice Property am is used in some maximum-size subset of mutually compatible activities of Sij Aij = optimal solution for activity selection from Sij Order activities in Aij in increasing order of finish time Let ak be the first activity in Aij = {ak, …} If ak = am Done! Otherwise, replace ak with am (resulting in a set Aij’) since fm ≤ fk the activities in Aij’ will continue to be compatible Aij’ will have the same size as Aij ⇒ am is used in some maximum-size subset Sij sm fm am fk CS 477/677 - Lecture 17

Why is the Theorem Useful? Dynamic programming Using the theorem Number of subproblems in the optimal solution Number of choices to consider 2 subproblems: Sik, Skj 1 subproblem: Smj (Sim = ∅) 1 choice: the activity am with the earliest finish time in Sij j – i – 1 choices Making the greedy choice (the activity with the earliest finish time in Sij) Reduces the number of subproblems and choices Allows solving each subproblem in a top-down fashion Only one subproblem left to solve! CS 477/677 - Lecture 17

Greedy Approach To select a maximum-size subset of mutually compatible activities from set Sij: Choose am ∈ Sij with earliest finish time (greedy choice) Add am to the set of activities used in the optimal solution Solve the same problem for the set Smj From the theorem By choosing am we are guaranteed to have used an activity included in an optimal solution ⇒ We do not need to solve the subproblem Smj before making the choice! The problem has the GREEDY CHOICE property CS 477/677 - Lecture 17

Characterizing the Subproblems The original problem: find the maximum subset of mutually compatible activities for S = S0, n+1 Activities are sorted by increasing finish time a0, a1, a2, a3, …, an+1 We always choose an activity with the earliest finish time Greedy choice maximizes the unscheduled time remaining Finish time of activities selected is strictly increasing CS 477/677 - Lecture 17 23

A Recursive Greedy Algorithm am fm Alg.: REC-ACT-SEL (s, f, i, n) m ← i + 1 while m ≤ n and sm < fi ►Find first activity in Si,n+1 do m ← m + 1 if m ≤ n then return {am} ⋃ REC-ACT-SEL(s, f, m, n) else return ∅ Activities are ordered in increasing order of finish time Running time: Θ(n) – each activity is examined only once Initial call: REC-ACT-SEL(s, f, 0, n) am fm am fm ai fi CS 477/677 - Lecture 17 24

Example k sk fk 0 -∞ 0 1 1 4 2 3 5 3 0 6 4 5 7 5 3 8 6 5 9 7 6 10 8 8 11 9 8 12 10 2 13 11 12 14 12 ∞ ∞+1 a1 m=1 a0 a2 a1 a3 a1 a4 m=4 a1 a5 a1 a4 a6 a1 a4 a7 a1 a4 a8 a1 a4 m=8 a9 a1 a4 a8 a10 a1 a4 a8 a11 a1 a4 a8 m=11 a1 a4 a8 a11 1 2 CS 477/677 - Lecture 18 CS 477/677 - Lecture 17 3 4 5 6 7 8 9 10 25 11 12 13 14 25

An Incremental Algorithm Alg.: GREEDY-ACTIVITY-SELECTOR(s, f, n) A ← {a1} i ← 1 for m ← 2 to n do if sm ≥ fi ► activity am is compatible with ai then A ← A ⋃ {am} i ← m ► ai is most recent addition to A return A Assumes that activities are ordered in increasing order of finish time Running time: Θ(n) – each activity is examined only once am fm am fm am fm ai fi CS 477/677 - Lecture 17

Steps Toward Our Greedy Solution Determined the optimal substructure of the problem Developed a recursive solution Proved that one of the optimal choices is the greedy choice Showed that all but one of the subproblems resulted by making the greedy choice are empty Developed a recursive algorithm that implements the greedy strategy Converted the recursive algorithm to an iterative one CS 477/677 - Lecture 17

Designing Greedy Algorithms Cast the optimization problem as one for which: we make a (greedy) choice and are left with only one subproblem to solve Prove the GREEDY CHOICE property: that there is always an optimal solution to the original problem that makes the greedy choice Prove the OPTIMAL SUBSTRUCTURE: the greedy choice + an optimal solution to the resulting subproblem leads to an optimal solution CS 477/677 - Lecture 17

Correctness of Greedy Algorithms Greedy Choice Property A globally optimal solution can be arrived at by making a locally optimal (greedy) choice Optimal Substructure Property We know that we have arrived at a subproblem by making a greedy choice Optimal solution to subproblem + greedy choice ⇒ optimal solution for the original problem CS 477/677 - Lecture 17

Readings Chapters 14, 15 CS 477/677 - Lecture 17