Presentation is loading. Please wait.

Presentation is loading. Please wait.

The Rate of Convergence of AdaBoost Indraneel Mukherjee Cynthia Rudin Rob Schapire.

Similar presentations


Presentation on theme: "The Rate of Convergence of AdaBoost Indraneel Mukherjee Cynthia Rudin Rob Schapire."— Presentation transcript:

1 The Rate of Convergence of AdaBoost Indraneel Mukherjee Cynthia Rudin Rob Schapire

2 AdaBoost (Freund and Schapire 97)

3

4 Basic properties of AdaBoost’s convergence are still not fully understood.

5 We address one of these basic properties: convergence rates with no assumptions.

6 AdaBoost is known for its ability to combine “weak classifiers” into a “strong” classifier AdaBoost iteratively minimizes “exponential loss” (Breiman 99, Frean and Downs, 1998; Friedman et al., 2000; Friedman, 2001; Mason et al., 2000; Onoda et al., 1998; Ratsch et al., 2001; Schapire and Singer, 1999)

7 AdaBoost is known for its ability to combine “weak classifiers” into a “strong” classifier AdaBoost iteratively minimizes “exponential loss” (Breiman 99, Frean and Downs, 1998; Friedman et al., 2000; Friedman, 2001; Mason et al., 2000; Onoda et al., 1998; Ratsch et al., 2001; Schapire and Singer, 1999)

8 AdaBoost is known for its ability to combine “weak classifiers” into a “strong” classifier AdaBoost iteratively minimizes “exponential loss” (Breiman 99, Frean and Downs, 1998; Friedman et al., 2000; Friedman, 2001; Mason et al., 2000; Onoda et al., 1998; Ratsch et al., 2001; Schapire and Singer, 1999)

9 AdaBoost is known for its ability to combine “weak classifiers” into a “strong” classifier AdaBoost iteratively minimizes “exponential loss” (Breiman 99, Frean and Downs, 1998; Friedman et al., 2000; Friedman, 2001; Mason et al., 2000; Onoda et al., 1998; Ratsch et al., 2001; Schapire and Singer, 1999)

10 AdaBoost is known for its ability to combine “weak classifiers” into a “strong” classifier AdaBoost iteratively minimizes “exponential loss” (Breiman 99, Frean and Downs, 1998; Friedman et al., 2000; Friedman, 2001; Mason et al., 2000; Onoda et al., 1998; Ratsch et al., 2001; Schapire and Singer, 1999) Exponential loss:

11

12

13 Known: AdaBoost converges asymptotically to the minimum of the exponential loss (Collins et al 2002, Zhang and Yu 2005) Convergence rates under strong assumptions: “weak learning” assumption holds, hypotheses are better than random guessing (Freund and Schapire 1997, Schapire and Singer 1999) assume that a finite minimizer exists (Rätsch et al 2002, many classic results) Conjectured by Schapire (2010) that fast convergence rates hold without any assumptions. Convergence rate is relevant for consistency of AdaBoost (Bartlett and Traskin 2007).

14 Known: AdaBoost converges asymptotically to the minimum of the exponential loss (Collins et al 2002, Zhang and Yu 2005) Convergence rates under assumptions: “weak learning” assumption holds, hypotheses are better than random guessing (Freund and Schapire 1997, Schapire and Singer 1999) assume that a finite minimizer exists (Rätsch et al 2002, many classic results) Conjectured by Schapire (2010) that fast convergence rates hold without any assumptions. Convergence rate is relevant for consistency of AdaBoost (Bartlett and Traskin 2007).

15 Known: AdaBoost converges asymptotically to the minimum of the exponential loss (Collins et al 2002, Zhang and Yu 2005) Convergence rates under assumptions: “weak learning” assumption holds, hypotheses are better than random guessing (Freund and Schapire 1997, Schapire and Singer 1999) assume that a finite minimizer exists (Rätsch et al 2002, many classic results) Conjectured by Schapire (2010) that fast convergence rates hold without any assumptions. Convergence rate is relevant for consistency of AdaBoost (Bartlett and Traskin 2007).

16 Known: AdaBoost converges asymptotically to the minimum of the exponential loss (Collins et al 2002, Zhang and Yu 2005) Convergence rates under assumptions: “weak learning” assumption holds, hypotheses are better than random guessing (Freund and Schapire 1997, Schapire and Singer 1999) assume that a finite minimizer exists (Rätsch et al 2002, many classic results) Conjectured by Schapire (2010) that fast convergence rates hold without any assumptions. Convergence rate is relevant for consistency of AdaBoost (Bartlett and Traskin 2007).

17 Outline Convergence Rate 1: Convergence to a target loss “Can we get within of a ‘reference’ solution?” Convergence Rate 2: Convergence to optimal loss “Can we get within of an optimal solution?”

18 Main Messages Usual approaches assume a finite minimizer – Much more challenging not to assume this! Separated two different modes of analysis – comparison to reference, comparison to optimal – different rates of convergence are possible in each Analysis of convergence rates often ignore the “constants” – we show they can be extremely large in the worst case

19 Convergence Rate 1: Convergence to a target loss “Can we get within of a “reference” solution?” Convergence Rate 2: Convergence to optimal loss “Can we get within of an optimal solution?” Based on a conjecture that says...

20

21 radius B 

22

23  

24  

25   This happens at:

26 radius B   This happens at:

27 radius B   This happens at:

28

29

30 Best known previous result is that it takes at most order rounds (Bickel et al).

31 Intuition behind proof of Theorem 1 Old fact: if AdaBoost takes a large step, it makes a lot of progress:

32 radius B    measures progress measures distance

33 Old Fact: First lemma says: Intuition behind proof of Theorem 1

34 Old Fact: First lemma says: Second lemma says: Combining: Intuition behind proof of Theorem 1

35 Lemma: There are simple datasets for which the number of rounds required to achieve loss L * is at least (roughly) the norm of the smallest solution achieving loss L *

36

37 Lemma: There are simple datasets for which the norm of the smallest solution achieving loss L * is exponential in the number of examples.

38

39

40

41 Number of rounds 10 100 1000 10000 1e+05 Rate on a Simple Dataset (Log scale)

42 Outline Convergence Rate 1: Convergence to a target loss “Can we get within of a “reference” solution?” Convergence Rate 2: Convergence to optimal loss “Can we get within of an optimal solution?”

43

44 Better dependence on than Theorem 1, actually optimal. Doesn’t depend on the size of the best solution within a ball Can’t be used to prove the conjecture because in some cases C>2 m. (Mostly it’s much smaller.)

45 Main tool is the “decomposition lemma” – Says that examples fall into 2 categories, Zero loss set Z Finite margin set F. – Similar approach taken independently by (Telgarsky, 2011)

46 + + + + + + + - - - - - + + + + - - - - - -

47 + + + + + + + - - - - - + + + + - - - - - - F

48 + + + + + + + - - - - - + + + + - - - - - - Z

49 For any dataset, there exists a partition of the training examples into Z and F s.t. these hold simultaneously: Decomposition Lemma

50 + + + + + + + - - - - - + + + + - - - - - -

51 For any dataset, there exists a partition of the training examples into Z and F s.t. these hold simultaneously: Decomposition Lemma

52 + + + + + + + - - - - - + + + + - - - - - - F

53 + + + + - - - - - - F

54 + + + + - - - - - - F

55 For any dataset, there exists a partition of the training examples into Z and F s.t. these hold simultaneously: Decomposition Lemma

56 We provide a conjecture about dependence on m. This would give optimal dependence on m and simultaneously.

57 To summarize Two rate bounds, one depends on the size of the best solution within a ball and has dependence. The other depends only on but C can be doubly exponential in m. Many lower bounds and conjectures in the paper.

58 To summarize Two rate bounds, one depends on the size of the best solution within a ball and has dependence. The other depends only on but C can be doubly exponential in m. Many lower bounds and conjectures in the paper. Thank you

59 Old Fact: First lemma says: Intuition behind proof of Theorem 2 Second lemma says:

60 Old Fact: First lemma says: Intuition behind proof of Theorem 2 Second lemma says:

61 see notes


Download ppt "The Rate of Convergence of AdaBoost Indraneel Mukherjee Cynthia Rudin Rob Schapire."

Similar presentations


Ads by Google