Presentation is loading. Please wait.

Presentation is loading. Please wait.

Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Efficient Optimal Linear Boosting of a Pair of Classifiers.

Similar presentations


Presentation on theme: "Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Efficient Optimal Linear Boosting of a Pair of Classifiers."— Presentation transcript:

1 Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Efficient Optimal Linear Boosting of a Pair of Classifiers Victor Boyarshinov and Malik Magdon-Ismail, IEEE Transaction on neural networks, Vol. 18, No. 2, 2007, pp. 317-328. Presenter : Wei-Shen Tai Advisor : Professor Chung-Chian Hsu 2007/5/16

2 N.Y.U.S.T. I. M. Intelligent Database Systems Lab Outline Introduction Optimal fat separators in two dimension  Mirrored-radial coordinates  Maximizing the Margin Subject to Minimum Weight Leave-one-out error Discussions Comments

3 N.Y.U.S.T. I. M. Intelligent Database Systems Lab Motivation MODEL aggregation  Enhancing the statistical performance of a set of weak classifiers to obtain a stronger classifier.  For example, boosting and bagging. Combinatorial problem of boosting a pair of classifiers  What is the optimal linear combination of this pair of classifiers?

4 N.Y.U.S.T. I. M. Intelligent Database Systems Lab Objective Optimal linear separation (l)  the optimal linear boosted classifier for g 1 and g 2 classifier (for whatever optimal means) in terms of its corresponding optimal linear classifier in the 2-D space.

5 N.Y.U.S.T. I. M. Intelligent Database Systems Lab A (linearly) boosted classification function  A (linearly) boosted classification function g(z)  g(z) = w 0 +w 1 g 1 (z) + w 2 g 2 (z) The corresponding classifier is  sign(g(z)) = sign(w 0 +w 1 g 1 (z) + w 2 g 2 (z))  Each data point z i  d can be mapped onto this 2-D feature space by { z i, y i }  A linearly boosted classifier corresponds exactly to a linear separator in this 2-D space.

6 N.Y.U.S.T. I. M. Intelligent Database Systems Lab Optimal linear separation in  2 Efficient algorithms for exact linear separation  Applicable to the case where the data points are not linearly separable. Definition 1.2: Two sets A and B are linearly separable iff  v, v 0 such that v T x + v 0 > 0,  x  A and v T x + v 0 < 0,  x  B. The pair (v, v 0 ) defines an (oriented) separating hyperplane. If two sets and are linearly separable, the margin of a separating hyperplane is the minimum distance of a data point to the hyperplane. The maximum margin separating hyperplane is a separating hyperplane with maximum possible margin. The weight (or error) for is the total weight summed over the points misclassified by  A separator l* is optimal if it has minimum weight

7 N.Y.U.S.T. I. M. Intelligent Database Systems Lab Optimal fat separator Definition 1.3 (Optimal fat separator)  A hyperplane l = (v, v 0 ) is an optimal fat separator for A and B if it is optimal and is also a maximum margin separator for A’(l) and B’(l).  Analogously define the optimal fat separator with respect to the new separable set that would be obtained if instead of removing the misclassified points, we flip the classes of these points—all our results apply here as well. Definition 2.1: A separator set Q  A  B  A set with the following property: if the points in Q are deleted, the remaining points are linearly separable. Definition 2.4: For hyperplane l, the positive separator set Q + (l)  Contains all misclassified points except the positive points (in A) that lie on l.

8 N.Y.U.S.T. I. M. Intelligent Database Systems Lab The best positive separator hyperplane Lemma 2.6: The optimal positive separator set over all hyperplanes passing through a candidate central point. Mirrored-radial coordinate  A point s(x) as the projection of x onto the upper hemisphere of the unit circle, through the origin a +.  The mirrored-radial coordinate θ(x) is then the angle of s(x), i.e., θ (s(x)).  Find l with minimum weight

9 N.Y.U.S.T. I. M. Intelligent Database Systems Lab Maximizing the margin subject to minimum weight Definition 2.12: For the set A  B, the margin of separator set Q,mar(Q), is the margin of the optimal fat separator for A  B \ Q.  Theorem 2.13: An optimal separator set Q* (1) For any other separator set W(Q*) ≦ W(Q), (2) and if W(Q*) =W(Q), then mar(Q*) ≧ mar(Q).  Convex hull Convex hull

10 N.Y.U.S.T. I. M. Intelligent Database Systems Lab Convex hull Definition  A set of points X in a real vector space V is the minimal convex set containing X.  To show this exists, it is necessary to see that every X is contained in at least one convex set (the whole space V, for example), and any intersection of convex sets containing X is also a convex set containing X.

11 N.Y.U.S.T. I. M. Intelligent Database Systems Lab Leave-one-out error Estimation of the accuracy of learning algorithms  Let e i denote the error of C (i) applied to the input point x i  Three types of x i Type I: x i is classified correctly by all distinct optimal fat separators constructed for. Such an makes no contribution to the leave-one-out error. Type II: x i is misclassified by all optimal fat separators constructed for X (i). Such an x i contributes w i to the leave-one-out error. Type III: There are distinct optimal separator sets for X (i). Let N c of these occurrences result in fat separators that classify correctly and N e of them misclassify.

12 N.Y.U.S.T. I. M. Intelligent Database Systems Lab Discussions Optimal linear boosting of a pair of classification functions  A significant improvement over the brute force exponential algorithm.  A smarter way to enumerate these lines to result in a speed up of a factor close to n.  Extended to maximize the margin among all optimal separator set. Limitation and future work  Only applies to the 2-D case. This is a significant limitation for linear separation.  Extend the approach further to obtain an optimal boosting of an arbitrary number of classification functions.

13 N.Y.U.S.T. I. M. Intelligent Database Systems Lab Comments Advantage  A model aggregation method for combining different classifiers.  Mirrored-radial coordinates save the search space for finding optimal separator hyperplane. Drawback  It lacks extra diagrams to explain lemmas and theorems in detail.  There is no experiment to demonstrate the performance of this method with other related methods in this paper. Application  Model aggregation for classification related applications.


Download ppt "Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Efficient Optimal Linear Boosting of a Pair of Classifiers."

Similar presentations


Ads by Google