Ensemble Based Systems in Decision Making Advisor: Hsin-His Chen Reporter: Chi-Hsin Yu Date: 2008.07.02 IEEE CIRCUITS AND SYSTEMS MAGAZINE 2006, Q3 Robi.

Slides:



Advertisements
Similar presentations
Random Forest Predrag Radenković 3237/10
Advertisements

CHAPTER 2: Supervised Learning. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Learning a Class from Examples.
My name is Dustin Boswell and I will be presenting: Ensemble Methods in Machine Learning by Thomas G. Dietterich Oregon State University, Corvallis, Oregon.
Combining Classification and Model Trees for Handling Ordinal Problems D. Anyfantis, M. Karagiannopoulos S. B. Kotsiantis, P. E. Pintelas Educational Software.
No Free Lunch Theorem Suppose we make no prior assumptions about the nature of the classification task. Can we expect any classification method to be superior.
Data Mining Classification: Alternative Techniques
Ensemble Methods An ensemble method constructs a set of base classifiers from the training data Ensemble or Classifier Combination Predict class label.
Paper presentation for CSI5388 PENGCHENG XI Mar. 23, 2005
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Decision Tree Algorithm
Ensemble Learning: An Introduction
Three kinds of learning
Bagging LING 572 Fei Xia 1/24/06. Ensemble methods So far, we have covered several learning methods: FSA, HMM, DT, DL, TBL. Question: how to improve results?
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Machine Learning: Ensemble Methods
Rotation Forest: A New Classifier Ensemble Method 交通大學 電子所 蕭晴駿 Juan J. Rodríguez and Ludmila I. Kuncheva.
Ensemble Learning (2), Tree and Forest
For Better Accuracy Eick: Ensemble Learning
3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Wei FENG Geo-Resources and Environment Lab, Bordeaux INP (Bordeaux Institute of Technology), France Supervisor:
Ensembles of Classifiers Evgueni Smirnov
Machine Learning CS 165B Spring 2012
Predicting Income from Census Data using Multiple Classifiers Presented By: Arghya Kusum Das Arnab Ganguly Manohar Karki Saikat Basu Subhajit Sidhanta.
Bias and Variance Two ways to measure the match of alignment of the learning algorithm to the classification problem involve the bias and variance. Bias.
CS 391L: Machine Learning: Ensembles
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
ECE 8443 – Pattern Recognition Objectives: Bagging and Boosting Cross-Validation ML and Bayesian Model Comparison Combining Classifiers Resources: MN:
1 Part II: Practical Implementations.. 2 Modeling the Classes Stochastic Discrimination.
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
ECE738 Advanced Image Processing Face Detection IEEE Trans. PAMI, July 1997.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Ensemble Methods: Bagging and Boosting
Ensembles. Ensemble Methods l Construct a set of classifiers from training data l Predict class label of previously unseen records by aggregating predictions.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
Ensemble Based Systems in Decision Making Yu-Mei, Chang Speech Lab, CSIE National Taiwan Normal University Robi Polikar Third Quarter 2006 IEEE circuits.
CLASSIFICATION: Ensemble Methods
Classifier ensembles: Does the combination rule matter? Ludmila Kuncheva School of Computer Science Bangor University, UK
Stefan Mutter, Mark Hall, Eibe Frank University of Freiburg, Germany University of Waikato, New Zealand The 17th Australian Joint Conference on Artificial.
Learning with AdaBoost
Ensemble Methods.  “No free lunch theorem” Wolpert and Macready 1995.
Ensemble Methods in Machine Learning
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Occam’s Razor No Free Lunch Theorem Minimum.
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Classification Ensemble Methods 1
1 January 24, 2016Data Mining: Concepts and Techniques 1 Data Mining: Concepts and Techniques — Chapter 7 — Classification Ensemble Learning.
Classification and Prediction: Ensemble Methods Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Finding τ → μ−μ−μ+ Decays at LHCb with Data Mining Algorithms
Validation methods.
Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn.
… Algo 1 Algo 2 Algo 3 Algo N Meta-Learning Algo.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Bagging and Boosting Cross-Validation ML.
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
Computational Intelligence: Methods and Applications Lecture 15 Model selection and tradeoffs. Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Mining Concept-Drifting Data Streams Using Ensemble Classifiers Haixun Wang Wei Fan Philip S. YU Jiawei Han Proc. 9 th ACM SIGKDD Internal Conf. Knowledge.
1 Machine Learning Lecture 8: Ensemble Methods Moshe Koppel Slides adapted from Raymond J. Mooney and others.
Machine Learning: Ensemble Methods
Machine Learning: Ensembles
Ensemble Learning Introduction to Machine Learning and Data Mining, Carla Brodley.
Neuro-Computing Lecture 5 Committee Machine
A “Holy Grail” of Machine Learing
Combining Base Learners
Introduction to Data Mining, 2nd Edition
Machine Learning Ensemble Learning: Voting, Boosting(Adaboost)
Ensemble learning.
Model Combination.
Ensemble learning Reminder - Bagging of Trees Random Forest
INTRODUCTION TO Machine Learning 3rd Edition
CS 391L: Machine Learning: Ensembles
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
Presentation transcript:

Ensemble Based Systems in Decision Making Advisor: Hsin-His Chen Reporter: Chi-Hsin Yu Date: IEEE CIRCUITS AND SYSTEMS MAGAZINE 2006, Q3 Robi Polikar, Dept. of Electrical and Computer Engineering (ECE) at Rowan

Outlines Introduction Ensemble Based Systems Creating an Ensemble Combining Classifiers/Learners Current & Emerging Areas Software Conclusions

Introduction (1) Ensemble system ◦ Multiple classifiers system ◦ Meta-learning algorithm

Introduction (2) Reasons for using ensemble based systems ◦ Statistical reasons ◦ Large volumes of data ◦ Too little data ◦ Divide and conquer ◦ Data fusion

Ensemble Based Systems (1) Two major steps for creating ensemble systems ◦ The specific procedure used for generating individual classifiers  Partitioning feature space (subspace)  Partitioning data/resampling data ◦ The strategy employed for combining the classifiers  Classifier selection  Classifier fusion

Ensemble Based Systems (2) Diversity, the cornerstone of ensemble systems ◦ To make each classifier as unique as possible  Individual classifiers make errors on different instances.  Decision boundaries are adequately different from those of others.  Different classifiers can handle different kind of instances – general instances, informative instances. ◦ The methods to archive diversity  Use different training datasets  Use different training parameters  Use different feature subsets

Ensemble Based Systems (3) Popular data split techniques ◦ Bootstrap resampling ◦ jackknife (k-fold data split)  Ex. in cross validation

Ensemble Based Systems (4) Measures of diversity ◦ Pair-wise measures, defined between two classifiers  For T classifiers, we can calculate T(T-1)/2 pair-wise diversity measures ◦ Given two hypotheses hi and hj,

Ensemble Based Systems (5) But … ◦ There is no diversity measure that consistently correlates with higher accuracy. No Free Lunch Theorem for optimization (D.H. Wolpert and W.G. Macready, 1997) ◦ No matter what algorithm we use, there is at least one target function for which random guessing is a better algorithm. ◦ Learning algorithm -> Turing machine -> language -> distribution So … ◦ Prior knowledge, data distribution, amount of training data, cost function are important. ◦ Try another algorithms if the performance is unacceptable.

Creating an Ensemble (1) Bootstrap aggregation

Creating an Ensemble (2) Adaptive boosting

Creating an Ensemble (3)

Creating an Ensemble (4) For certain instance, not for a dataset

Creating an Ensemble (5) Classifier selection Dynamic combination rules

Combining Classifiers (1) Combination rules are often grouped as ◦ Trainable vs. non-trainable ◦ Combination rules that apply  to class labels  to class-specific continuous outputs

Combining Classifiers (2) Combining class labels ◦ Majority voting  Unanimous voting  Simple majority ( > 50% )  Plurality voting (majority voting) ◦ Weighted majority voting ◦ Borda count  Devised in 1770 by Jean Charles de Borda  Each voter (classifier) rank-orders the candidates  (1 st,2 nd,…,N th ) place -> (N-1, N-2, …, 0) votes ◦ Behavior knowledge space (BKS)

Combining Classifiers (3)

Combining Classifiers (4) Combining continuous outputs C: Class number

Combining Classifiers (5) Approaches to combine continuous outputs ◦ Algebraic combiners  Mean rule, weighted average, trimmed mean, min/max/median rule, product rule, generalized means ◦ Decision templates ◦ Dempster-Shafer-based combination But, which one is better? ◦ No Free Lunch Theorem

Current & Emerging Areas Incremental learning Data fusion Feature selection Confidence estimation Error correcting output codes (ECOC) Output [ ] is closest to ω 5 code word with a Hamming distance of 1.

Software Entool ◦ A Matlab Toolbox for Ensemble Modeling PRTools ◦ Weka ◦ Java, Open source (GNU license) RapidMiner (formerly YALE) ◦ Java, Open source (GNU/OEM license)

Conclusions Ensemble-based systems ◦ It Enjoyed a growing attention. ◦ Its diversity is important. ◦ “No Free Lunch Theorem”

Thanks!!