T-BAG: Bootstrap Aggregating the TAGE Predictor Ibrahim Burak Karsli, Resit Sendag University of Rhode Island.

Slides:



Advertisements
Similar presentations
André Seznec Caps Team IRISA/INRIA 1 Looking for limits in branch prediction with the GTL predictor André Seznec IRISA/INRIA/HIPEAC.
Advertisements

H-Pattern: A Hybrid Pattern Based Dynamic Branch Predictor with Performance Based Adaptation Samir Otiv Second Year Undergraduate Kaushik Garikipati Second.
A Statistician’s Games * : Bootstrap, Bagging and Boosting * Please refer to “Game theory, on-line prediction and boosting” by Y. Freund and R. Schapire,
Ensemble Methods An ensemble method constructs a set of base classifiers from the training data Ensemble or Classifier Combination Predict class label.
TAGE-SC-L Branch Predictors
Longin Jan Latecki Temple University
Autocorrelation and Linkage Cause Bias in Evaluation of Relational Learners David Jensen and Jennifer Neville.
MCS Multiple Classifier Systems, Cagliari 9-11 June Giorgio Valentini Random aggregated and bagged ensembles.
Ensemble Learning what is an ensemble? why use an ensemble?
A Brief Introduction to Adaboost
Ensemble Learning: An Introduction
Introduction to Boosting Aristotelis Tsirigos SCLT seminar - NYU Computer Science.
Bagging LING 572 Fei Xia 1/24/06. Ensemble methods So far, we have covered several learning methods: FSA, HMM, DT, DL, TBL. Question: how to improve results?
The moment generating function of random variable X is given by Moment generating function.
Examples of Ensemble Methods
Prefetching On-time and When it Works Sequential Prefetcher With Adaptive Distance (SPAD) Ibrahim Burak Karsli Mustafa Cavus
Machine Learning CS 165B Spring 2012
Particle Filtering in Network Tomography
Zhangxi Lin ISQS Texas Tech University Note: Most slides are from Decision Tree Modeling by SAS Lecture Notes 6 Ensembles of Trees.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
CS 391L: Machine Learning: Ensembles
1 A 64 Kbytes ITTAGE indirect branch predictor André Seznec INRIA/IRISA.
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Benk Erika Kelemen Zsolt
1 A New Case for the TAGE Predictor André Seznec INRIA/IRISA.
End-biased Samples for Join Cardinality Estimation Cristian Estan, Jeffrey F. Naughton Computer Sciences Department University of Wisconsin-Madison.
Ensembles. Ensemble Methods l Construct a set of classifiers from training data l Predict class label of previously unseen records by aggregating predictions.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
ISQS 6347, Data & Text Mining1 Ensemble Methods. ISQS 6347, Data & Text Mining 2 Ensemble Methods Construct a set of classifiers from the training data.
E NSEMBLE L EARNING : A DA B OOST Jianping Fan Dept of Computer Science UNC-Charlotte.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 5: Credibility: Evaluating What’s Been Learned.
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Classification Ensemble Methods 1
Data Analytics CMIS Short Course part II Day 1 Part 3: Ensembles Sam Buttrey December 2015.
1 January 24, 2016Data Mining: Concepts and Techniques 1 Data Mining: Concepts and Techniques — Chapter 7 — Classification Ensemble Learning.
Classification and Prediction: Ensemble Methods Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn.
André Seznec Caps Team IRISA/INRIA 1 A 256 Kbits L-TAGE branch predictor André Seznec IRISA/INRIA/HIPEAC.
Competition II: Springleaf Sha Li (Team leader) Xiaoyan Chong, Minglu Ma, Yue Wang CAMCOS Fall 2015 San Jose State University.
… Algo 1 Algo 2 Algo 3 Algo N Meta-Learning Algo.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Single Correlator Based UWB Receiver Implementation through Channel Shortening Equalizer By Syed Imtiaz Husain and Jinho Choi School of Electrical Engineering.
Regression Tree Ensembles Sergey Bakin. Problem Formulation §Training data set of N data points (x i,y i ), 1,…,N. §x are predictor variables (P-dimensional.
Temporal Stream Branch Predictor (TS Predictor) Yongming Shen, Michael Ferdman.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Matrix Multiplication in Hadoop
FAT predictor Sabareesh Ganapathy, Prasanna Venkatesh Srinivasan, Maribel Monica.
1 Machine Learning Lecture 8: Ensemble Methods Moshe Koppel Slides adapted from Raymond J. Mooney and others.
Multiperspective Perceptron Predictor Daniel A. Jiménez Department of Computer Science & Engineering Texas A&M University.
Combining Bagging and Random Subspaces to Create Better Ensembles
Ensemble Classifiers.
Machine Learning: Ensemble Methods
Multilayer Perceptron based Branch Predictor
Zaman Faisal Kyushu Institute of Technology Fukuoka, JAPAN
Multiperspective Perceptron Predictor with TAGE
Chapter 13 – Ensembles and Uplift
COMP61011 : Machine Learning Ensemble Models
Ensemble Learning Introduction to Machine Learning and Data Mining, Carla Brodley.
ECE 5424: Introduction to Machine Learning
Ungraded quiz Unit 6.
A “Holy Grail” of Machine Learing
Combining Base Learners
Looking for limits in branch prediction with the GTL predictor
Direct or Remotely sensed
Scaled Neural Indirect Predictor
Ensembles.
TAGE-SC-L Again MTAGE-SC
5th JILP Workshop on Computer Architecture Competitions
Ensemble learning Reminder - Bagging of Trees Random Forest
CS639: Data Management for Data Science
Presentation transcript:

T-BAG: Bootstrap Aggregating the TAGE Predictor Ibrahim Burak Karsli, Resit Sendag University of Rhode Island

Bootstrap Aggregating Statistical method introduced by Breiman in 1996 Use ensemble of predictors –sub-predictors could be the same or different Train each slightly differently and independently Each predictor trained with resampled (with replacement) data set (bootstrapping) Aggregate their predictions The IDEA is: Many weak learners make strong learner Theoretically proven to perform better than single learner in an ensemble

Offline Bagging

Online Bagging

TAGE Predictor Winner of CBP3 State-of-art branch predictor Many parameters to allow variety

T-BAG: Prediction x32 TAGE PC aggregation prediction

Predictor Aggregation Bagging in nature uses 10s to 100s of predictors, so we target unlimited track Submitted predictor uses 32 TAGE predictors Keep track of successes of last 16 predictions with a sliding window for each predictor Aggregate the predictions using weighted sum

PC & resolveDir T-BAG: Update x32 TAGE Update Count

Random Update Each predictor is updated on each sample k times in a row where k is a random number generated by multinomial distribution Max k = 2 (because ctr width is 3bits) For submission, update on each sample 20%, 60%, 20% of the time, 0, 1, 2 times, respectively.

Sub Predictors 32 predictors Variability in min/max history lengths, number of tables, and use of PC in table indexing ctr 3-bits for all Each predictor’s size is about 15MB (submitted predictor 492MB) Min history varies between 3 and 13 Max history varies between 1,200 and 100,000 Number of tables varies between 20 and predictors use PC, the other 16 do NOT! –Use of PC in indexing tables for TAGE-like predictor is not significantly better!

Results AllSame_RandUpd -> misp/KI AllDifferent -> misp/KI AllDisfferent_RandUpd -> misp/KI

misp/KI ConfigurationBaseline TAGE32x-size TAGET-Bag32 AMEAN

Conclusion and Future Work Simple idea Different types of predictors Implementation with storage budget

Q&A