In the paper Figure 7, we claimed: “the best value for R depends on the amount of training data available.” Here are the results for Gun-Point Dataset.

Slides:



Advertisements
Similar presentations
Is Random Model Better? -On its accuracy and efficiency-
Advertisements

Random Forest Predrag Radenković 3237/10
Model generalization Test error Bias, variance and complexity
P449. p450 Figure 15-1 p451 Figure 15-2 p453 Figure 15-2a p453.
More Methodology; Nearest-Neighbor Classifiers Sec 4.7.
1 Fair Use Agreement This agreement covers the use of this presentation, please read carefully. You may freely use these slides for teaching, if You send.
CBF Dataset Two-Pat Dataset Euclidean DTW Increasingly Large Training.
Distinguishing Photographic Images and Photorealistic Computer Graphics Using Visual Vocabulary on Local Image Edges Rong Zhang,Rand-Ding Wang, and Tian-Tsong.
Evaluation.
Three kinds of learning
Steep learning curves Reading: DH&S, Ch 4.6, 4.5.
05/06/2005CSIS © M. Gibbons On Evaluating Open Biometric Identification Systems Spring 2005 Michael Gibbons School of Computer Science & Information Systems.
DATA MINING AND MACHINE LEARNING Addison Euhus and Dan Weinberg.
Graphing Examples Categorical Variables
Module 04: Algorithms Topic 07: Instance-Based Learning
GA-Based Feature Selection and Parameter Optimization for Support Vector Machine Cheng-Lung Huang, Chieh-Jen Wang Expert Systems with Applications, Volume.
GATree: Genetically Evolved Decision Trees 전자전기컴퓨터공학과 데이터베이스 연구실 G 김태종.
Categorical data. Decision Tree Classification Which feature to split on? Try to classify as many as possible with each split (This is a good split)
Today Ensemble Methods. Recap of the course. Classifier Fusion
Stefan Mutter, Mark Hall, Eibe Frank University of Freiburg, Germany University of Waikato, New Zealand The 17th Australian Joint Conference on Artificial.
Semi-Supervised Time Series Classification & DTW-D REPORTED BY WANG YAWEN.
Randomization in Privacy Preserving Data Mining Agrawal, R., and Srikant, R. Privacy-Preserving Data Mining, ACM SIGMOD’00 the following slides include.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
START OF DAY 5 Reading: Chap. 8. Support Vector Machine.
CS685 : Special Topics in Data Mining, UKY The UNIVERSITY of KENTUCKY Validation.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 5: Credibility: Evaluating What’s Been Learned.
Fall Detection in Homes of Older Adults Using the Microsoft Kinect ERIK E. STONE AND MARJORIE SKUBIC IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS,
Mole Quantity Conversions Grams  Moles  Atoms. First, fold your paper in half like this. Not this.
CSCI 347, Data Mining Evaluation: Cross Validation, Holdout, Leave-One-Out Cross Validation and Bootstrapping, Sections 5.3 & 5.4, pages
By: Will Schubert. Determine how the height you drop a ball from affects the height the ball bounces back up. Problem.
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Computational Intelligence: Methods and Applications Lecture 33 Decision Tables & Information Theory Włodzisław Duch Dept. of Informatics, UMK Google:
Week8 Fatemeh Yazdiananari.  Fixed the issues with classifiers  We retrained SVMs with the new UCF101 histograms  On temporally untrimmed videos: ◦
Bell Work: 8/25/15 Students want to determine which type of plant will best survive in a hot, dry environment. They place four different types of plants.
Still Life Drawing Today you will be using your sketchbook.
Next, this study employed SVM to classify the emotion label for each EEG segment. The basic idea is to project input data onto a higher dimensional feature.
Ensemble Learning, Boosting, and Bagging: Scaling up Decision Trees (with thanks to William Cohen of CMU, Michael Malohlava of 0xdata, and Manish Amde.
Mustafa Gokce Baydogan, George Runger and Eugene Tuv INFORMS Annual Meeting 2011, Charlotte A Bag-of-Features Framework for Time Series Classification.
Planting Science How does light intensity affect photosynthesis?
51 ttl What are the chances that a person picked at random will have a score of 9 or above?
Data Science Credibility: Evaluating What’s Been Learned
JMP Discovery Summit 2016 Janet Alvarado
Machine Learning for the Quantified Self
Predicting E. Coli Promoters Using SVM
Statistical Techniques
Supervised Learning Seminar Social Media Mining University UC3M
Urban Sound Classification with a Convolution Neural Network
Introduction Feature Extraction Discussions Conclusions Results
Project 2 k-NN 2018/11/10.
Evaluation and Its Methods
Discriminative Frequent Pattern Analysis for Effective Classification
iSRD Spam Review Detection with Imbalanced Data Distributions
Mole Quantity Conversions
WISDM Activity Recognition & Biometrics Applications of Classification
Mole Quantity Conversions
CSCI N317 Computation for Scientific Applications Unit Weka
Predictors Observed rainfall up to May.
Cross-validation Brenda Thomson/ Peter Fox Data Analytics
Ensemble learning Reminder - Bagging of Trees Random Forest
Model generalization Brief summary of methods
Machine Learning in Practice Lecture 27
Classification with CART
Analysis for Predicting the Selling Price of Apartments Pratik Nikte
COMP6321 MACHINE LEARNING PROJECT PRESENTATION
Graphing & Statistics Honors Biology.
Symmetry.
Evaluation and Its Methods
Happiness and Stocks Ali Javed, Tim Stevens
… 1 2 n A B V W C X 1 2 … n A … V … W … C … A X feature 1 feature 2
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
Presentation transcript:

In the paper Figure 7, we claimed: “the best value for R depends on the amount of training data available.” Here are the results for Gun-Point Dataset and another dataset, Two_ Pat, which we randomly pick half size instances repeatedly. The observation is that with fewer objects in the dataset, the accuracy decreases and peaks at larger window size. Gun PointTwo_Pat

Name# class# features# instancesEvaluationData type JF2220,0002,000/18,000real Letter261620,0005,000/15,000mixed Pen Digits101610,9927,494/3,498real Forest Cover Type754581,01211,340/569,672real Iris fold CVreal Ionosphere fold CVreal Voting Records fold CVBoolean Australian Credit fold CV6 numerial/8 categorical German Credit2241,00010-fold CVreal Leaf /242time series Two_Pat41285,0001,000/4,000time series Face161312,2311,113/1,118time series In the paper Table 4, we list the datasets used in the paper, here we present additional datasets and show all the experiments that could not fit in the paper due the limit of space.

Number of instances seen before interruption, S accuracy(%) Random Train Random Test SimpleRank Train SimpleRank Test Number of instances seen before interruption, S accuracy(%) Random Test SimpleRank Test JF, 2 classed, 20,000 instances, 2,000/18,000

Number of instances seen before interruption, S accuracy(%) Random Train Random Test SimpleRank Train SimpleRank Test Letter, 26 classes, 20,000 instances, 5,000/15, Number of instances seen before interruption, S accuracy(%) Random Test SimpleRank Test

Number of instances seen before interruption, S accuracy(%) Random Train Random Test SimpleRank Train SimpleRank Test Number of instances seen before interruption, S accuracy(%) Random Test SimpleRank Test Pen digits, 10 classed, 10,992 instances, 7,494/3,498

Number of instances seen before interruption, S accuracy(%) Random Train Random Test SimpleRank Train SimpleRank Test Number of instances seen before interruption, S accuracy(%) Random Test SimpleRank Test Forest Cover Type, 7 classes, 581,012 instances, 11,340/569,672

Number of instances seen before interruption, S accuracy(%) Random Test SimpleRank Test Australian Credit, 2 classes, 690 instances, 10-fold Cross Validation Number of instances seen before interruption, S data instances accuracy(%) RandomTrain RandomTest SimpleRankTrain SimpleRankTest DROP1 DROP2 DROP3

Number of instances seen before interruption, S accuracy(%) Random Train Random Test SimpleRank Train SimpleRank Test Number of instances seen before interruption, S accuracy(%) Random Test SimpleRank Test Number of instances seen before interruption, S accuracy(%) Random Test SimpleRank Test BestDrop Test data instances accuracy(%) RandomTrain RandomTest SimpleRankTrain SimpleRankTest DROP1 DROP2 DROP3 Ionosphere, 2 classes, 351 instances, 10-fold Cross Validation

Iris, 3 classes, 150 instances, 10-fold Cross Validation Number of instances seen before interruption, S accuracy(%) Random Train Random Test SimpleRank Train SimpleRank Test Number of instances seen before interruption, S accuracy(%) Random Test SimpleRank Test data instances accuracy(%) RandomTrain RandomTest SimpleRankTrain SimpleRankTest DROP1 DROP2 DROP3

Number of instances seen before interruption, S accuracy(%) Random Train Random Test SimpleRank Train SimpleRank Test Number of instances seen before interruption, S accuracy(%) Random Test SimpleRank Test Voting records Number of instances seen before interruption, S accuracy(%) Random Test SimpleRank Test BestDrop Test data instances accuracy(%) RandomTrain RandomTest SimpleRankTrain SimpleRankTest DROP1 DROP2 DROP3

Number of instances seen before interruption, S accuracy(%) Random Train Random Test SimpleRank Train SimpleRank Test Number of instances seen before interruption, S accuracy(%) Random Test SimpleRank Test German Credit, 2 classes, 1,000 instances, 10-fold Cross Validation data instances accuracy(%) RandomTrain RandomTest SimpleRankTrain SimpleRankTest DROP1 DROP2 DROP3

Two_Pat, 4 classes, 5,000 instances, 1,000/4,000 split

Leaf Dataset, 6 classes, 442 instances

Number of instances seen before interruption, accuracy(%) Face Dataset Random, Euclidean distance Random, Fixed R = 4 SimpleRank, Fixed R = 4 SimpleRank, AdaptiveR Random, Euclidean distance Random, Fixed R = 3 SimpleRank, Fixed R = 3 SimpleRank, AdaptiveR 3% 4% Face dataset, 16 classes, 2,231 instances, 1,113/1,118 split