Machine Learning Applying Our Knowledge Gold Team Week 3: Random Forests and SVM Results.

Slides:



Advertisements
Similar presentations
Robo-Lab Team 5 Derek Evers, James Bettes, Keith Pearson, Amanda Brown, and Bryce Bell.
Advertisements

Artificial Grammar Learning (AGL) First developed by Reber in 1967 Standard procedure: Subjects are shown a series of letter strings that follow particular.
Classification of the aesthetic value of images based on histogram features By Xavier Clements & Tristan Penman Supervisors: Vic Ciesielski, Xiadong Li.
How fast does a raven or crow fly?
Lesson learnt from the UCSD datamining contest Richard Sia 2008/10/10.
Weekly Report Start learning GPU Ph.D. Student: Leo Lee date: Sep. 18, 2009.
U.S. SENATE BILL CLASSIFICATION & VOTE PREDICTION Alessandra Paulino Rick Pocklington Serhat Selcuk Bucak.
Team Question Mark Jeff Brown Dante Kappotis Robert Vanderley Anthony Biasella A little information on: HERPES presents.
Reduce Instrumentation Predictors Using Random Forests Presented By Bin Zhao Department of Computer Science University of Maryland May
Machine Learning: Final Presentation James Dalphond James McCauley Andrew Wilkinson Phil Kovac Data Set: Yeast GOLD TEAM.
CS Instance Based Learning1 Instance Based Learning.
LIMITS The Limit of a Function LIMITS Objectives: In this section, we will learn: Limit in general Two-sided limits and one-sided limits How to.
Difficulties with Nonlinear SVM for Large Problems  The nonlinear kernel is fully dense  Computational complexity depends on  Separating surface depends.
Source Code Basics. Code For a computer to execute instructions, it needs to be in binary Each instruction is given a number Known as “operation code”
Data Mining – Algorithms: OneR Chapter 4, Section 4.1.
Dynamically altering the learning trajectories of novices with pedagogical agents Carole R. Beal, USC Ronald H. Stevens, UCLA Cognition & Student Learning.
CHAPTER 2 LIMITS AND DERIVATIVES. 2.2 The Limit of a Function LIMITS AND DERIVATIVES In this section, we will learn: About limits in general and about.
Chasing those perfect Hash functions… by DANIEL “3ICE” BEREZVAI ELTE.3ICE.HU/ /ALG2/GY/ALG2 GY 2 HF.
REU WEEK 8 Nancy Zanaty, UCF. Past approach modified  Previously: Classifying individual images in a timeseries as “ADHD” or “Non ADHD” as a test of.
An Example of Course Project Face Identification.
Weka Project assignment 3
New Story of the Hare and Tortoise.
Why Categorize in Computer Vision ?. Why Use Categories? People love categories!
Data Mining – Algorithms: Prism – Learning Rules via Separating and Covering Chapter 4, Section 4.4.
1 Comparison of Principal Component Analysis and Random Projection in Text Mining Steve Vincent April 29, 2004 INFS 795 Dr. Domeniconi.
Algorithms and their Applications CS2004 ( ) Dr Stephen Swift 1.2 Introduction to Algorithms.
Weka: a useful tool in data mining and machine learning Team 5 Noha Elsherbiny, Huijun Xiong, and Bhanu Peddi.
2002/12/201 Tortoise-hare race Once upon a time a tortoise and a hare had an argument about who was faster. They decided to settle the argument with a.
MINING MULTI-LABEL DATA BY GRIGORIOS TSOUMAKAS, IOANNIS KATAKIS, AND IOANNIS VLAHAVAS Published on July, 7, 2010 Team Members: Kristopher Tadlock, Jimmy.
Physical Database Design The last phase of database design. It is to determine how to store the database. RDBMSs usually support a number of alternative.
Homework #1 J. H. Wang Oct. 5, 2015.
The Scientific Method QuestionHypothesizeExperimentConclusion.
Data Mining – Algorithms: Decision Trees - ID3 Chapter 4, Section 4.3.
Multivariate Dyadic Regression Trees for Sparse Learning Problems Xi Chen Machine Learning Department Carnegie Mellon University (joint work with Han Liu)
LIMITS AND DERIVATIVES The Limit of a Function LIMITS AND DERIVATIVES In this section, we will learn: About limits in general and about numerical.
1 CSC 4510, Spring © Paula Matuszek CSC 4510 Support Vector Machines (SVMs)
U N I V E R S I T Y O F S O U T H F L O R I D A Hadoop Alternative The Hadoop Alternative Larry Moore 1, Zach Fadika 2, Dr. Madhusudhan Govindaraju 2 1.
Dec 21, 2006For ICDM Panel on 10 Best Algorithms Support Vector Machines: A Survey Qiang Yang, for ICDM 2006 Panel Partially.
Machine Learning in Practice Lecture 19 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
Exponents And how they make life easier! Exponents Exponents are used to write numbers in scientific notation. Exponents are powers of ten. 10 x 10 =
Error Error is the difference between a measured or calculated (reported) value and the true value. Scientists recognize that errors are present in their.
Weka. Weka A Java-based machine vlearning tool Implements numerous classifiers and other ML algorithms Uses a common.
Computational Intelligence: Methods and Applications Lecture 24 SVM in the non-linear case Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
Adiabatic Quantum Computing Josh Ball with advisor Professor Harsh Mathur Problems which are classically difficult to solve may be solved much more quickly.
Machine Learning in Practice Lecture 2 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Machine Learning in Practice Lecture 9 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Next, this study employed SVM to classify the emotion label for each EEG segment. The basic idea is to project input data onto a higher dimensional feature.
How significant is your life?. Do you make a difference?
Introduction to Machine Learning Prof. Nir Ailon Lecture 5: Support Vector Machines (SVM)
1 SCIENTIFIC METHOD systematic approach to problem solving *distinguishes science from other subjects.
A distributed PSO – SVM hybrid system with feature selection and parameter optimization Cheng-Lung Huang & Jian-Fan Dun Soft Computing 2008.
Research Methodology Proposal Prepared by: Norhasmizawati Ibrahim (813750)
Machine Learning in Practice Lecture 9 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Multiplication Timed Tests.
SVM With Stochastic Parameter Selection For Bovine Leather Defect Classification Roberto Viana, Ricardo Rodrigues Marco A. Alvarez And Hemerson Pistori.
Click to edit Master subtitle style Run ons, Comma Splices, Fused Sentences Created by: Alia K. and Scotty S.
Graduate Seminar In Machine Learning
Unit 5: Hypothesis Testing
BLAST Anders Gorm Pedersen & Rasmus Wernersson.
CS 8520: Artificial Intelligence
Statistical Learning Dong Liu Dept. EEIS, USTC.
TESTs about a population mean
Decision tree ensembles in biomedical time-series classifaction
Support Vector Machine _ 2 (SVM)
Classification Breakdown
Assignment 7 Due Application of Support Vector Machines using Weka software Must install libsvm Data set: Breast cancer diagnostics Deliverables:
Support Vector Machines 2
THE ASSISTIVE SYSTEM SHIFALI KUMAR BISHWO GURUNG JAMES CHOU
Presentation transcript:

Machine Learning Applying Our Knowledge Gold Team Week 3: Random Forests and SVM Results

The Good, The Bad, and The Ugly We started the week with implementing Fast Random Forests and libSVM into Weka. This wasn't too difficult. Phil has a powerful computer and we decided to run all the data sets as is. It still took a couple days to do all the runs we needed Collating the data, after we received it all, took quite a while. The Good: The Ugly: The Bad:

Collating the Results We set up experimenter to look at all the different test runs. To get the percent of correctly classified instances. To figure out the number of correctly classified and the number of incorrectly classified instance we opened the results as a.csv in Excel, and calculated the number correct over folds.

Results of Week 3: Random Forests

Summary of Results We concluded that the Primary data set performed the best over all with fast random forest with k = 0.5x and the number of trees didn't seem to matter. We concluded that primary ran the best on SVM with poly kernel, where C nor the exponent mattered. The percent correct was 83.18%. Also we noticed that with the RBF kernel all results we're generally bad. We noticed that FastRandomForest was indeed significantly faster than Random forest but SVM seemed to perform around the same speed as regular SMO. Something Interesting: