Using decision trees to build an a framework for multivariate time- series classification 1 Present By Xiayi Kuang.

Slides:



Advertisements
Similar presentations
The Software Infrastructure for Electronic Commerce Databases and Data Mining Lecture 4: An Introduction To Data Mining (II) Johannes Gehrke
Advertisements

DECISION TREES. Decision trees  One possible representation for hypotheses.
Random Forest Predrag Radenković 3237/10
CHAPTER 9: Decision Trees
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Huffman code and ID3 Prof. Sin-Min Lee Department of Computer Science.
Introduction Training Complexity, Pruning CART vs. ID3 vs. C4.5
K-NEAREST NEIGHBORS AND DECISION TREE Nonparametric Supervised Learning.
Data Mining Classification: Alternative Techniques
Ensemble Methods An ensemble method constructs a set of base classifiers from the training data Ensemble or Classifier Combination Predict class label.
A Quick Overview By Munir Winkel. What do you know about: 1) decision trees 2) random forests? How could they be used?
Real-Time Human Pose Recognition in Parts from Single Depth Images Presented by: Mohammad A. Gowayyed.
Personal Driving Diary: Constructing a Video Archive of Everyday Driving Events IEEE workshop on Motion and Video Computing ( WMVC) 2011 IEEE Workshop.
Ensemble Learning: An Introduction
Three kinds of learning
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
ICS 273A Intro Machine Learning
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
© Prentice Hall1 DATA MINING Introductory and Advanced Topics Part II Margaret H. Dunham Department of Computer Science and Engineering Southern Methodist.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Birch: An efficient data clustering method for very large databases
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
Machine Learning CS 165B Spring 2012
A Generic Approach for Image Classification Based on Decision Tree Ensembles and Local Sub-windows Raphaël Marée, Pierre Geurts, Justus Piater, Louis Wehenkel.
Issues with Data Mining
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
Skewing: An Efficient Alternative to Lookahead for Decision Tree Induction David PageSoumya Ray Department of Biostatistics and Medical Informatics Department.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
Human pose recognition from depth image MS Research Cambridge.
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
MACHINE LEARNING 10 Decision Trees. Motivation  Parametric Estimation  Assume model for class probability or regression  Estimate parameters from all.
Chapter 6 Classification and Prediction Dr. Bernard Chen Ph.D. University of Central Arkansas.
Multi-Speaker Modeling with Shared Prior Distributions and Model Structures for Bayesian Speech Synthesis Kei Hashimoto, Yoshihiko Nankaku, and Keiichi.
Patch Based Prediction Techniques University of Houston By: Paul AMALAMAN From: UH-DMML Lab Director: Dr. Eick.
1 CSCI 3202: Introduction to AI Decision Trees Greg Grudic (Notes borrowed from Thomas G. Dietterich and Tom Mitchell) Intro AIDecision Trees.
Decision Trees Example of a Decision Tree categorical continuous class Refund MarSt TaxInc YES NO YesNo Married Single, Divorced < 80K> 80K Splitting.
Konstantina Christakopoulou Liang Zeng Group G21
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Lucent Technologies - Proprietary 1 Interactive Pattern Discovery with Mirage Mirage uses exploratory visualization, intuitive graphical operations to.
ECE 471/571 – Lecture 20 Decision Tree 11/19/15. 2 Nominal Data Descriptions that are discrete and without any natural notion of similarity or even ordering.
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
1 An Efficient Optimal Leaf Ordering for Hierarchical Clustering in Microarray Gene Expression Data Analysis Jianting Zhang Le Gruenwald School of Computer.
WHAT IS DATA MINING?  The process of automatically extracting useful information from large amounts of data.  Uses traditional data analysis techniques.
DECISION TREES Asher Moody, CS 157B. Overview  Definition  Motivation  Algorithms  ID3  Example  Entropy  Information Gain  Applications  Conclusion.
Ensemble Learning, Boosting, and Bagging: Scaling up Decision Trees (with thanks to William Cohen of CMU, Michael Malohlava of 0xdata, and Manish Amde.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Mustafa Gokce Baydogan, George Runger and Eugene Tuv INFORMS Annual Meeting 2011, Charlotte A Bag-of-Features Framework for Time Series Classification.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
Feature learning for multivariate time series classification Mustafa Gokce Baydogan * George Runger * Eugene Tuv † * Arizona State University † Intel Corporation.
Ensemble Classifiers.
Machine Learning: Ensemble Methods
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
k-Nearest neighbors and decision tree
Introduction to Machine Learning and Tree Based Methods
Prepared by: Mahmoud Rafeek Al-Farra
Trees, bagging, boosting, and stacking
Chapter 6 Classification and Prediction
Pattern Recognition Sergios Theodoridis Konstantinos Koutroumbas
Basic machine learning background with Python scikit-learn
Classification and Prediction
Roberto Battiti, Mauro Brunato
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
©Jiawei Han and Micheline Kamber
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
Presentation transcript:

Using decision trees to build an a framework for multivariate time- series classification 1 Present By Xiayi Kuang

Outline Introduction – Goal, context & challenges Proposed approach – Overview – Clustering stage – Classification stage Experiments – Visual Surveillance Context Video sequence 2

Introduction Goal To build a flexible and generic system that can model possibly complex spatio-temporal events 3

Introduction Context 4 Computer vision applications – Ex: Automated Visual surveillance Medical domain – Ex: electronic graph classification Gesture Recognition

Introduction Challenges with MTS classification – Machine learning techniques cannot be directly applied to time-series data – Large dataset with many temporal attributes are hard to handle because of the number of combinatorial possibilities between the features – Invariance of the events with respect to times 5

Proposed Approach Complex spatio-temporal Event Multivariate time-series data: 6 Time- steps t1t2t3t…tn-1tn Attribute 1 f1 1 f2 1 f3 1 f… 1 fn-1 1 fn 1 Attribute 2 f1 2 f2 2 f3 2 f… 2 fn-1 2 fn 2 Attribute … f1 … f2 … f3 … f… … fn-1 … fn … Attribute m-1 f1 m-1 f2 m-1 f3 m-1 f… m-1 fn-1 m-1 fn m-1 Attribute m f1 m f2 m f3 m f… m fn-1 m fn m Pattern 1Pattern 2 before= event X

Proposed approach Framework Overview 7 Local Patterns (LP) extraction Temporal relation between LP

Proposed approach Theory: The Decision Tree (DT) and the ensemble of randomized trees (ERT) Top down induction Decision Trees – At each node, the ‘best split’ is chosen according to a specific distance measure (splitting criteria) – Clustering trees: Each node and leaf is a cluster – Classification trees: Each leaf is labeled with a class Ensemble of randomized trees – The ‘best split’ is chosen among X trials – 10 to 100 trees are built and we combine their predictions with a simple majority vote – Reduce overfitting and increase robustness and accuracy 8 Node 1Node 2Node 4Leaf Node3Node 5LeafNode 6Leaf

Proposed approach Clustering trees Goal: – To extract local patterns at each time-step How? – Each time-step in the training sample are clustered independently Algorithm – Multiple clustering trees – Split function at each node: Choose one attribute and one threshold that maximize a splitting criteria – Splitting criteria: Minimizing the intra-cluster distance and maximizing the inter-cluster distance  2 choices: Supervized clustering: The distance metric is the class entropy (each time-step of one sequence inherit the class label of its sequence) Unsupervized clustering: The distance can be the Euclidian distance (usual one) And then? – Each node is a pattern and is tagged – Each frame is labeled with the appropriate patterns 9

Proposed approach Clustering trees Time- steps t1t2t3t…tn-1tn Channel 1 f1 1 f2 1 f3 1 f… 1 fn-1 1 fn 1 Channel 2 f1 2 f2 2 f3 2 f… 2 fn-1 2 fn 2 Channel … f1 … f2 … f3 … f… … fn-1 … fn … Channel m-1 f1 m-1 f2 m-1 f3 m-1 f… m-1 fn-1 m-1 fn m-1 Channel m f1 m f2 m f3 m f… m fn-1 m fn m 10 t1 pattern 1 pattern 4 pattern X t2 pat. 2 pat. 8 pat. 9 pat. Y t3 pat. 1 pat. 4 pat. X pat. Y t… pat. 1 pat. 3 pat. 8 pat. X pat. Y tn-1 pat. 2 pat. 3 pat. 4 pat. 9 pat. Y tn pat. 1 pat. 4 pat. X pat. Y Preprocessing Input Data Clustering stage Multivariate time-series

Proposed approach Classification trees Goal: – To model the temporal relation between the local patterns Algorithm: – Ensemble of randomized trees Split function at each node: Splitting criteria: – Maximizing the Normalized information gain over X randomly selected tries We combine the trees prediction with a simple majority vote 11

Proposed approach Classification trees 12 t1 pattern 1 pattern 4 pattern X t2 pat. 2 pat. 8 pat. 9 pat. Y t3 pat. 1 pat. 4 pat. X pat. Y t… pat. 1 pat. 3 pat. 8 pat. X pat. Y tn-1 pat. 2 pat. 3 pat. 4 pat. 9 pat. Y tn pat. 1 pat. 4 pat. X pat. Y Event A Event B

Application Visual surveillance events Additional Challenges in automated visual surveillance – Segmentation and tracking are often not robust – Events are semantically complex and often subjective – Few training data available for interesting events – Events are extremely variable in length 13

Application video sequences Dataset: 7 events, no ground truth. Features: position, speed and size of the blobs 14

15

Application video sequences Results 16

17 Thanks 12/03/2008