Distributed Representative Reading Group. Research Highlights 1Support vector machines can robustly decode semantic information from EEG and MEG 2Multivariate.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

CSI :Florida A BAYESIAN APPROACH TO LOCALIZED MULTI-KERNEL LEARNING USING THE RELEVANCE VECTOR MACHINE R. Close, J. Wilson, P. Gader.
Complex Networks for Representation and Characterization of Images For CS790g Project Bingdong Li 9/23/2009.
Classification using intersection kernel SVMs is efficient Joint work with Subhransu Maji and Alex Berg Jitendra Malik UC Berkeley.
Software Quality Ranking: Bringing Order to Software Modules in Testing Fei Xing Michael R. Lyu Ping Guo.
What is Statistical Modeling
Assuming normally distributed data! Naïve Bayes Classifier.
Inferring individual perceptual experience from MEG: Robust statistics approach Andrey Zhdanov 1,4, Talma Hendler 1,2, Leslie Ungerleider 3, Nathan Intrator.
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Properties of Machine Learning Applications for Use in Metamorphic Testing Chris Murphy, Gail Kaiser, Lifeng Hu, Leon Wu Columbia University.
Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007.
Presented by Zeehasham Rasheed
Data mining and statistical learning - lecture 13 Separating hyperplane.
5/30/2006EE 148, Spring Visual Categorization with Bags of Keypoints Gabriella Csurka Christopher R. Dance Lixin Fan Jutta Willamowski Cedric Bray.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
05/06/2005CSIS © M. Gibbons On Evaluating Open Biometric Identification Systems Spring 2005 Michael Gibbons School of Computer Science & Information Systems.
Data Mining – Intro.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Multiclass object recognition
How To Do Multivariate Pattern Analysis
Thien Anh Dinh1, Tomi Silander1, Bolan Su1, Tianxia Gong
Theta-Coupled Periodic Replay in Working Memory Lluís Fuentemilla, Will D Penny, Nathan Cashdollar, Nico Bunzeck, Emrah Düzel Current Biology, 2010,20(7):
Comparison of Boosting and Partial Least Squares Techniques for Real-time Pattern Recognition of Brain Activation in Functional Magnetic Resonance Imaging.
Word category and verb-argument structure information in the dynamics of parsing Frisch, Hahne, and Friedericie (2004) Cognition.
ROC 1.Medical decision making 2.Machine learning 3.Data mining research communities A technique for visualizing, organizing, selecting classifiers based.
Learning Visual Bits with Direct Feature Selection Joel Jurik 1 and Rahul Sukthankar 2,3 1 University of Central Florida 2 Intel Research Pittsburgh 3.
ADVANCED CLASSIFICATION TECHNIQUES David Kauchak CS 159 – Fall 2014.
ArrayCluster: an analytic tool for clustering, data visualization and module finder on gene expression profiles 組員:李祥豪 謝紹陽 江建霖.
Document Categorization Problem: given –a collection of documents, and –a taxonomy of subject areas Classification: Determine the subject area(s) most.
Transcription of Text by Incremental Support Vector machine Anurag Sahajpal and Terje Kristensen.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Comparison of Bayesian Neural Networks with TMVA classifiers Richa Sharma, Vipin Bhatnagar Panjab University, Chandigarh India-CMS March, 2009 Meeting,
Current work at UCL & KCL. Project aim: find the network of regions associated with pleasant and unpleasant stimuli and use this information to classify.
Urban Building Damage Detection From Very High Resolution Imagery By One-Class SVM and Shadow Information Peijun Li, Benqin Song and Haiqing Xu Peking.
Automatic Image Annotation by Using Concept-Sensitive Salient Objects for Image Content Representation Jianping Fan, Yuli Gao, Hangzai Luo, Guangyou Xu.
MURI: Integrated Fusion, Performance Prediction, and Sensor Management for Automatic Target Exploitation 1 Dynamic Sensor Resource Management for ATE MURI.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Classification Techniques: Bayesian Classification
Data Mining – Intro. Course Overview Spatial Databases Temporal and Spatio-Temporal Databases Multimedia Databases Data Mining.
1 Pattern Recognition Pattern recognition is: 1. A research area in which patterns in data are found, recognized, discovered, …whatever. 2. A catchall.
1 Chapter 6. Classification and Prediction Overview Classification algorithms and methods Decision tree induction Bayesian classification Lazy learning.
Acclimatizing Taxonomic Semantics for Hierarchical Content Categorization --- Lei Tang, Jianping Zhang and Huan Liu.
Maximum Entropy Models and Feature Engineering CSCI-GA.2590 – Lecture 6B Ralph Grishman NYU.
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
Chapter 3: Maximum-Likelihood Parameter Estimation l Introduction l Maximum-Likelihood Estimation l Multivariate Case: unknown , known  l Univariate.
MVPD – Multivariate pattern decoding Christian Kaul MATLAB for Cognitive Neuroscience.
Multi-Speaker Modeling with Shared Prior Distributions and Model Structures for Bayesian Speech Synthesis Kei Hashimoto, Yoshihiko Nankaku, and Keiichi.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
Neural Text Categorizer for Exclusive Text Categorization Journal of Information Processing Systems, Vol.4, No.2, June 2008 Taeho Jo* 報告者 : 林昱志.
Image Classification for Automatic Annotation
Support Vector Machines and Gene Function Prediction Brown et al PNAS. CS 466 Saurabh Sinha.
Ensemble Methods in Machine Learning
Classification Ensemble Methods 1
Data Mining and Decision Support
Chapter 6. Classification and Prediction Classification by decision tree induction Bayesian classification Rule-based classification Classification by.
Chapter 6. Classification and Prediction Classification by decision tree induction Bayesian classification Rule-based classification Classification by.
Supervised Machine Learning: Classification Techniques Chaleece Sandberg Chris Bradley Kyle Walsh.
FMRI and Behavioral Studies of Human Face Perception Ronnie Bryan Vision Lab
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
Next, this study employed SVM to classify the emotion label for each EEG segment. The basic idea is to project input data onto a higher dimensional feature.
Ping-Tsun Chang Intelligent Systems Laboratory NTU/CSIE Using Support Vector Machine for Integrating Catalogs.
Does the brain compute confidence estimates about decisions?
Support Feature Machine for DNA microarray data
Asymmetric Gradient Boosting with Application to Spam Filtering
Ganapathy Mani, Bharat Bhargava, Jason Kobes*
Deep Learning Hierarchical Representations for Image Steganalysis
Araceli Ramirez-Cardenas, Maria Moskaleva, Andreas Nieder 
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Support Vector Machines
Adaptive multi-voxel representation of stimuli, rules and responses
Presentation transcript:

Distributed Representative Reading Group

Research Highlights 1Support vector machines can robustly decode semantic information from EEG and MEG 2Multivariate decoding techniques allow for detection of subtle, but distributed, effects 3Semantic categories and individual words have distributed spatiotemporal representations 4Representations are consistent between subjects and stimulus modalities 5A scalable hierarchical tree decoder further improves decoding performance

why do reported results vary from study to study,? Partly due to the statistical analysis (traditional univariate techniques ) of high-dimensional neuroimaging data require correction for multiple comparisons to control for false positives insensitive to subtle, but widespread, effects within the brain yield differing results depending on the specific responses elicited by the particular experiment performed

Why chose SVM Robust to high-dimensional data attempt to find a separating boundary which maximizes the margin between these classes reduces over-fitting and allows for good generalization when classifying novel data allows for a multivariate examination of the spatiotemporal dynamics

Why hierarchical tree decoding Single multiclass decoder can distinguish individual word representations well, it doesn’t, directly incorporate a prior knowledge about semantic classes and the features which best discriminate these categories. To combine information from the classifier models generated to decode semantic category and individual words, a hierarchical tree framework which attempts to decode word properties sequentially were implemented Given an unknown word, the tree decoder First classifies it as either a large (target) or small (nontarget) object Second classified as living or nonliving object Finally as an individual word within the predicted semantic category Advantages: allows the appropriate features to be used to decode each word property, narrowing the search space before individual words are decoded. such a tree construct is easily scalable and could allow for the eventual decoding of larger libraries of words.

Experiment visual (SV) and auditory version (SA) language tasks Task: Subjects were instructed to press a button if the presented word represented an object larger than 1 foot in any dimension Stimuli: representing objects larger than 1 foot : smaller than 1 foot = 1:1 living objects (animals and animal parts) and nonliving objects (man-made items)= 1:1 How to present stimuli: Half of the trials presented a novel word which was only shown only once during the experiment while the other half of the trials presented 1 of 10 repeated words (each shown multiple times during the experiment).

Decoding framework Features: Average amplitude in six 50-ms time windows were sampled from every channel and concatenated into a large feature vector for each trial Decoding living versus non living 200, 300, 400, 500, 600, and 700 ms poststimulus Decoding individual words 250, 300, 350, 400, 450, and 500 ms poststimulus

Decoding accuracy Compared to Naive Bayes classifier, SVM is better able to handle high dimension data

SVM weights show weights show important times and locations for decoding

Decoding is not based on low-level stimulus properties It is possible that the generated classifiers are utilizing neural activity related to low-level visual or auditory stimulus properties when decoding individual words To test this, we performed a shuffling based on stimulus properties to evaluate this potential confounding factor.

inter-modality and inter-subject decoding show shared neural representations inter-modality: train the classifier with one modality and test the classifier with the other modality inter-subject : classifier was trained on data from all but one subject within a single modality and then the remaining subject was used as test data.

Hierarchical tree decoding improves decoding performance A three-level hierarchical tree decoder was utilized to first decode the large/small distinction (utilizing amplitude and spectral features), then the living/nonliving object category (utilizing 200–700 ms amplitude features), and finally the individual word (utilizing 250–500 ms amplitude features).

Thanks for your attention!