TRADING OFF PREDICTION ACCURACY AND POWER CONSUMPTION FOR CONTEXT- AWARE WEARABLE COMPUTING Presented By: Jeff Khoshgozaran.

Slides:



Advertisements
Similar presentations
Statistical Machine Learning- The Basic Approach and Current Research Challenges Shai Ben-David CS497 February, 2007.
Advertisements

Introduction to Support Vector Machines (SVM)
Support Vector Machine
Generative Models Thus far we have essentially considered techniques that perform classification indirectly by modeling the training data, optimizing.
Mustafa Cayci INFS 795 An Evaluation on Feature Selection for Text Clustering.
1 VLDB 2006, Seoul Mapping a Moving Landscape by Mining Mountains of Logs Automated Generation of a Dependency Model for HUG’s Clinical System Mirko Steinle,
Classification / Regression Support Vector Machines
Fast Algorithms For Hierarchical Range Histogram Constructions

Adapting Ocean Surveys to the Observed Fields Characteristics Maria-João Rendas I3S, CNRS-UNSA.
Support Vector Machines
Machine learning continued Image source:
THE JIGSAW CONTINUOUS SENSING ENGINE FOR MOBILE PHONE APPLICATIONS Hong Lu,† Jun Yang,! Zhigang Liu,! Nicholas D. Lane,† Tanzeem Choudhury,† Andrew T.
Robust Multi-Kernel Classification of Uncertain and Imbalanced Data
Accelerometer-based Transportation Mode Detection on Smartphones
1 Prediction-based Strategies for Energy Saving in Object Tracking Sensor Networks Yingqi Xu, Wang-Chien Lee Proceedings of the 2004 IEEE International.
Support Vector Machines (SVMs) Chapter 5 (Duda et al.)
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
Ensemble Learning: An Introduction
Chapter 8 Operating System Support
Introduction to Boosting Aristotelis Tsirigos SCLT seminar - NYU Computer Science.
What is Learning All about ?  Get knowledge of by study, experience, or being taught  Become aware by information or from observation  Commit to memory.
Lecture 10: Support Vector Machines
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
1 Music Classification Using SVM Ming-jen Wang Chia-Jiu Wang.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Ensemble Learning (2), Tree and Forest
An Introduction to Support Vector Machines Martin Law.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
EWatch: A Wearable Sensor and Notification Platform Paper By: Uwe Maurer, Anthony Rowe, Asim Smailagic, Daniel P. Siewiorek Presenter: Ke Gao.
1 ENTROPY-BASED CONCEPT SHIFT DETECTION PETER VORBURGER, ABRAHAM BERNSTEIN IEEE ICDM 2006 Speaker: Li HueiJyun Advisor: Koh JiaLing Date:2007/11/6 1.
Support Vector Machines Mei-Chen Yeh 04/20/2010. The Classification Problem Label instances, usually represented by feature vectors, into one of the predefined.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
CPU Scheduling CSCI 444/544 Operating Systems Fall 2008.
An Introduction to Support Vector Machines (M. Law)
1 Chapter 6. Classification and Prediction Overview Classification algorithms and methods Decision tree induction Bayesian classification Lazy learning.
Evaluation of Techniques for Classifying Biological Sequences Authors: Mukund Deshpande and George Karypis Speaker: Sarah Chan CSIS DB Seminar May 31,
START OF DAY 5 Reading: Chap. 8. Support Vector Machine.
CS 478 – Tools for Machine Learning and Data Mining SVM.
Dr. Sudharman K. Jayaweera and Amila Kariyapperuma ECE Department University of New Mexico Ankur Sharma Department of ECE Indian Institute of Technology,
Linear Models for Classification
Support vector machine LING 572 Fei Xia Week 8: 2/23/2010 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A 1.
Support Vector Machines Tao Department of computer science University of Illinois.
CZ5225: Modeling and Simulation in Biology Lecture 7, Microarray Class Classification by Machine learning Methods Prof. Chen Yu Zong Tel:
Adaptive Sleep Scheduling for Energy-efficient Movement-predicted Wireless Communication David K. Y. Yau Purdue University Department of Computer Science.
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
A Brief Maximum Entropy Tutorial Presenter: Davidson Date: 2009/02/04 Original Author: Adam Berger, 1996/07/05
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Advanced Gene Selection Algorithms Designed for Microarray Datasets Limitation of current feature selection methods: –Ignores gene/gene interaction: single.
Learning A Better Compiler Predicting Unroll Factors using Supervised Classification And Integrating CPU and L2 Cache Voltage Scaling using Machine Learning.
Next, this study employed SVM to classify the emotion label for each EEG segment. The basic idea is to project input data onto a higher dimensional feature.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Mustafa Gokce Baydogan, George Runger and Eugene Tuv INFORMS Annual Meeting 2011, Charlotte A Bag-of-Features Framework for Time Series Classification.
Predictive Automatic Relevance Determination by Expectation Propagation Y. Qi T.P. Minka R.W. Picard Z. Ghahramani.
CS 9633 Machine Learning Support Vector Machines
Jacob R. Lorch Microsoft Research
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Erasmus University Rotterdam
Alan Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani
Vijay Srinivasan Thomas Phan
Pawan Lingras and Cory Butz
Support Vector Machines Introduction to Data Mining, 2nd Edition by
Pattern Recognition CS479/679 Pattern Recognition Dr. George Bebis
Hyperparameters, bias-variance tradeoff, validation
Design of Hierarchical Classifiers for Efficient and Accurate Pattern Classification M N S S K Pavan Kumar Advisor : Dr. C. V. Jawahar.
Xin Qi, Matthew Keally, Gang Zhou, Yantao Li, Zhen Ren
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Modeling IDS using hybrid intelligent systems
Presentation transcript:

TRADING OFF PREDICTION ACCURACY AND POWER CONSUMPTION FOR CONTEXT- AWARE WEARABLE COMPUTING Presented By: Jeff Khoshgozaran

Motivation Wearable devices sensing user information Context-Aware Mobile Computing Previous work Power consumption in full power mode Quickly depletes a critically constrained resource High sampling rate to provide accuracy Computational and space-intensive solutions Lack of scalability for knee and hip-worn sensors

eWATCH A context-aware wearable platform Several sensors including two-axis accelerometer Three power states for sensing and classifying data Full Power Active CPU, active peripherals (~ 30 ms) Idle State Core clock turned off, active peripherals Waiting for the next sample For SR=6 Hz, time interval~166ms Low Power State is active most of the time Inactive CPU and peripherals Active real-time clock Scheduling next wake up Using selective sample algorithm

Time/Frequency-Domain-Based Classification 5 second windows for computing features Time-based Using means, variances, median, etc. Frequency-based Using FFT on values of both accelerometer axes separately Human movement is periodic Frequency-based Approaches good for classifying accelerometer data Less expressive for very low sampling rates

Battery Lifetime Vs. Sampling Rate More costly computation and more dimensions While important, computation is not the dominant factor in reducing energy consumption

Using SVM for Classification A multi-class SVM used for actual classification Detect and exploit complex patterns in data Good for representing complex patterns Good for excluding unstable patterns (= overfitting) Computationally expensive training Very efficient classification (hardware friendly) Guassian Radial Basis Function used as kernel to classify non-linear data The class of kernel methods implicitly defines the class of possible patterns by introducing a notion of similarity between data Implicit and non-linear embedding of data in high- dimensional spaces Separated by a hyperplane in feature space

Power-optimized Classification Experiments Training data captured by 3 test participants Each activity recorded for 10 minutes Data was split into different recorded activities Data was partitioned into blocks of 5 seconds Used to extract time/frequency domain features Labeled examples used for training multi class SVM Prediction accuracy & power consumption computed

Results For all but extremely low frequency ranges, frequency based features perform superiorly. Optimum sampling frequency of 6 Hz 85% Increase

Selective Sampling vs. Prediction Accuracy What: Further reduce energy consumption How: Selective Sampling Why: Human activity: a continuous process Person more likely to continue an activity than to change to another at a point in time Selective sampling schedules classification Reduces number of observations Saving energy from continuous monitoring to few points in time Objective: keeping accuracy as high as possible A t is the user’s activity at time t

Selective Sampling (cont.) Select a set of observation times to maximize correct prediction of user’s activity for times when no sampling/classification is made Minimize the expected loss: Conditional plans Maximum # of observations Expected loss over all activity sequences a Selected observation times for a Minimize uncertainty : Sequence of decisions: depending on observations so far, decides when next observation should be made Entropy and dynamic programming used to find optimal

4 Schemes to Select the Conditional Plan Uniform Spacing Selects observation times at equally spaced intervals Random Spacing B random length observation times selected at random Exponential Backoff Maintains a maximum step size ∆ max If cur. act=last detected act., multiply ∆ max by α else ∆ max =1 Actual step size ∆ chosen uniformly at random from [1, ∆ max ] Next observation made at t+ ∆ Entropy-based Minimizing uncertainty using the entropy criterion Taking transition probabilities of states into account More frequent sampling for activities with short durations

Selective Sampling Experiments Four new objects performing hour-long activities Subjects were indirectly asked to perform representative tasks at random times User activities manually annotated by an observer Resulting in pairs sampled at 6Hz Data then partitioned into sequences of 5 seconds These blocks labeled with annotations and classified using pre-trained classifiers in the frequency domain

Results Continuous Sampling Competitive for low frequencies Factor of 2 improvement Using annotated data as “exact classification” No SVM (focusing on sampling) Using classifier output instead of annotations Error=Sampling + Classification Overall error dominated by classification from SVM and not by sampling Classification accuracy lower than previous experiments due to 1.new subjects 2. noisy real-world environment Roughly similar behavior to above 6Hz factor of 2.5 improvement

Conclusion High efficiency and accuracy for low range frequency of 1-10 Hz. Competitive classification accuracy for the highly erratic and ambiguous (but convenient) wrist- based sensing Four selective sampling strategies to further reduce the resource usage

Comments Using FFT for each dimension separately looses the correlation of among dimensions Semi-controlled user behavior for test data generation Authors assume continuous state change in a close set of predefined activities i.e., at any given time, one of these activities are taking place