Page 0 of 5 Dynamical Invariants of an Attractor and potential applications for speech data Saurabh Prasad Intelligent Electronic Systems Human and Systems.

Slides:



Advertisements
Similar presentations
1 Seminar of computational geometry Lecture #1 Convexity.
Advertisements

Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
Lect.3 Modeling in The Time Domain Basil Hamed
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Machine Learning Lecture 8 Data Processing and Representation
The General Linear Model Or, What the Hell’s Going on During Estimation?
Model Assessment, Selection and Averaging
Forecasting using Non Linear Techniques in Time Series Analysis – Michel Camilleri – September FORECASTING USING NON-LINEAR TECHNIQUES IN TIME SERIES.
Basic geostatistics Austin Troy.
Principal Component Analysis
Lecture Notes for CMPUT 466/551 Nilanjan Ray
Speech Recognition in Noise
Object Class Recognition Using Discriminative Local Features Gyuri Dorko and Cordelia Schmid.
Lecture #1COMP 527 Pattern Recognition1 Pattern Recognition Why? To provide machines with perception & cognition capabilities so that they could interact.
Part II – TIME SERIES ANALYSIS C2 Simple Time Series Methods & Moving Averages © Angel A. Juan & Carles Serrat - UPC 2007/2008.
ICA Alphan Altinok. Outline  PCA  ICA  Foundation  Ambiguities  Algorithms  Examples  Papers.
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
NonLinear Dimensionality Reduction or Unfolding Manifolds Tennenbaum|Silva|Langford [Isomap] Roweis|Saul [Locally Linear Embedding] Presented by Vikas.
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Principles of the Global Positioning System Lecture 11 Prof. Thomas Herring Room A;
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Nonlinear Mixture Autoregressive Hidden Markov Models for Speech Recognition S. Srinivasan, T. Ma, D. May, G. Lazarou and J. Picone Department of Electrical.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
PATTERN RECOGNITION AND MACHINE LEARNING
Oceanography 569 Oceanographic Data Analysis Laboratory Kathie Kelly Applied Physics Laboratory 515 Ben Hall IR Bldg class web site: faculty.washington.edu/kellyapl/classes/ocean569_.
Page 0 of 14 Dynamical Invariants of an Attractor and potential applications for speech data Saurabh Prasad Intelligent Electronic Systems Human and Systems.
Reconstructed Phase Space (RPS)
Department of Mechanical Engineering The University of Strathclyde, Glasgow Hybrid Systems: Modelling, Analysis and Control Yan Pang Department of Mechanical.
1 SUPPORT VECTOR MACHINES İsmail GÜNEŞ. 2 What is SVM? A new generation learning system. A new generation learning system. Based on recent advances in.
1 Multivariate Linear Regression Models Shyh-Kang Jeng Department of Electrical Engineering/ Graduate Institute of Communication/ Graduate Institute of.
Nonlinear Dynamical Invariants for Speech Recognition S. Prasad, S. Srinivasan, M. Pannuri, G. Lazarou and J. Picone Department of Electrical and Computer.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1.
ECE 8443 – Pattern Recognition LECTURE 08: DIMENSIONALITY, PRINCIPAL COMPONENTS ANALYSIS Objectives: Data Considerations Computational Complexity Overfitting.
CCN COMPLEX COMPUTING NETWORKS1 This research has been supported in part by European Commission FP6 IYTE-Wireless Project (Contract No: )
LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.
1 OUTPUT ANALYSIS FOR SIMULATIONS. 2 Introduction Analysis of One System Terminating vs. Steady-State Simulations Analysis of Terminating Simulations.
Data in Motion Michael Hoffman (Leicester) S Muthukrishnan (Google) Rajeev Raman (Leicester)
S. Srinivasan, S. Prasad, S. Patil, G. Lazarou and J. Picone Intelligent Electronic Systems Center for Advanced Vehicular Systems Mississippi State University.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
1 Challenge the future Chaotic Invariants for Human Action Recognition Ali, Basharat, & Shah, ICCV 2007.
Dynamic Models, Autocorrelation and Forecasting ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
Validation methods.
September 28, 2000 Improved Simultaneous Data Reconciliation, Bias Detection and Identification Using Mixed Integer Optimization Methods Presented by:
Madhulika Pannuri Intelligent Electronic Systems Human and Systems Engineering Center for Advanced Vehicular Systems An overview of work done so far in.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 10: PRINCIPAL COMPONENTS ANALYSIS Objectives:
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
ECE-7000: Nonlinear Dynamical Systems 3. Phase Space Methods 3.1 Determinism: Uniqueness in phase space We Assume that the system is linear stochastic.
CIVET seminar Presentation day: Presenter : Park, GilSoon.
On the Optimality of the Simple Bayesian Classifier under Zero-One Loss Pedro Domingos, Michael Pazzani Presented by Lu Ren Oct. 1, 2007.
[Chaos in the Brain] Nonlinear dynamical analysis for neural signals Jaeseung Jeong, Ph.D Department of Bio and Brain Engineering, KAIST.
Madhulika Pannuri Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Correlation Dimension.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Pyramid Vector Quantization
(5) Notes on the Least Squares Estimate
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Boosting and Additive Trees (2)
Machine Learning Basics
Object Modeling with Layers
Computational Intelligence: Methods and Applications
Descriptive Statistics vs. Factor Analysis
Computer Vision Chapter 4
EE513 Audio Signals and Systems
Connecting Data with Domain Knowledge in Neural Networks -- Use Deep learning in Conventional problems Lizhong Zheng.
Toward Functional Classification of Neuronal Types
Presentation transcript:

Page 0 of 5 Dynamical Invariants of an Attractor and potential applications for speech data Saurabh Prasad Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Dynamical Invariants from a Time Series

Page 1 of 5 Dynamical Invariants of an Attractor and potential applications for speech data Estimating the correlation integral from a time series Correlation Integral of an attractor’s trajectory : Correlation sum of a system’s attractor is a (probabilistic) measure quantifying the average number of neighbors in a neighborhood of radius along the trajectory. where represents the i’th point on the trajectory, is a valid norm and is the Heaviside’s unit step function (serving as a count function here) Correlation Dimension : In the limit that we have an infinitely large data set and a very small neighborhood For small epsilons, the correlation integral is expected to grow exponentially with the true dimension of the attractor Hence,

Page 2 of 5 Dynamical Invariants of an Attractor and potential applications for speech data Correlation Integral and dimension estimation: Practical considerations Temporal correlations vs. geometric correlations : Choosing neighbors in a small neighborhood about a point forces the inclusion of temporally correlated points  This results in biasing the estimator, yielding a lower dimension estimate Theiler’s correction : The solution is simple – exclude temporally correlated points from analysis. ‘w’ – Theiler’s correction factor An optimal value for w may be found by a space-time separation plot of the data- set.

Page 3 of 5 Dynamical Invariants of an Attractor and potential applications for speech data Some results on the Lorentz attractor

Page 4 of 5 Dynamical Invariants of an Attractor and potential applications for speech data Fitting the current discussion into the scope of Pattern Recognition RPS Lyapunov Spectra K 2 Entropy Correlation Dimension Input Time Series (scalar / vector) Measure discriminability in a space comprised of all possible combinations of these “features” Higher Order Statistics

Page 5 of 5 Dynamical Invariants of an Attractor and potential applications for speech data A note on Vector-Embedding Vector Embedding: x(t) = [s(t) s(t -  ) s(t - 2  ) … s(t - (d - 1)  )] x(t) = [s 1 (t) s 2 (t)….s m (t) s 1 (t -  ) s 2 (t -  ) ….s m (t -  ) … s 1 (t - (d - 1)  ) s 2 (t - (d - 1)  )…s m (t - (d -1)  )] It is typically assumed that the delay coordinates chosen are such that components of the embedded vectors are uncorrelated. For scalar embedding, an optimal choice of  ensures this. For vector embedding, a strong correlation between components of the (observed) vector stream may hurt the embedding procedure. PCA/SVD based de-correlation may help remove correlations in second order statistics. If correlations in higher order statistics of the data stream are removed, it is hoped that it will provide a more meaningful reconstruction.