Madhulika Pannuri Intelligent Electronic Systems Human and Systems Engineering Center for Advanced Vehicular Systems An overview of work done so far in.

Slides:



Advertisements
Similar presentations
Ch2 Data Preprocessing part3 Dr. Bernard Chen Ph.D. University of Central Arkansas Fall 2009.
Advertisements

Object Specific Compressed Sensing by minimizing a weighted L2-norm A. Mahalanobis.
From
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Lecture Slides Elementary Statistics Eleventh Edition and the Triola.
April 2, 2015Applied Discrete Mathematics Week 8: Advanced Counting 1 Random Variables In some experiments, we would like to assign a numerical value to.
Clustering short time series gene expression data Jason Ernst, Gerard J. Nau and Ziv Bar-Joseph BIOINFORMATICS, vol
Page 0 of 8 Time Series Classification – phoneme recognition in reconstructed phase space Sanjay Patil Intelligent Electronics Systems Human and Systems.
Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch. 2-1 Statistics for Business and Economics 7 th Edition Chapter 2 Describing Data:
Evaluating Hypotheses
Development of Empirical Models From Process Data
B a c kn e x t h o m e Classification of Variables Discrete Numerical Variable A variable that produces a response that comes from a counting process.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 6-1 Chapter 6 The Normal Distribution and Other Continuous Distributions.
QMS 6351 Statistics and Research Methods Chapter 7 Sampling and Sampling Distributions Prof. Vera Adamchik.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 6-1 Chapter 6 The Normal Distribution and Other Continuous Distributions.
Applications in GIS (Kriging Interpolation)
Radial Basis Function Networks
Elec471 Embedded Computer Systems Chapter 4, Probability and Statistics By Prof. Tim Johnson, PE Wentworth Institute of Technology Boston, MA Theory and.
©2003/04 Alessandro Bogliolo Background Information theory Probability theory Algorithms.
Page 0 of 14 Dynamical Invariants of an Attractor and potential applications for speech data Saurabh Prasad Intelligent Electronic Systems Human and Systems.
Reconstructed Phase Space (RPS)
Sampling Distributions
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 6-1 Chapter 6 The Normal Distribution and Other Continuous Distributions.
Chapter 7 Random-Number Generation
Module 1: Statistical Issues in Micro simulation Paul Sousa.
Chapter 2 Describing Data.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 6 Normal Probability Distributions 6-1 Review and Preview 6-2 The Standard Normal.
Sampling Distributions Chapter 7. The Concept of a Sampling Distribution Repeated samples of the same size are selected from the same population. Repeated.
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Digital Image Processing CCS331 Relationships of Pixel 1.
Language Model Grammar Conversion Wesley Holland Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering.
Nonlinear Dynamical Invariants for Speech Recognition S. Prasad, S. Srinivasan, M. Pannuri, G. Lazarou and J. Picone Department of Electrical and Computer.
Applications of Neural Networks in Time-Series Analysis Adam Maus Computer Science Department Mentor: Doctor Sprott Physics Department.
Maximum Likelihood Estimation Methods of Economic Investigation Lecture 17.
Expectation for multivariate distributions. Definition Let X 1, X 2, …, X n denote n jointly distributed random variable with joint density function f(x.
Chapter 2 Signals and Spectra (All sections, except Section 8, are covered.)
Madhulika Pannuri Intelligent Electronic Systems Human and Systems Engineering Center for Advanced Vehicular Systems An overview of work done so far in.
STATISTICAL COMPLEXITY ANALYSIS Dr. Dmitry Nerukh Giorgos Karvounis.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Modern Navigation Thomas Herring MW 11:00-12:30 Room
Low-Dimensional Chaotic Signal Characterization Using Approximate Entropy Soundararajan Ezekiel Matthew Lang Computer Science Department Indiana University.
Some figures adapted from a 2004 Lecture by Larry Liebovitch, Ph.D. Chaos BIOL/CMSC 361: Emergence 1/29/08.
Introduction to Chaos by: Saeed Heidary 29 Feb 2013.
CCN COMPLEX COMPUTING NETWORKS1 This research has been supported in part by European Commission FP6 IYTE-Wireless Project (Contract No: )
Introduction to Digital Signals
Page 0 of 8 Lyapunov Exponents – Theory and Implementation Sanjay Patil Intelligent Electronics Systems Human and Systems Engineering Center for Advanced.
1 8. One Function of Two Random Variables Given two random variables X and Y and a function g(x,y), we form a new random variable Z as Given the joint.
ECE-7000: Nonlinear Dynamical Systems Overfitting and model costs Overfitting  The more free parameters a model has, the better it can be adapted.
S. Srinivasan, S. Prasad, S. Patil, G. Lazarou and J. Picone Intelligent Electronic Systems Center for Advanced Vehicular Systems Mississippi State University.
Chapter 8: Simple Linear Regression Yang Zhenlin.
1 Challenge the future Chaotic Invariants for Human Action Recognition Ali, Basharat, & Shah, ICCV 2007.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
1 EE571 PART 3 Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic Engineering Eastern.
Chapter 31Introduction to Statistical Quality Control, 7th Edition by Douglas C. Montgomery. Copyright (c) 2012 John Wiley & Sons, Inc.
Language Model Grammar Conversion Wesley Holland, Julie Baca, Dhruva Duncan, Joseph Picone Center for Advanced Vehicular Systems Mississippi State University.
Topics 1 Specific topics to be covered are: Discrete-time signals Z-transforms Sampling and reconstruction Aliasing and anti-aliasing filters Sampled-data.
CZ5211 Topics in Computational Biology Lecture 4: Clustering Analysis for Microarray Data II Prof. Chen Yu Zong Tel:
1 Lecture Plan Modelling Profit Distribution from Wind Production (Excel Case: Danish Wind Production and Spot Prices) Reasons for copula.
Outline Time series prediction Find k-nearest neighbors Lag selection Weighted LS-SVM.
[Chaos in the Brain] Nonlinear dynamical analysis for neural signals Jaeseung Jeong, Ph.D Department of Bio and Brain Engineering, KAIST.
Page 0 of 5 Dynamical Invariants of an Attractor and potential applications for speech data Saurabh Prasad Intelligent Electronic Systems Human and Systems.
Nonlinear time series analysis Delay embedding technique –reconstructs dynamics from single variable Correlation sum –provides an estimator of fractal.
Chapter 6 The Normal Distribution and Other Continuous Distributions
Language Model Classes
SIGNALS PROCESSING AND ANALYSIS
Dimension Review Many of the geometric structures generated by chaotic map or differential dynamic systems are extremely complex. Fractal : hard to define.
Correlation and Regression
Overview of Language Model Classes and Release Progress
Test Drop Rules: If not:
Capacity Dimension of The Hénon Attractor
Discrete Random Variables: Joint PMFs, Conditioning and Independence
Presentation transcript:

Madhulika Pannuri Intelligent Electronic Systems Human and Systems Engineering Center for Advanced Vehicular Systems An overview of work done so far in 2006 RESEARCH PRESENTATION :

Page 1 of 36 Research Presentation Abstract Language Model ABNF: The LanguageModel ABNF reads in a set of productions in the form of ABNF and converts them to BNF. These productions are passed to LanguageModel BNF in the form they were received. Optimum time Delay estimation: Computes optimum time delay for re-constructing phase space. Auto Mutual Information is used for finding optimum time delay. Dimension Estimation: The dimension of an attractor is a measure of its geometric scaling properties and is ‘the most basic property of an attractor’. Though there are many dimensions, we will be implementing ‘Correlation Dimension’ and ‘Lyapunov Dimension’.

Page 2 of 36 Research Presentation Language Model ABNF Why conversion from ABNF to BNF? ABNF balances compactness and simplicity, with reasonable representational power. Differences: The differences between BNF and ABNF involve naming rules, repetition, alternatives, order-independence and value ranges. Removing the meta symbols makes it easy to convert it to finite state machines (IHD). Problem: Converting any given ABNF production to BNF is impossible. So, generalize the algorithm so that most of the expressions can be converted. The steps for conversion involve removing meta symbols one after the other. There were complications like multiple nesting. Ex: * (+(a, n) | (b | s))

Page 3 of 36 Research Presentation Concatenation: Removing concatenation requires two new rules with new names to be introduced. The last two productions resulting from conversion are not valid CFG’s but shorter than the original. Alternation: The production is replaced with a pair of productions. They may or may not be legal CFG’s. Kleene Star: New variable is introduced and the production is replaced. Gives rise to epsilon transition. Kleene Plus: Similar to Kleene Plus. No null transition. Language Model ABNF

Page 4 of 36 Research Presentation Optimum time delay estimation Why calculating time delay ? If we have a trajectory from a chaotic system and we only have data from one of the system variables, there's a neat theorem that says that a copy of the attractor of the system can be reconstructed by lagging the time series to embed it in more dimensions. In other words, if we have a point F( x, y, z, t) which is along some strange attractor, and we can only measure F(z,t), we can plot F (z, z + N, z + 2N, t), and the resulting object will be topologically identical to the original attractor..!!! Method of time delays provides a relatively simple way of constructing an attractor from single experimental time series. Can I choose some value randomly ?? NO. Choosing optimum time delay is not trivial as the dynamical properties of the reconstructed attractor are to be amenable for subsequent analysis. For an infinite amount of noise free data, we are free to choose arbitrarily, the time delay.

Page 5 of 36 Research Presentation For smaller values of tau, s(t) and s(t+tau) are very close to each other in numerical value, and hence they are not independent of each other. For large values of tau, s(t) and s(t+ tau) are completely independent of each other, and any connection between them in the case of chaotic attractor is random because of butterfly effect. We need a criterion for intermediate choice that is large enough so that s(t) and s(t+tau) are independent but not so large that s(t) and s(t+tau) are completely independent in statistical sense. Time delay is a multiple of sampling time (data is available at these times only). There are four common methods for determining an optimum time delay: 1.Visual Inspection of reconstructed attractors: The simplest way to choose tau. Consider successively larger values of tau and then visually inspect the phase portrait of the resulting attractor. Choose the tau value which appears to give the most spread out attractor. Disadvantage: This produces reasonable results with simple systems only. Optimum time delay estimation

Page 6 of 36 Research Presentation Methods to estimate optimum time delay 2. Dominant period relationship: We use the property that time delay is one quarter of the dominant period. Advantage: Quick and easy method for determining tau. Disadvantages: Can be used for low dimensional systems only. Many complex systems do not possess a single dominant frequency. 3.The Autocorrelation function: The auto correlation function r k, compares two data points in the time series separated by delay k, and is defined as, The delay for the attractor reconstruction, tau, is then taken at a specific threshold The behavior of correlation is inconsistent.

Page 7 of 36 Research Presentation Mutual Information Method 4. Minimum auto mutual Information method: The Mutual Information is given by: When the mutual information is minimum, the attractor is as spread out as much as possible. This condition for the choice of delay time is known as ‘minimum mutual information criterion’. Practical Implementation: To calculate the mutual information, the 2D reconstruction of an attractor is partitioned into a grid of Nc columns and Nr rows. Discrete probability density functions for X(i) and X(i+tau) are generated by summing the data points in each row and column of the grid respectively and dividing by the total number of attractor points. The joint probability of occurrence P(k, l) of the attractor in any particular box is calculated by counting the number of discrete points in the box and dividing by the total number of points on the attractor trajectory. The value of tau which gives the first minimum is the attractor reconstruction delay.

Page 8 of 36 Research Presentation Method used to calculate Mutual Information

Page 9 of 36 Research Presentation Ami plots for sine using IFC

Page 10 of 36 Research Presentation Lorenz time series

Page 11 of 36 Research Presentation Ami for lorentz time series (IFC)

Page 12 of 36 Research Presentation Variation in ami with variation in number of bins

Page 13 of 36 Research Presentation Effect of noise

Page 14 of 36 Research Presentation ami vs auto correlation Mutual information is a function of both linear and non linear dependencies between two variables. Where as correlation is function of linear dependencies only. The plots below show mutual information and correlation for Lorenz time series. As seen correlation function has a mean at tau ~ 30 and mutual information has a first minimum at tau ~ 12. Which one do I consider ?? The answer can be found simply by plotting reconstructed phase space with both the tau values and compare with the actual attractor.

Page 15 of 36 Research Presentation Reconstructed phase space (Lorenz) Two plots below show reconstructed phase space for tau value 15 ( in the case of ami) and 30 ( as shown by correlation). As can be seen from the plots below, the attractor looks clean and opened up at tau = 15. At tau = 30, the attractor is quite unrecognizable.

Page 16 of 36 Research Presentation Dimension Estimation (Next task) Many definitions of dimension exist. Out of all, we are interested in four types. Though we will not be implementing all the four, it would be nice to know them. 1. Box Counting Dimension or Capacity Dimension: Simple and easily understood of all.

Page 17 of 36 Research Presentation Dimension estimation (next task) The Information Dimension: Instead of simply counting each cube which contains part of attractor, how much of the attractor is contained within each cube is calculated. This measure seeks the differences in the distribution density points covering the attractor. For the case with evenly distributed points, information dimension is same as box counting dimension. Correlation Dimension: Why correlation dimension ? Calculating box counting dimension and information dimension require a prohibitive amount of computation time. The correlation dimension may be defined as follows: More formally correlation dimension is defined using Heaviside function:

Page 18 of 36 Research Presentation Correlation Dimension The hyper sphere is constructed with one of the points as center. The number of points on the attractor with in the hypersphere are counted. (Heaviside function does this) The Heaviside function is equal to unity if the distance between two points is less than hyper sphere radius; otherwise value zero is returned. The calculation of correlation sum involves following the reference trajectory, stopping at each discrete point and counting the number of attractor points within the hypersphere. The cumulative sum is then divided by N(N – 1) to get the correlation dimension. The maximum value of correlation dimension is 1 and is attained when radius of the probing hypersphere is greater than the attractor’s largest diameter and all the points are counted

Page 19 of 36 Research Presentation Dimension estimation Lyapunov Dimension: Relationship between fractal dimension and their spectrum of Lyapunov exponents. The most useful of all the dimensions. Based on dynamics rather than on geometry. Defined in terms of Lyapunov exponents (which describe the dynamics of a system). The Lyapunov dimension may be defines as:

Page 20 of 36 Research Presentation Next Tasks These are the assignments I will be working on rest of month. Task 1: Applying mutual information and finding tau for various phones ( on speech ). Task 2: Finding the correlation dimension and testing it for various time series. Task 3: Finding Lyapunov dimension.

Page 21 of 36 Research Presentation Summary Langauage Model ABNF removes the complexities in ABNF by removing meta symbols. Mutual Information estimation is the best method to estimate optimum time delay. The invariance of the tau value with variation in the parameters like the number of bins makes it better than correlation. Lyapunov dimension and Correlation dimension are the most useful dimensions.

Page 22 of 36 Research Presentation Language Model ABNF: Java Speech Grammar Format Specification, Version 1, Sun Microsystems Developer Network, October 26, 1998 (see A. Hunt and S. McGlashan, Eds., W3C Speech Recognition Grammar Specification Version 1.0, March 16, 2004 (see J. Picone, et al., “A Public Domain C++ Speech Recognition Toolkit,” Center for Advanced Vehicular Systems, Mississippi State University, Mississippi State, MS, USA, January J.E. Hopcroft, R. Motwani and J. D. Ullman, Introduction to Automata Theory, Languages, and Computation, Addison Wesley, Boston, MA, D. Jurafsky and J. H. Martin, Speech and Language Processing, Pearson, Singapore, 2000 Optimum time delay estimation and dimension estimation: P. S. Addison, Fractals and chaos, Institute of Physics publishing, London, U.K.,1997. H. Kantz and T. Schreiber, Nonlinear Time Series Analysis, Second Edition, Cambridge University Press, U.K., A. H. Nayfeh, B. Balachandran, Applied Nonlinear Dynamics, Wiley-Interscience Publication, NewYork, USA, References