June 2003Neural Computation for Time Series1 Neural Computation and Applications in Time Series and Signal Processing Georg Dorffner Dept. of Medical Cybernetics.

Slides:



Advertisements
Similar presentations
A Brief Overview of Neural Networks By Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch.
Advertisements

Slides from: Doug Gray, David Poole
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Artificial Neural Networks (1)
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Navneet Goyal, BITS-Pilani Perceptrons. Labeled data is called Linearly Separable Data (LSD) if there is a linear decision boundary separating the classes.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Kostas Kontogiannis E&CE
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Machine Learning Neural Networks
Lecture 14 – Neural Networks
Supervised and Unsupervised learning and application to Neuroscience Cours CA6b-4.
x – independent variable (input)
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 15: Introduction to Artificial Neural Networks Martin Russell.
Neural Networks Marco Loog.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward.
MACHINE LEARNING 12. Multilayer Perceptrons. Neural Networks Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Artificial Neural Networks
LOGO Classification III Lecturer: Dr. Bo Yuan
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
CS 484 – Artificial Intelligence
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
CSC2535: 2013 Advanced Machine Learning Lecture 3a: The Origin of Variational Bayes Geoffrey Hinton.
Radial Basis Function (RBF) Networks
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
Radial Basis Function Networks
1 Introduction to Artificial Neural Networks Andrew L. Nelson Visiting Research Faculty University of South Florida.
Neural Networks.
Soft Computing Colloquium 2 Selection of neural network, Hybrid neural networks.
Neural Networks. Plan Perceptron  Linear discriminant Associative memories  Hopfield networks  Chaotic networks Multilayer perceptron  Backpropagation.
Predicting volatility: a comparative analysis between GARCH Models and Neural Network Models MCs Student: Miruna State Supervisor: Professor Moisa Altar.
Biointelligence Laboratory, Seoul National University
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Machine Learning Chapter 4. Artificial Neural Networks
Chapter 3 Neural Network Xiu-jun GONG (Ph. D) School of Computer Science and Technology, Tianjin University
11 CSE 4705 Artificial Intelligence Jinbo Bi Department of Computer Science & Engineering
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Artificial Intelligence Techniques Multilayer Perceptrons.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Artificial Intelligence Chapter 3 Neural Networks Artificial Intelligence Chapter 3 Neural Networks Biointelligence Lab School of Computer Sci. & Eng.
Non-Bayes classifiers. Linear discriminants, neural networks.
CS621 : Artificial Intelligence
Fundamentals of Artificial Neural Networks Chapter 7 in amlbook.com.
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
Neural Networks Demystified by Louise Francis Francis Analytics and Actuarial Data Mining, Inc.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Neural Networks 2nd Edition Simon Haykin
CHEE825 Fall 2005J. McLellan1 Nonlinear Empirical Models.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Today’s Lecture Neural networks Training
Machine Learning Supervised Learning Classification and Regression
Deep Feedforward Networks
CSE P573 Applications of Artificial Intelligence Neural Networks
Machine Learning Today: Reading: Maria Florina Balcan
Neuro-Computing Lecture 4 Radial Basis Function Network
Artificial Intelligence Chapter 3 Neural Networks
CSE 573 Introduction to Artificial Intelligence Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
Introduction to Radial Basis Function Networks
Artificial Intelligence Chapter 3 Neural Networks
Artificial Intelligence 10. Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
PYTHON Deep Learning Prof. Muhammad Saeed.
Artificial Intelligence Chapter 3 Neural Networks
Presentation transcript:

June 2003Neural Computation for Time Series1 Neural Computation and Applications in Time Series and Signal Processing Georg Dorffner Dept. of Medical Cybernetics and Artificial Intelligence, University of Vienna And Austrian Research Institute for Artificial Intelligence

June 2003Neural Computation for Time Series2 Neural Computation Originally biologically motivated (information processing in the brain) Simple mathematical model of the neuron  neural network Large number of simple „units“ Massively parallel (in theory) Complexity through the interplay of many simple elements Strong relationship to methods from statistics Suitable for pattern recognition

June 2003Neural Computation for Time Series3 A Unit Propagation rule: –Weighted sum –Euclidian distance Transfer function f: –Threshold fct. (McCulloch & Pitts) –Linear fct. –Sigmoid fct. –Gaussian fct.  y j f xjxj w1w1 w2w2 wiwi … Weight Unit (Neuron) Activation, Output (Net-) Input

June 2003Neural Computation for Time Series4 Multilayer Perceptron (MLP), Radial Basis Function Network (RBFN) 2 (or more) layers (= connections) Input Units Hidden Units (typically nonlinear) Output Units (typically linear) MLP: RBFN:

June 2003Neural Computation for Time Series5 MLP as Universal Function Approximator E.g,: 1 Input, 1 Output, 5 Hidden MLP can approximate arbitrary functions (Hornik et al. 1990) trough superposition of weighted sigmoids Similar is true for RBFN move (bias) Stretch, mirror

June 2003Neural Computation for Time Series6 Training (Model Estimation) Typical error function: „Backpropagation“ (application of chain rule): contribution of error function contribution of network Iterative optimisation based on gradient (gradient descent, conjugent gradient, quasi-Newton): (summed squared error) target all patterns all outputs

June 2003Neural Computation for Time Series7 Recurrent Perceptrons Recurrent connection = feedback loop From hidden layer („Elman“) or output layer („Jordan“) Learning: „backpropagation through time“ Input Zustands- bzw. Kontextlayer copy

June 2003Neural Computation for Time Series8 Time series processing Given: time-dependent observables Scalar: univariate; vector: multivariate Typical tasks: - Forecasting - Noise modeling - Pattern recognition - Modeling - Filtering - Source separation Time series (minutes to days) Signals (milliseconds to seconds)

June 2003Neural Computation for Time Series9 Examples Standard & Poor‘sSunspots Preprocessed: (returns) Preprocessed: (de-seasoned)

June 2003Neural Computation for Time Series10 Autoregressive models Forecasting: making use of past information to predict (estimate) the future AR: Past information = past observations past observations Expected value Noise, „random shock“ Best forecast: expected value

June 2003Neural Computation for Time Series11 Linear AR models Most common case: Simplest form: random walk Nontrivial forecast impossible

June 2003Neural Computation for Time Series12 MLP as NAR Neural network can approximate nonlinear AR model „time window“ or „time delay“

June 2003Neural Computation for Time Series13 Noise modeling Regression is density estimation of: (Bishop 1995) Likelihood: Distribution with expected value F(x i ) Target = future past

June 2003Neural Computation for Time Series14 Gaussian noise Likelihood: Maximization = minimization of -logL (constant terms can be deleted, incl. p(x)) Corresponds to summed squared error (typical backpropagation)

June 2003Neural Computation for Time Series15 Complex noise models Assumption: arbitrary distribution Parameters are time dependent (dependent on past): Likelihood: Probability density function for D

June 2003Neural Computation for Time Series16 Heteroskedastic time series Assumption: Noise is Gaussian with time-dependent variance ARCH model MLP is nonlinear ARCH (when applied to returns/residuals)

June 2003Neural Computation for Time Series17 Non-Gaussian noise Other parametric pdfs (e.g. t-distribution) Mixture of Gaussians (Mixture density network, Bishop 1994) Network with 3k outputs (or separate networks)

June 2003Neural Computation for Time Series18 Identifiability problem Mixture models (like neural networks) are not identifiable (parameters cannot be interpreted) No distinction between model and noise e.g. sunspot data:  Models have to be treated with care

June 2003Neural Computation for Time Series19 Recurrent networks: Moving Average Second model class: Moving Average models Past information: random shocks Recurrent (Jordan) network: Nonlinear MA However, convergence not guaranteed

June 2003Neural Computation for Time Series20 GARCH Extension of ARCH: Explains „volatility clustering“ Neural network can again be a nonlinear version Using past estimates: recurrent network

June 2003Neural Computation for Time Series21 State space models Observables depend on (hidden) time-variant state Strong relationship to recurrent (Elman) networks Nonlinear version only with additional hidden layers

June 2003Neural Computation for Time Series22 Symbolic time series Examples: –DNA –Text –Quantised time series (e.g. „up“ and „down“) Past information: past p symbols  probability distribution Markov chains Problem: long substrings are rare alphabet

June 2003Neural Computation for Time Series23 Fractal prediction machines Similar subsequences are mapped to points close in space Clustering = extraction of stochastic automaton

June 2003Neural Computation for Time Series24 Relationship to recurrent network Network of 2nd order

June 2003Neural Computation for Time Series25 Other topics Filtering: corresponds to ARMA models NN as nonlinear filters Source separation independent component analysis Relationship to stochastic automata

June 2003Neural Computation for Time Series26 Practical considerations Stationarity is an important issue Preprocessing (trends, seasonalities) N-fold cross-validation time-wise (validation set must be after training set Mean and standard deviation  model selection train validation test

June 2003Neural Computation for Time Series27 Summary Neural networks are powerful semi-parametric models for nonlinear dependencies Can be considered as nonlinear extensions of classical time series and signal processing techniques Applying semi-parametric models to noise modeling adds another interesting facet Models must be treated with care, much data necessary