數值方法 2008, Applied Mathematics NDHU 1  Nonlinear Recursive relations  Nonlinear recursive relations  Time series prediction  Hill-Valley classification.

Slides:



Advertisements
Similar presentations
FUNCTION FITTING Student’s name: Ruba Eyal Salman Supervisor:
Advertisements

EE-M /7: IS L7&8 1/24, v3.0 Lectures 7&8: Non-linear Classification and Regression using Layered Perceptrons Dr Martin Brown Room: E1k
數值方法 2008, Applied Mathematics NDHU 1 Nonlinear systems Newton’s method The steepest descent method.
Data Mining and Machine Learning
Support Vector Machines and Kernels Adapted from slides by Tim Oates Cognition, Robotics, and Learning (CORAL) Lab University of Maryland Baltimore County.
SVM—Support Vector Machines
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Hidden Markov Models. Room Wandering I’m going to wander around my house and tell you objects I see. Your task is to infer what room I’m in at every point.
Ch. 4: Radial Basis Functions Stephen Marsland, Machine Learning: An Algorithmic Perspective. CRC 2009 based on slides from many Internet sources Longin.
Decision making in episodic environments
Artificial Intelligence Statistical learning methods Chapter 20, AIMA (only ANNs & SVMs)
Classification: Support Vector Machine 10/10/07. What hyperplane (line) can separate the two classes of data?
Margins, support vectors, and linear programming, oh my! Reading: Bishop, 4.0, 4.1, 7.0, 7.1 Burges tutorial (on class resources page)
An Illustrative Example
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks / Fall 2004 Shreekanth Mandayam ECE Department Rowan University.
Support Vector Machines
Data Mining with Neural Networks (HK: Chapter 7.5)
General Mining Issues a.j.m.m. (ton) weijters Overfitting Noise and Overfitting Quality of mined models (some figures are based on the ML-introduction.
Introduction to Neural Networks Kenji Nakayama Kanazawa University, JAPAN 適応システム理論 ガイダンス.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks ECE /ECE Fall 2010 Shreekanth Mandayam ECE Department Rowan University.
Ranga Rodrigo April 5, 2014 Most of the sides are from the Matlab tutorial. 1.
數值方法 2008, Applied Mathematics NDHU1  An iterative approach for root finding  Newton method Lecture 3II Root Finding.
數值方法 2008, Applied Mathematics NDHU 1 Simpson rule Composite Simpson rule.
Soft Computing Colloquium 2 Selection of neural network, Hybrid neural networks.
數值方法 2008 Applied Mathematics, NDHU1  Bisection method for root finding  Binary search  fzero Lecture 4.
數值方法 2008 Applied Mathematics, NDHU1  Bisection method for root finding  Binary search  fzero Lecture 4.
軟體實作與計算實驗 1 Lecture 9 Classification Two mappings from position to color Linear combination of distances to centers RBF: Linear combination of exponential.
數值方法 2008, Applied Mathematics NDHU1  Looping for loop while loop  Break and return  Recursion Lecture 2 Iteration.
Applications of Neural Networks in Time-Series Analysis Adam Maus Computer Science Department Mentor: Doctor Sprott Physics Department.
數值方法 2008, Applied Mathematics NDHU 1 Numerical Integration.
Development of neural network emulations of model physics components for improving the computational performance of the NCEP seasonal climate forecasts.
Adavanced Numerical Computation 2008, AM NDHU1 Function approximation Linear models Line fitting Hyper-plane fitting Discriminate analysis Least Square.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
數值方法 2008 Applied Mathematics, NDHU1  Lagrange polynomial  Polynomial interpolation Lecture 4II.
數值方法 2008, Applied Mathematics NDHU1 Chaos time series.
HMMs and SVMs for Secondary Structure Prediction
Learning Chaotic Dynamics from Time Series Data A Recurrent Support Vector Machine Approach Vinay Varadan.
Supervised Machine Learning: Classification Techniques Chaleece Sandberg Chris Bradley Kyle Walsh.
數值方法 2008, Applied Mathematics NDHU 1 Numerical Differentiation.
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
Vincenzo Piuri, SIcon/02, Houston, TX, USA, November 2002 Requirements Accuracy –It measures the performance of the final solution according to a.
Landmark-Based Speech Recognition: Spectrogram Reading, Support Vector Machines, Dynamic Bayesian Networks, and Phonology Mark Hasegawa-Johnson
數值方法, Applied Mathematics NDHU 1 Linear system. 數值方法, Applied Mathematics NDHU 2 Linear system.
Tree Diagrams Objective: To calculate probability using a tree diagram. Lesley Soar Valley College Five balls are put into a bag. Three are red.
數值方法 2008, Applied Mathematics NDHU 1 Ordinary differential equations II Runge-Kutta method.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Big data classification using neural network
The Gradient Descent Algorithm
Adavanced Numerical Computation 2008, AM NDHU
Lecture 5 Polynomial Approximation
One-layer neural networks Approximation problems
Lecture 3 Taylor Series Expansion
A system of nonlinear equations
Intelligent Information System Lab
NEURAL NETWORK APPROACHES FOR AUTOMOBILE MPG PREDICTION
NEURAL NETWORKS Lab 1 dr Zoran Ševarac FON, 2013.
Data Mining with Neural Networks (HK: Chapter 7.5)
Decision making in episodic environments
Figure S1: Gene importance plot derived from Variable/ Feature selection using machine learning on the training dataset. MeanDecreaseGini is the measure.
Using Artificial Neural Networks and Support Vector Regression to Model the Lyapunov Exponent Adam Maus.
Emre O. Neftci  iScience  Volume 5, Pages (July 2018) DOI: /j.isci
Image recognition: Defense adversarial attacks
Generating Coherent Patterns of Activity from Chaotic Neural Networks
Identification of Wiener models using support vector regression
Forecasting Electricity Demand and Prices with Machine Learning
CS539 Project Report -- Evaluating hypothesis
Recursively Adapted Radial Basis Function Networks and its Relationship to Resource Allocating Networks and Online Kernel Learning Weifeng Liu, Puskal.
Fig. 1. A diagram of artificial neural network consisting of multilayer perceptron. This simple diagram is for a conceptual explanation. When the logistic.
Artificial Neural Networks ECE /ECE Fall 2006
Presentation transcript:

數值方法 2008, Applied Mathematics NDHU 1  Nonlinear Recursive relations  Nonlinear recursive relations  Time series prediction  Hill-Valley classification  Lorenz series Neural Recursions

數值方法 2008, Applied Mathematics NDHU 2 Nonlinear recursion Post-nonlinear combination of predecessors z[t]=tanh(a 1 z[t-1]+ a 2 z[t-2]+…+ a  z[t-  ])+ e[t], t= ,…,N

數值方法 2008, Applied Mathematics NDHU 3 Post-nonlinear Projection tanh y

數值方法 2008, Applied Mathematics NDHU 4 Data generation z[t]=tanh(a 1 z[t-1]+ a 2 z[t-2]+…+ a  z[t-  ])+ e[t], t= ,…,N tanh

數值方法 2008, Applied Mathematics NDHU 5 Nonlinear Recursion LM learning

數值方法 2008, Applied Mathematics NDHU 6 Prediction Use initial  -1 instances to generate the full time series (red) based on estimated post-nonlinear filter Post-nonlinear filter

數值方法 2008, Applied Mathematics NDHU 7 UCI Machine Learning Repository Nonlinear recursions: hill-valley Classify hill and valley

數值方法 2008, Applied Mathematics NDHU 8 Noise-free Hill Valley

數值方法 2008, Applied Mathematics NDHU 9 Noise-free Hill Valley

數值方法 2008, Applied Mathematics NDHU 10 Hill-Valley with noise

數值方法 2008, Applied Mathematics NDHU 11 Hill-Valley with noise

數值方法 2008, Applied Mathematics NDHU 12 Hill-Valley with noise

數值方法 2008, Applied Mathematics NDHU 13 Nonlinear recursion LM_RBF: z[t]=F(z[t-1], z[t-2], …,z[t-  ])+ e[t], t= ,…,N LM_RBF 1

數值方法 2008, Applied Mathematics NDHU 14 LM_RBF 1 LM_RBF 2

數值方法 2008, Applied Mathematics NDHU 15 Nonlinear Recursion LM learning MLPotts 1 (H)

數值方法 2008, Applied Mathematics NDHU 16 Nonlinear Recursion LM learning MLPotts 2 (V)

數值方法 2008, Applied Mathematics NDHU 17 Approximating error LM_RBF 1 (H) Green: Original Hill Blue: Approximated Hill Red: Approximating error

數值方法 2008, Applied Mathematics NDHU 18 Approximating error LM_RBF 2 (V) Green: Original Hill Blue: Approximated Hill Red: Approximating error

數值方法 2008, Applied Mathematics NDHU 19 Separable 2D diagram of Hill-Valley Noise-free data set 1 LM_RBF 2 (V) LM_RBF 1 (H) Mean Approximating Errors of Two Models

數值方法 2008, Applied Mathematics NDHU 20 Separable 2D diagram of Hill-Valley Noise-free data set 2 Mean Approximating Errors of Two Models LM_RBF 2 (V) LM_RBF 1 (H)

數值方法 2008, Applied Mathematics NDHU 21 Separable 2D diagram of Hill-Valley Dataset 1 (Noise) Mean Approximating Errors of Two Models LM_RBF 2 (V) LM_RBF 1 (H)

數值方法 2008, Applied Mathematics NDHU 22 Separable 2D diagram of Hill-Valley Dataset 2 (Noise) Mean Approximating Errors of Two Models LM_RBF 2 (V) LM_RBF 1 (H)

數值方法 2008, Applied Mathematics NDHU 23 Chaos Time Series Eric's Home Page

數值方法 2008, Applied Mathematics NDHU 24 Lorenz Attractor

數值方法 2008, Applied Mathematics NDHU 25 Lorenz >> NN=20000; >> [x,y,z] = lorenz(NN); >> plot3(x,y,z,'.');

數值方法 2008, Applied Mathematics NDHU 26

數值方法 2008, Applied Mathematics NDHU 27 Nonlinear Recursion Neural nonlinear recursive relations

數值方法 2008, Applied Mathematics NDHU 28 Prediction Use initial  -1 instances to generate the full time series (red) based on derived neural recursions

數值方法 2008, Applied Mathematics NDHU 29 ρ=14, σ=10, β=8/3 ρ=13, σ=10, β=8/3 ρ=15, σ=10, β=8/3ρ=28, σ=10, β=8/3 Rayleigh number

數值方法 2008, Applied Mathematics NDHU 30 Sunspot time series sunpot data load sunspot_train.mat; n=length(Y); plot(1712:1712+n-1,Y);

數值方法 2008, Applied Mathematics NDHU 31 Sunspot time series sunpot data (testing) load sunspot_test_1.mat; n=length(Y); plot(1952:1952+n-1,Y);

數值方法 2008, Applied Mathematics NDHU 32 S[1] S[2] S[3]. S[  -1] S[  ]. S[t-  ] S[t-  +1]. S[t-1] S[t] S[t+1]. x[t] y[t] x[  -1] y[  -1] Time-series 2 paired data

數值方法 2008, Applied Mathematics NDHU 33 aMaM a2a2 a1a1 x Network tanh ……. 1 b2b2 b1b1 bMbM 1 r1r1 r2r2 rMrM r M+1 y

數值方法 2008, Applied Mathematics NDHU 34 Paired data : tau=2 x=[]; len=3; for i=1:n-len p=Y(i:i+len-1)'; x=[x p]; y(i)=Y(i+len); end figure; plot3(x(2,:),x(1,:),y,'.');hold on; y x

數值方法 2008, Applied Mathematics NDHU 35 Learning and prediction M=input('keyin the number of hidden units:'); [a,b,r]=learn_MLP(x',y',M); y=eval_MLP2(x,r,a,b,M); hold on; plot(1712+2: length(y)-1,y,'r') MSE for training data ME for training data K

數值方法 2008, Applied Mathematics NDHU 36 Prediction

數值方法 2008, Applied Mathematics NDHU 37 Sunspot Source codes

數值方法 2008, Applied Mathematics NDHU 38 Paired data: tau=5 x=[]; len=5; for i=1:n-len p=Y(i:i+len-1)'; x=[x p]; y(i)=Y(i+len); end

數值方法 2008, Applied Mathematics NDHU 39 Learning and prediction M=input('keyin the number of hidden units:'); [a,b,r]=learn_MLP(x',y',M); y=eval_MLP2(x,r,a,b,M); hold on; plot(1712+2: length(y)-1,y,'r') ???

數值方法 2008, Applied Mathematics NDHU 40 Example Repeat numerical experiments in slide #34 with Summarize your numerical results in a table

數值方法 2008, Applied Mathematics NDHU 41 RBF y exp x..

數值方法 2008, Applied Mathematics NDHU 42 addpath('..\..\..\Methods\FunAppr\LM_RBF'); % x : Nxd % y : 1xN N=200;d=2; x=rand(N,d)*2-1; y=tanh(-x(:,1).^2-x(:,2).^2); eval(['load data\sunspot_train.mat']); %([ ],:) tx = X'; ty = Y; Net=LM_RBF(tx,ty); [yhat,D]=evaRBF(tx,Net); mean((ty-yhat).^2)/2

數值方法 2008, Applied Mathematics NDHU 43 eval(['load data\sunspot_test_1.mat']); tex1 = X'; tey1 = Y; [yhat,D]=evaRBF(tex1,Net); mean((tey1-yhat).^2)/2 One_step_look_ahead prediction