Presentation is loading. Please wait.

Presentation is loading. Please wait.

Institute of Computer Science, Prague 1 Neural Networks Marcel Jiřina.

Similar presentations


Presentation on theme: "Institute of Computer Science, Prague 1 Neural Networks Marcel Jiřina."— Presentation transcript:

1 Institute of Computer Science, Prague 1 Neural Networks Marcel Jiřina

2 Institute of Computer Science, Prague 2 Introduction Neural networks and their use to classification and other tasks ICS AS CR  Theoretical computer science  Neural networks, genetic alg. and nonlinear methods  Numeric algorithms..1 mil. eq.  Fuzzy sets, approximate reasoning, possibility th.  Applications: Nuclear science, Ecology, Meteorology, Reliability in machinery, Medical informatics …

3 Institute of Computer Science, Prague 3 Structure of talk NN classification Some theory Interesting paradigms NN and statistics NN and optimization and genetic algorithms About application of NN Conlusions

4 Institute of Computer Science, Prague 4 NN classification NE – not existing. Associated response can be arbitrary and then must be given - by teacher Feed-forward, recurrent Fixed structure - growing ApproximatorsAssociative memories GeneralPredictorsAuto- associative Hetero- associative Classifiers Teacher MLP-BP RBF GMDH NNSU Marks Klán Hopfield Perceptron(*)Hamming No teacher Kohonen Carpentier Grossberg (SOM) NEKohonen (NE) SignalsContinuous, real-valuedBinary, multi-valued (continuous)

5 Institute of Computer Science, Prague 5 Some theory Kolmogorov theorem Kůrková – Theorem Sigmoid transfer function 

6 Institute of Computer Science, Prague 6 MLP - BP Three layer - Single hidden layer MLP – 4 layer – 2 hidden Other paradigms have its own theory – another

7 Institute of Computer Science, Prague 7 Interesting paradigms Paradigm – general notion on structure, functions and algorithms of NN MLP - BP RBF GMDH NNSU All: approximators Approximator + thresholding = Classifier

8 Institute of Computer Science, Prague 8 MLP - BP MLP – error Back Propagation coefficients ,   (0,1) - Lavenberg-Marquart - Optimization tools MLP with jump transfer function - Optimization Feed – forward (in recall) Matlab, NeuralWorks, … Good when default is sufficient or when network is well tuned: Layers, neurons, , 

9 Institute of Computer Science, Prague 9 RBF Structure same as in MLP Bell-shaped transfer function (Gauss)  Number and positions of centers: random – cluster analysis  “broadness” of that bell  Size of individual bells  Learning methods Theory similar to MLP Matlab, NeuralWorks, … Good when default is sufficient or when network is well tuned : Layers mostly one hidden, # neurons, transfer function, proper cluster analysis (fixed No. of clusters, variable? Near – Far metric or criteria)

10 Institute of Computer Science, Prague 10 GMDH 1 (…5) Group Method Data Handling –Group – initially a pair of signals only “per partes” or successive polynomial approximator Growing network “parameterless” – parameter-barren –No. of new neurons in each layer only (processing time) –(output limits, stopping rule parameters) Overtraining – learning set is split to –Adjusting set –Evaluation set GMDH 2-5: neuron, growing network, learning strategy, variants

11 Institute of Computer Science, Prague 11 GMDH 2 – neuron Two inputs x 1, x 2 only –True inputs –Outputs from neurons of the preceding layer Full second order polynomial y = a x 1 2 + b x 1 x 2 + c x 2 2 + d x 1 + e x 2 + f y = neuron’s output n inputs => n(n-1)/2 neurons in the first layer Number of neurons grows exponentially Order of resulting polynomial grows exponentially: 2, 4, 8, 16, 32, … Ivakhnenko polynomials … some elements are missing

12 Institute of Computer Science, Prague 12 GMDH 3 – learning a neuron Matrix of data: inputs and desired value u 1, u 2, u 3, …, u n, y sample 1 …. sample m A pair of two u’s are neuron’s inputs x 1, x 2 m approximating equations, one for each sample a x 1 2 + b x 1 x 2 + c x 2 2 + d x 1 + e x 2 + f = y MatrixX  = Y  = (a, b, c, d, e, f) t  Each row of X is x 1 2 +x 1 x 2 +x 2 2 +x 1 +x 2 +1 LMS solution  = (X t X) -1 X t Y If X t X is singular, we omit this neuron

13 Institute of Computer Science, Prague 13 GMDH 4 - growing network x 1, x 2 y = desired output

14 Institute of Computer Science, Prague 14 GMDH 5 learn. strategy Problem: Number of neurons grows exponentially NN=n(n-1)2 Let the first layer of neurons grow unlimited In next rows:  [learning set split to adjusting set and evaluating set]  Compute parameters a,…f using adjusting set  Evaluate error using evaluating set and sort  Select some n best neurons and delete the others  Build the next layer OR  Stop learning if stopping condition is met.

15 Institute of Computer Science, Prague 15 GMDH 6 learn. Strategy 2 Select some n best neurons and delete the others Control parameter of GMDH network

16 Institute of Computer Science, Prague 16 GMDH 7 - variants Basic – full quadratic polynomial – Ivakh. poly Cubic, Fourth order simplified …  Reach higher order in less layers and less params Different stopping rules Different ratio of sizes of adjusting set and evaluating set

17 Institute of Computer Science, Prague 17 NNSU GA Neural Network with Switching Units learned by the use of Genetic Algorithm Approximator by lot of local hyper-planes; today also by local more general hyper-surfaces Feed-forward network Originally derived from MLP for optical implementation Structure looks like columns above individual inputs More … František

18 Institute of Computer Science, Prague 18 Learning and testing set Learning set  Adjusting (tuning) set  Evaluation set Testing set One data set – the splitting influences results Fair evaluation problem

19 Institute of Computer Science, Prague 19 NN and statistics MLP-BP mean squared error minimization  Sum of errors squared … MSE criterion  Hamming distance for (pure) classifiers No other statistical criteria or tests are in NN:  NN transforms data, generates mapping  statistical criteria or tests are outside NN (  2, K-S, C-vM,…) Is NN good for K-S test? … is y=sin(x) good for  2 test? Bayes classifiers, k-th nearest neighbor, kernel methods …

20 Institute of Computer Science, Prague 20 NN and optimization and genetic algorithms Learning is an optimization procedure Specific to given NN General optimization systems or methods Whole NN Parts – GMDH and NNSU - linear regression Genetic algorithm  Not only parameters, the structure, too  May be faster than iterations

21 Institute of Computer Science, Prague 21 About application of NN Soft problems  Nonlinear  Lot of noise  Problematic variables  Mutual dependence of variables Application areas  Economy  Pattern recognition  Robotics  Particle physics ……

22 Institute of Computer Science, Prague 22 Strategy when using NN For “soft problems” only NOT for  Exact function generation  periodic signals etc. First subtract all “systematics”  Nearly noise remains  Approximate this nearly noise  Add back all systematics Understand your paradigm  Tune it patiently or  Use “parameterless” paradigm

23 Institute of Computer Science, Prague 23 Conlusions Powerfull tool  Good when well used  Simple paradigm, complex behavior Special tool  Approximator  Classifier Universal tool  Very different problems  Soft problems


Download ppt "Institute of Computer Science, Prague 1 Neural Networks Marcel Jiřina."

Similar presentations


Ads by Google