Presentation is loading. Please wait.

Presentation is loading. Please wait.

ARTIFICIAL NEURAL NETWORKS. Overview EdGeneral concepts Areej:Learning and Training Wesley:Limitations and optimization of ANNs Cora:Applications and.

Similar presentations


Presentation on theme: "ARTIFICIAL NEURAL NETWORKS. Overview EdGeneral concepts Areej:Learning and Training Wesley:Limitations and optimization of ANNs Cora:Applications and."— Presentation transcript:

1 ARTIFICIAL NEURAL NETWORKS

2 Overview EdGeneral concepts Areej:Learning and Training Wesley:Limitations and optimization of ANNs Cora:Applications and examples

3 Biological Inspiration

4 Neuron

5 Activation Function

6 Errors Is proportional to

7 Layers – Famous Problem

8 Layers – Building Blocks

9 Layers - Idea

10 Layers – Solution P1

11 Layers – Solution P2 3 layers“feedforward”

12 LEARNING AND TRAINING

13 Learning methods Supervised learning. Learn to predict an output when given an input. Unsupervised learning. Discover a good internal representation of the input.

14 Types of Artificial Neural Networks Neural network topologies can vary widely, resulting in differences in architecture, function, and behaviour. Feed- forward ANN. Recurrent ANN.

15 Feed- forward Neural Networks It is the first and the simplest type of ANNs. This type of organisation is also referred to as bottom-up or top-down. It allows signals to travel in one direction only. From input to output through the hidden layers if there were any. There are no feedback (loops or circles). The output of any layer does not affect that same layer.

16 Teaching process of multilayer ANN Image source [2]

17

18 Backpropagation algorithm Image source [2]

19 Backpropagation algorithm Image source [2]

20 Backpropagation algorithm Image source [2]

21 Backpropagation algorithm Image source [2]

22 Backpropagation algorithm Image source [2]

23 Backpropagation algorithm Image source [2]

24 Backpropagation algorithm Image source [2]

25 Backpropagation algorithm Image source [2]

26 Backpropagation algorithm Image source [2]

27 Backpropagation algorithm Image source [2]

28 Backpropagation algorithm Image source [2]

29 Backpropagation algorithm Image source [2]

30

31

32

33

34

35

36 Recurrent Neural Networks It can be called as feed- back or interactive networks. It have signals traveling in both directions (loops and circles). Feedback networks are powerful and can get extremely complicated. Computations derived from earlier input are fed back into the network, which gives them a kind of memory.

37 LIMITATIONS & OPTIMISATIONS

38 Overfitting When performance on training data is high but performance on unseen data is poor. Effectively memorizing training data without learning generalized rules.

39 Split-sample testing Available test data is split into training set, validation set, and test set. Final weights are those used at stopping point. Image source: [1]

40 Regularization Reduces overfitting by adjusting the training process. Lx regularization adds a regularization term to the cost function. L2 regularization or weight decay: λ is the regularization parameter. Tends towards smaller overall weights. Less sensitive to noise.

41 Regularization (cont’) L1 regularization: Tends to maintain most useful weights and shrink others.

42 Regularization (cont’) Dropout involves training with only a subset of neurons active. For each training run, half of the neurons are randomly chosen to be active, the others unused. Final network is less dependent on any particular input or neuron. More testing data is better for generalization, but expensive. Training data can be artificially increased by adjusting existing data.

43 Learning slowdown Using Mean-Squared-Error for the cost function can result in very slow learning when the derivative of the output is small. Learning can be initially slow if weight initialisation is far from correct.

44 Learning slowdown (cont’) The Cross-Entropy cost function eliminates the output derivative from the weight gradients. Weights learn solely based on output value and not slope of output when using sigmoid functions.

45 Learning slowdown (cont’) Alternative cost function is log-likelihood function combined with softmax output layer. Output represents a probability distribution. Useful for classifiers.

46 Normalization Scales data so that each input lies in same range, reducing influence of one input feature over another. Speeds up training when the process starts with all input features on same scale. If applied to input in training set then must subsequently be applied to all future input.

47 Normalization (cont’) ●Z-Score or Gaussian normalization scales by mean (μ) and standard deviation (σ). Min-Max normalization scales input to lie in a fixed range. Sigmoidal normalization reduces the effects of outliers in the input.

48 Vanishing Gradient Problem Increasing number of hidden layers reduces required complexity of network but makes training harder. Vanishing gradient problem is where the gradient of weights in of early layers are small in multi-layer networks compared to later layers and thus learn slower. Alternative is exploding gradient problem. Unstable gradient problem can be avoided with pre- training. Layers are trained one at a time to produce a compressed form of the input.

49 APPLICATIONS & EXAMPLES

50 Application Areas 1. Function approximation 2. Pattern recognition 3.Database mining Problems 1. Car insurance policy – claim frequency 2. Door security -- Fingerprint 3. Medical data mining -- Diagnosing lung cancer

51 Structure Image source:[1]

52 Image source:[8] Car insurance policy –claim frequency

53 Door security--fingerprint Image source:[9]

54 Image source:[10]

55 References [1] L. L. Priddy and P.E. Keller (2005). Artificial Neural Networks: An introduction. Vol. 68: SPIE press. [2] M. Bernacki. (2005) Principles of training multi-layer neural network using backpropagation. Available: http://home.agh.edu.pl/~vlsi/AI/backp_t_en/backprop.html [3] S. Haykin and N. Network, "A comprehensive foundation," Neural Networks, vol. 2, 2004. [4] M. James. (2012), The Triumph of Deep Learning. Available: www.i-programmer.info/programming/artificial- intelligence/5206-the-triumph-of-deep-learning.html [5] J. McCaffrey. (2014), How To Standardize Data for Neural Networks. Visual Studio Magazine. Available: visualstudiomagazine.com/articles/2014/01/01/how-to-standardize-data-for-neural-networks.aspx [6] M. Nielsen. (2015). Neural Networks and Deep Learning. Available: neuralnetworksanddeeplearning.com/ [7] W. Wu, H. Maier, G. Dandy, and R. May, "Exploring the impact of data splitting methods on artificial neural network models," in International Conference on Hydroinformatics (10th: 2012: Hamburg, Germany), 2012. [8] http://www2.units.it/renatop/papers/pelessoni_picech.pdf [9] http://www.iasj.net/iasj?func=fulltext&aId=75336 [10] http://www.irdindia.in/journal_ijacect/pdf/vol1_iss1/11.pdf

56 Thank you


Download ppt "ARTIFICIAL NEURAL NETWORKS. Overview EdGeneral concepts Areej:Learning and Training Wesley:Limitations and optimization of ANNs Cora:Applications and."

Similar presentations


Ads by Google