N. Saoulidou & G. Tzanakos1 ANN Basics : Brief Review N. Saoulidou, Fermilab & G. Tzanakos, Univ. of Athens.

Slides:



Advertisements
Similar presentations
Multi-Layer Perceptron (MLP)
Advertisements

Slides from: Doug Gray, David Poole
NEURAL NETWORKS Backpropagation Algorithm
Artificial Neural Networks (1)
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 7: Learning in recurrent networks Geoffrey Hinton.
Artificial Neural Networks
Mehran University of Engineering and Technology, Jamshoro Department of Electronic Engineering Neural Networks Feedforward Networks By Dr. Mukhtiar Ali.
Artificial Neural Networks
Machine Learning Neural Networks
Simple Neural Nets For Pattern Classification
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Optimizing number of hidden neurons in neural networks
Data Mining Techniques Outline
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 1 Artificial Neural Network (ANN) Paradigms Overview: The Backpropagation.
September 30, 2010Neural Networks Lecture 8: Backpropagation Learning 1 Sigmoidal Neurons In backpropagation networks, we typically choose  = 1 and 
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 15: Introduction to Artificial Neural Networks Martin Russell.
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks / Fall 2004 Shreekanth Mandayam ECE Department Rowan University.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Chapter 6: Multilayer Neural Networks
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks / Spring 2002 Shreekanth Mandayam Robi Polikar ECE Department.
Artificial Neural Networks
Before we start ADALINE
Data Mining with Neural Networks (HK: Chapter 7.5)
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks ECE /ECE Fall 2006 Shreekanth Mandayam ECE Department Rowan University.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
BEE4333 Intelligent Control
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Multiple-Layer Networks and Backpropagation Algorithms
Artificial Neural Networks
Multi-Layer Perceptrons Michael J. Watts
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
ANNs (Artificial Neural Networks). THE PERCEPTRON.
1 Chapter 6: Artificial Neural Networks Part 2 of 3 (Sections 6.4 – 6.6) Asst. Prof. Dr. Sukanya Pongsuparb Dr. Srisupa Palakvangsa Na Ayudhya Dr. Benjarath.
Appendix B: An Example of Back-propagation algorithm
Matlab Matlab Sigmoid Sigmoid Perceptron Perceptron Linear Linear Training Training Small, Round Blue-Cell Tumor Classification Example Small, Round Blue-Cell.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski
 Diagram of a Neuron  The Simple Perceptron  Multilayer Neural Network  What is Hidden Layer?  Why do we Need a Hidden Layer?  How do Multilayer.
Lecture 3 Introduction to Neural Networks and Fuzzy Logic President UniversityErwin SitompulNNFL 3/1 Dr.-Ing. Erwin Sitompul President University
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Intelligence Methods Neural Networks Lecture 4 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
An informal description of artificial neural networks John MacCormick.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 24 Nov 2, 2005 Nanjing University of Science & Technology.
Non-Bayes classifiers. Linear discriminants, neural networks.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
פרקים נבחרים בפיסיקת החלקיקים אבנר סופר אביב
EEE502 Pattern Recognition
Computing the Sensitivity of a Layered Perceptron R.J. Marks II August 31, 2002.
COMP53311 Other Classification Models: Neural Network Prepared by Raymond Wong Some of the notes about Neural Network are borrowed from LW Chan’s notes.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
N. Saoulidou & G. Tzanakos1 Analysis for new Period 3 Events N. Saoulidou and G. Tzanakos University of Athens, Department of Physics’ Div. Of Nuclear.
Kim HS Introduction considering that the amount of MRI data to analyze in present-day clinical trials is often on the order of hundreds or.
Supervised Learning in ANNs
Other Classification Models: Neural Network
Learning in Neural Networks
CSE 473 Introduction to Artificial Intelligence Neural Networks
Derivation of a Learning Rule for Perceptrons
Data Mining Lecture 11.
FUNDAMENTAL CONCEPT OF ARTIFICIAL NETWORKS
Prof. Carolina Ruiz Department of Computer Science
network of simple neuron-like computing elements
Multi-Layer Perceptron
Parametric Methods Berlin Chen, 2005 References:
Prof. Carolina Ruiz Department of Computer Science
Presentation transcript:

N. Saoulidou & G. Tzanakos1 ANN Basics : Brief Review N. Saoulidou, Fermilab & G. Tzanakos, Univ. of Athens

N. Saoulidou & G. Tzanakos2 -Methods: Artificial Neural Networks- ANN can be trained by MC generated events A trained ANN provides multidimensional cuts for data that are difficult to deduce in the usual manner from 1-d or 2-d histogram plots. ANN has been used in HEP HEP Packages: JETNET SNNS MLP fit

N. Saoulidou & G. Tzanakos3 -ANN BASICS- X Y Signal Background Event sample characterized by two variables X and Y (left figure) A linear combination of cuts can separate “signal” from “background” (right fig.) Define “step function” Separate “signal” from “background” with the following function: “Signal (x, y)” IN “Signal (x, y)” OUT X a 1 x+b 1 y+c 1 a 2 x+b 2 y+c 2 a 3 x+b 3 y+c 3 Y Signal Background

N. Saoulidou & G. Tzanakos4 The diagram resembles a feed forward neural network with two input neurons, three neurons in the first hidden layer and one output neuron. Threshold produces the desired offset. Constants a i, b i are the weights w i,j (i and j are the neuron indices). -ANN BASICS- a1a1 a2a2 a3a3 b1b1 b2b2 b3b3 YX Thres. Output c 1 c2c2 c3 c3 Visualization of function C(x,y)

N. Saoulidou & G. Tzanakos5 -ANN basics : Schematic- OUTPUT LAYER XiXi X1X1 INPUT PARAMETERS INPUT LAYER HIDDEN LAYER WEIGHTS w kj Bayesian Probability neuron i neuron j neuron k w ik Bias... Biological Neuron...

N. Saoulidou & G. Tzanakos6 Output of t j each neuron in the first hidden layer : Transfer function is the sigmoid function : –For the standard backpropagation training procedure of neural networks, the derivative of the neuron transfer functions must exist in order to be able to minimize the network error (cost) function E. –Theorem 1 : Any continuous function of any number of variables on a compact set can be approximated to any accuracy by a linear combination of sigmoids –Theorem 2 : Trained with desired output 1 for signal and 0 for background the neural network function (output function t j ) approximates the Bayesian Probability of an event being a signal. -ANN BASICS-

N. Saoulidou & G. Tzanakos7 The ANN output is the Bayes a posteriori probability & in the proof no special assumption has been made on the a priori P(S) and P(B) probabilities (absolute normalization)…..TRUE BUT THEIR VALUES DO MATTER ………(They should be what nature gave us) ANN analysis : Minimization of an Error (Cost) Function -ANN Probability (review)-

N. Saoulidou & G. Tzanakos8 Bayesian a posteriori probability : ANN output : P(S/x) ANN training examples : P(x/S) & P(x/B) ANN number of Signal Training Examples P(S) ANN number of Background Training Examples P(B) The MLP (ann) analysis and the Maximum Likelihood Method ( Bayes Classifier ) are equivalent. (c 11 c 22 = cost for making the correct decision & c 12 c 21 = cost for making the wrong decision ) -ANN probability (review)-

N. Saoulidou & G. Tzanakos9 -ANN Probability cont.- P(S/x)=0.5 P(S/x)=0.1 Worse hypothetical case 1: One variable characterizing the populations, which is identical for S and B, therefore : P(S)=0.1 & P(B)=0.9 If we train with equal numbers for signal and background the ANN will wrongly compute P(S/x)=0.5 If we train with the correct ratio for signal and background the ANN will correctly compute P(S/x)=0.1, which is exactly what Bayes a posteriori probability would give also. ANN output

N. Saoulidou & G. Tzanakos10 -ANN Probability cont.- P(S/x) =1 Best hypothetical case : One variable characterizing the populations, which is completely separated (different) for S and B. P(S)=0.1 & P(B)=0.9 If we train with equal numbers for signal and background the ANN will compute P(S/x)=1. If we train with the correct ratio for signal and background the ANN will again compute P(S/x)=1. In this case it does not matter if we use the correct a priori probabilities or not. ANN output P(S/x) =1

N. Saoulidou & G. Tzanakos11 ANN Probability (final...) The MLP output approximates the Bayesian a posteriori probability and the a priori class probabilities P(S) and P(B) should be considered correctly. The more similar the characteristics of the populations are, the more important the a priori probabilities are, in calculation of the final a posteriori probability by the MLP. In addition the more close to the boundary surface (between the two populations) an event is, the more sensitive it’s a posteriori probability is to changes in the a priori probabilities.