Learning in Neural and Belief Networks - Feed Forward Neural Network 2001 년 3 월 28 일 20013329 안순길.

Slides:



Advertisements
Similar presentations
Introduction to Neural Networks Eric Wong Martin Ho Eddy Li Kitty Wong.
Advertisements

Slides from: Doug Gray, David Poole
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
Artificial Neural Network
Tuomas Sandholm Carnegie Mellon University Computer Science Department
Kostas Kontogiannis E&CE
Artificial Neural Networks
Artificial Neural Networks - Introduction -
Machine Learning Neural Networks
1 Introduction to Bio-Inspired Models During the last three decades, several efficient machine learning tools have been inspired in biology and nature:
Artificial Intelligence (CS 461D)
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Prof. Carla P. Gomes Module: Neural Networks: Concepts (Reading:
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Rutgers CS440, Fall 2003 Neural networks Reading: Ch. 20, Sec. 5, AIMA 2 nd Ed.
Neural Networks Marco Loog.
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
Connectionist Modeling Some material taken from cspeech.ucd.ie/~connectionism and Rich & Knight, 1991.
AN INTERACTIVE TOOL FOR THE STOCK MARKET RESEARCH USING RECURSIVE NEURAL NETWORKS Master Thesis Michal Trna
Artificial Neural Networks
Lecture 4 Neural Networks ICS 273A UC Irvine Instructor: Max Welling Read chapter 4.
Data Mining with Neural Networks (HK: Chapter 7.5)
Artificial Neural Networks
ICS 273A UC Irvine Instructor: Max Welling Neural Networks.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Some more Artificial Intelligence
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Neural Networks. Plan Perceptron  Linear discriminant Associative memories  Hopfield networks  Chaotic networks Multilayer perceptron  Backpropagation.
Artificial Neural Networks
Computer Science and Engineering
Introduction to Neural Networks. Neural Networks in the Brain Human brain “computes” in an entirely different way from conventional digital computers.
2101INT – Principles of Intelligent Systems Lecture 10.
Artificial Neural Network Yalong Li Some slides are from _24_2011_ann.pdf.
Machine Learning Dr. Shazzad Hosain Department of EECS North South Universtiy
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Introduction to Artificial Intelligence (G51IAI) Dr Rong Qu Neural Networks.
CS 478 – Tools for Machine Learning and Data Mining Perceptron.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Artificial Neural Networks Chapter 4 Perceptron Gradient Descent Multilayer Networks Backpropagation Algorithm 1.
Dr.Abeer Mahmoud ARTIFICIAL INTELLIGENCE (CS 461D) Dr. Abeer Mahmoud Computer science Department Princess Nora University Faculty of Computer & Information.
Learning Neural Networks (NN) Christina Conati UBC
Introduction to Neural Networks Freek Stulp. 2 Overview Biological Background Artificial Neuron Classes of Neural Networks 1. Perceptrons 2. Multi-Layered.
1 Perceptron as one Type of Linear Discriminants IntroductionIntroduction Design of Primitive UnitsDesign of Primitive Units PerceptronsPerceptrons.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Chapter 6 Neural Network.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Machine Learning Supervised Learning Classification and Regression
Fall 2004 Backpropagation CS478 - Machine Learning.
CS 388: Natural Language Processing: Neural Networks
Learning with Perceptrons and Neural Networks
Learning in Neural Networks
Artificial Intelligence (CS 370D)
Artificial neural networks:
Real Neurons Cell structures Cell body Dendrites Axon
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Machine Learning Today: Reading: Maria Florina Balcan
Perceptron as one Type of Linear Discriminants
Artificial Neural Networks
Artificial Intelligence Lecture No. 28
Lecture Notes for Chapter 4 Artificial Neural Networks
Neural Networks ICS 273A UC Irvine Instructor: Max Welling
David Kauchak CS158 – Spring 2019

Presentation transcript:

Learning in Neural and Belief Networks - Feed Forward Neural Network 2001 년 3 월 28 일 안순길

Contents How the Brain works Neural Networks Perceptrons

Introduction Two view points in this chapter  Computational view points : representing function using network  Biological view points : mathematical model for brain Neuron: computing elements Neural Networks: collection of interconnected neurons

How the Brain Works Cell body (soma) :provides the support functions and structure of the cell Axon : a branching fiber which carries signals away from the neurons Synapse : converts a electrical signal into a chemical signal Dendrites : consist of more branching fibers which receive signal from other nerve cells Action potential: electrical pulse Synapse  excitatory: increasing potential  synaptic connection: plasticity  inhibitory: decreasing potential A collection of simple cells can lead to thoughts, action, and consciousness.

Comparing brains with digital computers They perform quite different tasks, have different properties Speed (in Switching speed)  computer is a million times faster  brain is a billion times faster Brain  Perform a complex task  More fault-tolerant: graceful degradation  To be trained using an inductive learning algorithm

Neural Networks NN: nodes(unit), links(has a numeric weight)  Each link has a weight  Learning : updating the weights Two computational components  linear component: input function  nonlinear component: activation function

Notation

Simple computing elements Total weighted input By applying the activation function g

Three activation function

Threshold  To cause the neuron to fire  can be replaced with an extra input weight.  The input greater than threshold, output 1  Otherwise 0

Applying neural network in Logic Gates

Network structures(I) Feed-forward networks  Unidirectional links, no cycles  DAG(directed acyclic graph)  No links between units in the same layer, no links backward to a previous layer, no links that skip a layer.  Uniformly processing from input units to output units  No internal state

input units/ output units/ hidden units Perceptron: no hidden units Multilayer networks: one or more hidden units Specific parameterized structure: fixed structure and activation function Nonlinear regression: g(nonlinear function)

Network Structures(II) Recurrent Network  The Brain similar to Recurrent Network  Brain has backward link like Recurrent  Recurrent networks have internal states stored in the activation level  Unstable, oscillate, exhibit chaotic behavior  Long computation time  Need advanced mathematical method

Network Structures(III) Examples  Hopfield networks  Bidirectional connections with symmetric weights  Associative memory: most closely resembles the new stimulus  Boltzmann machines  Stochastic(probabilitic) activation function

Optimal Network Struture(I) Too small network: in capable of representation Too big network: not generalized well  Overfitting when there are too many parameters. Feed forward NN with one hidden layer  can approximate any continuous function Feed forward NN with 2 hidden layer  can approximate any function

Optimal Network Structures(II) NERF(Network Efficiently Representable Functions)  Function that can be approximated with a small number of units  Using genetic algorithm: running the whole NN training protocol  Hill-climbing search(modifying an existing network structure)  Start with a big network: optimal brain damage  Removing weights from fully connected model  Start with a small network: tiling algorithm  Start with single unit and add subsequent units  Cross-validation techniques

Perceptrons Perceptron: single-layer, feed-forward network  Each output unit is indep. of the others  Each weight only affects one of the outputs where,

What perceptrons can represent Boolean function AND, OR, and NOT Majority function: W j =1, t=n/2 ->1 unit, n weights  In case of decision tree: O(2 n ) nodes can only represent linearly separable functions. cannot represent XOR

Examples of Perceptrons Entire input space is divided in two along a boundary defined by In Figure 19.9(a): n=2 In Figure 19.10(a): n=3

Learning linearly separable functions(I) Bad news: not many problem in this set Good news: given enough training examples, there exists a perceptron algorithm learning them. Neural network learning algorithm  Current-best-hypothesis(CBH) scheme  Hypothesis: a network defined by the current values of the weights  Initial network: randomly assigned weight in [-0.5, 0.5]  Repeat the update phase to achieve convergence  Each epoch: updating all the weights for all the examples

Learning linearly separable functions(II) Learning  The error  Err=T-O  :Rosenblatt in 1960  : learning rate Error positive  Need to increase O Error negative  Need to decrease O

Algorithm

Perceptrons(Minsky and Papert, 1969)  Limits of linearly separable functions Gradient descent search through weight space  Weight space han no local minima Difference btw. NN and other attribute-based methods such as decision trees.  Real numbers in some fixed range vs. discrete set Dealing with discrete set  Local encoding: a single input, discrete attribute values  None=0.0, Some=0.5, Full=1.0 (WillWait)  Distributed encoding: one input unit for each attribute

Example

Summary(I) Neural network is made by seeing human ’ s brain  Brain still superior to Computer in Switching Speed  More fault-tolerant Neural network  nodes(unit), links(has a numeric weight)  Each link has a weight  Learning : updating the weights  Two computational components  linear component: input function  nonlinear component: activation function

Summary(II) In this text, We only consider  Feed-forward networks  Unidirectional links, no cycles  DAG(directed acyclic graph)  No links between units in the same layer, no links backward to a previous layer, no links that skip a layer.  Uniformly processing from input units to output units  No internal state

Summary(III) Network size decides Representation Power  Overfitting when there are too many parameters. Feed forward NN with one hidden layer  can approximate any continuous function Feed forward NN with 2 hidden layer  can approximate any function

Summary(IV) Perceptron: single-layer, feed-forward network  Each output unit is indep. of the others  Each weight only affects one of the outputs  Only available in linear separable functions If Problem Space is flat, Neural Network is very available. In other words, if we make it easy in algorithm perspective, Neural network also do Basically, Back Propagation only guarantee Local Optimality in neural network