Simultaneous Recurrent Neural Networks for Static Optimization Problems By: Amol Patwardhan Adviser: Dr. Gursel Serpen August, 1999 The University of.

Slides:



Advertisements
Similar presentations
KULIAH II JST: BASIC CONCEPTS
Advertisements

Presentation of Designing Efficient Irregular Networks for Heterogeneous Systems-on-Chip by Christian Neeb and Norbert Wehn and Workload Driven Synthesis.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 7: Learning in recurrent networks Geoffrey Hinton.
Neural Networks  A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Mehran University of Engineering and Technology, Jamshoro Department of Electronic Engineering Neural Networks Feedforward Networks By Dr. Mukhtiar Ali.
Vector Processing. Vector Processors Combine vector operands (inputs) element by element to produce an output vector. Typical array-oriented operations.
CS 678 –Boltzmann Machines1 Boltzmann Machine Relaxation net with visible and hidden units Learning algorithm Avoids local minima (and speeds up learning)
Kostas Kontogiannis E&CE
CS 678 –Relaxation and Hopfield Networks1 Relaxation and Hopfield Networks Totally connected recurrent relaxation networks Bidirectional weights (symmetric)
Structure learning with deep neuronal networks 6 th Network Modeling Workshop, 6/6/2013 Patrick Michl.
1 Neural networks 3. 2 Hopfield network (HN) model A Hopfield network is a form of recurrent artificial neural network invented by John Hopfield in 1982.
Machine Learning Neural Networks
Neural Networks Basic concepts ArchitectureOperation.
© 2006 Pearson Addison-Wesley. All rights reserved14 A-1 Chapter 14 Graphs.
1 Chapter 5 Advanced Search. 2 Chapter 5 Contents l Constraint satisfaction problems l Heuristic repair l The eight queens problem l Combinatorial optimization.
Artificial Neural Networks ML Paul Scheible.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Radial Basis Functions
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Prof. Carla P. Gomes Module: Neural Networks: Concepts (Reading:
Ant Colony Optimization Optimisation Methods. Overview.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
D Nagesh Kumar, IIScOptimization Methods: M1L4 1 Introduction and Basic Concepts Classical and Advanced Techniques for Optimization.
Radial Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
HOW TO SOLVE IT? Algorithms. An Algorithm An algorithm is any well-defined (computational) procedure that takes some value, or set of values, as input.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Network Aware Resource Allocation in Distributed Clouds.
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Swarm Computing Applications in Software Engineering By Chaitanya.
Artificial Neural Networks An Overview and Analysis.
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Hybrid AI & Machine Learning Systems Using Ne ural Networks and Subsumption Architecture By Logan Kearsley.
Swarm Intelligence 虞台文.
NEURAL NETWORKS FOR DATA MINING
Classification / Regression Neural Networks 2
Spring 2015 Mathematics in Management Science Network Problems Networks & Trees Minimum Networks Spanning Trees Minimum Spanning Trees.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 21 Oct 28, 2005 Nanjing University of Science & Technology.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
1 Chapter 5 Advanced Search. 2 Chapter 5 Contents l Constraint satisfaction problems l Heuristic repair l The eight queens problem l Combinatorial optimization.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
The Application of The Improved Hybrid Ant Colony Algorithm in Vehicle Routing Optimization Problem International Conference on Future Computer and Communication,
Optimization with Neural Networks Presented by: Mahmood Khademi Babak Bashiri Instructor: Dr. Bagheri Sharif University of Technology April 2007.
Mobile Agent Migration Problem Yingyue Xu. Energy efficiency requirement of sensor networks Mobile agent computing paradigm Data fusion, distributed processing.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Adaptive Hopfield Network Gürsel Serpen Dr. Gürsel Serpen Associate Professor Electrical Engineering and Computer Science Department University of Toledo.
Neural Networks - Berrin Yanıkoğlu1 Applications and Examples From Mitchell Chp. 4.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Artificial Neural Networks for Data Mining. Copyright © 2011 Pearson Education, Inc. Publishing as Prentice Hall 6-2 Learning Objectives Understand the.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
A PID Neural Network Controller
Evolutionary Computation Evolving Neural Network Topologies.
Bayesian Neural Networks
Ch7: Hopfield Neural Model
Classification / Regression Neural Networks 2
Prof. Carolina Ruiz Department of Computer Science
Artificial Intelligence Methods
network of simple neuron-like computing elements
The use of Neural Networks to schedule flow-shop with dynamic job arrival ‘A Multi-Neural Network Learning for lot Sizing and Sequencing on a Flow-Shop’
Boltzmann Machine (BM) (§6.4)
A Dynamic System Analysis of Simultaneous Recurrent Neural Network
August 8, 2006 Danny Budik, Itamar Elhanany Machine Intelligence Lab
The Network Approach: Mind as a Web
Simulated Annealing & Boltzmann Machines
CSC 578 Neural Networks and Deep Learning
Prof. Carolina Ruiz Department of Computer Science
Presentation transcript:

Simultaneous Recurrent Neural Networks for Static Optimization Problems By: Amol Patwardhan Adviser: Dr. Gursel Serpen August, 1999 The University of Toledo

Drawbacks of conventional computing systems:- Perform poorly on complex problems Lack the computational power Do not utilize the inherent parallelism of problems Advantages of Artificial Neural Networks:- Perform well even on complex problems Very fast computational cycles if implemented in hardware Can take advantage of the inherent parallelism of problems Driving Force for the Research

Earlier Efforts to solve Optimization Problems Many ANN algorithms with feedforward and recurrent architectures have been used to solve unconstrained and combinatorial optimization problems The Hopfield network and its derivatives including Boltzmann machine and MFA seem to be most prominent and extensively applied ANN algorithms to solve these static optimization problems. However HN and their derivatives do not scale well with the increase in the size of the optimization problem.

Statement of Thesis Simultaneous Recurrent Neural Network, a trainable and recurrent ANN, to address the scaling problem Artificial Neural Network algorithms currently experience for static optimization problems Can we use ….

Research Approach A neural network simulator is developed for simulation of Simultaneous Recurrent Neural Network An extensive simulation study is conducted on two well known static optimization problems - Traveling Salesman Problem - Graph Path Search Problem Simulation results are analyzed

Significance of Research A powerful and efficient optimization tool Optimizer can solve real-life size and complex static optimization problems Will require a fraction of time if implemented in hardware Applications in many fields like - Routing in computer networks - VLSI circuit design - Planning in operational and logistic systems - Power distribution systems - Wireless and satellite communication systems

Hopfield Network and Static Optimization Problems Most widely used ANN algorithms Offer a computationally simple way for a class of optimization problems HN dynamics minimizes a quadratic Lyapunov function Employed as fixed-point attractors Performance greatly depends on constraint weight parameters

Shortcoming of the Hopfield Network Constraint weight parameters are set empirically All weights and connections are specified in advance Difficult to guess weights for large-scale problems Lack mechanism to incorporate the experience gained Quality of solution not good for large scale TSP Do not scale well with increase in the problem size Acceptable solution for Graph Path Search Problem can not be found

Why Simultaneous Recurrent Neural Network Hopfield Network do not employ any learning that can benefit from prior relaxations A relaxation based neural search algorithm, which can learn from its own experience is needed Simultaneous Recurrent Neural Network is a ….. - Recurrent algorithm - has relaxation search capability - has ability to learn

Simultaneous Recurrent Neural Network Feedforward Mapping  (.,W) Outputs Y Feedback Path Inputs X Simultaneous Recurrent Neural Network is a feedforward network with simultaneous feedback from outputs of the network to its inputs without any time delay

Simultaneous Recurrent Neural Network Follows a trajectory in the state space to relax to a fixed point The network is provided with the external inputs and initial outputs are typically assumed randomly The output of previous iteration is fedback to the network along with the external inputs to compute the output of next iteration The network iterates until it reaches a stable equilibrium point

Training of SRN Methods available for training of SRN in literature  Backpropagation Through Time (BTT) which requires the knowledge of desired outputs throughout the trajectory path  Error critics (EC) has no quarantee of yielding exact results in equilibrium  Truncation did not provide satisfactory results and needs to be further tested Recurrent Backpropagation requires only knowledge of desired outputs at the end of trajectory path and hence chosen to train SRN

Traveling Salesman Problem

Network Topology for Traveling Salesman Problem Hidden Layers Cost Matrix Output Array N  N nodes N  N nodes N  N nodes Input LayerHidden Layer(s)Output Layer Path Specification

Error Computation for TSP Constraints used for training TSP  Asymmetry of the path traveled  Column inhibition  Row inhibition  Cost of the path traveled  Values of the solution matrix Source Cities Destination Cities Output Matrix

Graph Path Search Problem Source Destination

Network Topology for Graph Path Search Problem Hidden Layers Cost Matrix Output Array N  N nodes N  N nodes N  N nodes Input LayerHidden Layer(s)Output Layer Path Specification N  N nodes Adjacency Matrix

Error Computation for GPSP Constraints used for training GPSP  Asymmetry of the sub-graph  Column inhibition  Row inhibition  Source and target vertex inhibition  Column/row excitation  Row/column excitation  Cost of the solution path  Number of vertices in the path Sources Vertices Destinations Vertices Output Matrix

Simulation:- Software Environment Language: C, MATLAB 5.2 GUI: Xwindows11 Plotting of Graphs: C program calling MATLAB functions for plotting of graphs Hardware Environment Sun OS 5.7 Sun Ultra machine 300MHz Physical Memory (RAM) 1280 MB Virtual Memory (Swap) 1590 MB

Simulation:- GUI for Simulator

Simulation:- Initialization Randomly initialize weights and initial outputs (Range: ) Randomly initialize cost matrix for TSP (Range: ) Randomly initialize adjacency matrix ( 0.0 or 1.0) depending on the connectivity level parameter for GPSP For TSP, values along the diagonal of the cost matrix are clamped to 1.0 to avoid self looping. For GPSP, values along the diagonal of the adjacency matrix and cost matrix are clamped to 1.0 to avoid self looping. Values of constraint weight parameters are set depending on the size of the problem

Simulation:- Initialization for TSP Initial values and Increments per 30 relaxation of constraint weight parameters for the TSP

Simulation:- Training Error function vs. Simulation Time for TSP

Simulation:- Results Convergence criteria of network is checked after every 100 relaxations Criteria: 95% of active nodes have value greater than 0.9

Simulation:- Results Plot of Normalized Distance between cities after the convergence of network to an acceptable solution vs. Problem Size

Simulation:- Results Plot of Number of Relaxations required for a solution and values of Constraint Weight Parameters g c and g r after 300 Relaxations vs. Problem Size

Simulation:- Initialization for GPSP Initial values and Increments per 30 relaxation of constraint weight parameters for the GPSP

Simulation:- Results for GPSP Convergence criteria of network is checked after every 100 relaxations Criteria: Active nodes have value greater than 0.8

Simulation:- Results for GPSP Plot of Number of Relaxations required for a solution and values of Constraint Weight Parameters g i after 300 Relaxations vs. Problem Size

Conclusions The SRN with the RBP was able to find “good quality” solutions, in the range of , for large-scale (40 to 500 city) Traveling Salesman Problem Solutions were obtained within acceptable computation efforts Normalized Distance between cities remained almost consistent as the problem size was varied from 40 to 500 cities The simulator developed does not require weights to be predetermined before simulation as is the case with the HN and its derivatives The initial and incremental values of constraint weight parameters play very important role in the training of the network

Conclusions (continued) Computational effort and memory requirement increased proportional to the square of the problem size The SRN with the RBP was able to find a solution for large-scale Graph Path Search Problem in the range of 40 to 500 vertices The solutions were obtained within acceptable computation efforts and time The computation effort required for the GPSP is 1.1 to 1.2 times more than that of the TSP The number of relaxations required increased with the increase in the problem size The GPSP was very sensitive to the constraint weight parameters

Conclusions (continued) Thus we can say that …. Simultaneous Recurrent Neural Network with Recurrent Backpropagation training algorithm scaled well for large-scale static optimization problems like the Traveling Salesman Problem and the Graph Path Search Problem within acceptable computation effort bounds.

Recommendations for Future Study The feasibility of the hardware implementation of the network and algorithm for the TSP should be thought over More number of simulations should be done for the GPSP to find the effect of change in each constraint weight parameter on the solution The effect of incorporating a stochastic or probabilistic component into the learning for network dynamics can also be studied to find the better approach Simulation study on weighted GPSP should be done for more practical use

Questions ?