Optimal Nonlinear Neural Network Controllers for Aircraft Joint University Program Meeting October 10, 2001 Nilesh V. Kulkarni Advisors Prof. Minh Q. Phan.

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Non-Gaussian Statistical Timing Analysis Using Second Order Polynomial Fitting Lerong Cheng 1, Jinjun Xiong 2, and Lei He 1 1 EE Department, UCLA *2 IBM.
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Lect.3 Modeling in The Time Domain Basil Hamed
Neural Network Control of a Hypersonic Inlet Joint University Program Meeting April 05, 2001 Nilesh V. Kulkarni Advisors Prof. Minh Q. Phan Dartmouth College.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Optimal Control of the Magneto-hydrodynamic Generator at the Scramjet Inlet Joint University Program Meeting June 12, 2002 Nilesh V. Kulkarni Advisors.
B.Macukow 1 Lecture 12 Neural Networks. B.Macukow 2 Neural Networks for Matrix Algebra Problems.
Robust control Saba Rezvanian Fall-Winter 88.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Training an Adaptive Critic Flight Controller
Visual Recognition Tutorial
Neural Nets Using Backpropagation Chris Marriott Ryan Shirley CJ Baker Thomas Tannahill.
280 SYSTEM IDENTIFICATION The System Identification Problem is to estimate a model of a system based on input-output data. Basic Configuration continuous.
Supervised learning 1.Early learning algorithms 2.First order gradient methods 3.Second order gradient methods.
Speaker Adaptation for Vowel Classification
Neural Networks Marco Loog.
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Silvia Ferrari Princeton University
Performance Optimization of the Magneto-hydrodynamic Generator at the Scramjet Inlet Nilesh V. Kulkarni Advisors: Prof. Minh Q. Phan Dartmouth College.
Radial Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Collaborative Filtering Matrix Factorization Approach
Radial Basis Function Networks
Approximating the Algebraic Solution of Systems of Interval Linear Equations with Use of Neural Networks Nguyen Hoang Viet Michal Kleiber Institute of.
1 Hybrid methods for solving large-scale parameter estimation problems Carlos A. Quintero 1 Miguel Argáez 1 Hector Klie 2 Leticia Velázquez 1 Mary Wheeler.
AUTOMATIC CONTROL THEORY II Slovak University of Technology Faculty of Material Science and Technology in Trnava.
Artificial Neural Networks
Classification Part 3: Artificial Neural Networks
Adaptive Critic Design for Aircraft Control Silvia Ferrari Advisor: Prof. Robert F. Stengel Princeton University FAA/NASA Joint University Program on Air.
Book Adaptive control -astrom and witten mark
Integrating Neural Network and Genetic Algorithm to Solve Function Approximation Combined with Optimization Problem Term presentation for CSC7333 Machine.
Chapter 9 Neural Network.
MEGN 537 – Probabilistic Biomechanics Ch.7 – First Order Reliability Methods Anthony J Petrella, PhD.
SWAN 2006, ARRI_ Control of Distributed-Information Nonlinear Stochastic Systems Prof. Thomas Parisini University of Trieste.
Using Neural Networks to Predict Claim Duration in the Presence of Right Censoring and Covariates David Speights Senior Research Statistician HNC Insurance.
1 RECENT DEVELOPMENTS IN MULTILAYER PERCEPTRON NEURAL NETWORKS Walter H. Delashmit Lockheed Martin Missiles and Fire Control Dallas, TX 75265
Optimal Feedback Control of the Magneto-hydrodynamic Generator for a Hypersonic Vehicle Nilesh V. Kulkarni Advisors: Prof. Minh Q. Phan Dartmouth College.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
Neural and Evolutionary Computing - Lecture 9 1 Evolutionary Neural Networks Design  Motivation  Evolutionary training  Evolutionary design of the architecture.
INVENTORY CONTROL AS IDENTIFICATION PROBLEM BASED ON FUZZY LOGIC ALEXANDER ROTSHTEIN Dept. of Industrial Engineering and Management, Jerusalem College.
CHAPTER 5 S TOCHASTIC G RADIENT F ORM OF S TOCHASTIC A PROXIMATION Organization of chapter in ISSO –Stochastic gradient Core algorithm Basic principles.
Scheduling Optimization in Wireless MESH Networks with Power Control and Rate Adaptation SECON 2006 Antonio Capone and Giuliana Carello Keon Jang 2007.
Model Reference Adaptive Control (MRAC). MRAS The Model-Reference Adaptive system (MRAS) was originally proposed to solve a problem in which the performance.
September Bound Computation for Adaptive Systems V&V Giampiero Campa September 2008 West Virginia University.
Non-Bayes classifiers. Linear discriminants, neural networks.
1 MODELING MATTER AT NANOSCALES 4. Introduction to quantum treatments The variational method.
Engineering Optimization Chapter 3 : Functions of Several Variables (Part 1) Presented by: Rajesh Roy Networks Research Lab, University of California,
EE459 Neural Networks Backpropagation
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
Stochastic Optimal Control of Unknown Linear Networked Control System in the Presence of Random Delays and Packet Losses OBJECTIVES Develop a Q-learning.
Hazırlayan NEURAL NETWORKS Backpropagation Network PROF. DR. YUSUF OYSAL.
Adaptive Optimal Control of Nonlinear Parametric Strict Feedback Systems with application to Helicopter Attitude Control OBJECTIVES  Optimal adaptive.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
دانشگاه صنعتي اميركبير دانشكده مهندسي پزشكي استاد درس دكتر فرزاد توحيدخواه بهمن 1389 کنترل پيش بين-دکتر توحيدخواه MPC Stability-2.
Effects of System Uncertainty on Adaptive-Critic Flight Control Silvia Ferrari Advisor: Prof. Robert F. Stengel Princeton University FAA/NASA Joint University.
CHEE825 Fall 2005J. McLellan1 Nonlinear Empirical Models.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
(COEN507) LECTURE III SLIDES By M. Abdullahi
Towards Adaptive Optimal Control of the Scramjet Inlet Nilesh V. Kulkarni Advisors: Prof. Minh Q. Phan Dartmouth College Prof. Robert F. Stengel Princeton.
Nonlinear balanced model residualization via neural networks Juergen Hahn.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Fall 2004 Backpropagation CS478 - Machine Learning.
Deep Feedforward Networks
DYNAMICS & CONTROL PDR 1 TEAM 4
One-layer neural networks Approximation problems
第 3 章 神经网络.
Dr. Unnikrishnan P.C. Professor, EEE
Collaborative Filtering Matrix Factorization Approach
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Presentation transcript:

Optimal Nonlinear Neural Network Controllers for Aircraft Joint University Program Meeting October 10, 2001 Nilesh V. Kulkarni Advisors Prof. Minh Q. Phan Dartmouth College, Hanover, NH Prof. Robert F. Stengel Princeton University, NJ

Presentation Outline: Research goals. Definition of the problem. Parametric optimization approach. Modified Approach. Neural Network implementation. Linear System Implementation. Nonlinear System Implementation. Conclusions

Research Goals  To come up with a control approach that is: Optimal or approaching optimality in the limit Applicable to both linear and non-linear systems Data-based (no need for an explicit analytical model of the system) Adaptive to account for slowly time-varying dynamics and operating conditions.  Application to aircraft.

General Problem Statement: For the system dynamics: Find a control law: To maximize a performance index (minimize a cost function)

Approaches:  Dynamic optimization approaches: Calculus of variations approach. Euler-Lagrange equations. Dynamic programming. Hamilton-Jacobi-Bellman equation. Specialization to Adaptive critic designs.  Static optimization approach: Parametric Optimization. Cost-to-go approach.

Direct Parametric Optimization Approach: Methodology: Find the unknown coefficients,’G’, that minimize the cost-to-go function Disadvantages: This approach reduces to solving a static optimization problem which is highly nonlinear even for linear systems Easily gets stuck in spurious local minima even for the case of finding a linear optimal controller for a linear system Chance of finding a workable optimal controller using such an approach in practice is very limited.

Illustrative Example: For a simple linear time invariant system, and we can write, And so, As seen the cost-to-go function expressed with a single parameter ‘G’, is a highly nonlinear function of the parameter and as seen from several test examples was found to contain several minima.

Modified Approach: Reformulate the control law: Set up the cost-to-go function in terms of the ‘G’s’: ‘r’ represents the order of approximation of the cost-to-go function

Modified Approach… Find the G’s by imposing the stationarity conditions: and Solving the second set of equations is not as easy and even less implementable in terms of a control architecture. x(k), the present state of the system appears as a coefficient in the stationarity conditions. By solving the stationarity conditions for multiple x(k)’s, presents enough equations for solving for the unknown G’s without solving the second set of conditions.

Illustrative Example: For a simple linear time invariant system, we can write, And so, As seen the cost-to-go function now expressed with the parameters ‘G’s’, is a quadratic function of the parameter and therefore has a single minimum

Role of Neural Networks For a nonlinear system, the controller is typically nonlinear. Cost-to-go function is a nonlinear function. Being universal function approximators, Neural Networks present themselves as ideal tools for handling nonlinear systems in the proposed Cost-to-go design approach Neural networks present a straightforward approach for making the design adaptive even in the case of a nonlinear system.

Parameterize the cost-to-go function using a Neural Network (CGA Neural Network) Inputs to the CGA Network: x(k), u(k),…,u(k+r-1) Use the analytical model, or a computer simulation or the physical model to generate the future states. Use the ‘r’ control values and the ‘r’ future states to get the ideal cost-to-go function estimate. Use this to train the CGA Neural Network Formulation of the Control Architecture: NN Cost-to-go function Approximator

CGA Neural Network Training Actual System or Simulation Model Neural Net Cost-to-go Approximator x(k) u(k) +  V(k) V nn (k) V err Neural Network Cost-to-go Approximator Training u(k+1) u(k+r-1)

Formulation of the Control Architecture: NN Controller Instead of a single controller structure (G), we need ‘r’ controller structures. The outputs of the ‘r’ controller structures, generate u(k) through u(k+r-1). Parameterize the ‘r’ controller structures using an effective Neural Network. x(k) Neural Network Controller u(k) u(k+1) … u(k+r-1)

Neural Network Controller Training Gradient of V(k) with respect to the control inputs u(k),…, u(k+r-1) is calculated using back-propagation through the ‘CGA’ Neural Network. These gradients can be further back-propagated through the Neural Network controller to get, through Neural Network controller is trained so that x(k) Neural Network Controller Trained Cost-to-go Neural Network u(k) x(k) V(k) 1 u(k+1) … u(k+r-1)

Advantages of the formulation The modified parametric optimization simplifies the optimization problem. CGA Network training and the controller Network training is decoupled. Implementation is system independent. So the basic architecture remains the same for linear or nonlinear systems. Implementation is data-based. No explicit analytical model needed. Parameterization using Neural Networks makes the control architecture adaptive. Order of approximation ‘r’ in the definition of V(k) serves as a tuning parameter.

Implementation for Linear Systems:  Motivation: Linear systems provide an easy way to see the details of the implementation of the cost-to-go design. Provides a means for comparison of the results with existing solutions.

Optimal Control of Aircraft Lateral Dynamics Airplane State Variables: β - Side slip angle p- Roll rate r - Yaw rate φ - Roll angle Airplane Input Variables:  r - Rudder Deflection  a - Aileron Deflection Phoenix Hobbico Hobbistar 60 tm model

Order of approximation (‘r’) Evaluated cost (‘J’) Control gain Gain obtained using the new data-based approach: Optimal controller gains calculated using LQR optimal solution with perfect knowledge of the system:  Evaluated optimal cost =

 - State trajectories using the cost-to-go design (r = 35) *** - State trajectories using the cost-to-go design (r = 50) - State trajectories using LQR based optimal control Comparison of the state trajectories using the cost-to-go design and the LQR design

 - Control trajectories using the cost-to-go design (r = 35) *** - Control trajectories using the cost-to-go design (r = 50) - Control trajectories using LQR based optimal control Comparison of the control trajectories using the cost-to-go design and the LQR design

Nonlinear Control of Aircraft in an approach configuration h  h glide x glide x V  Aircraft in an approach configuration Equations of motion in the wind-axes system V nom (ft/s)  nom (deg)  nom (deg) T nom (lb) Nominal Flight Conditions M (slugs) T max (lb) S ref (ft 2 ) CL0CL0 CLαCLα CD0CD0 ee  Aircraft Parameters

Implementation Details Equations are written with a change of coordinates while maintaining the nonlinearity.  X and  h are transformed through a coordinate transformation, so that now they represent perturbations along and perpendicular to the approach slope and we can now ignore the dynamics of the perturbation along the approach slope.

Implementation Details… Equations of motion in terms of the nonlinear perturbation dynamics: Equations are discretized with a time step of 0.5 seconds. Specification of the cost function:

Control Architecture Forming a combined Neural Network Fix the weights of the CGA part of the Network Training inputs to the network: Random values of x(k) Train the Network so that it gives the output value of zero for all the input random x(k)

Controller Network Subnet 1 Subnet 2 Subnet r x(k) A Layer With Quadratic Neurons V(k) x(k+2) x(k+1) u(k) x(k+r) u(k+1) u(k+r-1) u(k) u(k+1) u(k+r-1) A Control Architecture Proposed to Simplify the Neural Network Training Problem CGA Network Bringing Structure to the CGA Network

Implementation of the quadratic layer 0 0 Layer with square neurons 1 1

Advantages of the new structure: Guaranteed positive definiteness. Replaced training of a complex function by the training of several simpler functions. A good quality control ability. Allow for hybrid architecture

u(k),…, u(k+3) Subnet 2 Subnet 3 Subnet 4 Subnet 5 Subnet 1 Subnet 3 Subnet 4 Subnet 5 Quadratic Layerof Neurons x(k) u(k) u(k),…,u(k+4) x(k+1) x(k+3) x(k+4) x(k+5) u(k+5),…,u(k+9) u(k),…,u(k+9) V(k) x(k+6) x(k+10) u(k), u(k+1) u(k),…, u(k+2) u(k+3),…, u(k+5) u(k+4),…, u(k+7) u(k+5),…, u(k+8) Implementation of the Hybrid CGA Network of order ‘r = 10’, using trained subnets of order 1 through 5

Internal Structure of the Neural Network Controller showing the separate Controller Subnets

Neural Network Controller training using the trained CGA Network 1 x(k) Controller Network (G 1,G 2,…,G r ) Trained CGA Network u (k) x(k) V(k) est u(k+1) u(k+r-1) Combination Network with the controller Network before the Critic Network

Cumulative value of V(k) getting minimized with training of the Neural Network Controller Weights

. Aircraft response after an Initial perturbation with and without control.  -Open loop dynamics,  - Response with the Optimal Nonlinear Neural Network Controller, ‘O’ – Response with the LQR

Neural Network Controller Outputs

A comparison of the cost function, J, as a function of the order of approximation, ‘r’ of V(k)

Nonlinear Optimization- Global or Local

Conclusions: New Neural Network Control Architecture for optimal control. Applicable to both linear and nonlinear systems Data based. Systematic training procedure. Confirmation on a Nonlinear Aircraft Model.