Optimal resampling using machine learning Jesse McCrosky.

Slides:



Advertisements
Similar presentations
14 de Fevereiro de 2004, Instituto Sistemas e Robótica Emotion-Based Decision and Learning Bruno Damas.
Advertisements

Slides from: Doug Gray, David Poole
Partially Observable Markov Decision Process (POMDP)
Algorithms + L. Grewe.
Tracking Unknown Dynamics - Combined State and Parameter Estimation Tracking Unknown Dynamics - Combined State and Parameter Estimation Presenters: Hongwei.
Variance reduction techniques. 2 Introduction Simulation models should be coded such that they are efficient. Efficiency in terms of programming ensures.
5/15/2015 Mobile Ad hoc Networks COE 499 Localization Tarek Sheltami KFUPM CCSE COE 1.
CHE 185 – PROCESS CONTROL AND DYNAMICS
1 Reinforcement Learning Introduction & Passive Learning Alan Fern * Based in part on slides by Daniel Weld.
Artificial Learning Approaches for Multi-target Tracking Jesse McCrosky Nikki Hu.
Mutual Information Mathematical Biology Seminar
Using Interfaces to Analyze Compositionality Haiyang Zheng and Rachel Zhou EE290N Class Project Presentation Dec. 10, 2004.
Speaker Adaptation for Vowel Classification
Dynamic Face Recognition Committee Machine Presented by Sunny Tang.
1 Hybrid Agent-Based Modeling: Architectures,Analyses and Applications (Stage One) Li, Hailin.
Particle Filtering for Non- Linear/Non-Gaussian System Bohyung Han
Information Fusion Yu Cai. Research Article “Comparative Analysis of Some Neural Network Architectures for Data Fusion”, Authors: Juan Cires, PA Romo,
Radial Basis Function Networks
Rohit Ray ESE 251. What are Artificial Neural Networks? ANN are inspired by models of the biological nervous systems such as the brain Novel structure.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
CS Reinforcement Learning1 Reinforcement Learning Variation on Supervised Learning Exact target outputs are not given Some variation of reward is.
Presenter: Shant Mandossian EFFECTIVE TESTING OF HEALTHCARE SIMULATION SOFTWARE.
IE 594 : Research Methodology – Discrete Event Simulation David S. Kim Spring 2009.
Graph-based consensus clustering for class discovery from gene expression data Zhiwen Yum, Hau-San Wong and Hongqiang Wang Bioinformatics, 2007.
A Neural Network Approach to Predicting Stock Performance John Piefer ECE/CS 539 Project Presentation.
Computational Stochastic Optimization: Bridging communities October 25, 2012 Warren Powell CASTLE Laboratory Princeton University
1 Neural Networks- a brief intro Dr Theodoros Manavis
SLR w SI = Simple Linear Regression with Seasonality Indices
Modeling & Simulation: An Introduction Some slides in this presentation have been copyrighted to Dr. Amr Elmougy.
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
Neural Networks for Protein Structure Prediction Brown, JMB 1999 CS 466 Saurabh Sinha.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
Reinforcement Learning for the game of Tetris using Cross Entropy
« Particle Filtering for Joint Data- Channel Estimation in Fast Fading Channels » Tanya BERTOZZI Didier Le Ruyet, Gilles Rigal and Han Vu-Thien.
Major objective of this course is: Design and analysis of modern algorithms Different variants Accuracy Efficiency Comparing efficiencies Motivation thinking.
Model-based Bayesian Reinforcement Learning in Partially Observable Domains by Pascal Poupart and Nikos Vlassis (2008 International Symposium on Artificial.
Simulation is the process of studying the behavior of a real system by using a model that replicates the behavior of the system under different scenarios.
Fuzzy Reinforcement Learning Agents By Ritesh Kanetkar Systems and Industrial Engineering Lab Presentation May 23, 2003.
Evolutionary Algorithms for Finding Optimal Gene Sets in Micro array Prediction. J. M. Deutsch Presented by: Shruti Sharma.
Non-Bayes classifiers. Linear discriminants, neural networks.
Akram Bitar and Larry Manevitz Department of Computer Science
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Improving the Genetic Algorithm Performance in Aerial Spray Deposition Management University of Georgia L. Wu, W.D. Potter, K. Rasheed USDA Forest Service.
Active learning Haidong Shi, Nanyi Zeng Nov,12,2008.
An Artificial Neural Network Approach to Surface Waviness Prediction in Surface Finishing Process by Chi Ngo ECE/ME 539 Class Project.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
Nonlinear differential equation model for quantification of transcriptional regulation applied to microarray data of Saccharomyces cerevisiae Vu, T. T.,
CHEE825 Fall 2005J. McLellan1 Nonlinear Empirical Models.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Perceptrons Michael J. Watts
By Jyh-haw Yeh Department of Computer Science Boise State University.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate its.
Lecture 2 Introduction to Neural Networks and Fuzzy Logic President UniversityErwin SitompulNNFL 2/1 Dr.-Ing. Erwin Sitompul President University
School of Industrial and Systems Engineering, Georgia Institute of Technology 1 Defuzzification Filters and Applications to Power System Stabilization.
Fall 2004 Backpropagation CS478 - Machine Learning.
Market-Risk Measurement
Deep Feedforward Networks
Policy Gradient in Continuous Time
1 Department of Engineering, 2 Department of Mathematics,
Objective of This Course
1 Department of Engineering, 2 Department of Mathematics,
1 Department of Engineering, 2 Department of Mathematics,
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
by Xiang Mao and Qin Chen
Artificial Intelligence Chapter 3 Neural Networks
Chapter 7: Sampling Distributions
Artificial Intelligence Chapter 3 Neural Networks
Akram Bitar and Larry Manevitz Department of Computer Science
Presentation transcript:

Optimal resampling using machine learning Jesse McCrosky

Outline Introduction Policies  Costs  States Methodology  Data collection  Learning Results  Rho values  Simulation comparison

Introduction SERP has a parameter, rho, to control amount of resampling  Using correct value can substantially improve performance What is correct value?  Depends on filter state

Policies A policy is a mapping between states and actions  State is state of filter – discussed later  Actions are rho values to use An optimal policy minimizes the expected value of some cost function

Costs The one-step cost is simply:  Could also consider computation time and variance of estimate So for one iteration from time k to k+1, the optimal policy is: But this policy is greedy  Might result in a very low cost at time k+1, but result in very high costs later

Time Horizon Cost The one-step policy just defined is optimal for the last iteration of the filter The optimal i th step policy is:

Time Horizon Cost (continued) Time horizon cost combines cost of current iteration, plus next i iterations  Future iterations are discounted For general use, we want infinite time horizon: This should converge (hopefully quickly)

States Filter state is aggregate of discretized value of six state elements  Expected number of targets  Variance of expected number of targets  Change in median weight since last iteration  Boxplot state (3 elements)  Would also like to use variance of target position estimate, but difficult to define and calculate with unknown number of targets

Boxplot Boxplot state consists of three elements:  u1 – u4, u1 – u3, and u1 – u2 Where u1 is the highest weight u2 is the 75% median weight u3 is the 50% median weight u4 is the 25% median weight

Methedology Two approaches  Dynamic programming  Artificial learning Both cases involve trying various rho values in various states many times and finding lowest average cost

Dynamic Programming Create artificial particle set for each possible state and try each rho value Advantage: guaranteed data for every state Disadvantage: particle sets are artificial and may be unrealistic or non-representative Not used in paper  Outside area of interest for conference

Artificial Learning Evolve filter naturally with real signal and record costs in state vs. rho table Then use neural network to approximate optimal rho for states not encountered in training Advantages: uses “real” data, neural network can compensate for some bad data Disadvantages: slow to train, depends on quality of neural network

Data Collection The Rho Optimizer will generate a signal and attempt to filter it Each iteration, record the filter state, rho value, and cost Other details: i th step optimizer, choosing rho value, epochs

Learning Optimizer uses collected data to train a neural network After training the network will output a (approximately) correct value for any states in the data and a (hopefully) correct value for other states

Results Currently have one step results only 2000 epochs of 200 iterations each Each state variable discretized into 10 cells, except sigma-numtargets which has 5 10 possible rhoint values, from 1000 to

Graphs Using data collection stage data only, no neural network Optimal rho value on y-axis vs. discrete index of state component on x-axis Optimal rho value is average of optimal rhos for all states which match values in graph’s state component  Graph are biased by states encountered in simulation Because of bias and flattening, graphs are for novelty purposes only

Boxplot - 1

Boxplot - 2

Boxplot - 3

Delta Median Weight Note: index 2 corresponds with deltamedian = 0

Expected number of targets

Variance of expected number of targets

Graphs - Conclusions Correlation looks good in some elements Bad graphs maybe OK, for example, boxplot may look better if plotted in 4 dimensions Some surprises, does expected number of targets not matter? Real test of results will be simulation comparison between optimal and constant rhos

Simulation No simulations yet  Neural network still training as we speak Next week – Nikki will present her PHD simulations and I will present results of comparative simulations