Joseph Xu Soar Workshop 2012 1 Learning Modal Continuous Models.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Active Appearance Models
DECISION TREES. Decision trees  One possible representation for hypotheses.
Random Forest Predrag Radenković 3237/10
Unsupervised Learning Clustering K-Means. Recall: Key Components of Intelligent Agents Representation Language: Graph, Bayes Nets, Linear functions Inference.
Reinforcement Learning
Lazy vs. Eager Learning Lazy vs. eager learning
Middle Term Exam 03/01 (Thursday), take home, turn in at noon time of 03/02 (Friday)
Chapter 4: Linear Models for Classification
The loss function, the normal equation,
Prénom Nom Document Analysis: Parameter Estimation for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Data Mining Techniques Outline
CS 590M Fall 2001: Security Issues in Data Mining Lecture 3: Classification.
Instance Based Learning
1 Learning from Behavior Performances vs Abstract Behavior Descriptions Tolga Konik University of Michigan.
Statistical Methods Chichang Jou Tamkang University.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Project  Now it is time to think about the project  It is a team work Each team will consist of 2 people  It is better to consider a project of your.
1 Hybrid Agent-Based Modeling: Architectures,Analyses and Applications (Stage One) Li, Hailin.
Generative Models Rong Jin. Statistical Inference Training ExamplesLearning a Statistical Model  Prediction p(x;  ) Female: Gaussian distribution N(
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Data Mining CS 341, Spring 2007 Lecture 4: Data Mining Techniques (I)
INSTANCE-BASE LEARNING
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Chapter 5 Data mining : A Closer Look.
Classification and Prediction: Regression Analysis
Crash Course on Machine Learning
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
More Machine Learning Linear Regression Squared Error L1 and L2 Regularization Gradient Descent.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
Data Mining Joyeeta Dutta-Moscato July 10, Wherever we have large amounts of data, we have the need for building systems capable of learning information.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
New SVS Implementation Joseph Xu Soar Workshop 31 June 2011.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
1 SUPPORT VECTOR MACHINES İsmail GÜNEŞ. 2 What is SVM? A new generation learning system. A new generation learning system. Based on recent advances in.
Learning from observations
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
CSE 446 Logistic Regression Winter 2012 Dan Weld Some slides from Carlos Guestrin, Luke Zettlemoyer.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
Learning from observations
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Interactive Learning of the Acoustic Properties of Objects by a Robot
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Chapter1: Introduction Chapter2: Overview of Supervised Learning
Ensemble Methods in Machine Learning
John Lafferty Andrew McCallum Fernando Pereira
Learning Integrated Symbolic and Continuous Action Models Joseph Xu & John Laird May 29, 2013.
Classification Ensemble Methods 1
Regress-itation Feb. 5, Outline Linear regression – Regression: predicting a continuous value Logistic regression – Classification: predicting a.
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
Data Mining and Decision Support
CSC321: Introduction to Neural Networks and Machine Learning Lecture 15: Mixtures of Experts Geoffrey Hinton.
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
Search Control.. Planning is really really hard –Theoretically, practically But people seem ok at it What to do…. –Abstraction –Find “easy” classes of.
Non-parametric Methods for Clustering Continuous and Categorical Data Steven X. Wang Dept. of Math. and Stat. York University May 13, 2010.
CS 8751 ML & KDDInstance Based Learning1 k-Nearest Neighbor Locally weighted regression Radial basis functions Case-based reasoning Lazy and eager learning.
Linear Models & Clustering Presented by Kwak, Nam-ju 1.
Chapter 7. Classification and Prediction
Data Mining, Neural Network and Genetic Programming
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Machine Learning Basics
K Nearest Neighbor Classification
Joseph Xu Soar Workshop 31 June 2011
Instance Based Learning
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Presentation transcript:

Joseph Xu Soar Workshop Learning Modal Continuous Models

Setting: Continuous Environment Input to the agent is a set of objects with continuous properties – Position, rotation, scaling,... Output is fixed-length vector of continuous numbers Agent runs in lock-step with environment Fully observable 2 Output Input EnvironmentAgent px py rx ry A B pz rz px pypz 0.0 rx ryrz AB

Levels of Problem Solving 3 Motor Babbling Continuous Sampling Methods (RRT) Symbolic Model Free Methods (RL) Continuous Model Symbolic Abstraction Symbolic Planning Symbolic Model Slower Task Completion Specific Solutions Faster Task Completion General Solutions Problem Solving Method Knowledge Required Characteristics None Goal Recognition

Continuous Model Learning Learn a function x: current continuous state vector u: current output vector y: state vector in next time step 4 xuy Continuous Output

Locally Weighted Regression Motor Command left voltage: -0.6 right voltage: 1.2 ? xu k nearest neighbors Weighted Linear Regression 5

Problems with LWR Euclidean distance doesn’t capture relational similarity Averages over neighbors exhibiting different types of interactions 6 Query Neighbor

Problems with LWR 7 Query Neighbor Prediction Euclidean distance doesn’t capture relational similarity Averages over neighbors exhibiting different types of interactions

Modal Models Object behavior can be categorized into different Modes – Behavior within a single mode is usually simple and smooth (inertia, gravity, etc...) – Behaviors across modes can be discontinuous and complex (collisions, drops) – Modes can often be distinguished by discrete spatial relationships between objects Learn two-level models composed of: – A classifier that determines the active mode using spatial relationships – A set of linear functions (initial hypothesis), one for each model 8 Mode Classifier Mode 1 model Mode 2 model Mode 3 model ScenePrediction

Unsupervised Learning of Modes From Data 9 Environment Mode 2 time Mode 1 Expectation Maximization Learned Mode 1 Learned Mode 2 Continuous Features Training Data 0.5, 1.1, -0.2, 4,

Expectation Maximization Expectation Assuming your current model parameters are correct, what is the likelihood that the model m generated data point i ? Maximization Assuming each data point was generated by the most probable model, modify each model’s parameters to maximize likelihood of generating data Iterate until convergence to local maximum 10

Learning Classifier 11 Spatial Relations Training Data 0.5, 1.1, -0.2, 4, time Scene left-of(A,B) = 1 right-of(A,B) = 0 on-top(A,B) = 0 touch(A,B) = 0 A B class attributes Expectation Maximization Learned Mode 1 Learned Mode 2

Learning Classifier Classifier Training Data attributesclass touch(A, B) left-of(A, B) mode 1mode Use linear model for items in same model

Prediction Accuracy Experiment 2 Block Environment – Agent has two outputs (dx, dy) which control the x and y offsets of the controlled block at every times tep – The pushed block can’t be moved except by pushing it with the controlled block – Blocks are always axis-aligned, there’s no momentum Training – Instantiate Soar agent in a variety of spatial configurations – Run 10 time steps, each step is a training example Testing – Instantiate Soar agent in some configuration – Check accuracy of prediction for next time step 13

Prediction Accuracy – Pushed Block 14

Classification Performance 15

Prediction Performance Without Classification Errors 16

Levels of Problem Solving 17 Motor Babbling Continuous Sampling Methods (RRT) Symbolic Model Free Methods (RL) Continuous Model Symbolic Abstraction Symbolic Planning Symbolic Model Slower Task Completion Specific Solutions Faster Task Completion General Solutions Problem Solving Method Knowledge Required Characteristics None Goal Recognition

Symbolic Abstraction Lump continuous states sharing symbolic properties into a single symbolic state Should be Predictable – Planning requires accurate model (ex. STRIPS operators) – Tends to require more states, more symbolic properties Should be General – Fast planning and transferrable solutions – Tends to require fewer states, fewer symbolic properties 18 C2 C1 S1 S2 C1 S1: intersect(C1, C2) S2: ~intersect(C1, C2)

Symbolic Abstraction Hypothesis: contiguous regions of continuous space that share a single behavioral mode is a good abstract state – Planning within modes is simple because of linear behavior – Combinatorial search occurs at symbolic level Spatial predicates used in continuous model decision tree are a reasonable approximation 19

Abstraction Experiment 3 blocks, goal is to push c2 to t Demonstrate a solution trace to agent Agent stores sequence of abstract states in solution in epmem Agent tries to follow plan in analogous task Abstraction should include predicates about c1, c2, t, avoid predicates about d1, d2, d3 20 C2 C1 t d1 d2 d3 C2 C1 C2 C1 t d1 d2 d3 C2 C1

Generalization Performance Tasks Total (16 average)

Conclusions For continuous environments with interacting objects, modal models are more general and accurate than uniform model The relationships that distinguish between modes serve as useful symbolic abstraction over continuous state All this work takes Soar toward being able to autonomously learn and improve behavior in continuous environments 22

Evaluation Coal Scaling issues: linear regression is exponential in number of objects Linear modes is insufficient for more complex physics such as bouncing -> catastrophic failure Nuggets Modal model learning is more accurate and general than uniform models Abstraction learning results are promising, but preliminary 23