Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3

Slides:



Advertisements
Similar presentations
Lecture 18: Temporal-Difference Learning
Advertisements

CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 27 – Overview of probability concepts 1.
A Tutorial on Learning with Bayesian Networks
Technische Universität München Face Model Fitting based on Machine Learning from Multi-band Images of Facial Components Institute for Informatics Technische.
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Department of Computer Science Undergraduate Events More
Intelligent Agents Russell and Norvig: 2
Dynamic Bayesian Networks (DBNs)
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
Hidden Markov Models Reading: Russell and Norvig, Chapter 15, Sections
Introduction of Probabilistic Reasoning and Bayesian Networks
Visual Event Detection & Recognition Filiz Bunyak Ersoy, Ph.D. student Smart Engineering Systems Lab.
Modeling Human Reasoning About Meta-Information Presented By: Scott Langevin Jingsong Wang.
Yiannis Demiris and Anthony Dearden By James Gilbert.
System Architecture Intelligently controlling image processing systems.
What Are Partially Observable Markov Decision Processes and Why Might You Care? Bob Wall CS 536.
Cognitive Computer Vision
1 Reasoning Under Uncertainty Over Time CS 486/686: Introduction to Artificial Intelligence Fall 2013.
Albert Gatt Corpora and Statistical Methods Lecture 8.
Region labelling Giving a region a name. Image Processing and Computer Vision: 62 Introduction Region detection isolated regions Region description properties.
Report on Intrusion Detection and Data Fusion By Ganesh Godavari.
A Bayesian algorithm for tracking multiple moving objects in outdoor surveillance video Department of Electrical Engineering and Computer Science The University.
Part 2 of 3: Bayesian Network and Dynamic Bayesian Network.
Representing Uncertainty CSE 473. © Daniel S. Weld 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one.
CPSC 322, Lecture 31Slide 1 Probability and Time: Markov Models Computer Science cpsc322, Lecture 31 (Textbook Chpt 6.5) March, 25, 2009.
CPSC 422, Lecture 14Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14 Feb, 4, 2015 Slide credit: some slides adapted from Stuart.
Cognitive Processes PSY 334 Chapter 2 – Perception.
CS 188: Artificial Intelligence Fall 2009 Lecture 19: Hidden Markov Models 11/3/2009 Dan Klein – UC Berkeley.
Cognitive Computer Vision 3R400 Kingsley Sage Room 5C16, Pevensey III
Learning Models of Relational Stochastic Processes Sumit Sanghai.
Computer vision: models, learning and inference Chapter 6 Learning and Inference in Vision.
Intrusion and Anomaly Detection in Network Traffic Streams: Checking and Machine Learning Approaches ONR MURI area: High Confidence Real-Time Misuse and.
Learning to classify the visual dynamics of a scene Nicoletta Noceti Università degli Studi di Genova Corso di Dottorato.
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
Chapter 8 Prediction Algorithms for Smart Environments
Chapter 11 LEARNING FROM DATA. Chapter 11: Learning From Data Outline  The “Learning” Concept  Data Visualization  Neural Networks The Basics Supervised.
Probabilistic Context Free Grammars for Representing Action Song Mao November 14, 2000.
 The most intelligent device - “Human Brain”.  The machine that revolutionized the whole world – “computer”.  Inefficiencies of the computer has lead.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
The Science of Prediction Location Intelligence Conference April 4, 2006 How Next Generation Traffic Services Will Impact Business Dr. Oliver Downs, Chief.
K. J. O’Hara AMRS: Behavior Recognition and Opponent Modeling Oct Behavior Recognition and Opponent Modeling in Autonomous Multi-Robot Systems.
Recognizing Activities of Daily Living from Sensor Data Henry Kautz Department of Computer Science University of Rochester.
Report on Intrusion Detection and Data Fusion By Ganesh Godavari.
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
Using Inactivity to Detect Unusual behavior Presenter : Siang Wang Advisor : Dr. Yen - Ting Chen Date : Motion and video Computing, WMVC.
Estimating Component Availability by Dempster-Shafer Belief Networks Estimating Component Availability by Dempster-Shafer Belief Networks Lan Guo Lane.
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
C. Lawrence Zitnick Microsoft Research, Redmond Devi Parikh Virginia Tech Bringing Semantics Into Focus Using Visual.
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Lecture notes 9 Bayesian Belief Networks.
Cognitive Systems Foresight Language and Speech. Cognitive Systems Foresight Language and Speech How does the human system organise itself, as a neuro-biological.
Slides for “Data Mining” by I. H. Witten and E. Frank.
August 30, 2004STDBM 2004 at Toronto Extracting Mobility Statistics from Indexed Spatio-Temporal Datasets Yoshiharu Ishikawa Yuichi Tsukamoto Hiroyuki.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
1 Chapter 17 2 nd Part Making Complex Decisions --- Decision-theoretic Agent Design Xin Lu 11/04/2002.
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
Cognitive Computer Vision 3R400 Kingsley Sage Room 5C16, Pevensey III
HIERARCHICAL TEMPORAL MEMORY WHY CANT COMPUTERS BE MORE LIKE THE BRAIN?
A Forest of Sensors: Using adaptive tracking to classify and monitor activities in a site Eric Grimson AI Lab, Massachusetts Institute of Technology
Cognitive Computer Vision
Chapter 6: Temporal Difference Learning
Artificial Intelligence Lecture No. 5
CSc4730/6730 Scientific Visualization
Chapter 6: Temporal Difference Learning
Probability and Time: Markov Models
Computing probabilities using Expect problem-solving Trees: A worked example Jim Blythe USC/ISI.
Presentation transcript:

Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3

Lecture 9 Recognising visual behaviour – Bottom-up and top-down vision – HIVIS (case study) – Dynamic Bayesian Networks – Dynamic Decision Networks – Bayes Automated Taxi (case study) In this lecture we shall see how some of the techniques we have seen thus far can be used to build real Cognitive Vision Systems

Visual behaviour In computational terms, visual behaviour can be defined as a functional description of the spatial and/or temporal dynamics of a visual object or set of objects in an environment Functional description may be characterised by, for example: – “simple” tabular form – set of visual prototypes (facial models …) – statistical models (HMMs, VLMMs …) Recognising visual behaviour means finding the fit between the model and observation data

Visual behaviour © Johnson & Hogg Green circles indicate match with a “normal” trajectory. Red circles indicate “unusual” behaviour

Visual behaviour toy example due to Frey & Jojic involving changing structure over time PacMan moves forward at each time step with probability 0.8

Bottom-up / top-down vision A generalised view Scene Interpretation …… CONTROL POLICY (WITH STATE MEMORY) FEATURE COMBINATION d1d1 d2d2 dNdN Image Data Driven Task Based Control

Case study: traffic surveillance HIVIS (Buxton and Howarth) The task is to identify traffic behaviours (such as overtaking) Imagery taken from a roundabout in Germany Buxton and Howarth took two different approaches – HIVIS MONITOR (bottom-up) – HIVIS WATCHER (with top-down control)

HIVIS MONITOR on-road-surface on-entry-road stationary significant orientation change in-right-turn-region on-roundabout t1t1 t2t2 t3t3 t4t4 t5t5 t6t6 t7t7 t8t8 Space is divided into regions (see right) so we can specify relationships on objects

HIVIS MONITOR These spatial relationship primitives were then used to build a set of activity primitives using a temporal logic e.g.

HIVIS MONITOR We can use these logics to build up knowledge of individual object properties and object pairs (like overtaking) This is fine, but there are drawbacks: – We examine all objects without considering whether they are relevant to the behaviours we want to detect (e.g. vehicles passing in the background) – We can only recognise the behaviours after they have taken place. This is OK for an off-line analysis system (post-event analysis) but not useful for a “live” system where we want to be able to detect behaviours as they emerge (for control and prediction)

HIVIS WATCHER In HIVIS WATCHER, a set of (computationally low-cost) Visual Routines are used to generate pre-attentive objects A high level Control Policy is used to select objects that warrant attention (attentive selection) - In this case, mutual proximity of objects. This control policy determines a “watch” parameter (how useful are these objects to watch) Deictic (“pointing”) markers are assigned to selected objects. Behaviour primitives (“episodes”) are formed by applying BBN rules on objects referenced by the markers

HIVIS WATCHER Deictic (“pointing”) reference Pre-attentive selection assigns markers to objects that warrant attention For each object, we determine information about the relative position, speed and heading of the object BBNs are then used to combine data into likely episodes

HIVIS WATCHER Combining spatial relationships into episodes time pairs watch agents s o deictic state overtake follow queue unknown episode

Dynamic Bayesian Networks Recall from previous lectures that BBNs have Conditional Probability tables for each node This static approach can be extended to reflect the fact that external factors can influence the BBN (structure and CPT values) in a manner that may not be convenient to model by adding additional nodes to the network, resulting in a DBN So, using HIVIS as an example, knowledge of the likely episode at time t influences our belief in the likely episode at time t+1 (like a Markov assumption). Such an assumption can help us, for example, to maintain temporal continuity

Dynamic Decision Networks DDNs are similar in concept, but include the notion of utility U of an action resulting from a decision D Action p(Rain)=0.3p(Sunny=0.7) Walk to work Drive to work3010

Graphical representation of DBNs and DDNs Reproduced from Forbes/Huang/Kanazawa/Russell 1995 Decisions D are made by some agent (cf. to-down control) and inform the state evolution process. When “looking ahead” we use Utility U to evaluate the cost associated of being in a state

Bayes Automated Taxi (BAT) Forbes, Huang, Kanazawa and Russell 1995 Forbes et al was concerned with autonomous vehicles for driving on a normal (unadapted) highway using vision Developed a driving simulator for the BATmobile to test DDN based decision making module Use a set of high-level decision tree structures to decide which actions to pursue … Reproduced from Forbes/Huang/Kanazawa/Russell 1995

Bayes Automated Taxi (BAT) Reproduced from Forbes/Huang/Kanazawa/Russell 1995 This a part of the probabilistic network for one vehicle Smaller nodes with thicker outlines are sensor observations

Bayes Automated Taxi (BAT) Reproduced from Forbes/Huang/Kanazawa/Russell 1995

Further reading “Conceptual descriptions from monitoring and watching image sequences”, R. J. Howarth and H. Buxton, Image and Vision Computing 18, 2000 “The BATmobile: Towards a Bayesian Automated Taxi”, J. Forbes, T. Huang, K, Kanazawa, S. Russell, 1995

Summary Visual behaviour recognition involves modelling the spatial and temporal structures of scales in terms of objects The top-down or task-driven approach has many advantages in computational terms Dynamic Bayesian Nets and Dynamic Decision Nets are useful formalisms for real-world applications

Next time … Task based visual control