EMPATH: A Neural Network that Categorizes Facial Expressions Matthew N. Dailey and Garrison W. Cottrell University of California, San Diego Curtis Padgett.

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Introduction to Support Vector Machines (SVM)
Chapter 2.
Object recognition and scene “understanding”
Fast Readout of Object Identity from Macaque Inferior Tempora Cortex Chou P. Hung, Gabriel Kreiman, Tomaso Poggio, James J.DiCarlo McGovern Institute for.
HMAX Models Architecture Jim Mutch March 31, 2010.
Kostas Kontogiannis E&CE
Lesions of Retinostriate Pathway Lesions (usually due to stroke) cause a region of blindness called a scotoma Identified using perimetry note macular sparing.
Classification Neural Networks 1
Test on Friday!. Lesions of Retinostriate Pathway Lesions (usually due to stroke) cause a region of blindness called a scotoma Identified using perimetry.
Current Trends in Image Quality Perception Mason Macklem Simon Fraser University
Visual Pathways W. W. Norton Primary cortex maintains distinct pathways – functional segregation M and P pathways synapse in different layers Ascending.
Searching for the NCC We can measure all sorts of neural correlates of these processes…so we can see the neural correlates of consciousness right? So what’s.
Lecture 14 – Neural Networks
Visual Cognition II Object Perception. Theories of Object Recognition Template matching models Feature matching Models Recognition-by-components Configural.
Cognitive Processes PSY 334 Chapter 2 – Perception June 30, 2003.
Pattern Recognition Pattern - complex composition of sensory stimuli that the human observer may recognize as being a member of a class of objects Issue.
Visual Cognition II Object Perception. Theories of Object Recognition Template matching models Feature matching Models Recognition-by-components Configural.
Visual Expertise Is a General Skill Maki Sugimoto University of California, San Diego November 20, 2000.
Exemplar-based accounts of “multiple system” phenomena in perceptual categorization R. M. Nosofsky and M. K. Johansen Presented by Chris Fagan.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
October 14, 2010Neural Networks Lecture 12: Backpropagation Examples 1 Example I: Predicting the Weather We decide (or experimentally determine) to use.
Artificial Neural Networks
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
October 7, 2010Neural Networks Lecture 10: Setting Backpropagation Parameters 1 Creating Data Representations On the other hand, sets of orthogonal vectors.
An aside: peripheral drift illusion illusion of motion is strongest when reading text (such as this) while viewing the image in your periphery. Blinking.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Artificial Neural Networks for Secondary Structure Prediction CSC391/691 Bioinformatics Spring 2004 Fetrow/Burg/Miller (slides by J. Burg)
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Facial Feature Detection
Multiclass object recognition
Pattern Recognition Vidya Manian Dept. of Electrical and Computer Engineering University of Puerto Rico INEL 5046, Spring 2007
Speech Perception 4/6/00 Acoustic-Perceptual Invariance in Speech Perceptual Constancy or Perceptual Invariance: –Perpetual constancy is necessary, however,
December 5, 2012Introduction to Artificial Intelligence Lecture 20: Neural Network Application Design III 1 Example I: Predicting the Weather Since the.
1 Backprop, 25 years later… Garrison W. Cottrell Gary's Unbelievable Research Unit (GURU) Computer Science and Engineering Department Temporal Dynamics.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Machine Learning Chapter 4. Artificial Neural Networks
Classification / Regression Neural Networks 2
CSC321: Neural Networks Lecture 13: Learning without a teacher: Autoencoders and Principal Components Analysis Geoffrey Hinton.
Methodology of Simulations n CS/PY 399 Lecture Presentation # 19 n February 21, 2001 n Mount Union College.
Learning to perceive how hand-written digits were drawn Geoffrey Hinton Canadian Institute for Advanced Research and University of Toronto.
1 Ying-li Tian, Member, IEEE, Takeo Kanade, Fellow, IEEE, and Jeffrey F. Cohn, Member, IEEE Presenter: I-Chung Hung Advisor: Dr. Yen-Ting Chen Date:
Classification Course web page: vision.cis.udel.edu/~cv May 12, 2003  Lecture 33.
A Two-level Pose Estimation Framework Using Majority Voting of Gabor Wavelets and Bunch Graph Analysis J. Wu, J. M. Pedersen, D. Putthividhya, D. Norgaard,
Human vision Jitendra Malik U.C. Berkeley. Visual Areas.
Talk #3: EMPATH A Neural Network Model of Human Facial Expression Recognition Gary Cottrell Joint work with Matt Dailey, Curt Padgett, and Ralph Adolphs.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
Face Image-Based Gender Recognition Using Complex-Valued Neural Network Instructor :Dr. Dong-Chul Kim Indrani Gorripati.
SAD, ANGER, FEAR, DISGUST, HAPPY, CONTEMPT AND OTHERS ARE EVOLUTIONARILY DICTATED ADAPTIVE SURVIVAL MECHANISMS Darwin Tompkins Ekman Izard.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
1 Backprop: Representations, Representations, Representations Garrison W. Cottrell Gary's Unbelievable Research Unit (GURU) Computer Science and Engineering.
Categorical Perception of Expressions in Chimpanzees (Pan troglodytes) Jennifer M.B. Fugate March 15, 2007.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Neural Networks Lecture 4 out of 4. Practical Considerations Input Architecture Output.
EXAMPLES ABSTRACT RESEARCH QUESTIONS DISCUSSION FUTURE DIRECTIONS Very little is known about how and to what extent emotions are conveyed through avatar.
Do Expression and Identity Need Separate Representations?
CSC2535: Computation in Neural Networks Lecture 11 Extracting coherent properties by maximizing mutual information across space or time Geoffrey Hinton.
Learning in Neural Networks
Visual Computation and Learning Lab
Cognitive Processes PSY 334
Gary Cottrell & * 07/16/96 The Face of Fear: A Neural Network Model of Human Facial Expression Recognition Garrison W. Cottrell Gary's Unbelievable Research.
Classification Neural Networks 1
Jason A. Cromer, Jefferson E. Roy, Earl K. Miller  Neuron 
Creating Data Representations
Backprop, Representations, Representations, Representations,
Itzhak Fried, Katherine A MacDonald, Charles L Wilson  Neuron 
Jason A. Cromer, Jefferson E. Roy, Earl K. Miller  Neuron 
Modeling IDS using hybrid intelligent systems
Volume 50, Issue 1, Pages (April 2006)
Volume 27, Issue 1, Pages (January 2017)
Presentation transcript:

EMPATH: A Neural Network that Categorizes Facial Expressions Matthew N. Dailey and Garrison W. Cottrell University of California, San Diego Curtis Padgett California Institute of Technology

Facial Expression Recognition (Theory 1) Categorical Perception Categories are discrete entities Sharp categorical boundaries Discrimination of similar pairs of expressive faces is enhanced when near category boundaries

Facial Expression Recognition (Theory 2) Graded and expressions considered points in a continuous, low-dimensional space e.g. “Surprise” between “Happiness” and “Fear”

Historical Research (Categorical) Ekman and Friesen (1976) 10-step photos between pairs of caricatures Ekman 1999 essay on basic emotions Harnad, 1987 Categorical Perception Beale and Keil (1995) morph image sequence with famous faces Etcoff and Magee (1992) facial expression recognition tied to perceptual mechanism

Historical Research (Continuous) Schlosberg (1952) category ratings and subjects “errors” predicted accurately by arranging categories around an ellipse Russell (1980) structure theory of emotions Russell and Bullock (1986) emotion categories best thought of as fuzzy sets Russel et al. (1989),Katsikitis (1997), Schiano et al. (2000) continuous multidimensional perceptual space for facial expression perception

Young et al.’s (1997) “Megamix” Experiments Experiment 1: Subjects identify the emotional category in 10%, 30%, 50%, 70%, and 90% morphs between all pairs of the 6 prototypical expressions  6-way forced choice identification Experiment 2: Same as experiment 1 with the addition of the “neutral” face  7-way forced choice

Young et al.’s (1997) “Megamix” Experiments Experiment 3: Discriminate pairs of stimuli along the six transitions  Sequential discrimination task (ABX)  Simultaneous discrimination task (same-different) Experiment 4: Determine what expression is “mixed- in” to a faint morph  Given a morph or prototype stimulus, indicate the most apparent, second-most apparent, and third-most apparent emotion

“Megamix” Experiment Results Results from experiments 1-3 support the categorical view of facial expression perception Results from experiment 4 showed that subjects were significantly likely to detect mixed-in emotion at 30%. This supports the continuous, dimentional accounts of facial expression perception Rather than settling the issue of categorical vs. continuous theories they found evidence to support BOTH theories Until now, no computational model has ever been able to simultaneously explain these seemingly contradictory data

The Model Three layer neural network  Perceptual analysis  Object representation  Categorization Feedforward network (no backpropagation at later levels) Input is 240 x 292 grayscale face image

Perceptual Analysis Layer Neurons whose response properties are similar to complex cells in the visual cortex This is modeled by “Gabor Filters” Basically, these units do nonlinear edge detection at five different scales and eight different orientations

Object Representation Layer Extract small set of features from high dimensional data Equal to an “image compression” network that extracts global representations of the data Principal components analysis is used to model this layer 50 linear hidden units

Categorization Layer Simple perceptron with six outputs (one for each “basic” emotion) The network is set up so that the output can be interpreted as probabilities (i.e. they are all positive and sum to 1)

The Model

Experiments & Results Same experiments as the Young et al. “Megamix” experiments Results  The model and humans find the same expressions difficult or easy to interpret  When presented with morphs between pairs of expressions, the model and humans place similar sharp category boundaries between prototypes  The model and humans are similarly sensitive to mixed-in expressions in morph stimuli

More Results Network generalization to unseen faces, compared to human agreement on the same face (six-way forced choice)

More Results

Conclusion This model was able to simulate both the categorical and continuous nature of facial classification consistent with the human experiments conducted by Young et al. Categorical or Continuous?  Conclusion leans toward both theories being complimentary instead of mutually exclusive  “tapping different computational levels of processing”  Which method is dictated by the task and the data