Presentation is loading. Please wait.

Presentation is loading. Please wait.

What can computational models tell us about face processing?

Similar presentations


Presentation on theme: "What can computational models tell us about face processing?"— Presentation transcript:

1 What can computational models tell us about face processing?
Gary Cottrell & * 07/16/96 What can computational models tell us about face processing? Garrison W. Cottrell Gary's Unbelievable Research Unit (GURU) Computer Science and Engineering Department Institute for Neural Computation UCSD Collaborators on this work (and earlier versions of this work): Ralph Adolphs, Kristin Branson, Andy Calder, Matthew Dailey, Michael Fleming, Carrie Joyce, Janet Metcalfe, Nam Nguyen, Curt Padgett, Maki Sugimoto, Matt Tong, Brian Tran APA Talk, August 2000: The face of fear: *

2 What can computational models tell us about face processing?
Gary Cottrell & * 07/16/96 What can computational models tell us about face processing? Garrison W. Cottrell Gary's Unbelievable Research Unit (GURU) Computer Science and Engineering Department Institute for Neural Computation UCSD Collaborators on this work (and earlier versions of this work): Ralph Adolphs, Kristin Branson, Andy Calder, Matthew Dailey, Michael Fleming, Carrie Joyce, Janet Metcalfe, Nam Nguyen, Curt Padgett, Maki Sugimoto, Matt Tong, Brian Tran APA Talk, August 2000: The face of fear: *

3 What can computational models tell us about face processing?
Gary Cottrell & * 07/16/96 What can computational models tell us about face processing? Garrison W. Cottrell Gary's Unbelievable Research Unit (GURU) Computer Science and Engineering Department Institute for Neural Computation UCSD Collaborators on this work (and earlier versions of this work): Ralph Adolphs, Kristin Branson, Andy Calder, Matthew Dailey, Michael Fleming, Carrie Joyce, Janet Metcalfe, Nam Nguyen, Curt Padgett, Maki Sugimoto, Matt Tong, Brian Tran APA Talk, August 2000: The face of fear: *

4 What can computational models tell us about face processing?
Gary Cottrell & * 07/16/96 What can computational models tell us about face processing? Garrison W. Cottrell Gary's Unbelievable Research Unit (GURU) Computer Science and Engineering Department Institute for Neural Computation UCSD Collaborators on this work (and earlier versions of this work): Ralph Adolphs, Kristin Branson, Andy Calder, Matthew Dailey, Michael Fleming, Carrie Joyce, Janet Metcalfe, Nam Nguyen, Curt Padgett, Maki Sugimoto, Matt Tong, Brian Tran APA Talk, August 2000: The face of fear: *

5 What can computational models tell us about face processing?
Gary Cottrell & * 07/16/96 What can computational models tell us about face processing? Garrison W. Cottrell Gary's Unbelievable Research Unit (GURU) Computer Science and Engineering Department Institute for Neural Computation UCSD Collaborators on this work (and earlier versions of this work): Ralph Adolphs, Kristin Branson, Andy Calder, Matthew Dailey, Michael Fleming, Carrie Joyce, Janet Metcalfe, Nam Nguyen, Curt Padgett, Maki Sugimoto, Matt Tong, Brian Tran APA Talk, August 2000: The face of fear: *

6 Why model? Models rush in where theories fear to tread.
Gary Cottrell & * 07/16/96 Why model? Models rush in where theories fear to tread. Models can be manipulated in ways people cannot Models can be analyzed in ways people cannot. APA Talk, August 2000: The face of fear: *

7 Models rush in where theories fear to tread
Gary Cottrell & * 07/16/96 Models rush in where theories fear to tread Theories are high level descriptions of the processes underlying behavior. They are often not explicit about the processes involved. They are difficult to reason about if no mechanisms are explicit -- they may be too high level to make explicit predictions. Theory formation itself is difficult. Using machine learning techniques, one can often build a working model of a task for which we have no theories or algorithms (e.g., expression recognition). A working model provides an “intuition pump” for how things might work, especially if they are “neurally plausible” (e.g., development of face processing - Dailey and Cottrell). A working model may make unexpected predictions (e.g., the Interactive Activation Model and SLNT). APA Talk, August 2000: The face of fear: *

8 Models can be manipulated in ways people cannot
Gary Cottrell & * 07/16/96 Models can be manipulated in ways people cannot We can see the effects of variations in cortical architecture (e.g., split (hemispheric) vs. non-split models (Shillcock and Monaghan word perception model)). We can see the effects of variations in processing resources (e.g., variations in number of hidden units in Plaut et al. models). We can see the effects of variations in environment (e.g., what if our parents were cans, cups or books instead of humans? I.e., is there something special about face expertise versus visual expertise in general? (Sugimoto and Cottrell, Joyce and Cottrell, Tong & Cottrell)). We can see variations in behavior due to different kinds of brain damage within a single “brain” (e.g. Juola and Plunkett, Hinton and Shallice). APA Talk, August 2000: The face of fear: *

9 Models can be analyzed in ways people cannot
Gary Cottrell & * 07/16/96 Models can be analyzed in ways people cannot In the following, I specifically refer to neural network models. We can do single unit recordings. We can selectively ablate and restore parts of the network, even down to the single unit level, to assess the contribution to processing. We can measure the individual connections -- e.g., the receptive and projective fields of a unit. We can measure responses at different layers of processing (e.g., which level accounts for a particular judgment: perceptual, object, or categorization? (Dailey et al. J Cog Neuro 2002). APA Talk, August 2000: The face of fear: *

10 How (I like) to build Cognitive Models
Gary Cottrell & * 07/16/96 How (I like) to build Cognitive Models I like to be able to relate them to the brain, so “neurally plausible” models are preferred -- neural nets. The model should be a working model of the actual task, rather than a cartoon version of it. Of course, the model should nevertheless be simplifying (i.e. it should be constrained to the essential features of the problem at hand): Do we really need to model the (supposed) translation invariance and size invariance of biological perception? As far as I can tell, NO! Then, take the model “as is” and fit the experimental data: 0 fitting parameters is preferred over 1, 2 , or 3. APA Talk, August 2000: The face of fear: *

11 The other way (I like) to build Cognitive Models
Gary Cottrell & * 07/16/96 The other way (I like) to build Cognitive Models Same as above, except: Use them as exploratory models -- in domains where there is little direct data (e.g. no single cell recordings in infants or undergraduates) to suggest what we might find if we could get the data. These can then serve as “intuition pumps.” Examples: Why we might get specialized face processors Why those face processors get recruited for other tasks APA Talk, August 2000: The face of fear: *

12 Gary Cottrell & * 07/16/96 Today: Talk #1: The Face of Fear: A Neural Network Model of Human Facial Expression Recognition Joint work with Matt Dailey APA Talk, August 2000: The face of fear: *

13 A Good Cognitive Model Should:
Gary Cottrell & * 07/16/96 A Good Cognitive Model Should: Be psychologically relevant (i.e. it should be in an area with a lot of real, interesting psychological data). Actually be implemented. If possible, perform the actual task of interest rather than a cartoon version of it. Be simplifying (i.e. it should be constrained to the essential features of the problem at hand). Fit the experimental data. Make new predictions to guide psychological research. APA Talk, August 2000: The face of fear: *

14 Gary Cottrell & * 07/16/96 The Issue: Are Similarity and Categorization Two Sides of the Same Coin? Some researchers believe perception of facial expressions is a new example of categorical perception: Like the colors of a rainbow, the brain separates expressions into discrete categories, with: Sharp boundaries between expressions, and… Higher discrimination of faces near those boundaries. A good answer to these questions would tell us a lot about how vision works in the brain. So our work, in modeling expression recognition, is cool. APA Talk, August 2000: The face of fear: *

15 Percent of subjects who pressed this button Gary Cottrell & * 07/16/96
APA Talk, August 2000: The face of fear: *

16 Gary Cottrell & * 07/16/96 The Issue: Are Similarity and Categorization Two Sides of the Same Coin? Some researchers believe the underlying representation of facial expressions is NOT discrete: There are two (or three) underlying dimensions, e.g., intensity and valence (found by MDS). Our perception of expressive faces induces a similarity structure that results in a circle in this space A good answer to these questions would tell us a lot about how vision works in the brain. So our work, in modeling expression recognition, is cool. APA Talk, August 2000: The face of fear: *

17 The Face Processing System
Gary Cottrell & * 07/16/96 The Face Processing System PCA . Happy Sad Afraid Angry Surprised Disgusted Neural Net Pixel (Retina) Level Object (IT) Perceptual (V1) Category Gabor Filtering APA Talk, August 2000: The face of fear: *

18 The Face Processing System
Gary Cottrell & * 07/16/96 The Face Processing System PCA . Bob Carol Ted Alice Neural Net Pixel (Retina) Level Object (IT) Perceptual (V1) Category Gabor Filtering APA Talk, August 2000: The face of fear: *

19 The Face Processing System
Gary Cottrell & * 07/16/96 The Face Processing System . Bob Carol Ted Cup Can Book PCA Gabor Filtering Neural Net Pixel (Retina) Level Perceptual (V1) Level Object (IT) Level Category Level Feature level APA Talk, August 2000: The face of fear: *

20 The Face Processing System
Gary Cottrell & * 07/16/96 The Face Processing System LSF PCA HSFPCA . Gabor Filtering Bob Carol Ted Cup Can Book Neural Net Pixel (Retina) Level Object (IT) Perceptual (V1) Category APA Talk, August 2000: The face of fear: *

21 Gary Cottrell & * 07/16/96 The Gabor Filter Layer Basic feature: the 2-D Gabor wavelet filter (Daugman, 85): These model the processing in early visual areas Convolution * Magnitudes Subsample in a 29x36 grid APA Talk, August 2000: The face of fear: *

22 Principal Components Analysis
Gary Cottrell & * 07/16/96 Principal Components Analysis The Gabor filters give us 40,600 numbers We use PCA to reduce this to 50 numbers PCA is like Factor Analysis: It finds the underlying directions of Maximum Variance PCA can be computed in a neural network through a competitive Hebbian learning mechanism Hence this is also a biologically plausible processing step We suggest this leads to representations similar to those in Inferior Temporal cortex APA Talk, August 2000: The face of fear: *

23 Gary Cottrell & * 07/16/96 How to do PCA with a neural network (Cottrell, Munro & Zipser, 1987; Cottrell & Fleming 1990; Cottrell & Metcalfe 1990; O’Toole et al. 1991) A self-organizing network that learns whole-object representations (features, Principal Components, Holons, eigenfaces) Holons (Gestalt layer) Input from Perceptual Layer The next layer up extracts covariances from the previous layer -- essentially doing a principal components analysis (you can think factor analysis if you don’t know what PCA is). That is, the features are sensitive to the directions of maximum variation in the gabor filter responses. When you look at what the Optimal stimulus for these units are, (something it is easy to do in neural nets, relatively difficult in brains), you see these rather ghostly looking faces ... APA Talk, August 2000: The face of fear: *

24 Gary Cottrell & * 07/16/96 How to do PCA with a neural network (Cottrell, Munro & Zipser, 1987; Cottrell & Fleming 1990; Cottrell & Metcalfe 1990; O’Toole et al. 1991) A self-organizing network that learns whole-object representations (features, Principal Components, Holons, eigenfaces) Holons (Gestalt layer) Input from Perceptual Layer The next layer up extracts covariances from the previous layer -- essentially doing a principal components analysis (you can think factor analysis if you don’t know what PCA is). That is, the features are sensitive to the directions of maximum variation in the gabor filter responses. When you look at what the Optimal stimulus for these units are, (something it is easy to do in neural nets, relatively difficult in brains), you see these rather ghostly looking faces ... APA Talk, August 2000: The face of fear: *

25 Gary Cottrell & * 07/16/96 How to do PCA with a neural network (Cottrell, Munro & Zipser, 1987; Cottrell & Fleming 1990; Cottrell & Metcalfe 1990; O’Toole et al. 1991) A self-organizing network that learns whole-object representations (features, Principal Components, Holons, eigenfaces) Holons (Gestalt layer) Input from Perceptual Layer The next layer up extracts covariances from the previous layer -- essentially doing a principal components analysis (you can think factor analysis if you don’t know what PCA is). That is, the features are sensitive to the directions of maximum variation in the gabor filter responses. When you look at what the Optimal stimulus for these units are, (something it is easy to do in neural nets, relatively difficult in brains), you see these rather ghostly looking faces ... APA Talk, August 2000: The face of fear: *

26 Gary Cottrell & * 07/16/96 How to do PCA with a neural network (Cottrell, Munro & Zipser, 1987; Cottrell & Fleming 1990; Cottrell & Metcalfe 1990; O’Toole et al. 1991) A self-organizing network that learns whole-object representations (features, Principal Components, Holons, eigenfaces) Holons (Gestalt layer) Input from Perceptual Layer The next layer up extracts covariances from the previous layer -- essentially doing a principal components analysis (you can think factor analysis if you don’t know what PCA is). That is, the features are sensitive to the directions of maximum variation in the gabor filter responses. When you look at what the Optimal stimulus for these units are, (something it is easy to do in neural nets, relatively difficult in brains), you see these rather ghostly looking faces ... APA Talk, August 2000: The face of fear: *

27 Gary Cottrell & * 07/16/96 How to do PCA with a neural network (Cottrell, Munro & Zipser, 1987; Cottrell & Fleming 1990; Cottrell & Metcalfe 1990; O’Toole et al. 1991) A self-organizing network that learns whole-object representations (features, Principal Components, Holons, eigenfaces) Holons (Gestalt layer) Input from Perceptual Layer The next layer up extracts covariances from the previous layer -- essentially doing a principal components analysis (you can think factor analysis if you don’t know what PCA is). That is, the features are sensitive to the directions of maximum variation in the gabor filter responses. When you look at what the Optimal stimulus for these units are, (something it is easy to do in neural nets, relatively difficult in brains), you see these rather ghostly looking faces ... APA Talk, August 2000: The face of fear: *

28 Gary Cottrell & * 07/16/96 How to do PCA with a neural network (Cottrell, Munro & Zipser, 1987; Cottrell & Fleming 1990; Cottrell & Metcalfe 1990; O’Toole et al. 1991) A self-organizing network that learns whole-object representations (features, Principal Components, Holons, eigenfaces) Holons (Gestalt layer) Input from Perceptual Layer The next layer up extracts covariances from the previous layer -- essentially doing a principal components analysis (you can think factor analysis if you don’t know what PCA is). That is, the features are sensitive to the directions of maximum variation in the gabor filter responses. When you look at what the Optimal stimulus for these units are, (something it is easy to do in neural nets, relatively difficult in brains), you see these rather ghostly looking faces ... APA Talk, August 2000: The face of fear: *

29 Gary Cottrell & * 07/16/96 How to do PCA with a neural network (Cottrell, Munro & Zipser, 1987; Cottrell & Fleming 1990; Cottrell & Metcalfe 1990; O’Toole et al. 1991) A self-organizing network that learns whole-object representations (features, Principal Components, Holons, eigenfaces) ... Holons (Gestalt layer) Input from Perceptual Layer The next layer up extracts covariances from the previous layer -- essentially doing a principal components analysis (you can think factor analysis if you don’t know what PCA is). That is, the features are sensitive to the directions of maximum variation in the gabor filter responses. When you look at what the Optimal stimulus for these units are, (something it is easy to do in neural nets, relatively difficult in brains), you see these rather ghostly looking faces APA Talk, August 2000: The face of fear: *

30 Holons They act like face cells (Desimone, 1991):
Gary Cottrell & * 07/16/96 Holons They act like face cells (Desimone, 1991): Response of single units is strong despite occluding eyes, e.g. Response drops off with rotation Some fire to my dog’s face A novel representation: Distributed templates -- each unit’s optimal stimulus is a ghostly looking face (template-like), but many units participate in the representation of a single face (distributed). For this audience: Neither exemplars nor prototypes! Explain holistic processing: Why? If stimulated with a partial match, the firing represents votes for this template: Units “downstream” don’t know what caused this unit to fire. (more on this later…) These cells, or holons, as janet metcalfe and I called them, act like face cells. The respond to faces even when the eyes, for example Are occluded, the response drops off with rotation of the stimulus, and they generalize to other species. The representation is like distributed grandmother cells - each receptive field looks like a face, but each unit participates in some, but not all of the faces. Finally, there is a processing advantage to these representations: they do pattern completion. If part of a holon is matched, no one listening to this cell knows which half. APA Talk, August 2000: The face of fear: *

31 Gary Cottrell & * 07/16/96 The Final Layer: Classification (Cottrell & Fleming 1990; Cottrell & Metcalfe 1990; Padgett & Cottrell 1996; Dailey & Cottrell, 1999; Dailey et al. 2002) The holistic representation is then used as input to a categorization network trained by supervised learning. Output: Cup, Can, Book, Greeble, Face, Bob, Carol, Ted, Happy, Sad, Afraid, etc. Categories ... Holons Input from Perceptual Layer The last, and essential layer, is the classifier. This is a neural network trained to categorize the inputs using error correction learning. My colleagues and I have used these successfully for many years to classify visual inputs by class, identity, and expression. Excellent generalization performance demonstrates the sufficiency of the holistic representation for recognition APA Talk, August 2000: The face of fear: *

32 The Final Layer: Classification
Gary Cottrell & * 07/16/96 The Final Layer: Classification Categories can be at different levels: basic, subordinate. Simple learning rule (~delta rule). It says (mild lie here): add inputs to your weights (synaptic strengths) when you are supposed to be on, subtract them when you are supposed to be off. This makes your weights “look like” your favorite patterns – the ones that turn you on. When no hidden units => No back propagation of error. When hidden units: we get task-specific features (most interesting when we use the basic/subordinate distinction) APA Talk, August 2000: The face of fear: *

33 Outline An overview of our facial expression recognition system.
Gary Cottrell & * 07/16/96 Outline An overview of our facial expression recognition system. The internal representation shows the model’s prototypical representations of Fear, Sadness, etc. How our model accounts for the “categorical” data How our model accounts for the “two-dimensional” data Discussion Conclusions for part 1 APA Talk, August 2000: The face of fear: *

34 Facial Expression Database
Gary Cottrell & * 07/16/96 Facial Expression Database Ekman and Friesen quantified muscle movements (Facial Actions) involved in prototypical portrayals of happiness, sadness, fear, anger, surprise, and disgust. Result: the Pictures of Facial Affect Database (1976). 70% agreement on emotional content by naive human subjects. 110 images, 14 subjects, 7 expressions. Anger, Disgust, Neutral, Surprise, Happiness (twice), Fear, and Sadness This is actor “JJ”: The easiest for humans (and our model) to classify APA Talk, August 2000: The face of fear: *

35 The Final Layer: Classification
Gary Cottrell & * 07/16/96 The Final Layer: Classification The final layer is trained based on the putative emotion being expressed by the actor (what the majority of Ekman’s subjects responded when given a 6-way forced choice) Uses a very simple learning rule: The delta rule It says: add inputs to your weights (synaptic strengths) when you are supposed to be on, subtract them when you are supposed to be off. This makes your weights “look like” your favorite patterns – the ones that turn you on. Also: no hidden units => No back propagation of error APA Talk, August 2000: The face of fear: *

36 Results (Generalization)
Gary Cottrell & * 07/16/96 Results (Generalization) Expression Network % Correct Human % Agreement Happiness 100.0% 98.7% Surprise 92.4% Disgust 92.3% Anger 89.2% 88.9% Sadness 82.9% Fear 66.7% 87.7% Average 89.9% 91.6% Kendall’s tau (rank order correlation): .667, p=.0441 Note: This is an emergent property of the model! APA Talk, August 2000: The face of fear: *

37 Correlation of Net/Human Errors
Gary Cottrell & * 07/16/96 Correlation of Net/Human Errors Like all good Cognitive Scientists, we like our models to make the same mistakes people do! Networks and humans have a 6x6 confusion matrix for the stimulus set. This suggests looking at the off-diagonal terms: The errors Correlation of off-diagonal terms: r = [F (1,28) = 13.3; p = ] Again, this correlation is an emergent property of the model: It was not told which expressions were confusing. APA Talk, August 2000: The face of fear: *

38 Outline An overview of our facial expression recognition system.
Gary Cottrell & * 07/16/96 Outline An overview of our facial expression recognition system. The internal representation shows the model’s prototypical representations of Fear, Sadness, etc. How our model accounts for the “categorical” data How our model accounts for the “two-dimensional” data Discussion Conclusions for part 1 APA Talk, August 2000: The face of fear: *

39 Examining the Net’s Representations
Gary Cottrell & * 07/16/96 Examining the Net’s Representations We want to visualize “receptive fields” in the network. But the Gabor magnitude representation is noninvertible. We can learn an approximate inverse mapping, however. We used linear regression to find the best linear combination of Gabor magnitude principal components for each image pixel. Then projecting each unit’s weight vector into image space with the same mapping visualizes its “receptive field.” APA Talk, August 2000: The face of fear: *

40 Examining the Net’s Representations
Gary Cottrell & * 07/16/96 Examining the Net’s Representations The “y-intercept” coefficient for each pixel is simply the average pixel value at that location over all faces, so subtracting the resulting “average face” shows more precisely what the units attend to: Apparently local features appear in the global templates. APA Talk, August 2000: The face of fear: *

41 Outline An overview of our facial expression recognition system.
Gary Cottrell & * 07/16/96 Outline An overview of our facial expression recognition system. The internal representation shows the model’s prototypical representations of Fear, Sadness, etc. How our model accounts for the “categorical” data How our model accounts for the “two-dimensional” data Discussion Conclusions for part 1 APA Talk, August 2000: The face of fear: *

42 Morph Transition Perception
Gary Cottrell & * 07/16/96 Morph Transition Perception Morphs help psychologists study categorization behavior in humans Example: JJ Fear to Sadness morph: 0% % % % % % % Young et al. (1997) Megamix: presented images from morphs of all 6 emotions (15 sequences) to subjects in random order, task is 6-way forced choice button push. APA Talk, August 2000: The face of fear: *

43 Results: classical Categorical Perception: sharp boundaries…
Gary Cottrell & * 07/16/96 Results: classical Categorical Perception: sharp boundaries… iness Fear Sadness Disgust Anger Happ- Surprise 6-WAY ALTERNATIVE FORCED CHOICE PERCENT CORRECT DISCRIMINATION …and higher discrimination of pairs of images when they cross a perceived category boundary APA Talk, August 2000: The face of fear: *

44 Results: Non-categorical RT’s
Gary Cottrell & * 07/16/96 Results: Non-categorical RT’s “Scalloped” Reaction Times iness Fear Sadness Disgust Anger Happ- Surprise BUTTON PUSH REACTION TIME APA Talk, August 2000: The face of fear: *

45 Results: More non-categorical effects
Gary Cottrell & * 07/16/96 Results: More non-categorical effects Young et al. Also had subjects rate 1st, 2nd, and 3rd most apparent emotion. At the 70/30 morph level, subjects were above chance at detecting mixed-in emotion. These data seem more consistent with continuous theories of emotion. APA Talk, August 2000: The face of fear: *

46 Modeling Megamix 1 trained neural network = 1 human subject.
Gary Cottrell & * 07/16/96 Modeling Megamix 1 trained neural network = 1 human subject. 50 networks, 7 random examples of each expression for training, remainder for holdout. Identification = average of network outputs Response time = uncertainty of maximal output (1.0 - ymax). Mixed-in expression detection: record 1st, 2nd, 3rd largest outputs. Discrimination: 1 – correlation of layer representations We can then find the layer that best accounts for the data APA Talk, August 2000: The face of fear: *

47 Modeling Six-Way Forced Choice
Gary Cottrell & * Modeling Six-Way Forced Choice 07/16/96 iness Fear Sadness Disgust Anger Happ- Surprise Overall correlation r=.9416, with NO FIT PARAMETERS! APA Talk, August 2000: The face of fear: *

48 Model Discrimination Scores
Gary Cottrell & * 07/16/96 Model Discrimination Scores iness Fear Sadness Disgust Anger Happ- Surprise HUMAN MODEL OUTPUT LAYER R=0.36 MODEL OBJECT LAYER R=0.61 PERCENT CORRECT DISCRIMINATION The model fits the data best at a precategorical layer: The layer we call the “object” layer; NOT at the category level APA Talk, August 2000: The face of fear: *

49 Gary Cottrell & * 07/16/96 Discrimination Classically, one requirement for “categorical perception” is higher discrimination of two stimuli at a fixed distance apart when those two stimuli cross a category boundary Indeed, Young et al. found in two kinds of tests that discrimination was highest at category boundaries. The result that we fit the data best at a layer before any categorization occurs is significant: In some sense, the category boundaries are “in the data,” or at least, in our representation of the data. APA Talk, August 2000: The face of fear: *

50 Outline An overview of our facial expression recognition system.
Gary Cottrell & * 07/16/96 Outline An overview of our facial expression recognition system. The internal representation shows the model’s prototypical representations of Fear, Sadness, etc. How our model accounts for the “categorical” data How our model accounts for the “non-categorical” data Discussion Conclusions for part 1 APA Talk, August 2000: The face of fear: *

51 Reaction Time: Human/Model
Gary Cottrell & * 07/16/96 Reaction Time: Human/Model iness Fear Sadness Disgust Anger Happ- Surprise MODEL REACTION TIME (1 - MAX_OUTPUT) HUMAN SUBJECTS REACTION TIME Correlation between model & data: .6771, p<.001 APA Talk, August 2000: The face of fear: *

52 Mix Detection in the Model
Gary Cottrell & * 07/16/96 Mix Detection in the Model Can the network account for the continuous data as well as the categorical data? YES. APA Talk, August 2000: The face of fear: *

53 Human/Model Circumplexes
Gary Cottrell & * Human/Model Circumplexes 07/16/96 These are derived from similarities between images using non-metric Multi-dimensional scaling. For humans: similarity is correlation between 6-way forced-choice button push. For networks: similarity is correlation between 6-category output vectors. APA Talk, August 2000: The face of fear: *

54 Outline An overview of our facial expression recognition system.
Gary Cottrell & * 07/16/96 Outline An overview of our facial expression recognition system. How our model accounts for the “categorical” data How our model accounts for the “two-dimensional” data The internal representation shows the model’s prototypical representations of Fear, Sadness, etc. Discussion Conclusions for part 1 APA Talk, August 2000: The face of fear: *

55 Discussion Our model of facial expression recognition:
Gary Cottrell & * 07/16/96 Discussion Our model of facial expression recognition: Performs the same task people do On the same stimuli At about the same accuracy Without actually “feeling” anything, without any access to the surrounding culture, it nevertheless: Organizes the faces in the same order around the circumplex Correlates very highly with human responses. Has about the same rank order difficulty in classifying the emotions APA Talk, August 2000: The face of fear: *

56 Gary Cottrell & * 07/16/96 Discussion The discrimination correlates with human results most accurately at a precategorization layer: The discrimination improvement at category boundaries is in the representation of data, not based on the categories. These results suggest that for expression recognition, the notion of “categorical perception” simply is not necessary to explain the data. Indeed, most of the data can be explained by the interaction between the similarity of the representations and the categories imposed on the data: Fear faces are similar to surprise faces in our representation – so they are near each other in the circumplex. APA Talk, August 2000: The face of fear: *

57 Conclusions from this part of the talk
Gary Cottrell & * 07/16/96 Conclusions from this part of the talk The best models perform the same task people do Concepts such as “similarity” and “categorization” need to be understood in terms of models that do these tasks Our model simultaneously fits data supporting both categorical and continuous theories of emotion The fits, we believe, are due to the interaction of the way the categories slice up the space of facial expressions, and the way facial expressions inherently resemble one another. It also suggests that the continuous theories are correct: “discrete categories” are not required to explain the data. We believe our results will easily generalize to other visual tasks, and other modalities. This is contrary to some folks’ belief that evidence for sharp category boundaries, scalloped response time curves, and good discrimination at boundaries implies categorical perception of stimuli. APA Talk, August 2000: The face of fear: *

58 Outline for the next two parts
Gary Cottrell & * 07/16/96 Outline for the next two parts Why would a face area process BMW’s?: Behavioral & brain data Model of expertise results Why is this model wrong, and what can we do to fix it? APA Talk, August 2000: The face of fear: *

59 Are you a perceptual expert?
Gary Cottrell & * 07/16/96 Are you a perceptual expert? Take the expertise test!!!** “Identify this object with the first name that comes to mind.” **Courtesy of Jim Tanaka, University of Victoria APA Talk, August 2000: The face of fear: *

60 “2002 BMW Series 7” - Expert! “Car” - Not an expert Gary Cottrell & *
07/16/96 “Car” - Not an expert “2002 BMW Series 7” - Expert! APA Talk, August 2000: The face of fear: *

61 “Indigo Bunting” - Expert!
Gary Cottrell & * 07/16/96 “Bird” or “Blue Bird” - Not an expert “Indigo Bunting” - Expert! APA Talk, August 2000: The face of fear: *

62 “Jerk” or “Megalomaniac” - Democrat!
Gary Cottrell & * 07/16/96 “Face” or “Man” - Not an expert “George Dubya”- Expert! “Jerk” or “Megalomaniac” - Democrat! APA Talk, August 2000: The face of fear: *

63 How is an object to be named?
Gary Cottrell & * 07/16/96 How is an object to be named? Animal Superordinate Level Bird Basic Level (Rosch et al., 1971) Indigo Bunting Subordinate Species Level APA Talk, August 2000: The face of fear: *

64 Entry Point Recognition
Gary Cottrell & * 07/16/96 Entry Point Recognition Animal Semantic analysis Entry Point Bird Visual analysis Entry Point Downward Shift Hypothesis Indigo Bunting Fine grain visual analysis APA Talk, August 2000: The face of fear: *

65 Dog and Bird Expert Study
Gary Cottrell & * 07/16/96 Dog and Bird Expert Study • Each expert had a least 10 years experience in their respective domain of expertise. • None of the participants were experts in both dogs and birds. • Participants provided their own controls. Tanaka & Taylor, 1991 APA Talk, August 2000: The face of fear: *

66 Object Verification Task
Gary Cottrell & * 07/16/96 Object Verification Task Robin Bird Animal Plant Dog Sparrow Superordinate Basic Subordinate YES NO YES NO APA Talk, August 2000: The face of fear: *

67 Gary Cottrell & * 07/16/96 Dog and bird experts recognize objects in their domain of expertise at subordinate levels. 900 Novice Domain Downward Shift Expert Domain 800 Mean Reaction Time (msec) 700 600 Superordinate Basic Subordinate Animal Bird/Dog Robin/Beagle APA Talk, August 2000: The face of fear: *

68 Is face recognition a general form of perceptual expertise?
Gary Cottrell & * 07/16/96 Is face recognition a general form of perceptual expertise? George W. Bush Indigo Bunting 2002 Series 7 BMW APA Talk, August 2000: The face of fear: *

69 Gary Cottrell & * 07/16/96 Face experts recognize faces at the individual level of unique identity 1200 Faces Objects 1000 Mean Reaction Time (msec) Downward Shift 800 600 Superordinate Basic Subordinate Tanaka, 2001 APA Talk, August 2000: The face of fear: *

70 Event-related Potentials and Expertise
Gary Cottrell & * 07/16/96 Event-related Potentials and Expertise Face Experts Object Experts N170 Tanaka & Curran, 2001; see also Gauthier, Curran, Curby & Collins, 2003, Nature Neuro. Bentin, Allison, Puce, Perez & McCarthy, 1996 Novice Domain Expert Domain APA Talk, August 2000: The face of fear: *

71 Neuroimaging of face, bird and car experts
Gary Cottrell & * 07/16/96 Neuroimaging of face, bird and car experts Cars-Objects Birds-Objects Fusiform Gyrus Car Experts Fusiform Gyrus “Face Experts” Bird Experts Gauthier et al., 2000 Fusiform Gyrus APA Talk, August 2000: The face of fear: *

72 How to identify an expert?
Gary Cottrell & * 07/16/96 How to identify an expert? Behavioral benchmarks of expertise • Downward shift in entry point recognition • Improved discrimination of novel exemplars from learned and related categories Neurological benchmarks of expertise • Enhancement of N170 ERP brain component • Increased activation of fusiform gyrus APA Talk, August 2000: The face of fear: *

73 End of Tanaka Slides Gary Cottrell & * 07/16/96
APA Talk, August 2000: The face of fear: *

74 Kanwisher showed the FFA is specialized for faces
Gary Cottrell & * 07/16/96 Kanwisher showed the FFA is specialized for faces But she forgot to control for what??? APA Talk, August 2000: The face of fear: *

75 Greeble Experts (Gauthier et al. 1999)
Gary Cottrell & * 07/16/96 Greeble Experts (Gauthier et al. 1999) Subjects trained over many hours to recognize individual Greebles. Activation of the FFA increased for Greebles as the training proceeded. APA Talk, August 2000: The face of fear: *

76 The “visual expertise mystery”
Gary Cottrell & * 07/16/96 The “visual expertise mystery” If the so-called “Fusiform Face Area” (FFA) is specialized for face processing, then why would it also be used for cars, birds, dogs, or Greebles? Our view: the FFA is an area associated with a process: fine level discrimination of homogeneous categories. But the question remains: why would an area that presumably starts as a face area get recruited for these other visual tasks? Surely, they don’t share features, do they? Read the slide Sugimoto & Cottrell (2001), Proceedings of the Cognitive Science Society APA Talk, August 2000: The face of fear: *

77 Solving the mystery with models
Gary Cottrell & * 07/16/96 Solving the mystery with models Main idea: There are multiple visual areas that could compete to be the Greeble expert - “basic” level areas and the “expert” (FFA) area. The expert area must use features that distinguish similar looking inputs -- that’s what makes it an expert Perhaps these features will be useful for other fine-level discrimination tasks. We will create Basic level models - trained to identify an object’s class Expert level models - trained to identify individual objects. Then we will put them in a race to become Greeble experts. Then we can deconstruct the winner to see why they won. The main idea here is that there are multiple visual areas. Basic level areas and the FFA. The FFA has a harder task -- fine level discrimination within a homogeneous class, so it must have features that distinguish similar inputs. Perhaps these features are useful for other tasks like this. So, to test this idea, we created two kinds of models -- basic level categorizers and experts. Then we race them to see who can distinguish greebles first. Sugimoto & Cottrell (2001), Proceedings of the Cognitive Science Society APA Talk, August 2000: The face of fear: *

78 Model Database A network that can differentiate faces, books, cups and
Gary Cottrell & * 07/16/96 Model Database We start with the cottrell and metcalfe face images,. The dailey and cottrel object images, and greebles. One kind of network is trained to say Cup can book face. Another network is trained to say cup can book bob carol ted and alice. A network that can differentiate faces, books, cups and cans is a “basic level network.” A network that can also differentiate individuals within ONE class (faces, cups, cans OR books) is an “expert.” APA Talk, August 2000: The face of fear: *

79 Gary Cottrell & * 07/16/96 Model can Bob Ted cup (Experts) book Pretrain two groups of neural networks on different tasks. Compare the abilities to learn a new individual Greeble classification task. Carol Greeble1 Greeble2 Greeble3 can cup book face (Non-experts) Hidden layer APA Talk, August 2000: The face of fear: *

80 Expertise begets expertise
Gary Cottrell & * Expertise begets expertise 07/16/96 Amount Of Training Required To be a Greeble Expert This is a graph of the results. The X-axis represents the amount of training on the first task - either basic level processing, (cup can book face) or expert level processing. We trained four kinds of experts - a face expert, book expert, can expert and cup experts. The Y axis shows the amount of training required to become a greeble expert. Notice that the more the pretraining, the faster the learning of the greeble task. This is somewhat strange -- usually neural networks show overtraining effects -- here, we are getting positive transfer. So, if these nets were competing to solve the greeble task the FFA would win. Or, if our parents were cans, the FCA would win. Training Time on first task Learning to individuate cups, cans, books, or faces first, leads to faster learning of Greebles (can’t try this with kids!!!). The more expertise, the faster the learning of the new task! Hence in a competition with the object area, FFA would win. If our parents were cans, the FCA (Fusiform Can Area) would win. APA Talk, August 2000: The face of fear: *

81 Gary Cottrell & * 07/16/96 Entry Level Shift: Subordinate RT decreases with training (rt = uncertainty of response = 1.0 -max(output)) Human data Network data --- Subordinate Basic These graphs show the downward entry level shift in reaction times for the humans and the face network we get the shift in our networks. RT # Training Sessions APA Talk, August 2000: The face of fear: *

82 How do experts learn the task?
Gary Cottrell & * 07/16/96 How do experts learn the task? Expert level networks must be sensitive to within-class variation: Representations must amplify small differences Basic level networks must ignore within-class variation. Representations should reduce differences So, the mystery remains, why is it better to be an expert? We can now deconstruct the networks to see why. The main idea is, An expert’s job is to be sensitive to variation WITHIN a class -- it has to respond differently to all of the faces. This variability may generalize to new inputs. So we looked at the variance of the hidden layer responses to stimuli/. APA Talk, August 2000: The face of fear: *

83 Observing hidden layer representations
Gary Cottrell & * 07/16/96 Observing hidden layer representations Principal Components Analysis on hidden unit activation: PCA of hidden unit activations allows us to reduce the dimensionality (to 2) and plot representations. We can then observe how tightly clustered stimuli are in a low-dimensional subspace We expect basic level networks to separate classes, but not individuals. We expect expert networks to separate classes and individuals. APA Talk, August 2000: The face of fear: *

84 Gary Cottrell & * 07/16/96 Subordinate level training magnifies small differences within object representations 1 epoch 80 epochs 1280 epochs Face greeble Basic APA Talk, August 2000: The face of fear: *

85 Greeble representations are spread out prior to Greeble Training
Gary Cottrell & * 07/16/96 Greeble representations are spread out prior to Greeble Training greeble Face Basic APA Talk, August 2000: The face of fear: *

86 Variability Decreases Learning Time
Gary Cottrell & * 07/16/96 Variability Decreases Learning Time (r = ) Greeble Learning Time Greeble Variance Prior to Learning Greebles APA Talk, August 2000: The face of fear: *

87 Examining the Net’s Representations
Gary Cottrell & * 07/16/96 Examining the Net’s Representations We want to visualize “receptive fields” in the network. But the Gabor magnitude representation is noninvertible. We can learn an approximate inverse mapping, however. We used linear regression to find the best linear combination of Gabor magnitude principal components for each image pixel. Then projecting each hidden unit’s weight vector into image space with the same mapping visualizes its “receptive field.” APA Talk, August 2000: The face of fear: *

88 Two hidden unit receptive fields
Gary Cottrell & * 07/16/96 Two hidden unit receptive fields AFTER TRAINING AS A FACE EXPERT AFTER FURTHER TRAINING ON GREEBLES HU 16 HU 36 NOTE: These are not face-specific! APA Talk, August 2000: The face of fear: *

89 Controlling for the number of classes
Gary Cottrell & * 07/16/96 Controlling for the number of classes We obtained 13 classes from hemera.com: 10 of these are learned at the basic level. 10 faces, each with 8 expressions, make the expert task 3 (lamps, ships, swords) are used for the novel expertise task. APA Talk, August 2000: The face of fear: *

90 Results: Pre-training
Gary Cottrell & * 07/16/96 Results: Pre-training New initial tasks of similar difficulty: In previous work, the basic level task was much easier. These are the learning curves for the 10 object classes and the 10 faces. APA Talk, August 2000: The face of fear: *

91 Number of training epochs on faces or objects
Gary Cottrell & * 07/16/96 Results As before, experts still learned new expert level tasks faster Number of epochs To learn swords After learning faces Or objects Number of training epochs on faces or objects APA Talk, August 2000: The face of fear: *

92 Outline Why would a face area process BMW’s? Why this model is wrong…
Gary Cottrell & * 07/16/96 Outline Why would a face area process BMW’s? Why this model is wrong… APA Talk, August 2000: The face of fear: *

93 Gary Cottrell & * 07/16/96 background I have a model of face and object processing that accounts for a lot of data… APA Talk, August 2000: The face of fear: *

94 The Face and Object Processing System
Gary Cottrell & * 07/16/96 The Face and Object Processing System PCA Gabor Filtering Book Can Cup Face Bob Sue Jane Expert Level Classifier Pixel (Retina) Level Gestalt Perceptual (V1) Category Hidden Layer Basic FFA LOC APA Talk, August 2000: The face of fear: *

95 Effects accounted for by this model
Gary Cottrell & * 07/16/96 Effects accounted for by this model Categorical effects in facial expression recognition (Dailey et al., 2002) Similarity effects in facial expression recognition (ibid) Why fear is “hard” (ibid) Rank of difficulty in configural, featural, and inverted face discrimination (Zhang & Cottrell, 2006) Holistic processing of identity and expression, and the lack of interaction between the two (Cottrell et al. 2000) How a specialized face processor may develop under realistic developmental constraints (Dailey & Cottrell, 1999) How the FFA could be recruited for other areas of expertise (Sugimoto & Cottrell, 2001; Joyce & Cottrell, 2004, Tran et al., 2004; Tong et al, 2005) The other race advantage in visual search (Haque & Cottrell, 2005) Generalization gradients in expertise (Nguyen & Cottrell, 2005) Priming effects (face or character) on discrimination of ambiguous chinese characters (McCleery et al, under review) Age of Acquisition Effects in face recognition (Lake & Cottrell, 2005; L&C, under review) Memory for faces (Dailey et al., 1998, 1999) Cultural effects in facial expression recognition (Dailey et al., in preparation) APA Talk, August 2000: The face of fear: *

96 Gary Cottrell & * 07/16/96 background I have a model of face and object processing that accounts for a lot of data… But it is fundamentally wrong in an important way! APA Talk, August 2000: The face of fear: *

97 Gary Cottrell & * 07/16/96 background I have a model of face and object processing that accounts for a lot of data… But it is fundamentally wrong in an important way! It assumes that the whole face is processed to the same level of detail everywhere…. APA Talk, August 2000: The face of fear: *

98 Gary Cottrell & * 07/16/96 background I have a model of face and object processing that accounts for a lot of data… But it is fundamentally wrong in an important way! It assumes that the whole face is processed to the same level of detail everywhere….and that the face is always aligned with “the brain” APA Talk, August 2000: The face of fear: *

99 Gary Cottrell & * 07/16/96 background Instead, people and animals actively sample the world around them with eye movements. The images our eyes take in “slosh around” on the retina like an MTV video. We only get patches of information from any fixation But somehow (see previous talk), we perceive a stable visual world. APA Talk, August 2000: The face of fear: *

100 Gary Cottrell & * 07/16/96 background So, a model of visual processing needs to be dynamically integrating information over time. We need to decide where to look at any moment: There are both bottom up (inherently interesting things ) and top down influences on this (task demands) APA Talk, August 2000: The face of fear: *

101 Gary Cottrell & * 07/16/96 Questions for today What form should a model take in order to model eye movements without getting into muscle control? Do people actually gain information from saccades in face processing? How is the information integrated? How do features change over time? APA Talk, August 2000: The face of fear: *

102 Gary Cottrell & * 07/16/96 answers What form should a model take in order to model eye movements without getting into muscle control? Probabilistic Do people actually gain information from saccades? Yes How is the information integrated? In a Bayesian way How do features change over time? Don’t know yet, but we hypothesize that the effective receptive fields grow over time APA Talk, August 2000: The face of fear: *

103 This (brief!) part of the talk
Gary Cottrell & * 07/16/96 This (brief!) part of the talk What form should a model take in order to model eye movements without getting into muscle control? Probabilistic Do people actually gain information from saccades?  Yes How is the information integrated?  In a Bayesian way: NIMBLE How do features change over time? Don’t know yet, but we hypothesize that the effective receptive fields grow over time APA Talk, August 2000: The face of fear: *

104 Gary Cottrell & * 07/16/96 Where to look? We have lots of ideas about this (but, obviously, none as good as Baldi’s surprise!), but for the purposes of this talk, we will just use bottom up salience: Figure out where the image is “interesting,” and look there “interesting” = where there are areas of high contour APA Talk, August 2000: The face of fear: *

105 Interest points created using Gabor filter variance
Gary Cottrell & * 07/16/96 Interest points created using Gabor filter variance APA Talk, August 2000: The face of fear: *

106 Given those fragments, how do we recognize people?
Gary Cottrell & * 07/16/96 Given those fragments, how do we recognize people? The basic idea is that we simply store the fragments with labels: Bob, Carol, Ted or Alice In the study set, not in the study set (Note that we have cast face recognition as a categorization problem.) When we look at a new face, we look at the labels of the most similar fragments, and vote, using kernel density estimation, comparing a Gaussian kernel and 1-nearest neighbor. This can be done easily in a (Naïve) Bayesian framework APA Talk, August 2000: The face of fear: *

107 A basic test: Face Recognition
Gary Cottrell & * 07/16/96 A basic test: Face Recognition Following (Duchaine & Nakayama, 2005): Images are displayed for 3 seconds -- which means about 10 fixations - which means about 10 fragments are stored, chosen at random using bottom-up salience 30 lures Normal human ROC area (A’) is from 0.9 to 1.0 NIMBLE results: APA Talk, August 2000: The face of fear: *

108 Multi-Class Memory Tasks
Gary Cottrell & * 07/16/96 Multi-Class Memory Tasks An identification task tests NIMBLE on novel images from multiple classes Face identification: 29 identities from the FERET dataset Changes in expression and illumination Object identification: 20 object classes from the COIL-100 dataset Changes in orientation, scale and illumination APA Talk, August 2000: The face of fear: *

109 Identification Results
Gary Cottrell & * 07/16/96 Identification Results [Belongie, Malik & Puzicha, 2002] report 97.6% accuracy on a 20-class, 4 training image set from COIL-100 APA Talk, August 2000: The face of fear: *

110 Do people really need more than one fixation to recognize a face?
Gary Cottrell & * 07/16/96 Do people really need more than one fixation to recognize a face? To answer this question, we ran a simple face recognition experiment: Subjects saw 32 faces, then were given an old/new recognition task with those 32 faces and 32 more. However, subjects were restricted to 1, 2, 3, or unlimited fixations on the face. APA Talk, August 2000: The face of fear: *

111 Gary Cottrell & * 07/16/96 Using a gaze-contingent display, we can limit the fixations on the face itself to 1, 2, 3, or unlimited fixations The average face mask prevents subjects from acquiring information about the face from peripheral vision APA Talk, August 2000: The face of fear: *

112 Fixations during training
Gary Cottrell & * 07/16/96 Fixations during training APA Talk, August 2000: The face of fear: *

113 Fixations during testing
Gary Cottrell & * 07/16/96 Fixations during testing APA Talk, August 2000: The face of fear: *

114 Overall Data Overall fixation maps during training
Gary Cottrell & * 07/16/96 Overall Data Overall fixation maps during training 1st Fixation 2nd Fixation 3rd Fixation 4+ Fixations Overall APA Talk, August 2000: The face of fear: *

115 Overall Data Overall fixation maps during testing
Gary Cottrell & * 07/16/96 Overall Data Overall fixation maps during testing 1st Fixation 2nd Fixation 3rd Fixation 4+ Fixations Overall APA Talk, August 2000: The face of fear: *

116 Gary Cottrell & * 07/16/96 Results This says you gain the most recognition by making two fixations, and a third doesn’t help. Hence. if a picture is worth a thousand words, each fixation is worth about 500 words! APA Talk, August 2000: The face of fear: *

117 Gary Cottrell & * Long term questions 07/16/96 Do people have visual routines for sampling from stimuli? Do people have different routines for different tasks on identical stimuli? Do the routines change depending on the information gained at any particular location? Do different people have different routines? How do those routines change with expertise? How do they change over the acquisition of expertise? APA Talk, August 2000: The face of fear: *

118 Gary Cottrell & * 07/16/96 END APA Talk, August 2000: The face of fear: *


Download ppt "What can computational models tell us about face processing?"

Similar presentations


Ads by Google