Human Emotion Synthesis David Oziem, Lisa Gralewski, Neill Campbell, Colin Dalton, David Gibson, Barry Thomas University of Bristol, Motion Ripper, 3CR.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

24/10/02AutoArch Overview Computer Vision and Media Group: Selected Previous Work David Gibson, Neill Campbell Colin Dalton Department of Computer Science.
Active Appearance Models
The Extended Cohn-Kanade Dataset(CK+):A complete dataset for action unit and emotion-specified expression Author:Patrick Lucey, Jeffrey F. Cohn, Takeo.
Automatic determination of skeletal age from hand radiographs of children Image Science Institute Utrecht University C.A.Maas.
Proposed concepts illustrated well on sets of face images extracted from video: Face texture and surface are smooth, constraining them to a manifold Recognition.
Face Recognition Method of OpenCV
3D Face Modeling Michaël De Smet.
Real-Time Human Pose Recognition in Parts from Single Depth Images Presented by: Mohammad A. Gowayyed.
 INTRODUCTION  STEPS OF GESTURE RECOGNITION  TRACKING TECHNOLOGIES  SPEECH WITH GESTURE  APPLICATIONS.
Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots Chao-Yeh Chen and Kristen Grauman University of Texas at Austin.
A 4-WEEK PROJECT IN Active Shape and Appearance Models
Introduction to Data-driven Animation Jinxiang Chai Computer Science and Engineering Texas A&M University.
Exchanging Faces in Images SIGGRAPH ’04 Blanz V., Scherbaum K., Vetter T., Seidel HP. Speaker: Alvin Date: 21 July 2004.
Direct Methods for Visual Scene Reconstruction Paper by Richard Szeliski & Sing Bing Kang Presented by Kristin Branson November 7, 2002.
CONTENT BASED FACE RECOGNITION Ankur Jain 01D05007 Pranshu Sharma Prashant Baronia 01D05005 Swapnil Zarekar 01D05001 Under the guidance of Prof.
Face Recognition using PCA (Eigenfaces) and LDA (Fisherfaces)
Real-time Combined 2D+3D Active Appearance Models Jing Xiao, Simon Baker,Iain Matthew, and Takeo Kanade CVPR 2004 Presented by Pat Chan 23/11/2004.
Feature Detection and Emotion Recognition Chris Matthews Advisor: Prof. Cotter.
ENTERFACE ’10 Amsterdam, July-August 2010 Hamdi Dibeklio ğ lu Ilkka Kosunen Marcos Ortega Albert Ali Salah Petr Zuzánek.
FACE RECOGNITION, EXPERIMENTS WITH RANDOM PROJECTION
Active Appearance Models Computer examples A. Torralba T. F. Cootes, C.J. Taylor, G. J. Edwards M. B. Stegmann.
MVL (Machine Vision Lab) UIC HUMAN MOTION VIDEO DATABASE Jezekiel Ben-Arie ECE Department University Of Illinois at Chicago Scripts, Queries, Recognition.
Gait Recognition Simon Smith Jamie Hutton Thomas Moore David Newman.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
EKMAN’S FACIAL EXPRESSIONS STUDY A Demonstration.
Learning the Appearance of Faces: A Unifying Approach for the Analysis and Synthesis of Images. Thomas Vetter Germany University of Freiburg
Recognizing Emotions in Facial Expressions
PhD Thesis. Biometrics Science studying measurements and statistics of biological data Most relevant application: id. recognition 2.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Database Construction for Speech to Lip-readable Animation Conversion Gyorgy Takacs, Attila Tihanyi, Tamas Bardi, Gergo Feldhoffer, Balint Srancsik Peter.
Facial Feature Detection
1 Mean shift and feature selection ECE 738 course project Zhaozheng Yin Spring 2005 Note: Figures and ideas are copyrighted by original authors.
Graphite 2004 Statistical Synthesis of Facial Expressions for the Portrayal of Emotion Lisa Gralewski Bristol University United Kingdom
General Tensor Discriminant Analysis and Gabor Features for Gait Recognition by D. Tao, X. Li, and J. Maybank, TPAMI 2007 Presented by Iulian Pruteanu.
EE 492 ENGINEERING PROJECT LIP TRACKING Yusuf Ziya Işık & Ashat Turlibayev Yusuf Ziya Işık & Ashat Turlibayev Advisor: Prof. Dr. Bülent Sankur Advisor:
Multimodal Information Analysis for Emotion Recognition
MULTIMEDIA TECHNOLOGY SMM 3001 MEDIA - VIDEO. In this chapter How digital video differs from conventional analog video How digital video differs from.
Face Recognition: An Introduction
CSC505 Particle Systems. CSC505 Object Representations So far we have represented (rendered) objects with –Lines –Polygons (triangles) –Curves These techniques.
CSE 185 Introduction to Computer Vision Face Recognition.
Epitomic Location Recognition A generative approach for location recognition K. Ni, A. Kannan, A. Criminisi and J. Winn In proc. CVPR Anchorage,
Intelligent Control and Automation, WCICA 2008.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Multimodal Interaction Dr. Mike Spann
Performance Comparison of Speaker and Emotion Recognition
AAM based Face Tracking with Temporal Matching and Face Segmentation Mingcai Zhou 1 、 Lin Liang 2 、 Jian Sun 2 、 Yangsheng Wang 1 1 Institute of Automation.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
Point Distribution Models Active Appearance Models Compilation based on: Dhruv Batra ECE CMU Tim Cootes Machester.
Face Recognition Summary –Single pose –Multiple pose –Principal components analysis –Model-based recognition –Neural Networks.
Similarity Measurement and Detection of Video Sequences Chu-Hong HOI Supervisor: Prof. Michael R. LYU Marker: Prof. Yiu Sang MOON 25 April, 2003 Dept.
Ekman’s Facial Expressions Study A Demonstration.
Facial Expressions and Emotions Mental Health. Total Participants Adults (30+ years old)328 Adults (30+ years old) Adolescents (13-19 years old)118 Adolescents.
Face detection and recognition Many slides adapted from K. Grauman and D. Lowe.
Out of sample extension of PCA, Kernel PCA, and MDS WILSON A. FLORERO-SALINAS DAN LI MATH 285, FALL
Interpreting Ambiguous Emotional Expressions Speech Analysis and Interpretation Laboratory ACII 2009.
Under Guidance of Mr. A. S. Jalal Associate Professor Dept. of Computer Engineering and Applications GLA University, Mathura Presented by Dev Drume Agrawal.
Do Expression and Identity Need Separate Representations?
CS 2750: Machine Learning Dimensionality Reduction
Recognition: Face Recognition
Simon Smith Jamie Hutton Thomas Moore David Newman
The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression By: Patrick Lucey, Jeffrey F. Cohn, Takeo.
Categorizing sex and identity from the biological motion of faces
Outline S. C. Zhu, X. Liu, and Y. Wu, “Exploring Texture Ensembles by Efficient Markov Chain Monte Carlo”, IEEE Transactions On Pattern Analysis And Machine.
Turning to the Masters: Motion Capturing Cartoons
Easy Generation of Facial Animation Using Motion Graphs
Outline Announcement Perceptual organization, grouping, and segmentation Hough transform Read Chapter 17 of the textbook File: week14-m.ppt.
EE 492 ENGINEERING PROJECT
Active Appearance Models theory, extensions & cases
Data-Driven Approach to Synthesizing Facial Animation Using Motion Capture Ioannis Fermanis Liu Zhaopeng
Presentation transcript:

Human Emotion Synthesis David Oziem, Lisa Gralewski, Neill Campbell, Colin Dalton, David Gibson, Barry Thomas University of Bristol, Motion Ripper, 3CR Research

Synthesising Facial Emotions – University of Bristol – 3CR Research Project Group Motion Ripper Project –Methods of motion capture. –Re-using captured motion signatures. –Synthesising new or extend motion sequences. –Tools to aid animation. Collaboration between University of Bristol CS, Matrix Media & Granada.

Synthesising Facial Emotions – University of Bristol – 3CR Research Introduction What is an emotion? Ekman outlined 6 different basic emotions. –joy, disgust, surprise, fear, anger and sadness. Emotional states relate to ones expression and movement. Synthesising video footage of an actress expressing different emotions.

Synthesising Facial Emotions – University of Bristol – 3CR Research

Video Textures Video textures or temporal textures are textures with motion. (Szummer’96) Schodl’00, reordered frames from the original to produce loops or continuous sequences. –Doesn’t produce new footage. Campbell’01, Fitzgibbon’01, Reissell’01, used Autoregressive process (ARP) to synthesis frames. Examples of Video Textures

Synthesising Facial Emotions – University of Bristol – 3CR Research Autoregressive Process Statistical model Calculating the model involves working out the parameter vector (a 1 …a n ) and w. n is known as the order of the sequence. y(t) = – a 1 y(t – 1) – a 2 y(t – 2) – … – a n y(t – n) + w.ε Parameter vector (a1,…,an)Noise Current value at time t

Synthesising Facial Emotions – University of Bristol – 3CR Research Autoregressive Process Statistical model Increasing dimensionality of y drastically increases the complexity in calculating (a 1 …a n ). y(t) = – a 1 y(t – 1) – a 2 y(t – 2) – … – a n y(t – n) + w.ε

Synthesising Facial Emotions – University of Bristol – 3CR Research Autoregressive Process PCA analysis of Sad footage in 2D Secondary mode Primary mode Principal Components Analysis is used to reduce number of dimensions in the original sequence.

Synthesising Facial Emotions – University of Bristol – 3CR Research Autoregressive Process PCA analysis of Sad footage in 2DGenerated sequence using an ARP Secondary mode Primary mode Non-Gaussian Distribution is incorrectly modelled by an ARP.

Synthesising Facial Emotions – University of Bristol – 3CR Research Face Modelling Campbell’01, synthesised a talking head. Cootes and Talyor’00, combined appearance model. –Isolates shape and texture. Requires labelled frames. –Must label important features on the face. Labelled points

Synthesising Facial Emotions – University of Bristol – 3CR Research Combined Appearance Shape space Hand Labelled video footage provides a point set which represents the shape space of the clip.

Synthesising Facial Emotions – University of Bristol – 3CR Research Combined Appearance Shape space Texture space Warping each frame into a standard pose, creates the texture space. The standard pose is the mean position of the points.

Synthesising Facial Emotions – University of Bristol – 3CR Research Combined Appearance Shape space Texture space Combined space Joining the shape and texture space and then re-analysing using PCA produces the combined space.

Synthesising Facial Emotions – University of Bristol – 3CR Research Combined Appearance Shape space Texture space Combined space Reconstruction of the original sequence from the combined space. Combined space

Synthesising Facial Emotions – University of Bristol – 3CR Research Secondary mode Primary mode Combined Appearance Combined Appearance sequence Original sequence in 2D Secondary mode Primary mode Change in distribution after applying The combined appearance technique

Synthesising Facial Emotions – University of Bristol – 3CR Research Secondary mode Primary mode Combined Appearance Generated Sequence Original sequence Secondary mode Primary mode ARP model Visually the generated plot appears to have been generated using the same stochastic process as the original.

Synthesising Facial Emotions – University of Bristol – 3CR Research Copying and ARP Combine the benefits of copying with ARP –New motion signatures. –Handles non-Gaussian distributions.

Synthesising Facial Emotions – University of Bristol – 3CR Research Copying and ARP Original input Reduced input PCA Important to reduce the complexity of the search process. Need around 30 to 40 dimensions in this example.

Synthesising Facial Emotions – University of Bristol – 3CR Research Copying and ARP Original input Reduced input Segmented input PCA Reduced segments PCA Temporal segments of between 15 to 30 frames. Need to reduce each segment to be able to train ARP’s.

Synthesising Facial Emotions – University of Bristol – 3CR Research Copying and ARP Original input Reduced input Segmented input Reduced segments PCA ARP Synthesised segments Many of the learned models are unstable % are usable.

Synthesising Facial Emotions – University of Bristol – 3CR Research Copying and ARP Original input Reduced input Segmented input Reduced segments PCA ARP Synthesised segments Segment selection Outputted Sequence

Synthesising Facial Emotions – University of Bristol – 3CR Research Example First mode Time t End of generated sequence. Possible segments. Compared section

Synthesising Facial Emotions – University of Bristol – 3CR Research First mode Time t Example Closest 3 segments are chosen.

Synthesising Facial Emotions – University of Bristol – 3CR Research First mode Time t Example The segment to be copied is randomly selected from the closest 3.

Synthesising Facial Emotions – University of Bristol – 3CR Research First mode Time t Example Segments are blended together using a small overlap and averaging the overlapping pixels.

Synthesising Facial Emotions – University of Bristol – 3CR Research Secondary mode Primary mode Secondary mode Primary mode Copying & ARP model PCA analysis of Sad footage in 2D Generated sequence Copying and ARP Potentially infinitely long. Includes new novel motions.

Synthesising Facial Emotions – University of Bristol – 3CR Research Results (Angry) Source FootageCopying with ARPCombined Appearance ARP Combined appearance produces higher resolution frames. Better motion from the copying and ARP approach

Synthesising Facial Emotions – University of Bristol – 3CR Research Results (Sad) Source FootageCopying with ARPCombined Appearance ARP Similar results as with the angry footage –Copied approach is less blurred due to the reduced variance.

Synthesising Facial Emotions – University of Bristol – 3CR Research Comparison Results - Combined appearance - Segment copying Simple objective comparison. –Randomly selected temporal segments.

Synthesising Facial Emotions – University of Bristol – 3CR Research Comparison Perceptually is it better to have good motion or higher resolution.

Synthesising Facial Emotions – University of Bristol – 3CR Research Combined appearanceSegment Copying with ARP

Synthesising Facial Emotions – University of Bristol – 3CR Research Other potential uses Self Organising Map Uses combined appearance –as each ARP model provides a minimal representation of the given emotion. Can navigate between emotions to create new interstates. Angry Sad Happy

Synthesising Facial Emotions – University of Bristol – 3CR Research Conclusions Both methods can produce synthesised clips of a given emotion. Combined appearance produces higher definition frames. Copying and ARPs generates more natural movements.

Synthesising Facial Emotions – University of Bristol – 3CR Research Questions