Let’s shake hands! On the coordination of gestures of humanoids Zsofi Ruttkay Herwin van Welbergen Balázs Varga.

Slides:



Advertisements
Similar presentations
Expressive Gestures for NAO NAO TechDay, 13/06/2012, Paris Le Quoc Anh - Catherine Pelachaud CNRS, LTCI, Telecom-ParisTech, France.
Advertisements

Virtual Reality Design Virtual reality systems are designed to produce in the participant the cognitive effects of feeling immersed in the environment.
ENTERFACE’08 Multimodal Communication with Robots and Virtual Agents.
Spike Train Statistics Sabri IPM. Review of spike train  Extracting information from spike trains  Noisy environment:  in vitro  in vivo  measurement.
Herwin van Welbergen Dennis Reidsma Stefan Kopp.  Beyond turn taking interaction ◦ Continuous perception and behavior generation  Interpersonal coordination.
A cross-linguistic comparison of the coordination between hand gestures and phonological prominence Giorgos Tserdanelis.
 Public speaking is speaking to a group of people in a structured, deliberate manner intended to inform, influence, or entertain the listeners.
ENTERFACE’08 Multimodal high-level data integration Project 2 1.
On the parameterization of clapping Herwin van Welbergen Zsófia Ruttkay Human Media Interaction, University of Twente.
Towards a Reactive Virtual Trainer Zsófia Ruttkay, Job Zwiers, Herwin van Welbergen, Dennis Reidsma HMI, Dept. of CS, University of Twente Amsterdam, The.
Unsupervised Clustering in Multimodal Multiparty Meeting Analysis.
1 IUT de Montreuil Université Paris 8 Emotion in Interaction: Embodied Conversational Agents Catherine Pelachaud.
Understanding Non- Verbal Communication MRS. DOBBINS.
Motor Control Theories
Building the Design Studio of the Future Aaron Adler Jacob Eisenstein Michael Oltmans Lisa Guttentag Randall Davis October 23, 2004.
Nonverbal Communication Voice Body Talk Environmental Cues.
Perspectives on Walking in an Environment Işık Barış Fidaner BM 526 Project.
Susan Bock Solutions 2007 Communication What is Communication? Communication Styles Communication Spectrum.
 Before a BCI can be used for control purposes, several training sessions are necessary ◦ Operant conditioning  Feed back, real-time changes to the.
Ch 14.  Rhythmic Movement: mvmt in time to sound  Dance: moving rhythmically usually to music during prescribed steps & gestures  Typically done in.
PED 246: Methods and Techniques of Dance and Gymnastics.
Managing and Teaching the Physical Education Lesson Chapter 7.
Rhythm is easy to feel but quite hard to define, so let’s start thinking about it.
Presentation Construct and deliver a crafted and controlled oral text AS Credits.
Recognition of meeting actions using information obtained from different modalities Natasa Jovanovic TKI University of Twente.
Expressive Emotional ECA ✔ Catherine Pelachaud ✔ Christopher Peters ✔ Maurizio Mancini.
Chapter 7. BEAT: the Behavior Expression Animation Toolkit
AsapRealizer 2.0: The Next Steps in Fluent Behavior Realization for ECAs Herwin van Welbergen, Ramin Yaghoubzadeh, Stefan Kopp Social Cognitive Systems.
Affective Interfaces Present and Future Challenges Introductory statement by Antonio Camurri (Univ of Genoa) Marc Leman (Univ of Gent) MEGA IST Multisensory.
APML, a Markup Language for Believable Behavior Generation Soft computing Laboratory Yonsei University October 25, 2004.
Characteristics of Poetry. Sensory appeal is words, phrases, or images that appeal to your senses. Interpretation of poetry is to make sense, or assign.
Lesson Planning SIOP.
Speech Perception 4/4/00.
Backchannels Through Gaze as Indicators of Persuasive Success E. Bevacqua, M. Mancini, C. Peters, C. Pelachaud University of Paris 8 Isabella Poggi Università.
Human Brain and Behavior Laborator y. Mathematics Chemistry Biomedical science Psychology Physics Biology Engineering Philosophy Complexity.
ENTERFACE 08 Project 1 “MultiParty Communication with a Tour Guide ECA” Mid-term presentation August 19th, 2008.
Conformance Test Experiments for Distributed Real-Time Systems Rachel Cardell-Oliver Complex Systems Group Department of Computer Science & Software Engineering.
Toward a Unified Scripting Language 1 Toward a Unified Scripting Language : Lessons Learned from Developing CML and AML Soft computing Laboratory Yonsei.
Perceptual Analysis of Talking Avatar Head Movements: A Quantitative Perspective Xiaohan Ma, Binh H. Le, and Zhigang Deng Department of Computer Science.
1 1. Representing and Parameterizing Agent Behaviors Jan Allbeck and Norm Badler 연세대학교 컴퓨터과학과 로봇 공학 특강 학기 유 지 오.
Österreichisches Forschnungsinstitut für Artificial Intelligence Representational Lego for ECAs Brigitte Krenn.
Feedback Elisabetta Bevacqua, Dirk Heylen,, Catherine Pelachaud, Isabella Poggi, Marc Schröder.
How (Not) to Give a Good Talk Steffen Koch, Daniel Maurer, Michael Stoll, Sebastian Volz, Andrés Bruhn with contributions by Michael Raschke and others.
1 Article 11 FLOOR General The evaluation of the exercise begins with the first movement of the gymnast. The duration of the floor exercise may.
Aiming Computing Technology at Enhancing the Quality of Life of People with ALS Some Sketches on Directions in Minimal Signaling Communication Communication.
SELF-MOTION PERCEPTION: ASSESSMENT BY REAL-TIME COMPUTER-GENERATED ANIMATIONS OVERALL GOAL: PERCEPTUAL TESTS FOR VESTIBULAR FUNCTION A NEW PROCEDURE FOR.
Review of Parnas’ Criteria for Decomposing Systems into Modules Zheng Wang, Yuan Zhang Michigan State University 04/19/2002.
ACE TESOL Diploma Program – London Language Institute OBJECTIVES You will understand: 1. Accent reduction techniques to use with your students You will.
Ta-when-tee-won. Voice Paralanguage – vocal qualities “assist” language Disfluencies – disruptions in the flow of words – Verbal Junk (um, uh, like, and.
It Starts with iGaze: Visual Attention Driven Networking with Smart Glasses It Starts with iGaze: Visual Attention Driven Networking with Smart Glasses.
Multimodal Plan Representation for Adaptable BML Scheduling Dennis Reidsma, Herwin van Welbergen, Job Zwiers.
5 Event Handling Interactive Programming Suggested Reading Interaction: Events and Event Handling, Supplemental Text for CPSC 203 Distributed this term.
Elements of Music By: Montana Miracle. Pitch  The highness or lowness of a tone.  The position of a note determines the element of music.  It may be.
ENTERFACE’08 Multimodal Communication with Robots and Virtual Agents mid-term presentation.
Chapter 8 Teaching Dance and Rhythms. Place for Dance in Elementary Schools Dance is part of the physical education curriculum. Dance can be a part of.
WP6 Emotion in Interaction Embodied Conversational Agents WP6 core task: describe an interactive ECA system with capabilities beyond those of present day.
Human Joint Transportation in a Multi-User Virtual Environment Stephan Streuber Astros.
Writing Lesson Plans KIN 229. Lesson Plans Refer to Handout on Lesson Plan Format Designing written lesson plans In each written lesson, include information.
Presented By Meet Shah. Goal  Automatically predicting the respondent’s reactions (accept or reject) to offers during face to face negotiation by analyzing.
Guidelines of Business Process Modeling Team: Alejandra Saavedra Andrea Rodriguez Ez Lawrence.
Modeling Expressivity in ECAs
Klara Nahrstedt Spring 2009
On Defining Cephalic Gesture Categories
Grounding by nodding GESPIN 2009, Poznan, Poland
Chapter 12 Speed, Agility, and Quickness
Chapter 4 – Communication Skills
Identifying Confusion from Eye-Tracking Data
Motor Control Theories
CHAPTER 8 The Nonverbal Code.
Presentation transcript:

Let’s shake hands! On the coordination of gestures of humanoids Zsofi Ruttkay Herwin van Welbergen Balázs Varga

Our goals Coordinating gesture to external signals Coordinating gesture to other modalities Comparison of synch phenomena of rhythmic motion and speech- accompanying gesture Define a synchronization language Create an adaptive real-time animation engine

Content Examples of coordination The multimodal coordination problem Existing solutions Our solution Coordinating gesture To what? How to adapt gesture for coordination? How to specify coordination? Conclusions Open issues Questions

Coordination example: Gesture-speech coordination Gestures and speech come from a single process of utterance formation (McNeill) => Gesture timing is not a slave of speech timing Time alignment is achieved while we speak The stroke of gesture precedes or ends at the phonological peak syllable of the speech Often we need to adjust the timing of gestures or speech to make the alignment fit Gestures can be speeded up or slowed down Gestures can be ‘paused’ using hold phases Speech can be stopped to finish complex gestures Connection words (uh…) can be used to slow down speech Etc

Coordination example: The virtual dancer: moving to the music ‘Beat moments’ in the animation should be aligned to beats in the music Annotate all beat moments in the animation Predict the beats in the music Locally speed up or slow down the animation to fit to the music There is a maximum stretch or skew in the dance motion

Coordination example: The virtual trainer: tutoring exercises on music Exercises are executed using several modalities Body movement Speech Music/metronome Sound (clap, foot tap) Challenges Coordination Monitoring user => real time (re)planning Exaggeration to point out details Speed up / slow down Feedback/correction …

Coordination example: Handshake: coordination between 2 humans Handshake is used for greeting, agreeing and accepting Complex coordination between two persons Guided by Social protocols Haptic feedback Visual feedback

Generalizing: the multimodal coordination problem ‘Behaviors’ on different modalities (speech, gesture, dance motion, music) Synchronization between behaviors at key time moments The timing of flexible behaviors can be adapted to achieve synchronization

Coordination: related work Classic approach in speech/gesture coordination: Speech leads, gesture follows MURML (Kopp et al.) No leading modality Planning in sequential chunks containing one piece of speech and one aligned gesture Co-articulation at the border of chunks BML (Kopp, Krenn, Marsella, Marshall, Pelachaud, Pirker, Thórisson, Vilhjalmsson) No leading modality Synchronized alignment points in behavior phases For now, aimed mainly at speech/gesture synchronization In development

Coordination: our previous work Virtual Dancer Synchronization between music (beats) and dance animation Linear time stretching/skewing Virtual Presenter Synchronization between speech, gesture, posture and sheet display Leading modality can change over time GESTYLE markup language with par/seq and wait constructs to define synchronization

Our multimodal synchronization model No leading modality, just align key moments Every phase of a behavior has a preferred length Stretching/skewing/skipping if necessary

Coordination of hand gestures to external signals What do we want to coordinate with? How do synchronization constraints effect movement? How to stretch/skew? How can we define synchronization? using BML scripts?

Ontology of coordination signals Flexibility Origin FixedFlexible World - pointing at a moving object - clapping to rhythm of music - Humanoid’s own modality - gesture aligned to speech which is taken as leading signal - gaze and hand coordination Other humanoid’s modality - back-channeling as listener to a speaker e.g. by head nods - hand shake - two hands involved in taking over an object

Clapping experiment Clapping and counting How is the synchrony between clap and count? How do the movement characteristics of clapping change with tempo? Time distribution Amplitude Left/right hand symmetry

Clapping experiment: setup Mocap analysis of two subjects Instructions: Clap and count from 21 to 31 Clap and count to the metronome

Clapping experiment: results The phonological synchrony rule was valid for counting while clapping: Clap before phonological peak of the count The clapping was speeded up by decreasing the path distance of the hand A pre-stroke hold can be used to slow down For our right-handed subjects, the right hand was moving ahead in phase compared to the left The standard deviation of the relative phase between the left and the right hand increased with the clapping frequency

Hand shake experiment Which movement phases can be identified? How are they coordinated? What gaze patterns can be seen? What movement characteristics can be identified in the different phases? Timing, duration Form How is above effected by Refusal or avoidance to shake hands Social relations between participants

Hand shake experiment: setup Motion capture of two subjects (P1, P2) shaking hands Annotation of gaze patterns Variations Basic Triggered P2 initiates P2 tries to avoid shaking hands P2 rejects

Modeling coordination in BML BML is an XML language defining multimodal synchronization BML events can be used to synchronize with other BML scripts/world events The BML T observer is introduced for synchronization with (repeated) outside world events

Coordination with events BML is designed to work in event driven systems is used to fire an event message is used to wait for an event If the event does not occur after a set time, wait can fire a no-event message After the event occurs, or the timeout is exceeded, the script continues

Coordination with events: handshake extendwithdrawpumpconnect extend withdrawpump connect wait timeout subject 1 subject 2

Coordination with the observer An observer observes a specific part of the world and provides timing information on that Example: beats in music Why observers instead of events? Explicit outside world trigger Multiple (repeated) trigger Timing of observer triggers can be predicted for easy planning Synchronization within behavior phases does not suffice

Coordination with the observer Clapping

Conclusions Gesture synchronization mechanisms are also found in rhythmic motion Adaptation of timing in gesture affects several movement characteristics Linear speedup/slowdown does not suffice Gesture coordination can be modeled using BML T

Open issues What modalities do we have to stretch/skew/skip? Can we generalize our findings from clapping/handshake? Do the semantics of a motion change if we change its timing? E.g. emotions, individual features

Questions

Easter eggs

Synchronization with observer Declaration: Synchronization to beat 1 Synchronization to closest beat/all beats