Multimodal Human-Computer Interaction New Interaction Techniques 22. 1

Slides:



Advertisements
Similar presentations
PUI 98 Leveraging Human Capabilities in Perceptual Interfaces George G. Robertson Microsoft Research George G. Robertson Microsoft Research.
Advertisements

National Technical University of Athens Department of Electrical and Computer Engineering Image, Video and Multimedia Systems Laboratory
Structured Design The Structured Design Approach (also called Layered Approach) focuses on the conceptual and physical level. As discussed earlier: Conceptual.
Cognitive Systems, ICANN panel, Q1 What is machine intelligence, as beyond pattern matching, classification and prediction. What is machine intelligence,
Breakout session B questions. Research directions/areas Multi-modal perception cognition and interaction Learning, adaptation and imitation Design and.
Empirical and Data-Driven Models of Multimodality Advanced Methods for Multimodal Communication Computational Models of Multimodality Adequate.
A multimodal dialogue-driven interface for accessing the content of recorded meetings Agnes Lisowska ISSCO/TIM/ETI University of Geneva IM2.MDM Work done.
John Hu Nov. 9, 2004 Multimodal Interfaces Oviatt, S. Multimodal interfaces Mankoff, J., Hudson, S.E., & Abowd, G.D. Interaction techniques for ambiguity.
Introduction to HCC and HCM. Human Centered Computing Philosophical-humanistic position regarding the ethics and aesthetics of a workplace Any system.
Stanford hci group / cs376 research topics in human-computer interaction Multimodal Interfaces Scott Klemmer 15 November 2005.
CS335 Principles of Multimedia Systems Multimedia and Human Computer Interfaces Hao Jiang Computer Science Department Boston College Nov. 20, 2007.
0 HCI Today Talk about HCI Success Stories Talk about HCI Success Stories Talk about Norman’s Paper Talk about Norman’s Paper Start talking about The human.
Support for Palm Pilot Collaboration Including Handwriting Recognition.
Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Cognitive Robots © 2014, SNU CSE Biointelligence Lab.,
Mobile Multimodal Applications. Dr. Roman Englert, Gregor Glass March 23 rd, 2006.
1. Human – the end-user of a program – the others in the organization Computer – the machine the program runs on – often split between clients & servers.
Chapter 11: Interaction Styles. Interaction Styles Introduction: Interaction styles are primarily different ways in which a user and computer system can.
Chapter 14: Artificial Intelligence Invitation to Computer Science, C++ Version, Third Edition.
Multimedia Specification Design and Production 2013 / Semester 2 / week 8 Lecturer: Dr. Nikos Gazepidis
Center for Human Computer Communication Department of Computer Science, OG I 1 Designing Robust Multimodal Systems for Diverse Users and Mobile Environments.
11.10 Human Computer Interface www. ICT-Teacher.com.
CP SC 881 Spoken Language Systems. 2 of 23 Auditory User Interfaces Welcome to SLS Syllabus Introduction.
Affective Interfaces Present and Future Challenges Introductory statement by Antonio Camurri (Univ of Genoa) Marc Leman (Univ of Gent) MEGA IST Multisensory.
Break-out Group # D Research Issues in Multimodal Interaction.
Stanford hci group / cs376 u Scott Klemmer · 16 November 2006 Speech & Multimod al.
User Interface Agents Roope Raisamo Department of Computer and Information Sciences University of Tampere
Human-Computer Interaction
Laws Of Interface Design. 2 of User Control The interface will allow the user to perceive that they are in control and will allow appropriate control.
User-System Interaction: from gesture to action Prof. dr. Matthias Rauterberg IPO - Center for User-System Interaction TU/e Eindhoven University of Technology.
School of something FACULTY OF OTHER Facing Complexity Using AAC in Human User Interface Design Lisa-Dionne Morris School of Mechanical Engineering
Intelligent Robot Architecture (1-3)  Background of research  Research objectives  By recognizing and analyzing user’s utterances and actions, an intelligent.
Mixed Reality: A Model of Mixed Interaction Céline Coutrix, Laurence Nigay User Interface Engineering Team CLIPS-IMAG Laboratory, University of Grenoble.
Cognitive Systems Foresight Language and Speech. Cognitive Systems Foresight Language and Speech How does the human system organise itself, as a neuro-biological.
Multi-Modal Dialogue in Personal Navigation Systems Arthur Chan.
CARE properties Chris Vandervelpen
Understanding Naturally Conveyed Explanations of Device Behavior Michael Oltmans and Randall Davis MIT Artificial Intelligence Lab.
Stanford hci group / cs376 u Jeffrey Heer · 19 May 2009 Speech & Multimodal Interfaces.
W3C Multimodal Interaction Activities Deborah A. Dahl August 9, 2006.
A Speech Interface to Virtual Environment Authors Scott McGlashan and Tomas Axling Swedish Institute of Computer Science.
Software Architecture for Multimodal Interactive Systems : Voice-enabled Graphical Notebook.
WP6 Emotion in Interaction Embodied Conversational Agents WP6 core task: describe an interactive ECA system with capabilities beyond those of present day.
Speech and multimodal Jesse Cirimele. papers “Multimodal interaction” Sharon Oviatt “Designing SpeechActs” Yankelovich et al.
Multimodal and Natural computer interaction Evelina Stanevičienė.
MULTIMODAL AND NATURAL COMPUTER INTERACTION Domas Jonaitis.
Perceptive Computing Democracy Communism Architecture The Steam Engine WheelFire Zero Domestication Iron Ships Electricity The Vacuum tube E=mc 2 The.
SIE 515 Universal Design Lecture 9.
CS 580 chapter 4 paradigms.
Characteristics of Graphical and Web User Interfaces
HUMAN MEDIA INTERACTION CREATIVE TECHNOLOGY FOR PLAY AND
Ten Myths of Multimodal Interaction
Human Computer Interaction (HCI)
Visual Information Retrieval
CHAPTER 1 Introduction BIC 3337 EXPERT SYSTEM.
11.10 Human Computer Interface
Overview of Year 1 Progress Angelo Cangelosi & ITALK team
Unit 2 User Interface Design.
Human-Computer Interaction Introduction to HCI
DESIGNING WEB INTERFACE Presented By, S.Yamuna AP/CSE
Software engineering USER INTERFACE DESIGN.
Multimodal Interfaces
Graphical User Interface Based Digital Sixth Sense
Speech & Multimodal Scott Klemmer · 16 November 2006.
Map of Human Computer Interaction
Professor John Canny Spring 2003
Characteristics of Graphical and Web User Interfaces
Human Computer Interaction
Human Computer Interaction Lecture 19 Universal Design
Chapter 9 System Control
Professor John Canny Spring 2004
Computer Vision Readings
Presentation transcript:

Multimodal Human-Computer Interaction New Interaction Techniques 22. 1 Multimodal Human-Computer Interaction New Interaction Techniques 22.1.2001 Roope Raisamo (rr@cs.uta.fi) Department of Computer and Information Sciences University of Tampere, Finland

Multimodal human-computer interaction A definition [Raisamo, 1999e, p. 2]: ”Multimodal interfaces combine many simultaneous input modalities and may present the information using synergistic representation of many different output modalities”

Multimodal interaction techniques Our definition of an interaction technique [Raisamo, 2000]: An interaction technique is a way to carry out an interactive task. It is defined in the binding, sequencing, and functional levels, and is based on using a set of input and output devices or technologies. In a multimodal interaction technique there are more than one inputs or outputs used for the same task.

Two views A Human-Centered View A System-Centered View common in psychology often considers human input channels, i.e., computer output modalities, and most often vision and hearing applications: a talking head, audio-visual speech recognition, ... A System-Centered View common in computer science a way to make computer systems more adaptable

Multimodal human- computer interaction

Senses and modalities [Silbernagel, 1979]

Design space for multimodal user interfaces [Nigay and Coutaz, 1993]

An architecture for multimodal user interfaces Adapted from [Maybury and Wahlster, 1998]

Modeling [Nigay and Coutaz, 1993]

Put – That – There [Bolt, 1980]

Potential benefits A list by Maybury and Wahlster [1998, p. 15]: Efficiency Redundancy Perceptability Naturalness Accuracy Synergy Mutual disambiguation of recognition errors [Oviatt, 1999a]

Common misconceptions A list by Oviatt [1999b]: If you build a multimodal system, user will interact multimodally. Speech and pointing is the dominant multimodal integration pattern. Multimodal input involves simultaneous signals. Speech is the primary input mode in any multimodal system that includes it. Multimodal language does not differ linguistically from unimodal language.

Common misconceptions Multimodal integration involves redundancy of content between modes. Individual error-prone recognition technologies combine multimodally to produce even greater unreliability. All users’ multimodal commands are integrated in a uniform way. Different input modes are capable of transmitting comparable content. Enhanced efficiency is the main advantage of multimodal systems.

Two paradigms for multimodal user interfaces Computer as a tool multiple input modalities are used to enhance direct manipulation behavior of the system the machine is a passive tool and tries to understand the user through all different input modalities that the system recognizes the user is always responsible for initiating the operations follows the principles of direct manipulation [Shneiderman, 1982; 1983]

Two paradigms for multimodal user interfaces Computer as a dialogue partner the multiple modalities are used to increase the anthropomorphism in the user interface multimodal output is important: talking heads and other human-like modalities speech recognition is a common input modality in these systems can often be described as an agent-based conversational user interface

Two hypotheses on combining modalities The combination of human output channels effectively increases the bandwidth of the humanmachine channel. This has been discovered in many empirical studies of multimodal human-computer interaction [Oviatt, 1999b].

Two hypotheses on combining modalities Adding extra output modality requires more neurocomputational resources and will lead to deteriorated output quality, resulting in reduced effective bandwidth. Two types of effects are usually observed: a slow-down of all output processes, and interference errors due to the fact that selective attention cannot be divided between the increased number of output channels. Two examples of this: writing when speaking, and speaking when driving a car.

Call for research A summary in [Raisamo, 1999e] pointed out that more research is needed to understand the following: How the brain works and which modalities can best be used to gain the synergy advantages that are possible with multimodal interaction? When a multimodal system is preferred to a unimodal system? Which modalities make up the best combination for a given interaction task? Which interaction devices to assign to these modalities in a given computing system? How to use these interaction devices, that is, which interaction techniques to select or develop for a given task?

Touch’n’Speak [Raisamo, 1998] Touch’n’Speak is a multimodal user interface framework that makes use of combined touch and speech input and different output modalities Input: touch buttons, touch lists, touch gestures in area selection (time, location, pressure), speech commands Output: graphical, textual, and auditory (non-speech) output, speech feedback The framework was used to implement a restaurant information system that provides information on restaurants in Cambridge, MA, USA.

A snapshot of Touch’n’Speak

Examples CHI2000 Video Proceedings: The Efficiency of Multimodal Interaction for a Map-Based Task (8:18) SIGGRAPH Video Review 76, CHI’92 Technical Video Program: Multi-Modal Natural Dialogue (10:25) SIGGRAPH Video Review 77, CHI’92 Technical Video Program: Combining Gestures and Direct Manipulation (9:56) CHI’99 Video Proceedings: Embodiment in Conversational Interfaces: Rea (2:08)

Homework Read Chapter 2 (Multimodal interaction) in [Raisamo, 1999e]. [Raisamo, 1999e] is available online at http://granum.uta.fi/pdf/951-44-4702-6.pdf A printable version is available online at http://www.cs.uta.fi/~rr/interact/dissertation.pdf