MULTIMODAL AND NATURAL COMPUTER INTERACTION Domas Jonaitis.

Slides:



Advertisements
Similar presentations
PUI 98 Leveraging Human Capabilities in Perceptual Interfaces George G. Robertson Microsoft Research George G. Robertson Microsoft Research.
Advertisements

Irek Defée Signal Processing for Multimodal Web Irek Defée Department of Signal Processing Tampere University of Technology W3C Web Technology Day.
Interaction Devices By: Michael Huffman Kristen Spivey.
Empirical and Data-Driven Models of Multimodality Advanced Methods for Multimodal Communication Computational Models of Multimodality Adequate.
Tailoring Tabletop Interfaces for Musical Control Liam O'Sullivan Electronic and Electrical Engineering, Trinity College Dublin, Ireland Trinity College.
Human Computer Interaction
John Hu Nov. 9, 2004 Multimodal Interfaces Oviatt, S. Multimodal interfaces Mankoff, J., Hudson, S.E., & Abowd, G.D. Interaction techniques for ambiguity.
Lecture 7 Date: 23rd February
Stanford hci group / cs376 research topics in human-computer interaction Multimodal Interfaces Scott Klemmer 15 November 2005.
ISTD 2003, Thoughts and Emotions Interactive Systems Technical Design Seminar work: Thoughts & Emotions Saija Gronroos Mika Rautanen Juha Sunnari.
CS335 Principles of Multimedia Systems Multimedia and Human Computer Interfaces Hao Jiang Computer Science Department Boston College Nov. 20, 2007.
CSD 5230 Advanced Applications in Communication Modalities 7/3/2015 AAC 1 Introduction to AAC Orientation to Course Assessment Report Writing.
1. Plan : 1. Models of interaction 2. Types of interaction 3. Existing technologies 4. Advances in HCI 5. Architecture 2.
Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Cognitive Robots © 2014, SNU CSE Biointelligence Lab.,
Closing conference of SYSIASS – June 17 th 2014 Multimodal Bio-signal based Control of Intelligent Wheelchair Professor Huosheng Hu Leader of Activity.
Copyright John Wiley & Sons, Inc. Chapter 3 – Interactive Technologies HCI: Developing Effective Organizational Information Systems Dov Te’eni Jane.
There are different types of translator. An Interpreter Interpreters translate one instruction at a time from a high level language into machine code every.
Multimedia Specification Design and Production 2013 / Semester 2 / week 8 Lecturer: Dr. Nikos Gazepidis
Center for Human Computer Communication Department of Computer Science, OG I 1 Designing Robust Multimodal Systems for Diverse Users and Mobile Environments.
11.10 Human Computer Interface www. ICT-Teacher.com.
ST01 - Introduction 1 Introduction Lecturer: Smilen Dimitrov Sensors Technology – MED4.
Fall UI Design and Implementation1 Lecture 20: HCI Research Topics.
Multimodal Information Access Using Speech and Gestures Norbert Reithinger
Stanford hci group / cs376 u Scott Klemmer · 16 November 2006 Speech & Multimod al.
User Interface Agents Roope Raisamo Department of Computer and Information Sciences University of Tampere
Turns human body into a touch screen finger input Interface. By, M.PRATHYUSHA 07P61A1261 IT-B.
Human Computer Interaction © 2014 Project Lead The Way, Inc.Computer Science and Software Engineering.
Human-Computer Interaction
School of something FACULTY OF OTHER Facing Complexity Using AAC in Human User Interface Design Lisa-Dionne Morris School of Mechanical Engineering
KAMI KITT ASSISTIVE TECHNOLOGY Chapter 7 Human/ Assistive Technology Interface.
1 Human Computer Interaction Week 5 Interaction Devices and Input-Output.
© 2003 Gina Joue & Brian Duffy Dr. Brian Duffy
Different Types of HCI CLI Menu Driven GUI NLI
CARE properties Chris Vandervelpen
Copyright John Wiley & Sons, Inc. Chapter 3 – Interactive Technologies HCI: Developing Effective Organizational Information Systems Dov Te’eni Jane.
Stanford hci group / cs376 u Jeffrey Heer · 19 May 2009 Speech & Multimodal Interfaces.
Software Architecture for Multimodal Interactive Systems : Voice-enabled Graphical Notebook.
What is Multimedia Anyway? David Millard and Paul Lewis.
Usability and User Interfaces CS 560. Human Computer Interaction Human Computer Interaction (HCI) is the academic and industrial discipline that studies.
NCP meeting Jan 27-28, 2003, Brussels Colette Maloney Interfaces, Knowledge and Content technologies, Applications & Information Market DG INFSO Multimodal.
Speech and multimodal Jesse Cirimele. papers “Multimodal interaction” Sharon Oviatt “Designing SpeechActs” Yankelovich et al.
Multimodal and Natural computer interaction Evelina Stanevičienė.
Perceptive Computing Democracy Communism Architecture The Steam Engine WheelFire Zero Domestication Iron Ships Electricity The Vacuum tube E=mc 2 The.
Ten Myths of Multimodal Interaction
Introducing virtual REALITY
Derek Hunt Education Commons
Information Computer Technology
Information Computer Technology
Fundamentals of Information Systems, Sixth Edition
Intelligent IVI with AI
Fundamentals of Information Systems
11.10 Human Computer Interface
Module 2… Talking with computers
GESTURE RECOGNITION TECHNOLOGY
Human Computer Interaction Lecture 20 Universal Design
DESIGNING WEB INTERFACE Presented By, S.Yamuna AP/CSE
Multimodal Interfaces
CEN3722 Human Computer Interaction Advanced Interfaces
Human Computer Interaction
Smart Learning concepts to enhance SMART Universities in Africa
Pervasive Computing Happening?
Multimodal Human-Computer Interaction New Interaction Techniques 22. 1
Speech & Multimodal Scott Klemmer · 16 November 2006.
There are different types of translator.
Map of Human Computer Interaction
Professor John Canny Spring 2003
universal design (web accessibility)
Human and Computer Interaction (H.C.I.) &Communication Skills
Chapter 9 System Control
Professor John Canny Spring 2004
Presentation transcript:

MULTIMODAL AND NATURAL COMPUTER INTERACTION Domas Jonaitis

CONCEPTION Multimodal human-computer interaction refers to the “interaction with the virtual and physical environment through natural modes of communication”, i. e. the modes involving the five human senses. This implies that multimodal interaction enables a more free and natural communication, interfacing users with automated systems in both input and output. Specifically, multimodal systems can offer a flexible, efficient and usable environment allowing users to interact through input modalities, such as speech, handwriting, hand gesture and gaze, and to receive information by the system through output modalities, such as speech synthesis, smart graphics and others modalities, opportunely combined.

AIM Multimodal interaction systems aim to support the recognition of naturally occurring forms of human language and behavior through the use of recognition-based technologies (Oviatt, 2003; Waibel et al., 1996)

HUMAN – COMPUTER INTERACTION A PROCESS WHEN INFORMATION TRANSFERRED 1 )User to machine (Human computer interaction (HCI)) 2) Machine to user(Computer human interaction (CHI))

MODALITIES Inputs: sensory organs Outputs: effectors

HUMAN - COMPUTER INTERACTION

ADVANTAGES OF MULTIMODELS Natural/realism making: making use of more (appropriate) senses New ways of interacting Flexible: different modalities excel at different tasks Wearable computers and small devices Helps the visually/physically impaired Faster, more efficient, higher information processing bandwidth Robust: mutual disambiguation of recognition errors Multimodal interfaces are more engaging Use more of users senses User perceive multiple things at once

EXAMPLES

MULTIMODAL INPUT

MULTIMODAL OUTPUT

LEVELS OF MULTIMODAL FUSION Data level Feature level Decision level

HUMAN COMPUTER INTERACTION ARCHITECTURE Multimodal HCI Systems The system based on combination of multiple modalities of interaction by simultaneous use of different channels Unimodal HCI Systems The system based on single channel of input

MODALITIES VISUAL-BASED HCI Facial Expression Analysis Body Movement Tracking (Large-scale) Gesture Recognition Gaze Detection (Eyes Movement Tracking)

MODALITIES AUDIO-BASED HCI Speech Recognition Speaker Recognition(fig 6). Auditory Emotion Analysis Human-Made Noise/Sign Detections (Gasp, Sigh, Laugh, Cry, etc.) Musical Interaction

MODALITIES SENSOR-BASED HCI 1. Pen-Based Interaction 2. Mouse & Keyboard 3. Joysticks 4. Motion Tracking Sensors and Digitizers 5. Haptic Sensors 6. Pressure Sensors 7. Taste/Smell Sensors

ARCHITECTURE OF MULTIMODAL USER INTERFACE

APPLIED COMPUTER VISION TECHNOLOGIES COMPUTER VISION GESTURE, MOTION RECOGNITION GOOGLE GLASS OCULUS, MICROSOFT HOLOLENS

APPLIED SOUND TECHNOLOGIES Speech recognition technologies Machine translation Siri, Google

APPLIED TOUCH TECHNOLOGIES Haptic technologies recreate the sense of touch by applying force, vibrations or motions to the user

BRAIN COMPUTER INTERFACE (BCI) Tries to understand human brain waves

SMELL/TASTE Electronic nose Electronic tongue (used to classify liquid contents)

PROBLEMS AND CHALENGES? Technological issues High price Portability Power consumption Lack of specialized software Intelligent agents require supercomputers Need powerful computers Virtual reality is tiring after long usage Human issues how do we handle conflicting information how to keep user attention what happens when everyone starts talking with their devices

POTENTIAL USAGE Can process and manipulating complex data Can help elderly or disabled people to communicate Can improve many routine things. Entertainment industries

CONCLUSIONS Although the literature on formal assessment of multimodal systems is still sparse, various studies have shown that multimodal interfaces maybe preferred by users over unimodal alternatives, can offer better flexibility and reliability, can offer interaction alternatives to better meet the needs of diverse users with a range of usage patterns and preferences. Multimodal interfaces can increase task efficiency. They permit the flexible use of input modes, including alternation and integrated use. They support improved efficiency, especially when manipulating graphical information. They can support shorter and simpler speech utterances than a speech-only interface, which results in fewer disfluencies and more robust speech recognition. They can support greater precision of spatial information than a speech-only interface, since pen input can be quite precise. They give users alternatives in their interaction techniques. They lead to enhanced error avoidance and ease of error resolution. They accommodate a wider range of users, tasks, and environmental situations. They are adaptable during continuously changing environmental conditions. They accommodate individual differences, such as permanent or temporary handicaps. They can help prevent overuse of any individual mode during extended computer usage.

FUTURE TECHNOLOGIES Self Driving car (Google)

NATURAL AND EVAPORATING INTERFACES

QUESTIONS? 1.What are the main human senses used for multimodal computer interaction ? 2.The conception of multimodal computer interaction? 3.The advantages of multimodal computer interaction? 4.Where can we implement this model? 5.The practical example of this model?