1 Voice Command Generation for Teleoperated Robot Systems Authors : M. Ferre, J. Macias-Guarasa, R. Aracil, A. Barrientos Presented by M. Ferre. Universidad.

Slides:



Advertisements
Similar presentations
GMD German National Research Center for Information Technology Darmstadt University of Technology Perspectives and Priorities for Digital Libraries Research.
Advertisements

Input Jeopardy The Keyboard Pointing Devices Pictures and Sounds Scanners and Readers Grab Bag $100100$100100$100100$100100$ $200200$200200$200200$200200$
For suggestions about how Talk Factory can be incorporated into lessons on investigating evaporation please see the associated lesson plans on the Talk.
User Interface. What is a User Interface  A user interface is a link between the user and the computer. It allows the user and the computer to communicate.
INSTRUCTOR:Dr.Veton Kepuska STUDENT:Dileep Narayan.Koneru YES/NO RECOGNITION SYSTEM.
Input to the Computer * Input * Keyboard * Pointing Devices
Talk Factory Primary Generic: instructions for use Supporting children‘s whole class discussions in primary school Copyright 2011 Open University.
Detection of Recognition Errors and Out of the Spelling Dictionary Names in a Spelled Name Recognizer for Spanish R. San-Segundo, J. Macías-Guarasa, J.
INCORPORATING MULTIPLE-HMM ACOUSTIC MODELING IN A MODULAR LARGE VOCABULARY SPEECH RECOGNITION SYSTEM IN TELEPHONE ENVIRONMENT A. Gallardo-Antolín, J. Ferreiros,
From the NXT top menu Connect desired hardware as indicated Enter a command in the box indicated from the menu provided Repeat for all 5 boxes.
Simplifying Wheelchair Mounted Robotic Arm Control with a Visual Interface Katherine M. Tsui and Holly A. Yanco University of Massachusetts, Lowell Computer.
Simplifying Wheelchair Mounted Robotic Arm Control with a Visual Interface Katherine M. Tsui and Holly A. Yanco University of Massachusetts Lowell Katherine.
CONTROLLING A HIFI WITH A CONTINUOUS SPEECH UNDERSTANDING SYSTEM ICSLP’ 98 CONTROLLING A HIFI WITH A CONTINUOUS SPEECH UNDERSTANDING SYSTEM J. Ferreiros,
1 / 31 CS 425/625 Software Engineering User Interface Design Based on Chapter 15 of the textbook [SE-6] Ian Sommerville, Software Engineering, 6 th Ed.,
CS335 Principles of Multimedia Systems Multimedia and Human Computer Interfaces Hao Jiang Computer Science Department Boston College Nov. 20, 2007.
T.Sharon-A.Frank 1 Internet Resources Discovery (IRD) Multimedia - Motivation & Definition.
© 2001 CUHK Recognition Software Building Block Overview 1. Task specification — What is the function of the system ? 2. Recognizer Design — Vocabulary.
Intermediate GNVQ ICT Input Devices and Methods Manual input devices are used by people to enter data by hand. Keyboard Pointing devices Joystick Digital.
BoardMaker Ver 6 Creating Interactive Activities for the Classroom.
GCSE Information Technology Input Devices and Methods 3 Manual input devices are used by people to enter data by hand. Keyboard Pointing devices Joystick.
Use of Multimedia in Engineering. Mechatronics engineering is based on the combination from three basic engineering field that is mechaninal, electronics.
Natural Language Processing and Speech Enabled Applications by Pavlovic Nenad.
Semantic Parsing for Robot Commands Justin Driemeyer Jeremy Hoffman.
Computer for Health Sciences
Knowledge Base approach for spoken digit recognition Vijetha Periyavaram.
Interactive Input Methods & Graphical User Input
Multimedia Specification Design and Production 2013 / Semester 2 / week 8 Lecturer: Dr. Nikos Gazepidis
CGMB214: Introduction to Computer Graphics
Unit 1_9 Human Computer Interface. Why have an Interface? The user needs to issue instructions Problem diagnosis The Computer needs to tell the user what.
CHAPTER FOUR COMPUTER SOFTWARE.
Activity 3: Multimodality HMI for Hands-free control of an intelligent wheelchair L. Wei, T. Theodovidis, H. Hu, D. Gu University of Essex 27 January 2012.
Introduction to Interactive Media Interactive Media Tools: Software.
Spoken dialog for e-learning supported by domain ontologies Dario Bianchi, Monica Mordonini and Agostino Poggi Dipartimento di Ingegneria dell’Informazione.
These slides are designed to accompany Software Engineering: A Practitioner’s Approach, 7/e (McGraw-Hill, 2009) Slides copyright 2009 by Roger Pressman.1.
 The most intelligent device - “Human Brain”.  The machine that revolutionized the whole world – “computer”.  Inefficiencies of the computer has lead.
1 Final Conference, 19th – 23rd January 2015 Geneva, Switzerland RP 15 Force estimation based on proprioceptive sensors for teleoperation in radioactive.
Designing Interface Components. Components Navigation components - the user uses these components to give instructions. Input – Components that are used.
MULTIMEDIA DEFINITION OF MULTIMEDIA
SPEECH CONTENT Spanish Expressive Voices: Corpus for Emotion Research in Spanish R. Barra-Chicote 1, J. M. Montero 1, J. Macias-Guarasa 2, S. Lufti 1,
Spoken Dialog Systems and Voice XML Lecturer: Prof. Esther Levin.
MULTIMEDIA INPUT / OUTPUT TECHNOLOGIES INTRODUCTION 6/1/ A.Aruna, Assistant Professor, Faculty of Information Technology.
Projective Virtual Reality Freund and Rossmann Summarized by Geb Thomas.
1 CP586 © Peter Lo 2003 Multimedia Communication Human Computer Interaction.
HARDWARE INPUT DEVICES GETTING DATA INTO THE COMPUTER.
Microsoft Assistive Technology Products Brought to you by... Jill Hartman.
E.g.: MS-DOS interface. DIR C: /W /A:D will list all the directories in the root directory of drive C in wide list format. Disadvantage is that commands.
CMP 131 Introduction to Computer Programming Violetta Cavalli-Sforza Week 3, Lecture 1.
SD1230 Unit 6 Desktop Applications. Course Objectives During this unit, we will cover the following course objectives: – Identify the characteristics.
Total Physical Response (TPR) 1. "Babies don't learn by memorizing lists; why should children or adults?" James J. Asher an emeritus professor of psychology.
Talking to Robots Using MS Speech SDK 5.1 in C# Sebastian van Delden USC Upstate
1 Galatea: Open-Source Software for Developing Anthropomorphic Spoken Dialog Agents S. Kawamoto, et al. October 27, 2004.
金聲玉振 Taiwan Univ. & Academia Sinica 1 Spoken Dialogue in Information Retrieval Jia-lin Shen Oct. 22, 1998.
Importance of user interface design – Useful, useable, used Three golden rules – Place the user in control – Reduce the user’s memory load – Make the.
Laboratory 2 Group 19 The Group of Destiny. User Interface - Debugging Objectives:  Display: Sensor data (telemetry) – including IR sensors, status of.
In computing, an input device is a piece of computer hardware equipment used to provide data and control signals to an information processing system such.
CIS 210 Systems Analysis and Development Week 7 Part II Designing Interfaces and Dialogues,
1 Interaction Devices CIS 375 Bruce R. Maxim UM-Dearborn.
Speech Processing 1 Introduction Waldemar Skoberla phone: fax: WWW:
© NCC Education Limited V1.0 Introduction to Computing  Unit 4: Human-Computer Interaction.
Knowledge Based Systems ExpertSystems Difficulties in Expert System Development u Scarce resources – new technology – demand for trained personnel u Development.
An AV Control Method Using Natural Language Understanding Author : M. Matsuda, T. Nonaka, and T. Hase Date : Speaker : Sian-Lin Hong IEEE Transactions.
Characteristics of Graphical and Web User Interfaces
2016 Maintenance Innovation Challenge
Natural Language Processing and Speech Enabled Applications
Google translate app demo
Software engineering USER INTERFACE DESIGN.
TestNav Text-to-Speech (TTS) Training Guide
Characteristics of Graphical and Web User Interfaces
TestNav8 Text-to-Speech (TTS) Training Guide
Presentation transcript:

1 Voice Command Generation for Teleoperated Robot Systems Authors : M. Ferre, J. Macias-Guarasa, R. Aracil, A. Barrientos Presented by M. Ferre. Universidad Politécnica de Madrid (Spain)

2 Commands by voice Voice is the most natural way for human communication Multimedia interfaces allow communication by voice Some telerobotics systems have been commanded by voice

3 The Speech Engine It is based on an isolated word recognition system It works in real time with a minimum hardware add-on. It converts the spoken utterances into text, defined into a dictionary previously.

4 Experiments to evaluate human interaction. Two kind of voice commands have been studied: –Low Level Commands : Connected with telerobot movements –High Level Commands : Connected with teleoperated task

5 Low Level Command Experiments. Description A video image with the robot end position is displayed A point is continuously drawn over it showing the ideal robot trajectory Trajectory quality is the average error between end robot position and last drawn point.

6 LLC - Operator Input Devices The 3 operator input devices are: –A master-arm, –A joystick, –A voice recognition system. All experiments have a similar level of proprioception

7 LLC - Master Arm Master-arm has 3 dof but only 2 are used, both confined within a horizontal plane Operator takes the M-A by the end M-A generates position references for the teleoperated robot

8 LLC - Joystick The joystick has 2 dof. It generates rate references for the telerobot: –Right movements involve telerobot right speeds, and so on.

9 LLC - Voice Commands Voice interface vocabulary is made up by 13 words: –Numbers from “1” to“12” –“STOP” Numbers indicated robot end speeds as a clockwise direction

10 LLC - Experiment Results X-axes represents participants Y-axes trajectory quality Statistical tests show: –There are not differences between operators –There are differences between input devices –Master-Arm and Joystick are similar

11 High Level Commands Two interfaces have been evaluated: –Natural Language Procedure: Voice commands are generated following a given imperative grammar structure –Menu: Different options shown graphically are chosen by voice

12 HLC - Natural Lan- guage Procedure Isolated words are recognized by the VISHA system An imperative grammar structure is defined for the operator sentences Semantic Analysis extracts the meaning D.I. and P.A. obtain the command for the telerobot

13 HLC - Menu Menu interface provides the operator with different options chosen by voice. Articulated words active the consecutive submenus until complete the desired command

14 HLC - Results NLP and Menu interface generate the commands whit a very different interaction. Recognition rate is a critical factor for the operator usefulness (>95 %) Firstly operator prefers NLP interface because it is spectacular Then operator prefers Menu interaction because it is more straightforward and comfortable for an isolated word recognition system

15 Voice Interaction Conclusions (I) Hand movements are the best way to guide a telerobot (LLC) In this experiments proprioception is more important than the mode control (rate or position) To guide a robot by voice could be attractive but it is inefficiently because orders are generated slowly

16 Voice Interaction Conclusions (II) HLC need a lot of information from the remote environment and the executed task Menu is the best voice interface using an isolated word recognition system

17 Applications and Future works A menu interface by voice has been applied in telerobotic system for live line maintenance, called ROBTET Future works: –Study differences between rate and position control guiding a telerobot –Compare NLP and Menu interfaces using a word recognition system working continuously