Didier Perroud Raynald Seydoux Frédéric Barras.  Abstract  Objectives  Modalities ◦ Project modalities ◦ CASE/CARE  Implementation ◦ VICI, Iphone,

Slides:



Advertisements
Similar presentations
Microsoft Research Faculty Summit 2003 Brad A. Myers & Jeffrey Nichols The Personal Universal Controller and.NET CF Pebbles Research Project Human Computer.
Advertisements

A Natural Interactive Game By Zak Wilson. Background This project was my second year group project at University and I have chosen it to present as it.
Video Game Design Lesson 1. Game Designer Person involved in the development of a video game Person involved in the development of a video game Usually.
ARIS The Augmented Rea l ity Studio. Outline  Background  Problem definition  Proposed solution  System design  Functionalities  Comparison with.
Social Activity Recognition Using a Wrist-Worn Accelerometer Ashton Maltie UNCC WiNS Lab Ashton Maltie UNCC WiNS Lab.
Page 1 SIXTH SENSE TECHNOLOGY Presented by: KIRTI AGGARWAL 2K7-MRCE-CS-035.
Virtual Dart: An Augmented Reality Game on Mobile Device Supervisor: Professor Michael R. Lyu Prepared by: Lai Chung Sum Siu Ho Tung.
The Science of Digital Media Microsoft Surface 7May Metropolia University of Applied Sciences Display Technologies Seminar.
The State of the Art in VoiceXML Chetan Sharma, MS Graduate Student School of CSIS, Pace University.
Slides prepared by Rose Williams, Binghamton University Chapter 1 Getting Started 1.1 Introduction to Java.
Class 6 LBSC 690 Information Technology Human Computer Interaction and Usability.
1 SWE Introduction to Software Engineering Lecture 22 – Architectural Design (Chapter 13)
LemGen (Linguistic EMulation and Generation ENgine) CS491 Project Chris Lemcke.
CSD 5230 Advanced Applications in Communication Modalities 7/3/2015 AAC 1 Introduction to AAC Orientation to Course Assessment Report Writing.
Support for Palm Pilot Collaboration Including Handwriting Recognition.
© 2001 CUHK Recognition Software Building Block Overview 1. Task specification — What is the function of the system ? 2. Recognizer Design — Vocabulary.
HUMANOID ANIMATION DRIVEN BY HUMAN VOICE Thesis Advisor : Dr. Donald P. Brutzman Second Reader : Dr. Xiaoping Yun A Thesis By Ozan APAYDIN, Turkish Navy.
A Web Services Based Streaming Gateway for Heterogeneous A/V Collaboration Hasan Bulut Computer Science Department Indiana University.
Mobile Multimodal Applications. Dr. Roman Englert, Gregor Glass March 23 rd, 2006.
Prototype & Design Computer Inputs. How to Prototype & Design Computer Inputs Step 1: Review Input Requirements Step 2: Select the GUI Controls Step 3:
Computerized Train Control System by: Shawn Lord Christian Thompson.
Augmented Reality and 3D modelling Done by Stafford Joemat Supervised by Mr James Connan and Mr Mehrdad Ghaziasgar.
HAND GESTURE BASED HUMAN COMPUTER INTERACTION. Hand Gesture Based Applications –Computer Interface A 2D/3D input device (Hand Tracking) Translation of.
VoiceXML Builder Arturo Ramirez ACS 494 Master’s Graduate Project May 04, 2001.
Alternate Version of STARTING OUT WITH C++ 4 th Edition Chapter 1 Introduction to Computers and Programming.
Antigone Engine Kevin Kassing – Period
JavaScript and The Document Object Model MMIS 656 Web Design Technologies Acknowledgements: 1.Notes from David Shrader, NSU GSCIS 2.Some material adapted.
Conversational Applications Workshop Introduction Jim Larson.
Playful Stimulation against Parkinson’s Disease -
Java Programming, 3e Concepts and Techniques Chapter 3 Section 65 – Manipulating Data Using Methods – Java Applet.
Rujchai Ung-arunyawee Department of Computer Engineering Khon Kaen University.
Presentation by: K.G.P.Srikanth. CONTENTS  Introduction  Components  Working  Applications.
ITCS 6010 SALT. Speech Application Language Tags (SALT) Speech interface markup language Extension of HTML and other markup languages Adds speech and.
Game Engine Programming. Game Engine Game Engine Rendering Engine (OGRE) Rendering Engine (OGRE) Physics Engine (Bullet) Physics Engine (Bullet) Input/Output.
Submitted by:- Vinay kr. Gupta Computer Sci. & Engg. 4 th year.
1 Computer Graphics Assistant Professor Dr. Sana’a Wafa Al-Sayegh 2 nd Semester ITGD3107 University of Palestine.
© 2008 Eyeblaster. All rights reserved Sales Training Presented by: Christopher Maurer ● Sales Engineer ● 15 th December 2009 Live Streaming and Augmented.
Visual C++ Programming: Concepts and Projects
Voice Recognition (Presentation 2) By: Priya Devi A. S/W Developer, Xsys technologies Bangalore.
Physically Realistic Interface for a User Inside VR Masahide Hashimoto Kenji Miyamoto Faculty of Engineering Hosei University (Japan)
GENESIS OF VIRTUAL REALITY  The term ‘Virtual reality’ (VR) was initially coined by Jaron Lanier, founder of VPL Research (1989)..
Real-Time Cyber Physical Systems Application on MobilityFirst Winlab Summer Internship 2015 Karthikeyan Ganesan, Wuyang Zhang, Zihong Zheng.
ENTERFACE 08 Project 1 “MultiParty Communication with a Tour Guide ECA” Mid-term presentation August 19th, 2008.
March 20, 2006 © 2005 IBM Corporation Distributed Multimodal Synchronization Protocol (DMSP) Chris Cross IETF 65 March 20, 2006 With Contribution from.
ENTERFACE ’08 Project 2 “Multimodal High Level Data Integration” Final Report August 29th, 2008.
MULTI-LINGUAL AND DEVICELESS COMPUTER ACCESS FOR DISABLED USERS C.Premnath and J.Ravikumar S.S.N. College of Engineering TamilNadu.
Controlling Computer Using Speech Recognition (CCSR) Creative Masters Group Supervisor : Dr: Mounira Taileb.
Using Google's Web Speech API with Moodle for language learning tasks
1 Galatea: Open-Source Software for Developing Anthropomorphic Spoken Dialog Agents S. Kawamoto, et al. October 27, 2004.
Augmented Reality and 3D modelling Done by Stafford Joemat Supervised by Mr James Connan.
ENTERFACE 08 Project #1 “ MultiParty Communication with a Tour Guide ECA” Final presentation August 29th, 2008.
1 EiffelMedia. 2 Overview Features of the library Documentation Demos Community Stats Roadmap.
Stanford hci group / cs376 u Jeffrey Heer · 19 May 2009 Speech & Multimodal Interfaces.
W3C Multimodal Interaction Activities Deborah A. Dahl August 9, 2006.
Software Architecture for Multimodal Interactive Systems : Voice-enabled Graphical Notebook.
Requirements Analysis Review Modeling Team Matt Weyant 26 Oct 1999.
Magic Wand Battle Game Team 53 Shanoon Martin, Jialin Sun, Manfei Wu.
Accelerometer based motion gestures for mobile devices Presented by – Neel Parikh Advisor Committee members Dr. Chris Pollett Dr. Robert Chun Dr. Mark.
Augmented Reality and 3D modelling Done by Stafford Joemat Supervised by Mr James Connan and Mehrdad Ghaziasgar.
BeNeFri University Vision assistance for people with serious sight problems Moreno Colombo Marin Tomić Future User Interfaces, BeNeFri University.
Faculty of Information Technology, Brno University of Technology, CZ
Applications Active Web Documents Active Web Documents.
Hand Gestures Based Applications
Enterprise Computing Collaboration System Example
Google translate app demo
Chapter 2: Input and output devices
Architecting and integrating sensor drivers
Multimodal FooBilliard
Multimodal Caricatural Mirror
Natural User Interaction with Perceptual Computing
Presentation transcript:

Didier Perroud Raynald Seydoux Frédéric Barras

 Abstract  Objectives  Modalities ◦ Project modalities ◦ CASE/CARE  Implementation ◦ VICI, Iphone, Voice recognition, Network  Demonstration  Conclusion

 Coordination between two persons to move a ball into a labyrinth  Rotation possible on the x and y axis  Gates can be opened with vocal and gestural commands

 Coordinate the following technologies: ◦ Augmented reality with tags ◦ Gesture detection ( with Iphone accelerometers) ◦ Voice recognition ( words) ◦ Collaborative environments ◦ Physic engine

 Inputs ◦ Hand rotation in x and y axis ( one axis per player)  direct manipulation of the labyrinth board ◦ Hand pumping for gates’ openings ◦ Voice recognition (words) for selecting gate to open and start the game  Outputs ◦ Image on the beamer ◦ Iphone vibrations

 CASE ◦ Semantic level of abstraction  CARE ◦ Gesture orientation: assignment ◦ Gesture pumping/Voice selection: complementary to open a gate ◦ Voice commands: assignment  Decision level fusion  Fission: image, vibration

 Blocks ◦ Webcam, Tag detection ◦ OpenGL, Physic engine  Multimodality Management ◦ state machine  Augmented reality application ◦ event based  Messages from the gateway ◦ Voice events ◦ Gesture events (orientation X and Y, shake)  Messages to the gateway ◦ Vibration events

 Handle the UIAccelerometer interface  Generate motionEvent when shaking  Messages to the gateway ◦ Orientations (X or Y) ◦ Shake  Messages from the gateway ◦ Vibrate

Windows speech API  SDK Features: ◦ API definition files ◦ Runtime component ◦ Control Panel applet ◦ Text-To-Speech engines in multiple languages. ◦ Speech Recognition engines in multiple languages. ◦ Redistributable components ◦ Sample application code. ◦ Sample engines ◦ Documentation.

Our System  A speech recognition engine  A grammar <grammar xmlns=" xmlns:xsi=" xsi:schemaLocation=" xml:lang="en-EN" version="1.0"> New game Pause Exit Open gate one Open gate two Close gate one Close gate two

 Recognition comparison before training / after training

 Live  Videos

 Problems with the physic engine ◦ Coordination user moves – physic moves  Voice recognition OK  High-level programing  Heterogeneity not a problem  Functional prototype

 Thank you