Faculty of Information Technology, Brno University of Technology, CZ

Slides:



Advertisements
Similar presentations
Jan SedmidubskyOctober 28, 2011MUFIN: Large-scale Similarity Search Leader: prof. Pavel Zezula Members: Dr. Michal Batko Dr. Vlastislav.
Advertisements

Descriptive schemes for facial expression introduction.
Concepts & Techniques for Accessible, Closed Captioned Web-Based Video 10th Annual Accessing Higher Ground: Accessible Media, Web and Technology Conference.
ENTERFACE’08 Multimodal Communication with Robots and Virtual Agents.
Personalized Abstraction of Broadcasted American Football Video by Highlight Selection Noboru Babaguchi (Professor at Osaka Univ.) Yoshihiko Kawai and.
Didier Perroud Raynald Seydoux Frédéric Barras.  Abstract  Objectives  Modalities ◦ Project modalities ◦ CASE/CARE  Implementation ◦ VICI, Iphone,
Department of Computer Science and Engineering, CUHK 1 Final Year Project 2003/2004 LYU0302 PVCAIS – Personal Video Conference Archives Indexing System.
TimeCleanser: A Visual Analytics Approach for Data Cleansing of Time-Oriented Data Theresia Gschwandtner, Wolfgang Aigner, Silvia Miksch, Johannes Gärtner,
1 Final Year Project 2003/2004 LYU0302 PVCAIS – Personal Video Conference Archives Indexing System Supervisor: Prof Michael Lyu Presented by: Lewis Ng,
1 Taking the Classroom Experience to the Web Innovative Ways to Use Streaming Media in E-Learning Geoff Spencer Higher Education Market Manager Accordent.
Introduction to machine learning
Evolution!!! By Mikhail & Nikhil Acknowledgements to Max The Kokusai Warrior for the XNA Game Engine, Rhemyst and Rymix for the Kinect DTW Gesture Recognizer.
Jan SedmidubskySeptember 23, 2014Motion Retrieval for Security Applications Jan Sedmidubsky Jakub Valcik Pavel Zezula Motion Retrieval for Security Applications.
Extraction of Adverse Drug Effects from Clinical Records E. ARAMAKI* Ph.D., Y. MIURA **, M. TONOIKE ** Ph.D., T. OHKUMA ** Ph.D., H. MASHUICHI ** Ph.D.,K.WAKI.
TINONS1 Nonlinear SP and Pattern recognition
Software Demo Presentation tools John Hines Jane Kaplan Kevin Krawzik.
Project JUTS – Urban Traffic Simulation System David Hartman DSS 2004.
M4 – Video Processing, Brno University of Technology1 M4 – Video Processing Igor Potůček, Michal Španěl, Ibrahim Abu Kteish, Olivier Lai Kan Thon, Pavel.
1 Multimodal Group Action Clustering in Meetings Dong Zhang, Daniel Gatica-Perez, Samy Bengio, Iain McCowan, Guillaume Lathoud IDIAP Research Institute.
Privacy Protection for Life-log Video Jayashri Chaudhari, Sen-ching S. Cheung, M. Vijay Venkatesh Department of Electrical and Computer Engineering Center.
Publishing Technology & Media Solutions.  The flips are back with new features.  Embed Audio & video with seamless streaming.
Break-out Group # D Research Issues in Multimodal Interaction.
Department of Computer Science and Engineering, CUHK 1 Final Year Project 2003/2004 LYU0302 PVCAIS – Personal Video Conference Archives Indexing System.
EDA385 Project Presentation The Sound Disguiser. Overview Sample input audio Read input from the rotary encoder Process the audio due to choosen mode.
Ontologies and Lexical Semantic Networks, Their Editing and Browsing Pavel Smrž and Martin Povolný Faculty of Informatics,
ENTERFACE 08 Project 2 “multimodal high-level data integration” Mid-term presentation August 19th, 2008.
Prof. Thomas Sikora Technische Universität Berlin Communication Systems Group Thursday, 2 April 2009 Integration Activities in “Tools for Tag Generation“
1 These materials are prepared only for the students enrolled in the course Distributed Software Development (DSD) at the Department of Computer Science.
Marketing Development Block 4 Dr. Uma Kanjilal. Stages of a Multimedia Project  Planning and costing- infrastructure, time, skills etc.  Designing and.
Beyond the PC Kiosks & Handhelds Albert Huang Larry Rudolph Oxygen Research Group MIT CSAIL.
Inking as I edit  Add a quick remark in free hand here below.
Gesture Recognition in a Class Room Environment Michael Wallick CS766.
CAMEO: Face Recognition Year 1 Progress and Year 2 Goals Fernando de la Torre, Carlos Vallespi, Takeo Kanade.
Automatic Video Editing Stanislav Sumec. Motivation  Multiple source video data – several cameras in the meeting room, several meeting rooms in teleconference,
1 Detecting Group Interest-level in Meetings Daniel Gatica-Perez, Iain McCowan, Dong Zhang, and Samy Bengio IDIAP Research Institute, Martigny, Switzerland.
ENTERFACE 08 Project #1 “ MultiParty Communication with a Tour Guide ECA” Final presentation August 29th, 2008.
1. Electronic Presentation Overview Terry Harnden.
ENGR 1181 College of Engineering Engineering Education Innovation Center Introduction to Technical Communication.
ENTERFACE’08 Multimodal Communication with Robots and Virtual Agents mid-term presentation.
  Computer vision is a field that includes methods for acquiring,prcessing, analyzing, and understanding images and, in general, high-dimensional data.
WP6 Emotion in Interaction Embodied Conversational Agents WP6 core task: describe an interactive ECA system with capabilities beyond those of present day.
Emotion-aware recommender systems Master’s thesis SupervisorMentorStudent Katrien VerbertKarsten SeippJeroen Reinenbergh.
XML scheme for configuration data V.Černý on behalf of the Bratislava group Initiated by Marco’s call:...a software layer for implementing the handling.
ELAN as a tool for oral history CLARIN Oral History Workshop Oxford Sebastian Drude CLARIN ERIC 18 April 2016.
Speech emotion detection General architecture of a speech emotion detection system: What features?
Learning Objectives Today we will Learn: The different methods of implementation The differences between user and technical documentation.
CIRP Annals - Manufacturing Technology 60 (2011) 1–4 Augmented assembly technologies based on 3D bare-hand interaction S.K. Ong (2)*, Z.B. Wang Mechanical.
Outline  What is MySmartEye ?  Motivation, objectives.  Implementation.  Programming techniques.  Future Work.  Demo.
Building Enterprise Applications Using Visual Studio®
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Automatic Network Protocol Analysis
Automatic Video Shot Detection from MPEG Bit Stream
Supervisor: Prof Michael Lyu Presented by: Lewis Ng, Philip Chan
Ubiquitous Computing and Augmented Realities
Online Shopping APP.
DVD Presentation TRANSCAT
Multimodal Human-Computer Interaction New Interaction Techniques 22. 1
ورود اطلاعات بصورت غيربرخط
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
University of Washington, Autumn 2018
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
AHED Automatic Human Emotion Detection
A HCL Proprietary Utility for
AHED Automatic Human Emotion Detection
Interactive media.
Demystifying Web Content Accessibility Guidelines
Cengizhan Can Phoebe de Nooijer
Jiwon Kim Steve Seitz Maneesh Agrawala
Arguments, Arrangement and Analysis
Intro to HoloLens Development
Presentation transcript:

Faculty of Information Technology, Brno University of Technology, CZ AMIDA Video Edit Demo Presented by: Pavel Zemčík Faculty of Information Technology, Brno University of Technology, CZ

Overview Purpose and principle Resources Block diagram View Conclusion AMIDA Video Edit Demo

Purpose and Principle The purpose is to automatically produce single video stream from several (synchronized) audio/video records. The output can be produced for „full length“ or „reduced“ timing of the meeting (record) Based on features acquired from the video sequence in real-time or „offline“ and also on other sources of information, such as sound/speech, etc. The information is processed through a set of rules (implemented in Prolog) and then the decision is made on the best camera/view and possibly effect The rules include the „technical ones“ and also the ones that lead into „visually appealing“ output AMIDA Video Edit Demo

Resources The resources used as input for decision making include: Position of heads and hands of participants Information about the speaker (audio and visual) Keyword detection input Scheme of the meeting (type and organization) Future extensions, e.g. Semantics of the meeting (whom/what we talk about) Object detection Emotion and gesture recognition AMIDA Video Edit Demo

Automatic Video Edit Block Diagram AMIDA Video Edit Demo

View of the output AMIDA Video Edit Demo

Conclusions The automatic editing system is feasible: it can produce visually appealing output human „touch“ better due to unknown semantics continuous improvement expected The editing process can be affected by: change of preferences of „whom/what“ to prefer choice of topic „focus on the one who uses keywords“ Online and offline versions available (differences in using or not using „future events“ and some simplifications in real-time version, especially in terms of feature extraction). AMIDA Video Edit Demo