Key Areas Aimed at Children 3 > 5 Early Intervention Help Child Recognise Facial Expressions Provide meta-data capture in order to create reports that can be shared between parents, therapists and teachers
What are we doing that is different Adding a new level of interaction with the Kinect motion detection Allowing for a digital report to be captured and shared between selected parties to monitor the child’s progress Allowing for a level of customisation to allow Therapists, Parents and Teachers a way to make the application personalised To enable a one on one relationship with subject matter through use of their own image
Main Technologies Kinect SDK > To provide an interactive interface for learning HTML / CSS / JS > to provide animations and UI layout Azure > to provide a centralised place for reporting and to enable Parent / Teacher / Therapist login Simple Motion API > Audio file interpretation for emotional response Windows Desktop Application > First prototype to run on a desktop, future builds can run in Metro 8 UI and Xbox Live
Meta-Data Collection and Reporting Reports are collected at points through-out the apps The reports are sent to the “Cloud” At the Bing Fund “Apps for Autism” presentations we heard “Customisation is the key to tools and one thing we don’t see enough of” Customisation enables the application to grow with the child and for them to see their own facial expressions
Your consent to our cookies if you continue to use this website.