42nd IEEE Conference on Decision and Control, Dec 9-12, 2003

Slides:



Advertisements
Similar presentations
Auditory Localisation
Advertisements

Figure 10.13, page 344 Earoverall view Copyright © 2002 Wadsworth Group. Wadsworth is an imprint of the Wadsworth Group, a division of Thomson Learning.
Sound Localization Superior Olivary Complex. Localization: Limits of Performance Absolute localization: localization of sound without a reference. Humans:
Chapter 2: Marr’s theory of vision. Cognitive Science  José Luis Bermúdez / Cambridge University Press 2010 Overview Introduce Marr’s distinction between.
University of North Carolina at Chapel Hill Spatial Sound Localization for Robots Nikunj Raghuvanshi.
Binaural Hearing Or now hear this! Upcoming Talk: Isabelle Peretz Musical & Non-musical Brains Nov. 12 noon + Lunch Rm 2068B South Building.
3-D Sound and Spatial Audio MUS_TECH 348. Psychology of Spatial Hearing There are acoustic events that take place in the environment. These can give rise.
Extracting the pinna spectral notches Vikas C. Raykar | Ramani Duraiswami University of Maryland, CollegePark B. Yegnanaryana Indian Institute of Technology,
3-D Sound and Spatial Audio MUS_TECH 348. Wightman & Kistler (1989) Headphone simulation of free-field listening I. Stimulus synthesis II. Psychophysical.
Localizing Sounds. When we perceive a sound, we often simultaneously perceive the location of that sound. Even new born infants orient their eyes toward.
Source Localization in Complex Listening Situations: Selection of Binaural Cues Based on Interaural Coherence Christof Faller Mobile Terminals Division,
Control & Sensing Research in the Intelligent Servosystems Lab P. S. Krishnaprasad Institute for Systems Research & Department of Electrical and Computer.
Active Contour Models (Snakes)
Development of sound localization
Tracking a moving object with real-time obstacle avoidance Chung-Hao Chen, Chang Cheng, David Page, Andreas Koschan and Mongi Abidi Imaging, Robotics and.
Hearing & Deafness (3) Auditory Localisation
Spectral centroid 6 harmonics: f0 = 100Hz E.g. 1: Amplitudes: 6; 5.75; 4; 3.2; 2; 1 [(100*6)+(200*5.75)+(300*4)+(400*3.2)+(500*2 )+(600*1)] / = 265.6Hz.
EE2F1 Speech & Audio Technology Sept. 26, 2002 SLIDE 1 THE UNIVERSITY OF BIRMINGHAM ELECTRONIC, ELECTRICAL & COMPUTER ENGINEERING Digital Systems & Vision.
Boston University--Harvard University--University of Illinois--University of Maryland A System Identification Problem from Lord Rayleigh P. S. Krishnaprasad.
Speech Segregation Based on Sound Localization DeLiang Wang & Nicoleta Roman The Ohio State University, U.S.A. Guy J. Brown University of Sheffield, U.K.
CSP Auditory input processing 1 Auditory input processing Lecturer: Smilen Dimitrov Cross-sensorial processing – MED7.
What’s Making That Sound ?
C enter for A uditory and A coustic R esearch Representation of Timbre in the Auditory System Shihab A. Shamma Center for Auditory and Acoustic Research.
On the Preprocessing and Postprocessing of HRTF Individualization Based on Sparse Representation of Anthropometric Features Jianjun HE, Woon-Seng Gan,
On the Preprocessing and Postprocessing of HRTF Individualization Based on Sparse Representation of Anthropometric Features Jianjun HE, Woon-Seng Gan,
Improved 3D Sound Delivered to Headphones Using Wavelets By Ozlem KALINLI EE-Systems University of Southern California December 4, 2003.
The auditory system Romain Brette Romain Brette Ecole Normale Supérieure.
Biologically-inspired Visual Landmark Navigation for Mobile Robots
LINKWITZ LAB What are the On-axis & Off-axis
Applied Psychoacoustics Lecture: Binaural Hearing Jonas Braasch Jens Blauert.
Issac Garcia-Munoz Senior Thesis Electrical Engineering Advisor: Pietro Perona.
Virtual Worlds: Audio and Other Senses. VR Worlds: Output Overview Visual Displays: –Visual depth cues –Properties –Kinds: monitor, projection, head-based,
3-D Sound and Spatial Audio MUS_TECH 348. Main Types of Errors Front-back reversals Angle error Some Experimental Results Most front-back errors are front-to-back.
Sounds in a reverberant room can interfere with the direct sound source. The normal hearing (NH) auditory system has a mechanism by which the echoes, or.
THE MANIFOLDS OF SPATIAL HEARING Ramani Duraiswami | Vikas C. Raykar Perceptual Interfaces and Reality Lab University of Maryland, College park.
Week 6: (March 15, 2011) Auditory Maps and orienting: need for Coordinate Transformations.
Audio Systems Survey of Methods for Modelling Sound Propagation in Interactive Virtual Environments Ben Tagger Andriana Machaira.
Modal Analysis of Rigid Microphone Arrays using Boundary Elements Fabio Kaiser.
3-D Sound and Spatial Audio MUS_TECH 348. Physical Modeling Problem: Can we model the physical acoustics of the directional hearing system and thereby.
L INKWITZ L AB S e n s i b l e R e p r o d u c t i o n & R e c o r d i n g o f A u d i t o r y S c e n e s Hearing Spatial Detail in Stereo Recordings.
Jens Blauert, Bochum Binaural Hearing and Human Sound Localization.
Figures for Chapter 14 Binaural and bilateral issues Dillon (2001) Hearing Aids.
This research was supported by Delphi Automotive Systems
Perceptual Analysis of Talking Avatar Head Movements: A Quantitative Perspective Xiaohan Ma, Binh H. Le, and Zhigang Deng Department of Computer Science.
Human Detection and Localization of Sounds in Complex Environments W.M. Hartmann Physics - Astronomy Michigan State University QRTV, UN/ECE/WP-29 Washington,
Boston University--Harvard University--University of Illinois--University of Maryland Distributed Sensor Fields and Uncertainty: Bio-mimetic Methods for.
A Model of Binaural Processing Based on Tree-Structure Filter-Bank
Auditory Neuroscience 1 Spatial Hearing Systems Biology Doctoral Training Program Physiology course Prof. Jan Schnupp HowYourBrainWorks.net.
3-D Sound and Spatial Audio MUS_TECH 348. Are IID and ITD sufficient for localization? No, consider the “Cone of Confusion”
On the manifolds of spatial hearing
PSYC Auditory Science Spatial Hearing Chris Plack.
Feel the beat: using cross-modal rhythm to integrate perception of objects, others, and self Paul Fitzpatrick and Artur M. Arsenio CSAIL, MIT.
Fletcher’s band-widening experiment (1940)
Boston University--Harvard University--University of Illinois--University of Maryland Distributed Sensing, Control, and Uncertainty (Maryland Overview)
SPATIAL HEARING Ability to locate the direction of a sound. Ability to locate the direction of a sound. Localization: In free field Localization: In free.
Fundamentals of Sensation and Perception
3-D Sound and Spatial Audio MUS_TECH 348. What do these terms mean? Both terms are very general. “3-D sound” usually implies the perception of point sources.
Sound Localization and Binaural Hearing
Auditory Localization in Rooms: Acoustic Analysis and Behavior
PSYCHOACOUSTICS A branch of psychophysics
Sound 101: What is it, Why is it, Where is it?
Kocaeli University Introduction to Engineering Applications
Volume 62, Issue 1, Pages (April 2009)
Distributed Sensing, Control, and Uncertainty
Hearing Spatial Detail
Volume 62, Issue 1, Pages (April 2009)
Localizing Sounds.
3 primary cues for auditory localization: Interaural time difference (ITD) Interaural intensity difference Directional transfer function.
Localization behavior in ITD-alone condition and head turn analysis.
Week 13: Neurobiology of Hearing Part 2
Presentation transcript:

42nd IEEE Conference on Decision and Control, Dec 9-12, 2003 A Biomimetic Apparatus for Sound Source Localization (A System Identification Problem from Lord Rayleigh) Amir A. Handzel¹, Sean B. Andersson², Martha Gebremichael³ and P. S. Krishnaprasad University of Maryland, College Park Department of Electrical and Computer Engineering & Institute for Systems Research 42nd IEEE Conference on Decision and Control, Dec 9-12, 2003 ¹ Beyond Genomics Inc. ² Division of Engineering and Applied Sciences, Harvard ³ Microsoft Research

Lord Rayleigh Around 1876, Rayleigh was creating the theory of sound. He was interested in the physics of sound propagation and the perception of sound. 1842-1919 “The problem of the whispering gallery”, Philosophical Magazine, pp 1001-1004, 1910

Sound following

Outline Sound Source Localization Model-free approaches A model and an identification problem Ambiguity and determining elevation References Demonstration Thanks to Shihab Shamma for inspiration. Vinay Shah did recent measurements and demos.

References A. A. Handzel and P. S. Krishnaprasad. “Bio-mimetic sound source localization”, IEEE Sensors Journal, 2(6), 607-617, 2002 A. A. Handzel, S. B. Andersson, M. Gebremichael, and P. S. Krishnaprasad. “A bio-mimetic apparatus for sound source localization”, Proc. 42nd IEEE Conf. on Decision and Control, December, 2003. S. B. Andersson, A.A. Handzel, V. Shah and P. S. Krishnaprasad. “Robot phonotaxis with dynamic sound source localization”, submitted (November 2003).

Barn Owl and Robot Can we capture the barn owl’s auditory acuity in a binaural robot? 2 degrees ~ microseconds resolution

Sound Localization in Nature Localization: spatial aspect of auditory sense Sensory organ arrangement: Vision -- spatial “topographic” Audition -- tonotopic, transduction to sound pressure in frequency bands special computation required, performed in dedicated brainstem circuits and cortex

Acoustic Cues for Localization Binaural/Inter-aural Level/Intensity Difference (ILD/IID) Time/Phase Difference (ITD/IPD) On-set difference/precedence effect Monaural: spectral-directional filtering by Pinna, mostly for elevation

Place Theory (L. Jeffress) J. Comp. Physiol. & Psychol Jeffress model and schematic of brainstem auditory circuits for detection of interaural time (ITD) differences; from Carr & Amagai (1996)

Stereausis (S. Shamma et. al.) J. Acoust. Soc. Am. (1989) 86:989-1006 AVCN Ipsi- center contra- lateral lateral Yj Characteristic frequency Ckk +1 Ckk Ckk -1 Xi Cij Characteristic frequency Ipsi-lateral cochlea AVCN Sound Contara- lateral cochlea or

Initial Motivation The above approaches are static, and do not take account of motion. But psychophysical experiments inspired by early suggestions of Hans Wallach (1940), show active horizontal head rotations improve localization, break inter-aural symmetry, and thus provide information on elevation (Perret & Noble 1997, Wightman & Kistler 1999). Formulating this as an identification problem leads to insight and algorithms. We will show how (head) movement can be helpful is resolving ambiguities. Applications arise in guiding a robot towards an acoustic source. First, for a model, we turn to Rayleigh.

Lord Rayleigh and Binaural Perception ILD and ITD both needed for azimuth. (What about elevation?) Rayleigh set up the problem of sound propagation around an acoustically hard sphere. He introduced head-related transfer functions (HRTF). HRTF computed from sound pressure field generated at a point on the sphere by a point source located at (θ, φ). For a sphere, one has to solve the Helmholtz equation. See section 385 of The Theory of Sound Edition

Coordinate Systems - Azimuthal q - Polar f - Elevation f - Azimuth Microphones at poles on horizontal plane

Pressure field proportional to Static Solution Pressure field proportional to Does not depend on azimuthal angle (f) Head Related Transfer Function (HRTF) Numerical (e.g. FMP), and empirical methods for non-spherical heads

A representation of HRTF information

Some Related Work on HRTFs Blauert, J. (1997). Spatial Hearing (Revised Edition) (MIT Press, Cambridge, MA). C. P. Brown and R. O. Duda (1998). “A structural model for binaural sound synthesis”, IEEE Trans. Speech and Audio Proc., 6(5):476-488. Duda, R. O. (1995). “Estimating azimuth and elevation from the interaural head related transfer function,” in Binaural and Spatial Hearing, R. Gilkey and T. Anderson, Eds. (Lawrence Erlbaum Associates, Hillsdale, N.J.) R. O. Duda

Feature Plane (cylinder) and Signatures ILD & IPD constitute an intermediate computational space for localization. At each frequency a source gives rise to a point in the ILD-IPD plane (cylinder). A (broadband) point source imprints a signature curve on this feature plane (cylinder) according to its location.

Symmetry of Static Localization Sound pressure and resulting inter-aural functions depend only on polar angle; azimuth invariant -- SO(2) symmetry Sources on same circle of directions have identical signatures. Hence the localization confusion. Introduce distance measures.

Symmetry and Rotations - Azimuthal q - Polar f - Elevation f - Azimuth

Breaking the Symmetry Azimuthal invariance, but polar rotations do change the localization functions. Key mathematical step: infinitesimal rotations act as derivative operator -- generate vector fields on signatures. Derivatives ‘modulated’ by Cos(f) -- thus elevation extracted from horizontal rotation! (Head movement helps)

Experimental Results Broad band source - sum of pure tones 43 Hz – 11 KHz in steps of 43Hz. Passed through anti-aliasing filter and sampled at 22KHz. Knowles FG-3329 microphones used on head of 22.6 cm maximum diameter. To determine ILD and IPD, each 512 point segment (23 ms) of data was passed through an FFT. Measured IPD and ILD were smoothed by a nine-point moving average. This yields empirically determined (discrete) signature curves on ILD-IPD space. Localization computations based on minimizing distance functions. Implementation of this step on mobile robot achieved as a table lookup.

Pumpkin head side-view (left) and top view (right). Minimum diameter 19 cm and maximum diameter 22.6 cm.

Plot on left displays smoothed ILD against theoretical ILD for source at 17.5 degrees in horizontal plane. Plot on right shows smoothed IPD against theoretical IPD for same source.

Plot on left shows distance functions for source at 15 deg and 17.5 deg. Plot on right shows distance functions for source at 72.5 deg and 75 deg.

Performance plots for IPD-ILD algorithm (left) and traditional ITD algorithm (right)

Implications and Applications Psychophysics: auditory displays, auditory component of virtual environments and hearing aids. Bio-mimetic active robot head References: A. A. Handzel and P. S. Krishnaprasad. “Bio-mimetic sound source localization”, IEEE Sensors Journal, 2(6), 607-617, 2002 A. A. Handzel, S. B. Andersson, M. Gebremichael, and P. S. Krishnaprasad. “A bio-mimetic apparatus for sound source localization”, Proc. 42nd IEEE Conf. on Decision and Control, Dec. 2003. S. B. Andersson, A.A. Handzel, V. Shah and P. S. Krishnaprasad. “Robot phonotaxis with dynamic sound source localization”, submitted (November 2003).

Front Back Demo Without front-back distinction With front-back distinction