PSYC 60041 Auditory Science Spatial Hearing Chris Plack.

Slides:



Advertisements
Similar presentations
Auditory Localisation
Advertisements

Sound Localization Superior Olivary Complex. Localization: Limits of Performance Absolute localization: localization of sound without a reference. Humans:
Psychoacoustics Perception of Direction AUD202 Audio and Acoustics Theory.
Binaural Hearing Or now hear this! Upcoming Talk: Isabelle Peretz Musical & Non-musical Brains Nov. 12 noon + Lunch Rm 2068B South Building.
3-D Sound and Spatial Audio MUS_TECH 348. Psychology of Spatial Hearing There are acoustic events that take place in the environment. These can give rise.
Hearing Detection Loudness Localization Scene Analysis Music Speech.
Reflections Diffraction Diffusion Sound Observations Report AUD202 Audio and Acoustics Theory.
WEEK 11 Revision class 1. Assignment Two – 20% Sound Measurements and Observations Due: Week 11 via electronic submission Weighting: 30% Learning.
Localizing Sounds. When we perceive a sound, we often simultaneously perceive the location of that sound. Even new born infants orient their eyes toward.
Minimum Audible Angle Measured in Young and Old CBA Mice Using Prepulse Inhibition of Startle Paul D. Allen, Jordan Bell, Navin Dargani, Catherine A. Moore,
1 Auditory Sensitivity, Masking and Binaural Hearing.
AUDITORY LOCALIZATION Lynn E. Cook, AuD Occupational Audiologist NNMC, Bethesda, MD.
Chapter 6 (Sections ) Sound. The speed of sound in a substance depends on: the mass of its constituent atoms, and the strength of the forces between.
Chapter 6: Masking. Masking Masking: a process in which the threshold of one sound (signal) is raised by the presentation of another sound (masker). Masking.
All you have is a pair of instruments (basilar membranes) that measure air pressure fluctuations over time Localization.
Exam and Assignment Dates Midterm 1 Feb 3 rd and 4 th Midterm 2 March 9 th and 10 th Final April 20 th and 21 st Idea journal assignment is due on last.
Hearing Detection Loudness Localization Scene Analysis Music Speech.
Sensory Systems: Auditory. What do we hear? Sound is a compression wave: When speaker is stationary, the air is uniformly dense Speaker Air Molecules.
Reading Assignment! We’ll discuss the chapter by Gregory in your book on Friday of next week.
Stochastic Properties of Neural Coincidence Detector cells Ram Krips and Miriam Furst.
Sound Transmission and Echolocation Sound transmission –Sound properties –Attenuation Echolocation –Decoding information from echos.
Hearing & Deafness (3) Auditory Localisation
AUDITORY PERCEPTION Pitch Perception Localization Auditory Scene Analysis.
STUDIOS AND LISTENING ROOMS
Spectral centroid 6 harmonics: f0 = 100Hz E.g. 1: Amplitudes: 6; 5.75; 4; 3.2; 2; 1 [(100*6)+(200*5.75)+(300*4)+(400*3.2)+(500*2 )+(600*1)] / = 265.6Hz.
A.Diederich– International University Bremen – USC – MMM – Spring Onset and offset Sounds that stop and start at different times tend to be produced.
Sound source segregation (determination)
Chapter 6: The Human Ear and Voice
Hearing.
Acoustics/Psychoacoustics Huber Ch. 2 Sound and Hearing.
The Auditory System Dr. Kline FSU. What is the physical stimulus for audition? Sound- vibrations of the molecules in a medium like air. The hearing spectrum.
SOUND IN THE WORLD AROUND US. OVERVIEW OF QUESTIONS What makes it possible to tell where a sound is coming from in space? When we are listening to a number.
Applied Psychoacoustics Lecture: Binaural Hearing Jonas Braasch Jens Blauert.
Cosc 6326/Psych6750X Audition and Auditory Displays.
Chapter 12: Auditory Localization and Organization
Sounds in a reverberant room can interfere with the direct sound source. The normal hearing (NH) auditory system has a mechanism by which the echoes, or.
 Space… the sonic frontier. Perception of Direction  Spatial/Binaural Localization  Capability of the two ears to localize a sound source within an.
Chapter 12: Sound Localization and the Auditory Scene.
Chapter 12: Sound Localization and the Auditory Scene.
Audio Systems Survey of Methods for Modelling Sound Propagation in Interactive Virtual Environments Ben Tagger Andriana Machaira.
How Can You Localize Sound? Ponder this: –Imagine digging two trenches in the sand beside a lake so that water can flow into them. Now imagine hanging.
‘Missing Data’ speech recognition in reverberant conditions using binaural interaction Sue Harding, Jon Barker and Guy J. Brown Speech and Hearing Research.
L INKWITZ L AB S e n s i b l e R e p r o d u c t i o n & R e c o r d i n g o f A u d i t o r y S c e n e s Hearing Spatial Detail in Stereo Recordings.
Jens Blauert, Bochum Binaural Hearing and Human Sound Localization.
Figures for Chapter 14 Binaural and bilateral issues Dillon (2001) Hearing Aids.
Hearing Research Center
This research was supported by Delphi Automotive Systems
Human Detection and Localization of Sounds in Complex Environments W.M. Hartmann Physics - Astronomy Michigan State University QRTV, UN/ECE/WP-29 Washington,
Reading Assignment! We’ll discuss the chapter by Gregory in your book on Thursday of next week.
Loudness level (phon) An equal-loudness contour is a measure of sound pressure (dB SPL), over the frequency spectrum, for which a listener perceives a.
Auditory & tactile displays EGR 412 Human Factors Engineering ISE
Listeners weighting of cues for lateral angle: The duplex theory of sound localization revisited E. A. MacPherson & J. C. Middlebrooks (2002) HST. 723.
Introduction to psycho-acoustics: Some basic auditory attributes For audio demonstrations, click on any loudspeaker icons you see....
Hearing Detection Loudness Localization Scene Analysis Music Speech.
3-D Sound and Spatial Audio MUS_TECH 348. Are IID and ITD sufficient for localization? No, consider the “Cone of Confusion”
Fletcher’s band-widening experiment (1940)
The role of reverberation in release from masking due to spatial separation of sources for speech identification Gerald Kidd, Jr. et al. Acta Acustica.
SPATIAL HEARING Ability to locate the direction of a sound. Ability to locate the direction of a sound. Localization: In free field Localization: In free.
Fundamentals of Sensation and Perception
3-D Sound and Spatial Audio MUS_TECH 348. What do these terms mean? Both terms are very general. “3-D sound” usually implies the perception of point sources.
Sound Localization and Binaural Hearing
Loudness level (phon) An equal-loudness contour is a measure of sound pressure (dB SPL), over the frequency spectrum, for which a listener perceives a.
PSYCHOACOUSTICS A branch of psychophysics
Precedence-based speech segregation in a virtual auditory environment
Loudness level (phon) An equal-loudness contour is a measure of sound pressure (dB SPL), over the frequency spectrum, for which a listener perceives a.
What is stereophony? Stereos = solid (having dimensions: length width, height) Phonics = study of sound stereophony (stereo) is an aural illusion – a.
Hearing Spatial Detail
Speech Perception (acoustic cues)
Localizing Sounds.
3 primary cues for auditory localization: Interaural time difference (ITD) Interaural intensity difference Directional transfer function.
Presentation transcript:

PSYC Auditory Science Spatial Hearing Chris Plack

Spatial Hearing Learning Outcomes –Understand the main cues used to localise sounds: IIDs and ITDs –Understand other cues to localisation including head movements and pinna effects –Understand the BMLD –Understand what is meant by “the precedence effect” and how it is measured

Some Definitions Binaural hearing - hearing with both ears. Localisation - the ability to locate a sound source in the environment. Lateralisation - the ability to localise in the head a sound presented over headphones.

Localisation

Localisation Our ears give us much less information about object location than our eyes. We have only got two spatial channels for hearing (our two ears) compared to arguably several million for vision (the receptors in each retina). However, we can hear sound sources that are beyond our line of sight (e.g., behind our heads), and this helps us to orient attention and can be important for survival.

Binaural Cues Our two ears can be used to localise sounds. A sound will tend to be more intense, and arrive earlier, at the ear closest to the sound source. Hence, we can determine the direction of a sound source based on: Interaural Intensity Differences (IIDs - also called ILDs, interaural level diferences) Interaural Time Differences (ITDs)

Intensity Cues Mainly because of the shadowing effect of the head, a sound to the right will be more intense in the right ear than in the left ear: LR

Interaural Intensity Differences The sound reaching the ear farthest from the source is less intense due to head shadowing mainly, and also to dissipation of intensity with distance according to inverse-square law (only useful for sounds close to head). Low-frequency sounds diffract around the head, high- frequency sounds don't, and thus a high- frequency shadow is cast over the farthest ear. Hence the IID is frequency dependent, being greatest for high frequencies.

Timing Cues Because sound travels relatively slowly (330 m/s), a sound from the right will arrive perceptibly earlier at the right ear than at the left ear: LR LR Time 1Time 2

Interaural Time Differences The interaural distance (approx. 23 cm) produces a maximum ITD of about 0.69 ms when the source is directly opposite one ear (90 o ). The ITD falls to zero as the source moves forward or backwards to be in front (0 o ) or behind (180 o ) the listener. Smallest detectable ITD is about 0.01 ms!

Ambiguities in ITDs For a continuous pure tone, ambiguities arise when the period of the tone is less than twice the ITD - closest peaks in the waveform may suggest wrong ear is leading. For a sound directly to the side, this occurs for a frequency of about 735 Hz. Ambiguities can be resolved if the tone is modulated, i.e., if there are envelope cues (including abrupt onsets).

Ambiguities in ITDs

Duplex Theory The duplex theory suggests that sound localisation is based on interaural time differences at low frequencies and interaural intensity differences at high frequencies. However, for fluctuating high-frequency signals the envelope can carry good timing information. It is now thought that, for most sounds (which have wideband spectra), ITDs may dominate at all frequencies.

Minimum Audible Angle Indicates the smallest change in sound source position that can be detected by the listener. Using sinusoidal signals, the MAA is smallest for frontal signals (1 o for frequencies below 1 kHz). Around 1.5 kHz the IIDs are small and ITDs become ambiguous resulting in an increase in MAA. Performance worsens markedly as the source moves away from a frontal position, but in the real world the listener can move their head!

The Cone of Confusion Interaural time and intensity differences are ambiguous. For example, we can’t tell the difference between a sound directly in front and a sound directly behind using IIDs or ITDs. Same IIDs and ITDs for sound source on surface of cone:

The Cone of Confusion Ambiguities can be resolved by: Head movements Spectral effects of pinna, head, and torso reflections

Head Movements …but what if sound is too brief? a) ITD = 0:b) ITD shows left leading: ? ?

Effects of Pinna The pinna modifies sound entering ear depending on direction, resolving ambiguities and providing cues to source elevation: 10 dB 15 ° - 15° Frequency (kHz)

Because of shape of concha, sounds at higher elevations have shorter reflected path lengths, hence notch at a higher frequency: Effects of Pinna

Hebrank & Wight (1974) JASA 56, p. 1829

Accurate vertical localisation only with broadband signals (and only those with energy > 4 kHz). Vertical localisation prevented by occlusion of the convolutions in the pinnae (horizontal localisation unaffected apart from front/back distinctions). Vertical localisation almost as good with single ear. Vertical localisation sensitive to manipulations of the source spectrum. Middlebrooks & Green (1991) Ann. Rev. Psychol. 42, p. 135 Evidence for Importance of Spectral Cues in Vertical Localisation

Distance Perception Loudness is an important cue for distance: In the direct field (little reverberation) the further away the source is, the quieter the sound. Better with familiar sounds (e.g. speech). Direct-to-reverberant energy ratio is another cue: The closer the sound the louder the direct sound will be compared with the early reflected sounds. e.g. Zahorik (2002) JASA 111, p 1832

Binaural Unmasking

Binaural Masking Level Difference Measure tone (signal) threshold in presence of broadband masker with identical signal and masker to both ears. Invert phase of tone (or masker) in one ear so that signal and masker are lateralised differently. Masked signal threshold is lower (binaural release from masking). The difference between in-phase and altered-phase thresholds is called the BMLD.

N o S o - masker & signal same phase at both ears - poor detection N o S π - masker same phase, signal π radians out of phase - good detection N m S m - masker and signal presented monaurally - poor detection N o S m - masker same to both ears, signal monaural - good detection

The BMLD is frequency dependent since it relies on ITDs:

ConditionBMLD (dB) N u S  3 N u S o 4 N  S m 6 N o S m 9 N  S o 13 N o S  15 N = noise masker, S = signal, u = uncorrelated noise, o = no phase shift, m = monaural,  = 180° phase shift

Huggins Pitch Present the same noise to both ears over headphones - noise is lateralised to the centre of the head. Now decorrelate a narrow band of noise between the ears (so that the band is different between the ears). This band “pops out” and is heard as having a pitch corresponding to the centre frequency of the band: Huggins pitch.

Frequency (Hz) Level Same in both ears Decorrelated between the ears NOISE 500 Huggins Pitch

Gockel, Carlyon, and Plack (2010). Can Huggins pitch harmonics be combined with diotic pure tone harmonics to produce a residue pitch? Mixed-mode conditions (1 HP + 1 NBN): Single-mode conditions (2 HP or 2 NBN):

Present two successive pairs of harmonics. Does pitch change follow analytic (spectral) or synthetic (residue) pitch? FrequenciesF0 (Hz)Harmonic Numbers st & 2 nd nd & 3 rd th & 5 th

Response of listeners for mixed-mode and single-mode conditions highly correlated: Suggests Huggins and diotic harmonics are processed by the same mechanism and combined after MSO.

The Precedence Effect

In a reverberant space (such as a room) sound from a source reflects off the walls, and arrives at the ear from different directions. Why don’t these reflections confuse the auditory system?

The Precedence Effect Direct sound follows shortest path and arrives first. The auditory system takes advantage of this by restricting analysis to the sound arriving first. I.e. the first arriving wavefront takes precedence.

The Precedence Effect For example, click from two loudspeakers separated by 80 o. If click is simultaneous, then heard between loudspeakers. Delay imposed on left loudspeaker. For 0-1 ms delay, sound image moves to right loudspeaker. For delays of 1-30 ms, image localised at right loudspeaker with no contribution from left (precedence effect). For delays > ms effect breaks up, and a direct sound and echo are heard.

Virtual Auditory Space

Sounds presented over headphones tend to be lateralised inside the head. However, if we record sounds using two microphones in the ear canals (or in the ear canals of a “dummy head”) then when this recording is presented over headphones it seems external and can be localised outside the head. The cues from the pinna, head, and torso help to give a recording a spacious quality when presented over headphones.

Dummy Head Recordings

Spatial Hearing Learning Outcomes –Understand the main cues used to localise sounds: IIDs and ITDs –Understand other cues to localisation including head movements and pinna effects –Understand the BMLD –Understand what is meant by “the precedence effect” and how it is measured