Streaming David Meredith Aalborg University. Sequential integration The connection of parts of an auditory spectrum over time to form concurrent streams.

Slides:



Advertisements
Similar presentations
Terms By: Katie Romano. A Ambit- the range of pitches Arch-form- symmetric in time & climaxes in the middle Attack- initial growth of sound Avant-garde-
Advertisements

The Pythagorean perception of music Music was considered as a strictly mathematical discipline, handling with number relationships, ratios and proportions.
Chapter 2: Rhythm and Pitch
Music Introduction to Humanities. Music chapter 9 Music is one of the most powerful of the arts partly because sounds – more than any other sensory stimulus.
Auditory scene analysis 2
Parts and Layers Electronic Counterpoint is built up in layers
Handel Close Analysis and cadences. Cadences A cadence is a type of musical ending. It’s like a full stop or comma at the end of a sentence. There are.
Chapter 1 Music Notation.
Synthesis. What is synthesis? Broad definition: the combining of separate elements or substances to form a coherent whole. (
A.Diederich– International University Bremen – USC – MMM – Spring Auditory scene.
Identifying Frequencies. Terms: LoudnessPitchTimbre.
1 All About Sound Rob Shaffer Stoklosa Middle School, Lowell June 11, 2013 Note: this lesson utilizes the Audacity computer program to create and analyze.
Pitch Perception.
ALL MUSIC HAS VALUE TO SOMEBODY. What is Music? The Organization of Sound in Time.
A.Diederich – International University Bremen – USC – MMM – Spring 2005 Development of music perception Dowling, W.J. (1999). Development of music perception.
Pitch-spelling algorithms David Meredith Aalborg University.
Music Perception. Why music perception? 1. Found in all cultures - listening to music is a universal activity. 2. Interesting from a developmental point.
Music and shape: Exploring cross-modal representation Examples, a model and analyses Professor Adam Ockelford University of Roehampton 11th March 2010.
MUSIC 1000A Lecture 2 Review and announcements Attend more than one concert before the concert report is due. Course objective Introduction to effect.
Chapter 6 (Sections ) Sound. The speed of sound in a substance depends on: the mass of its constituent atoms, and the strength of the forces between.
Auditory Scene Analysis (ASA). Auditory Demonstrations Albert S. Bregman / Pierre A. Ahad “Demonstration of Auditory Scene Analysis, The perceptual Organisation.
A.Diederich – International University Bremen – USC – MMM – Spring 2005 Scales Roederer, Chapter 5, pp. 171 – 181 Cook, Chapter 14, pp. 177 – 185 Cook,
A.Diederich– International University Bremen – USC – MMM – Spring 5 1 The Perception of Frequency cont'd.
A.Diederich – International University Bremen – USC – MMM – Spring 2005 Rhythm and timing  Clarke, E.F. Rhythm and timing in music. In Deutsch, D. Chapter.
Timbre (pronounced like: Tamber) pure tones are very rare a single note on a musical instrument is a superposition (i.e. several things one on top of.
Auditory Scene Analysis. Sequential vs. Simultaneous Organisation Sequential grouping involves connecting components over time to form streams Simultaneous.
Tone and Voice: A Derivation of the Rules of Voice- Leading from Perceptual Principles DAVID HURON Music Perception 2001, 19, 1-64.
AUDITORY PERCEPTION Pitch Perception Localization Auditory Scene Analysis.
A.Diederich– International University Bremen – USC – MMM – Spring Onset and offset Sounds that stop and start at different times tend to be produced.
What are harmonics? Superposition of two (or more) frequencies yields a complex wave with a fundamental frequency.
Streaming David Meredith Aalborg University. Sequential integration The connection of parts of an auditory spectrum over time to form concurrent streams.
Grouping David Meredith Aalborg University. Musical grouping structure Listeners automatically chunk or “segment” music into structural units of various.
Introduction to algorithmic models of music cognition David Meredith Aalborg University.
Beats and Tuning Pitch recognition Physics of Music PHY103.
Metre David Meredith Aalborg University. Theories of musical metre A theory of musical metre should be able to predict – points in time when listeners.
Audio Scene Analysis and Music Cognitive Elements of Music Listening
SOUND IN THE WORLD AROUND US. OVERVIEW OF QUESTIONS What makes it possible to tell where a sound is coming from in space? When we are listening to a number.
Hearing in Time Slides used for talk to accompany Roger go to yellow three … Sarah Hawkins* and Antje Heinrich** *Centre for Music and Science, University.
Area of Study 03: Texture and Melody “Melody” AQA GCSE Music.
Hearing: auditory coding mechanisms. Harmonics/ Fundamentals ● Recall: most tones are complex tones, consisting of multiple pure tones ● The lowest frequency.
Melodic Organization Motive Rhythmic Motive Melodic Motive
Tone and Voice: A Derivation of the Rules of Voice- leading from Perceptual Principles David Huron David Huron Music Perception, Vol. 19, No. 1 (2001)
Chapter 3 Scales and Melody.
What’s that scale?? 1 Note Grades should be available on some computer somewhere. The numbers are based on the total number of correct answers, so 100%
Music and the Mind Cognitive Elements of Music Listening Kevin D. Donohue Databeam Associate Professor Electrical and Computer Engineering University.
MUSC 1000 Intro to Music MWF10-10:50. Some General Questions: What is Music? Where do we listen to music? Are there any composers or bands you know or.
A Compression-Based Model of Musical Learning David Meredith DMRN+7, Queen Mary University of London, 18 December 2012.
Pitch Perception Or, what happens to the sound from the air outside your head to your brain….
Helen Jane Long Biography, Composition History, and Listening Guides Megan Koplin - Music 1010 Semester Project - 07/08/2013.
When the Brain is attending a cocktail party When the Brain is attending a cocktail party Rossitza Draganova.
Acoustic Illusions & Sound Segregation Reading Assignments for Quizette 3 Chapters 10 & 11.
Introduction to psycho-acoustics: Some basic auditory attributes For audio demonstrations, click on any loudspeaker icons you see....
Communication Process Trading Symbols.
Timbre and Memory An experiment for the musical mind Emily Yang Yu Music 151, 2008.
The Overtone Series Derivation of Tonic Triad – Tonal Model Timbre
Sound Questions.
Computing with Music: Creating Musical Patterns Campy, 2005 Quinnipiac University Computing with Music: Creating Musical Patterns John Peterson No Longer.
Chapter 12 Preview Objectives The Production of Sound Waves
Audio Scene Analysis and Music Cognitive Elements of Music Listening Kevin D. Donohue Databeam Professor Electrical and Computer Engineering University.
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Part I Elements.
Physics Mrs. Dimler SOUND.  Every sound wave begins with a vibrating object, such as the vibrating prong of a tuning fork. Tuning fork and air molecules.
GESTALT PRINCIPLES OF VISUAL DESIGN Gestalt principles of visual design are derived from the theories of gestalt psychology, which were applied to the.
Musical Structure. Sound As the pressure of air changes rapidly, the ear- brain detects the changes as sound (sound waves/vibrations). Some sounds are.
YEAR 11 MATHS REVISION Box Plots Cumulative Frequency with Box Plots.
Tone and Voice: A Derivation of the Rules of Voice-Leading from Perceptual Principles By David Huron Presentation by: Aaron Yang.
SOUND.
Precedence-based speech segregation in a virtual auditory environment
Auditory Illusions Phy103 Physics of Music Fall 2005
Integrating Segmentation and Similarity in Melodic Analysis
Speech Perception (acoustic cues)
Presentation transcript:

Streaming David Meredith Aalborg University

Sequential integration The connection of parts of an auditory spectrum over time to form concurrent streams (Bregman and Ahad, 1995, p. 7) – e.g., connection of tones played on a single instrument to form a melody – we hear a sound to continue even when it is joined by another sound to form a mixture Sequential integration continues until the sound changes suddenly (e.g., in frequency, timbre, amplitude, location) – cf. Lerdahl and Jackendoff’s GPR 3 in grouping Grouping is segmentation of streams into structural units We usually associate a different stream with each separate sound source – Brain attempts to analyse mixed sound that reaches ear into streams corresponding to various sources Each stream has its own independent rhythm and melody We are better at recognizing patterns and relationships between sounds when they are all in the same stream

Stream segregation in a cycle of six tones (Bregman and Ahad, 1995, p.8, Track 1) Based on experiment by Bregman and Campbell (1971) Six tones, 3 high alternating with 3 low, repeated several times When slow, all six tones integrate into a single stream When fast, split into two streams, one high, one low Hard to hear temporal relationships between high and low tones when played fast

Pattern recognition within and across streams (Bregman and Ahad, 1995, pp , Track 2) In two parts – In first part hear a three-tone standard containing notes in a single stream Then have to listen out for standard in the 6-tone comparison pattern – In second part hear a three-tone standard containing notes from different streams Again listen out for standard in 6-tone comparison Much harder to hear the standard in the comparison when it contains notes from more than one stream

Effect of speed and frequency on stream segregation (Bregman and Ahad, 1995, pp , Track 3) van Noorden (1975, 1977) used a “galloping” pattern consisting of two high tones with a lower tone in between (see above) If middle note is similar enough in pitch and timbre to outer notes, then integrate into a single stream with a galloping rhythm If middle note is different enough in pitch or timbre from outer notes, then splits into two streams, each with an isochronous rhythm Different rhythms and melodies make it easy to tell whether you are hearing 1 or 2 streams If keep frequency difference the same, then can split into two streams by increasing speed – need higher speed for a smaller frequency difference

Effect of repetition on streaming (Bregman and Ahad, 1995, p.13, Track 4) Pattern only splits after you’ve heard a few repetitions If we split a stimulus into multiple streams too easily, we would be too sensitive to changes and have an unstable perception of the auditory scene (Bregman, 1990, p.130) We therefore have a “damped, lazy” response We start out by assuming only one source and only add sources (streams) if there is enough evidence (e.g., lots of tones clustered into two different frequency regions)

Segregation of a melody from distractor tones (Bregman and Ahad, 1995, pp , Track 5) Sequence constructed by interleaving tones of a familiar melody with random distractor tones (see above) – first time you hear the sequence, each distractor tones is within 4 semitones of previous melody tone – in second and subsequent times, distractor tones become further and further away from melody tones Put your hand up when you first recognize the melody! The melody becomes easier to recognize the further away the distractor tones are in pitch Perceptual links between melody notes are stronger when distractor tones are in a different stream

Compound melody In Baroque music (particularly on non-sustaining instruments like the harpsichord), common for a single instrument to play part that rapidly alternates between different pitch ranges Part perceived to segregate into two streams (‘voices’) Known as compound melody or virtual polyphony Example above from Prelude in G major from Book 2 of Bach’s Das Wohltemperirte Klavier – Right hand segregates into two isochronous streams – Also see “fusion” between parallel tenor part and lower virtual part in compound right hand

Streaming in African xylophone music (Bregman and Ahad, 1995, pp , Track 7) Wegner (1990, 1993) identified some interesting instances of sequential integration and stream segregation in Ugandan amadinda music Two players play a repeating cycle of notes, the notes of one player being interleaved with those of the other The combined sequence is isochronous and each part is isochronous But the combined sequence is heard to split into two streams that are not isochronous Also the the two streams heard do not correspond to the separate parts played – each stream contains some notes from one part and some from the other

Segregating the two players’ parts in amadinda music (Bregman and Ahad, 1995, p. 19, Track 8) Can make each part in amadinda music separately audible by transposing one by an octave This puts the two parts in separate streams

Stream segregation based on timbre difference (Bregman and Ahad, 1995, p. 21, Track 10) Stream segregation can also be induced by using tones with different timbre, but the same pitch We assume tones with different timbres come from different sources, so tend to perceive them as belonging to different streams Here, middle tone has different timbre from outer tones, so segregates into a different stream at a moderate speed (despite having same pitch)

Effects of connectedness on segregation (Bregman and Ahad, 1995, pp , Track 12) Gestalt principle of “good continuation” also seems to influence streaming “Good continuation” says that we group elements that lie on a smooth curve Bregman and Dannenbring (1973) showed that connecting tones with glissandi helps to integrate them into the same stream even if they are widely separated in pitch

Effects of streaming on timing judgements (Bregman and Ahad, 1995, pp , Track 13) Saw earlier that it is hard to perceive temporal relationships between tones in different streams Here, galloping pattern has middle, lower tone temporally either exactly half-way between outer tones or slightly after half-way between the two outer tones When frequency difference is small, all tones in one stream, so easy to hear whether middle tone is exactly half way between outer tones When frequency difference is large, much harder to tell whether middle tone is exactly half-way between upper tones Demonstration based on experiment by van Noorden (1975)

Dependence of streaming on context (Bregman and Ahad, 1995, pp , Track 15) AB heard as being in the same stream if XY is far away in frequency AB can be made to be in different streams by bringing XY closer to them in frequency Can tell AB split into different streams because harder to hear AB in right- hand comparison than left-hand comparison Based on experiment by Bregman (1978)

Releasing a two-tone target by capturing interfering tones (Bregman and Ahad, 1995, pp , Track 16) Try to tell whether AB are in the same order in the comparison Hard when comparison has just two flanking tones Easy when comparison preceded by longer sequence of tones that capture flanking tones into a separate stream Need several repetitions to hear Xs as being in a different stream from AB/BA

X-Patterns (Bregman and Ahad, 1995, pp.31-32, Track 17) X-pattern has two interleaved, crossing, isochronous tone sequences, one ascending and one descending Remember if you can easily hear a standard in a comparison, this means the standard is in one stream in the comparison Here, it is harder to hear a complete ascending or descending sequence than a “bouncing” percept – implies integrate lower notes into one stream and upper notes into another stream Can make full ascending or descending sequence easier to hear by giving them different timbres Based on experiment by Tougas and Bregman (1985)

Temperley’s (2001) model of counterpoint Takes a piano-roll input, quantized to beats at the lowest metrical level Since also quantized in pitch domain, can represent as a grid of squares Red bar indicates onset, notes of same pitch cannot overlap Streaming done only using onset time, offset time and MIDI note number Predicts voice to which each note belongs

Temperley’s (2001) model of counterpoint: 4 well-formedness rules CWFR 1: “A stream must consist of a set of temporally contiguous squares on the plane.” (p.97) A stream doesn’t have to span a whole piece or movement

Temperley’s (2001) model of counterpoint: 4 well-formedness rules CWFR 2: “A stream may be only one square wide in the pitch dimension.” (p.98) Note that this means you cannot have a stream of fused chords!

Temperley’s (2001) model of counterpoint: 4 well-formedness rules CWFR 3: “Streams may not cross in pitch.” (p.98) Recall results on X- patterns – We prefer to hear a bouncing percept not a crossing one However, parts do sometimes cross, so maybe this should be a (strong) preference rule

Temperley’s (2001) model of counterpoint: 4 well-formedness rules CWFR 4: “Each note must be entirely included in a single stream.” (p.99) So not possible for part of a note to be in one stream and another part to be in a different stream Though it is possible for a single note to be entirely in more than one stream

Temperley’s (2001) model of counterpoint: 4 preference rules CPR 1 (Pitch proximity rule): “Prefer to avoid large leaps within streams.” (p.100) As discussed earlier (van Noorden, 1975, 1977), large differences in frequency tend to segregate tones into different streams

Temperley’s (2001) model of counterpoint: 4 preference rules CPR 2 (New stream rule): “Prefer to minimize the number of streams.” (p.101) As discussed earlier, we begin by assuming 1 stream and need evidence before hearing two or more streams (Bregman 1990, p. 130)

Temperley’s (2001) model of counterpoint: 4 preference rules CPR 3 (White square rule): “Prefer to minimize the number of white squares in streams.” (p.101) Musical voices frequently contain rests, but if the rests are too long, the connection between the tones becomes very weak Recall results of experiment with frequency glides connecting notes (Bregman and Dannenbring 1973)

Temperley’s (2001) model of counterpoint: 4 preference rules CPR 4 (Collision rule): “Prefer to avoid cases where a single square is included in more than one stream.” (p.101) This only happens rarely Example is tone at crossing point in an X-pattern (Tougas and Bregman 1985)

References Bregman, A. S. (1978). Auditory streaming: Competition among alternative organizations. Perception and Psychophysics, 23, 391–398. Bregman, A. S. (1990). Auditory Scene Analysis: The Perceptual Organization of Sound. MIT Press, Cambridge, MA. Bregman, A. S. and Ahad, P. A. (1995). Demonstrations of auditory scene analysis: The perceptual organization of sound. Audio CD. Bregman, A. S. and Campbell, J. (1971). Primary auditory stream segregation and perception of order in rapid sequences of tones. Journal of Experimental Psychology, 89, 244–49. Bregman, A. S. and Dannenbring, G. (1973). The effect of continuity on auditory stream segregation. Perception and Psychophysics, 13, 308–312. Bregman, A. S. and Rudnicky, A. (1975). Auditory segregation: Stream or streams? Journal of Experimental Psychology: Human Perception and Performance, 1, 263–267. Lerdahl, F. and Jackendoff, R. (1983). A Generative Theory of Tonal Music. MIT Press, Cambridge, MA. Temperley, D. (2001). The Cognition of Basic Musical Structures. MIT Press, Cambridge, MA. Tougas, Y. and Bregman, A. S. (1985). The crossing of auditory streams. Journal of Experimental Psychology: Human Perception and Performance, 11, 788–798. van Noorden, L. P. A. S. (1975). Temporal coherence in the perception of tone sequences. Ph.D. thesis, Eindhoven University of Technology, Eindhoven, The Netherlands. van Noorden, L. P. A. S. (1977). Minimum differences of level and frequency for perceptual fission of tone sequences abab. Journal of the Acoustical Society of America, 61, 1041–1045. Wegner, U. (1990). Xylophonmusik aus Buganda (Ostafrika). Number 1 in Musikbogen: Wege zum Verständnis fremder Musikkulturen. Florian Noetzel Verlag, Wilhelmshaven. (Cassette and book). Wegner, U. (1993). Cognitive aspects of amadinda xylophone music from Buganda: Inherent patterns reconsidered. Ethnomusicology, 37, 201–241.