Thursday 29 th November. Periods 3 + 4. Music Technology A2.

Slides:



Advertisements
Similar presentations
Advanced Mixing Advanced Audio and Mixing with Greg Hill AVGenius.com Copyright © 2011 AV Genius, LLC
Advertisements

Thursday 4 th October. Periods Music Technology A2.
EE2F2 - Music Technology 4. Effects. Effects (FX) Effects are applied to modify sounds in many ways – we will look at some of the more common Effects.
AVQ Automatic Volume and eQqualization control Interactive White Paper v1.6.
Thursday 18 th October. Periods Music Technology A2.
MUSC1010 – WEEK 9 Pro Tools Audio Techniques. ADDITIONAL SOUND SAMPLES You are invited to explore the extensive collection of free sound samples available.
Int 2 Multimedia Revision. Digitised Sound Analogue sound recorded from person, or real instruments.
Microphones The Basics.
EE2F2 - Music Technology 2. Stereo and Multi-track Recording.
MUSC1010 – WEEK 8 Sound Radiation and Pro Tools Audio.
Effects. Dynamic Range Processors Fixed Time Delay Effects Variable Time Delay Effects Reverberation Effects Time and Pitch Changing Effects Distortion.
1 Live Sound Reinforcement Microphones. 2 Live Sound Reinforcement A microphone is a transducer that changes sound waves into electrical signals and there.
I Power Higher Computing Multimedia technology Audio.
EE2F2 - Music Technology 9. Additive Synthesis & Digital Techniques.
EE2F2: Music Technology - Revision Two exam questions Music Recording Technology Mixing & multi-track recording Effects MIDI & Sequencers Virtual Studio.
Chapter 7 Principles of Analog Synthesis and Voltage Control Contents Understanding Musical Sound Electronic Sound Generation Voltage Control Fundamentals.
Audio signal processing. Audio signal processing, sometimes referred to as audio processing. It is the processing of a representation of auditory signals,
SYED SYAHRIL TRADITIONAL MUSICAL INSTRUMENT SIMULATOR FOR GUITAR1.
Input Transducers (Microphones).
EE2F2 - Music Technology 3. Mixing. Mixing Basics In the simplest terms, mixing is just adding two or more sounds together. Of course, things are rarely.
Dr. Tim Collins EE2F2 – Music Technology Dr. Tim Collins
Signal processing and Audio storage Equalization Effect processors Recording and playback.
1 Manipulating Digital Audio. 2 Digital Manipulation  Extremely powerful manipulation techniques  Cut and paste  Filtering  Frequency domain manipulation.
Reverberation parameters and concepts Absorption co-efficient
Using Sound in Games Alex Baumann Outline 3D Spatialization Getting and Editing Sounds Using Sounds in Games Music in Games Example Videos.
So far: Historical overview of speech technology  basic components/goals for systems Quick review of DSP fundamentals Quick overview of pattern recognition.
Live Sound Reinforcement
Microphone types, placements & Scenarios. Polar Patterns For any given scenario you must think; What type of microphone to use? What polar pattern? Where.
Joshua “Rock Star” Jenkins Jeff “Tremolo” Smith Jairo “the boss” Rojas
Audio Production Basics. Pros and Cons of Digital Recording Pros – Much easier for the average person. Much less cumbersome Will not deteriorate over.
1 Live Sound Reinforcement Equalizers and other signal processing equipment.
Thursday 27 th September. Periods Music Technology A2.
ADVANCED RADIO PRODUCTION Books: “Modern Radio Production” by Hausman, Benoit, Messere, & O’Donnell: Chapter 15 Pertemuan 12 Matakuliah: O Dasar-Dasar.
Thursday 13 th September. Periods Music Technology A2.
Mono and Stereo Miking Techniques. Choosing Microphones Limited collection: useful for broad range of applications  Neumannn KM 184’s (desert island.
EE Audio Signals and Systems Effects Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
COMP Representing Sound in a ComputerSound Course book - pages
Mics: The Sonic Lens. Recording Chain A system or “ecosystem” for recording and playback.
Synthesis advanced techniques. Other modules Synthesis would be fairly dull if we were limited to mixing together and filtering a few standard waveforms.
Inspire School of Arts and Science Jim White Compression.
Dynamics Processors Riding the wave. What are dynamics? “In music, dynamics normally refers to the volume of a sound or note” -
Dynamic Range and Dynamic Range Processors
Sound Quality.
Sound. Why does sound matter? Sound Design is extremely important to the success of a show and for the audience to have a “total” experience. Think of.
Games Development Practices Sound Effects & OpenAL CO2301 Games Development 1 Week 15.
Chapter 15 Recording and Editing Sound. 2Practical PC 5 th Edition Chapter 15 Getting Started In this Chapter, you will learn: − How sound capability.
Inspire School of Arts and Science Jim White. What is Reverb? Reverb or ‘reverberation’ is not simply just an effect which makes vocals sound nice! It.
The microphone is your primary tool in the sound chain from sound source to audio storage medium.
Revision CUS30109 Certificate III in music. Microphones - Condenser w phantom power - Dynamic - What each is used for - Polar patterns/ frequency response.
Virtual Instruments and Sequencing. Proses Pembuatan Lagu/Musik TrackingMixingMastering.
Signal Processing (time-based effects)
By Cheyenne Morgan Unit 1- Set up and operate a DAW.
EE Audio Signals and Systems Room Acoustics Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Loudness level (phon) An equal-loudness contour is a measure of sound pressure (dB SPL), over the frequency spectrum, for which a listener perceives a.
Effects. Effects in Music n All music that is recorded or amplified relies on effects to enhance certain characteristics of the sound. n Guitarists typically.
Digital Audio II. Volume Scaling The volume of a digitized audio source may be increased or decreased by multiplying each sample value by an appropriate.
Year 13 Music Technology Listening Exam - Kind of!
Microphones. How Microphones Work Sound is created when a vibrating object (such as a guitar string, drum skin etc..) causes the air around it to vibrate.
Microphones National 5/Higher Music Technology Kincorth Academy.
Time Based Processors. Reverb source:
Today’s lesson Discovering and understanding compression Working on club dance portfolios.
Objective Sound and Psychoacoustics. Viers on Location Sound Gathering production sound Nat, b-roll, interview, dialog “Technique will trump technology”
Chapter 15 Recording and Editing Sound
The final layer uses a swirling pad timbre to play block chords
Fairmont Park Baptist Church Technical Support Ministry
EE2F2: Music Technology - Revision
Audio Review RTV 440 S18.
Games Development Practices Sound Effects
“Music Technology from Scratch” Pages
Processing.
Presentation transcript:

Thursday 29 th November. Periods Music Technology A2

A2 exam written question revision ‘The development of music technology’ Must answer 1 question (choice of 2) 16 marks available (8% of A2) Include 16 different points Organise ideas in date order. After planning, answer can be bullet points or prose

A2 exam written question revision ‘The development of music technology’ Topics: Synthesisers Drum machines Samplers Audio effects Mixers audio processing – EQ / compression / expanders / gates / filters MIDI Recording media (tape / digital etc). Consumer media (Vinyl & MP3 etc) Multi-track recording (Digital and Analogue) Computer based recording – Cubase / Logic etc Electric Guitars and Amplification. Digital Synthesis / FM / Additive / Wavetable / Sample based

Sound wave

What is Audio Processing……?

Audio Processing: The ‘alteration’ or changing of a sound signal. Analogue processors operate directly on the electrical signal, while digital processors operate mathematically on the digital representation of that signal. Types: Effects (already covered) EQ (already covered) Compression / Expansion Gates Filters (already covered) More complex DSP processors like Pitch shift / Time stretch / Selective noise removal / Multiband compression.

Audio Processing: Compression ‘ reduces the volume of loud sounds or amplifies quiet sounds by narrowing or "compressing" an audio signal's dynamic range.’ Experiment using compression and use your ear to judge the effect – don’t just change the controls to what you think is correct. What do the controls do? What is ‘soft knee’ / ‘hard knee’ ? What is ‘ducking’? What is side chaining? What is de-essing? What is limiting?

Audio Processing: Gates ‘A noise gate is used when the level of the 'signal' is above the level of the 'noise‘ threshold. The threshold is set just above the level of the 'noise' and so when there is no 'signal' the gate is closed (no sound). A noise gate does not remove noise from the signal. When the gate is open both the signal and the noise will pass through but the signal in theory will be louder than the noise and will mask it partially. Gates typically feature the controls: 'attack', 'release', ‘threshold’ and 'hold' settings.’

Audio Processing: Expanders ‘Expanding’ is a type of gating – it increases the dynamic range. The opposite to compression. Unlike a compressor, the expander reduces gain for signals below the threshold (like a gate). The ratio still defines output change verses input change. A compressor keeps the loud parts from getting too loud, an expander makes the quiet parts quieter.

Audio Processing: More complex DSP. Multiband compression. Basically splits up an audio signal into ‘bands’ using filters. Ie 3 bands would commonly be the low, mid and high frequencies. The 3 bands are then compressed separately using conventional compressors. The 3 compressed signals are then combined or mixed together to re- form the original audio but in a compressed form. scgawUDt0oM Commonly used in mastering and broadcast (radio / tv etc) where very high (low dynamic range with no headroom) audio levels are needed.

Homework: ‘Reverberation has been used to enhance music since the birth of recording. Describe the different techniques and technologies that have been used to apply reverberation to recordings from the 1920s to the present day.’ Research the above exam question and make 20 valid points and explain each point. We will answer the question together in the next lesson and get maximum marks! Negative comments will be given for not doing the homework.

Reverb is added to recordings that have been made using close-mic techniques (1) or recordings that have been made in anechoic chambers / dry spaces (1). This is so that reverb levels can be controlled by an engineer or performer (1). Reverb is an auxiliary effect (1). Reverb time / RT60 (1) Reverb is often found in guitar amplifiers / FX pedals / synth in-built FX (1). Reverb can be a dedicated hardware / rackmount unit (1). Reverb can be added with a software plug-in (1). Natural reverberation (1) Captured using ambient mic’ing (1) techniques. More reverb is captured if a microphone is further away (1). Omni (1) or fig-8 (1) microphones pick up more reverb than cardioids. Ambient mic’s are added to close mic’ed signals and blended together (1). Award for suitable example of ambient mic’ing (1), eg spaced omnis, A- B pair, co-incident pair, X-Y pair, ORTF, M-S, Blumlein stereo, Decca tree, binaural, Sound field, PZM, boundary. Room acoustics determine the nature of the reverb (1) Echo chambers (1) Up until the 1950’s (1) reverb was made by playing a dry signal in a room (1). This was achieved by placing a loudspeaker at one end of a room (1) and an omni- directional (1) microphone at the other (1). The height and angle of the microphones affect the amount of reverb (1). The room would be constructed from reflective material like stone or concrete to increase the amount of reverb (1). The room needs to be sound-proof to prevent noise from getting in (1). The signal from this microphone would then be added to the dry mix (1). Often this reverb is Controlled in the mix with a keyed gate (1). Some studios used garages or stairwells (1) Plate reverb (1) In 1950’s (1), the EMT 140 Reverberation Unit (1) was released. Though a plate reverb unit is large, it is smaller than a dedicated echo chamber (1). It was very expensive (1). The unit consists of a large metal sheet (1). The sound is input by a speaker-like transducer attached to the sheet (1), causing the sheet to vibrate (1). The reflections from the edge of the sheet are picked-up by microphone-like transducers (1). Two microphone-like transducers are used for stereo (1). The reverb time is controlled by a damping pad (1) which is used to absorb the reflections more quickly (1). Plate reverbs give a very dense reverb (1). Even today, plate reverbs 6MT04_ are often used on vocals (1). ‘Reverberation has been used to enhance music since the birth of recording. Describe the different techniques and technologies that have been used to apply reverberation to recordings from the 1920s to the present day.’

Spring reverb (1)Also developed in the 1950’s (1). The spring reverb works in a similar fashion to the plate reverb (1) except a spring is used instead of a metal sheet. The main advantage of the spring reverb is that it’s a lot smaller(1). Spring reverb has a characteristic twangy sound (1). When a spring reverb is knocked it creates a thundering sound (1). Spring reverbs tend to have a mid-range boost (1) and a long reverb time (1) The reverb time can be adjusted by the tension of the spring (1). Digital reverb (1) Digital reverb appeared in recording studios in the 1980’s (1). Yamaha (1) and Lexicon (1) were the pioneer manufacturers. A signal is input into a computer algorithm (1) which consists of many delays /echos (1) with feedback (1) and filtering (1) which emulate the reverb characteristics of real spaces (1). The main advantage of digital reverb is that the character of the reverb can be easily changed by adjusting parameters on-screen (1), or even via MIDI (1). Award max (1) for naming a parameter in addition to reverb time. Also the early reflections are more prominent than plate or spring reverbs (1) so they sound more realistic (1). Digital reverb is very heavy on the CPU (1). As CPUs got quicker, the realism of the reverb improved (1). In the late 1990’s (1) digital reverbs became software plug-ins in sequencing packages. Convolution/Sampling reverb (1) Convolution reverb became a reality with increased processing power (1). It appeared in studios in the 2000’s (1). Sony pioneered this type of reverb (1). Convolution reverb takes a sample of the reverb of a space (1) called an impulse response (1). This is created by emitting an excitation/stimulus into the room (1). Such signals include a sine sweep/balloon pop/gun shoot (1). Frequently, the excitation signal must later be de-convolved from the impulse response (1) or cut from the start of the impulse response (1). It is then convolved with the dry signal to produce the reverb. It allows a sound to be placed in a specific venue without having to actually visit it (1), or to recreate specific hardware units (1). Award an example of a plug-in eg Space Designer (1). ‘Reverberation has been used to enhance music since the birth of recording. Describe the different techniques and technologies that have been used to apply reverberation to recordings from the 1920s to the present day.’

Homework (Write this down please) Prepare for a test on all the keywords you have written down in your glossary this term. The test is next week.