Homework 2: Radio Play

due on 11/25 (Tue) in the Homework Factory and in your /Library/220a/hw2/ directory

Overview

Compose a short 'musical radio play' (1-2 min) consisting of sounds spatialized into four channels which can be listened to binaurally via your web page (with headphones). Place an HTML file hw2.html that links to your code and sound files which are in your /Library/Web/220a/hw2/ subdirectory. Work in quad, convert the final version to ambisonic format and submit it as a web page which plays it via an ambisonic-to-binaural decoder. We'll use OmniTone for that. Make sure that your submission is timestamped on the Homework Factory.

Background

Binaural recording uses stereo mics pointed outwards from inside your ears to capture as closely as possible the exact sound pressure waves entering your ear canals. A binaural radio play produced by the BBC, The Revenge, demonstrates the possibilities. The binaural technique captures filter (transfer function) differences caused by body parts shadowing and reflecting sounds arriving from various directions: the ear flaps (pinnae), head, shoulders, etc. Played back over headphones or earbuds, binaural preserves the interaural loudness difference (ILD) and interaural time difference (ITD) cues which are basic to sound localization.

Early work in binaural recording was accompanied by predictions that its superior imaging would create a world where everyone would eventually listen through headphones. Playing binaurally-encoded sounds over stereo loudspeakers doesn't result in either good binaural or good stereo and that's one thing holding back wider use. For a position paper, see Jens Blauert's AES Heyser Lecture. He makes a provocative case for binaural as a part of an increasingly realistic synthetic world.

One artist whose work leverages the medium is Janet Cardiff. She composes site-specific 3D audio narratives with spine-tingling interplay of real and phantom presences binaurally-produced. Her telephonecall from SFMOMA 2001 is a benchmark piece which opens up the possibilities of what you might expect to compose with mobile devices. The composition led participants through the gallery, each directed by a camcorder which they were holding while watching with a pre-recorded self-guided tour. You'd turn a corner and someone would be there singing in the space (convincingly, so you could point to them) only they weren't there then, but at some other point in time, past, alternative present, or future.

Our approach starts with composing spatialized sound for the quad speaker arrangement. Pick a short text which might be a monologue, group dialog, or whatever you want, but it should constitute some sort of a script (feel free to write it from scratch). You'll first record your own voice (with your new mic) and then possibly other voices in combination depending on the text to be read.

Sound Panning

Here are two tutorial examples of spatialization - one will simulate rain falling throughout the stereo field, and the other will give you a template by which you can 'move' a sound from one point in stereo space to another. Walk through this tutorial. Its second example will require you to use either this sound file (.aiff format) or the same thing (.wav format). It depends what you write in the .read("choose_whichever_name_here") function call of the SndBuf object that you instantiate to read in your input soundfile. Replace "choose_whichever_name_here" with its proper path and name.

Composition - Spatial Sound

Requirements

  • There must be at least three sound sources that have some aspect of localization in them. These can be more atmospheric (e.g. wind that moves throughout the entirety of the radioplay), or more related directly to smaller objects of interest (e.g. voices that speak from different points in space, an instrument that has a particular 'home' or direction in space).
  • One of the sound sources must move. This can have direct meaning in the play (e.g. panning a monophonic recording of a car to create the illusion of it moving from the left to the right channels, or perhaps a bee or an insect that moves), or the meaning can be more abstract (e.g. some melodic string that oscillates between some group of channels).
  • You should conceptualize your radio play/creation in four channels - as if your listener had two speakers in front and two behind them, set up in a square (i.e. quad setup). The technique will spatialize sound in a horizontal plane anywhere around the listener. We will be listening to it as binaural audio when played from the browser. For technical instructions on how to utilize these four channels, and then mix then render in binaural, follow these instructions: How-to: convert to Ambisonics for Binaural Playback from OmniTone (2016)

Considerations

  • If you like, you can make this a purely literal radio play - with the tracks all recorded using a microphone (the one you can build in the MaxLab would be a great choice), with a focus on dialogue and more literal elements of a play (e.g. closing doors, etc.). However, you could also 'musicify' it, (e.g. a conversation between Peter Pan and Tinkerbell, where Tinkerbell is a sound/melody you synthesize - perhaps she flies around in the sound space! Or, perhaps you're simulating a rehearsal, where a director/conductor takes the center of the sound stage from one side, and rehearses a couple of instruments on different sides of the stage.) It's entirely up to you to select what it'll be.
  • If you create/use a function that plays a sound file and spork it at different times during your radio play, you could potentially add many sound effects and elements to add sonic interest to your creation.
  • Techniques you can use include localization with interaural intensity difference (IID), interaural time difference (ITD), Schroeder-style reverb (e.g., NRev), and processing for time and pitch transposition (e.g., SndBuf rate and PitShift)