Supported by the Danish Composers Society, I recently acquired a compact 10 channel speaker setup, for composing (and performing!) multi channel concerts...
The system consists of 10 small active studio monitors with a 5" kevlar membrane and a 1" dome tweeter. The monitors are bi-amped with a 40W amp for the bass and a 30W amp for the tweeter. It's not a big powerful system, but it's effective, precise and well suited for smaller venues of 60-90m2.
Sound localization is a big subject of research, and here are just a few key concepts. When we localize a sound source, we use a combination of three main indicators.
ILD (Interaural Level Difference)
A sound source will always be closer to one ear than the other (on less the source is on-axis). Since sound level decays with distance the sound level perceived by the first ear hit by the wavefront, will be louder than what is perceived by the second. And the head will also shadow away some of the higher frequencies. This level difference is used to determine the direction of a sound source.
ITD (Interaural Time Difference)
When a sounds wavefront approaches the ears, it will reach the ears with a small time delay, which creates a phase difference. This phase difference is detected by the brain and is used to calculate the direction. We can detect an ITD in the magnitude of tens of microseconds, which gives a directional precision of down to 2 degrees under the right circumstances.
ILD becomes close to useless with wavelengths longer that ear distance times 4, because the head isn't big enough to shade off the wave front at the second ear. ITD becomes useless at wavelengths shorter than ear distance times 2 because the phase difference gets so big that the brain can't calculate if the phase difference is 1/4 wavelength or 5/4 wavelength og 9/4 wavelength or more.
HRTF (Head Related Transfer Function)
The head related transfer function is the frequency "coloration" that happens when a wavefront hits the head, torso, pinnae, hair etc. from a certain angle. The coloration is different depending on th angle so from registering which frequencies are boosted and which are attenuated, the brain leans the directional fingerprint imbedded in the sounds perceived.
So since localizing sound sources are one of the core competences of the auditory system, it is if course, interesting to work with this in a artistic way. Deciding to go with 10 speakers is more of a practical and financial decision rather than an artistic decision, but it still radically expands the possibilities over a stereo or a 5.1 setup. And I found that it is just enough speakers to make an acceptably smooth circular panning. But if I could have had 1000 speakers sphericially distributed around the audience, I would prefer that:-)
Other sounds were made up from very simple basic synthetic waveforms (sines, saws, squares). A rather obvious finding is that sounds that has energy distributed over a broad spectrum, are much easier to locate than sounds with narrow spectral energy. A sine wave, for instance, that per definition only has energy at one frequency, is very hard to source locate. Almost all the spectral cues are gone, so you only have vague cues from ILD and/or ITD depending on the frequency.
On the other hand, short broad spectrum bursts are quite easy and fast to locate.
The patch below is a 10ch panning interface that can be automated for different predicitable or unpredictable patterns.
To be continued...