On Wed, Jan 26, 2011 at 05:33:05AM +0000, Augustine Leudar wrote:

> First, don't try and send HTML to this list, as you can see it will
> be removed.
> 
> Sorry - I  dont know what you mean - as far as  Iknow  I haven't sent any
> html to this list (at least not intentionally) - I assume you don't mean
> links as there was no link in my original message and
>  there are also many links in the messages on this list.

No, links are perfectly OK. But your messages (as they arrive
to readers of the list) end like this:

An HTML attachment was scrubbed...
URL: 
<https://mail.music.vt.edu/mailman/private/sursound/attachments/20110126/b08d7757/attachment.html>

which means your mail program is sending an HTML copy of your
text as well. Most mailing lists have a 'no HTML' policy, and
some, like this one, will actually remove it.
 
> I only got so far but as I understand it it uses sound
> pressure levels and phase differences to plot x,y,z spherical coordinates
> which are then reconstituted in the decoding - out of curiosity Why the need
> for the w coordinate - cant the sound pressure level be gleaned from the x,y
> and z ?

If they were just coordinates then you could find w from x,y,z.
But they are something different. 

Image the sound field created by some source, e.g. a sine wave
reproduced by a speaker. At each point in space there two physical
quantities involved:

* The pressure is going up and down at the frequency of the signal.
  This the W signal, which will be produced by an omnidirectional
  microphone.
* To make the pressure variations possible at all, some air must move. 
  Image a tiny volume of air around some point. It will be moving 
  back and forth. The velocity of this movement is a sine wave, and it
  will (in normal cases) be in phase with the pressure signal. But it
  also has a *direction* - it is a vector in 3-D space. So it can be
  split as the sum of three components along some chosen x,y,z axes.
  These components are the X,Y,Z signals. A figure-of-8 microphone
  produces a signal equal to the projection of the velocity vector
  on its (the microphone's) axis. Three of them are required to have
  the full vector.
  
For a single source at sufficient distance, the magnitude of the 
velocity is proportional to the pressure, they are in phase, and
the direction of the velocity points towards the source. Pressure
and velocity are closely related in that case, but even then you
can't find W from X,Y,Z alone. The reason is that if you put the
source in the exactly opposite direction, and invert its phase,
then it will produce exactly the same X,Y,Z - the inversion of
the signal and the direction cancel each other. So you need W
(which will be inverted) as a phase reference.

In the more general case, pressure and velocity are independent.
If you have the same signal reproduced by different sources
(or just a single source and a reflection on a wall) you get
standing waves, and there will be points were the pressures
cancel but velocity does not and vice versa.

Now pressure and velocity are the only two physical quantities
that define a sound field. The basic idea of Ambisonics is that
if you can somehow reconstruct them at the listener's position,
the the listener must hear the same thing as for the real source.

First order Ambisonics (the version using only W,X,Y,Z) can do
this in a small area, where 'small' is expressed in wavelengths.
So this area is large for low frequencies and gets smaller as
frequency goes up. Above a few hundred Hz it has become so small
that it is even too small for single listener. So for mid and
high frequencies, Ambisonics will use a different strategy which
is based on psychoacoustics. For installations where the listening
area has to be much larger, the 'psychoacoustic' strategy is 
normally used at all frequencies. 

Higher order Ambisonics (using more signals) can reconstruct the
sound field in a larger area. It still needs the psychoacoustic
approach for all but the LF range, but this improves as well as
order goes up. At third order and above it becomes very close to
VBAP in fact, with only the speakers nearest to a source direction
contributing in any significant way. The difference with VBAP is
that Ambisonics will not treat the directions corresponding to
the speakers as 'special'. This helps to conceal the speakers
being the source of the reproduced sound.

Ciao,

-- 
FA

There are three of them, and Alleline.

_______________________________________________
Sursound mailing list
Sursound@music.vt.edu
https://mail.music.vt.edu/mailman/listinfo/sursound

Reply via email to