AudioPannerNode and Spatialization

Hello list,

I get back to you to get some information about the AudioPannerNode (and 
more widely about spatialization).
At Ircam, one of the research team works on spatialization, and I have 
been asked to help building an interface from HRTF files.
For what we understood, the AudioPannerNode is
- a panning effect
- a distance related sound attenuation
- a beam directivity

1. Panning effect
The panning effect seems to use HRTF filters, and we have some audio 
sample libraries with those kind of filters (based on the shape of the 
user) :
     a. is it a 'default' human body which is used for rendering in 
AudioPannerNode?
     b. how could we use our own HRTF impulse files ?

2. Distance attenuation
For distance attenuation, in our model, the distance affects also the 
spectrum ( sources closer will typically boost a low frequency).
    a. how isit implemented in web audio api ?
    b. Is there a way to achieve this kind of rendering using 
AudioPannerNode ?

3. 'Beam' (or Sound is may be better word for that) directivity
We would like to understand the way it has been implemented, is it a 
lowpass filter first or second order ?
In our case (implemented in a sofware called 'the spat') the directive 
beam interacts with a room effect ( through ConvolverNode for instance). 
Is there a way to achieve this also ?

Thanks for all your anwsers, (we would like to test our spatialization 
effects (and models) through the web audio api, to have rich end user 
experiences).

Regards,





-- 
Samuel Goldszmidt
IRCAM - APM / CRI
01 44 78 14 78

Received on Thursday, 12 July 2012 10:55:01 UTC