- From: Samuel Goldszmidt <samuel.goldszmidt@ircam.fr>
- Date: Thu, 12 Jul 2012 12:54:21 +0200
- To: public-audio@w3.org
- Message-ID: <4FFEACDD.7040407@ircam.fr>
Hello list, I get back to you to get some information about the AudioPannerNode (and more widely about spatialization). At Ircam, one of the research team works on spatialization, and I have been asked to help building an interface from HRTF files. For what we understood, the AudioPannerNode is - a panning effect - a distance related sound attenuation - a beam directivity 1. Panning effect The panning effect seems to use HRTF filters, and we have some audio sample libraries with those kind of filters (based on the shape of the user) : a. is it a 'default' human body which is used for rendering in AudioPannerNode? b. how could we use our own HRTF impulse files ? 2. Distance attenuation For distance attenuation, in our model, the distance affects also the spectrum ( sources closer will typically boost a low frequency). a. how isit implemented in web audio api ? b. Is there a way to achieve this kind of rendering using AudioPannerNode ? 3. 'Beam' (or Sound is may be better word for that) directivity We would like to understand the way it has been implemented, is it a lowpass filter first or second order ? In our case (implemented in a sofware called 'the spat') the directive beam interacts with a room effect ( through ConvolverNode for instance). Is there a way to achieve this also ? Thanks for all your anwsers, (we would like to test our spatialization effects (and models) through the web audio api, to have rich end user experiences). Regards, -- Samuel Goldszmidt IRCAM - APM / CRI 01 44 78 14 78
Received on Thursday, 12 July 2012 10:55:01 UTC