- From: Dr. Markus Walther <walther@svox.com>
- Date: Fri, 17 Oct 2008 12:38:56 +0200
Eric Carlson wrote: >> >> Imagine e.g. an audio editor in a browser and the task "play this >> selection of the oscillogram"... >> >> Why should such use cases be left to the Flash 10 crowd >> (http://www.adobe.com/devnet/flash/articles/dynamic_sound_generation.html)? >> >> >> I for one want to see them become possible with open web standards! >> > I am anxious to see audio-related web apps appear too, I just don't > think that including 'start' and 'end' attributes won't make them > significantly easier to write. I did a tiny prototype of the above use case - audio editor in a browser - and it would have been significantly easier to write, if not Apple's Safari had a bad and still unfixed bug in the implementation of 'end' ... (http://bugs.webkit.org/show_bug.cgi?id=19305) > >> In addition, cutting down on number of HTTP transfers is generally >> advocated as a performance booster, so the ability to play sections of a >> larger media file using only client-side means might be of independent >> interest. >> > The 'start' and 'end' attributes, as currently defined in the spec, > only limit the portion of a file that is played - not the portion of a > file that is downloaded. I know that, but for me that's not the issue at all. The issue is _latency_. How long from a user action to audible playback - that's what's relevant to any end user. You can't do responsive audio manipulation in the browser without fast, low-latency client-side computation. All the server-side proposals miss this crucial point. For another use case, consider web-based tools for DJs, for mixing and combining audio clips. There's a lot of clips on the web. But if manipulating them is not realtime enough, people won't care. For another use case, consider web-based games with dynamic audio, etc. Robert O'Callahan wrote: > On Fri, Oct 17, 2008 at 5:24 AM, Dr. Markus Walther <walther at svox.com > <mailto:walther at svox.com>> wrote: > > Imagine e.g. an audio editor in a browser and the task "play this > selection of the oscillogram"... > > Why should such use cases be left to the Flash 10 crowd > (http://www.adobe.com/devnet/flash/articles/dynamic_sound_generation.html)? > > > If people go in that direction they won't be using cue ranges etc, > they'll be using dynamic audio generation, which deserves its own API. And I proposed the beginnings of such an API in several postings on this list under the topic 'audio canvas', but it seemingly met with little interest. Now Flash 10 has some of the things I proposed... maybe that's a louder voice? > OK, in principle you could use <audio> with data:audio/wav, but that > would be crazy. Then again, this is the Web so of course people will do > that. I did exactly that in my tiny audio-editor prototype for proof-of-concept purposes - I guess I must be crazy :-) Actually it was partly a workaround for browser bugginess, see above. Give me an API with at least float getSample(long samplePosition) putSample(long samplePosition, float sampleValue) play(long samplePositionStart, unsigned long numSamples), and sanity will be restored ;-) The current speed race w.r.t. the fastest JavaScript on the planet will then take care of the rest. Silvia Pfeiffer wrote: > > Linking to a specific time point or section in a media file is not > something that needs to be solved by HTML. It is in fact a URI issue > and is being developed by the W3C Media Fragments working group. > > If you use a URI such as http://example.com/mediafile.ogv#time=12-30 > in the src attribute of the video element, you will not even have to > worry about "start" and "end" attributes for the video element. Unless Media Fragments can be a) set dynamically for an already downloaded media file _without triggering re-download_, b) time specification can be accurate to the individual sample for the case of audio, c) W3C finishes this quickly enough and d) browsers take the W3C recommendation seriously, it is not an alternative for my use cases. It's all about dynamic audio and the future. By the time the spec hits the market, static media is not the only thing on the web anymore. Jonas Sicking wrote: > > The problem with relying on cues is that audio plays a lot faster than > we can gurentee that cue-callbacks will happen. So if you for example > create a audio file with a lot of sound effects back to back it is > possible that a fraction of a second of the next sound will play before > the cue-callback is able to stop it. If I understand this correctly, cue callback delay would potentially make it impossible to have precise audio intervals, needed for the above use cases. But _then_ replacing 'start' and 'end' with cue ranges and 'currentTime' is NOT possible, because they are no longer guaranteed to be equivalent in terms of precision. It seems the arguments are converging more towards keeping 'start' and 'end' in the spec. -- Markus
Received on Friday, 17 October 2008 03:38:56 UTC