- From: David Horner <dhorner@apollo.med-web.com>
- Date: Mon, 5 Apr 1999 00:18:20 -0500 (EST)
- To: Henrik Frystyk Nielsen <frystyk@w3.org>
- cc: www-lib@w3.org
However, in LoadToFile... HTLoadFile() is used to grab the requested data. I would like an interface to a stream that I could pull buffers(chunks) off. // In HTLoadFile form // Load data to file. status = HTLoadToFile(getme, request, outputfile); // open file with data from HTLoadToFile if(fopen(outputfile,"r")) { // Deal with data. } // What I would like to see in libwww. // Open a stream to pipe request data to. status = HTLoadToStream(url,output,request); // Read data as it comes. (maybe transfer from stream to chunk) while(somereadfunc(output)) { //Deal with data. } I think that it would be more efficient to just deal with the data as it comes down the stream. This method would reduce the number of complications (file errors, disk i/o errors). Does this exist?? Are my models correct?? Am I way off base? --Dave Horner On Sun, 4 Apr 1999, Henrik Frystyk Nielsen wrote: > > > > Dave wrote: > > > > Is the chunk created by: > > chunk = HTLoadAnchorToChunk(anchor, request); > > The size of the file being requested? > > Is there any way to grab the requested file by lines or by buffer? > > I have looked at the source and it looks like you malloc the whole > > file into memory?! > > Could this be a problem with big files (100's of megs)? > > Yes indeed - and also you have to wait for the whole thing to come in. > If you want to use "progressive rendering" of a document then you have > to use streams instead of the chunk version > > As an example of using streams, the sample app > > http://www.w3.org/Library/Examples/LoadToFile.c > > linked from > > http://www.w3.org/Library/Examples/ > A > in fact uses streams because the HTLoadFile() in > > http://www.w3.org/Library/src/HTAccess.html > > uses streams. > > -- > Henrik Frystyk Nielsen, > World Wide Web Consortium > http://www.w3.org/People/Frystyk >
Received on Monday, 5 April 1999 01:07:43 UTC