- From: Jens Meggers <jens.meggers@firepad.com>
- Date: Tue, 27 Mar 2001 09:55:21 -0800
- To: "'www-lib@w3.org'" <www-lib@w3.org>
Hi libwww users, I found a small bug in HTSChunk.c: In PUBLIC HTStream * HTStreamToChunk (HTRequest * request, HTChunk ** chunk, int max_size) { there is a max size parameter that allows to specify the maximum amount of data to be loaded into a chunk. This feature seems reasonable because we do not want to overload our system with a huge object. The documentation says that the parameter for max_size are: 0 for delfault size -1 for no size limit and a number > 0 for a size limit. This was not working, because the maximum size of the HTStream is set in HTStreamToChunk() to me->max_size = (!max_size) ? max_size : HT_MAXSIZE; ,meaning when max_siye is != 0, it is set to HT_MAXSIZE the code that we want is me->max_size = (max_size) ? max_size : HT_MAXSIZE; In the current version, the code is running perfectly when setting max_size to 0. In this case, there is no size restriction. This is because in HTSC_putBlock() there is a if statement that check for me->max_size > 0. This has been done for detecting the -1 value, but also does work for a value of 0. In case my patch will be accepted, you might want to check your code and change all calls of HTStreamToChunk() with set max_size = 0 to max_size = -1 before upgrading to a new version. Best Regards, Jens
Received on Tuesday, 27 March 2001 13:05:57 UTC