- From: Dan Lazin <dlazin@google.com>
- Date: Fri, 17 Sep 2021 16:37:55 -0400
- To: Laurent Le Meur <laurent.lemeur@edrlab.org>
- Cc: "public-epub-wg@w3.org" <public-epub-wg@w3.org>
- Message-Id: <492E5079-EB79-439D-98C1-998E1B8139C2@google.com>
Thanks, Laurent! To quickly answer "what does it bring compared to the current implementation for users": 1) "Page numbers" that are consistent across different reading systems (as opposed to each RS having its own proprietary calculation) 2) Guaranteed existence of stable page numbers (as opposed to being dependent on the publisher adding a page-list) Remember, the idea is only that the algorithm is a worst-case fallback. The ideal solution is that every RS implements the algorithm, but no RS ever has to use it because the publishers choose to add page-lists to every book instead. Page-lists also provide consistent cross-device (and likely cross-digital/print) page numbers ... but they're not mandatory. The algorithmic page numbers only come into play if the publisher doesn't add a page-list, so if they don't like the calculated page numbers, it's easy to fix :) > On Sep 17, 2021, at 1:28 PM, Laurent Le Meur <laurent.lemeur@edrlab.org> wrote: > > Following our discussion today, the feedback from Mickaƫl Menu, maintainer of Readium Mobile, about the idea to calculate positions ("calculated page numbers") from the raw text, and not the full html content: > > --- > - It would much more CPU and IO intensive to compute the positions from this, because we'd have to read (potentially streamed / encrypted) each resource to get its full size, compared to just looking at the ZIP metadata in the EPUB case > (nb Laurent: the size of the uncompressed resource is in bytes, not unicode code points). > - It might be quite complicated to match a position in the web view. > - It will likely give a poor experience for image-based resources. > (nb Laurent: I don't see what could give a good experience). > > This algorithm feels nice and less dependent of technical stuff like HTML structure, but what does it bring compared to the current implementation for the users? We will still have a varying number of positions per screen and they won't better match physical books. > > IMHO the TF could focus on improving on the downsides of the current implementations which often depend on the ZIP instead of the resources themselves. Meaning for the same publication, we won't get the same number of positions if the EPUB is exploded, if the compression status of resources change or if an EPUB is streamed from the web. > > One redeeming quality of the strategy suggested by the TF is that if the HTML structure changes but not the actual publication text, then we get the same positions. But not sure it matters in practice. > --- > > Best regards > Laurent > It's much more CPU and IO intensive to compute the positions from this, because we have to read (potentially streamed / encrypted) all the resources, compared to just looking at the ZIP metadata > Like you said, it might be quite complicated to match a position in the web view > It will likely give a poor experience for image-based resources > This algorithm feels nice and less dependent of technical stuff like HTML structure, but what does it bring compared to the current implementation for the users? We will still have a varying number of positions per screen and they won't better match physical books. > 5:26 <https://readium.slack.com/archives/C2JNGCN6R/p1631892401014900> > IMHO the TF could focus on improving on the downsides of the current implementations which often depend on the ZIP instead of the resources themselves. Meaning for the same publication, we won't get the same number of positions if the EPUB is exploded, if the compression status of resources change or if an EPUB is streamed from the web. > 5:27 <https://readium.slack.com/archives/C2JNGCN6R/p1631892476015700> > (You can probably read that I'm biased for the old Readium positions strategy which was using the "original length" of resources instead of the archive entry length ) > 5:29 <https://readium.slack.com/archives/C2JNGCN6R/p1631892542016500> > One redeeming quality of the strategy suggested by the TF is that if the HTML structure changes but not the actual publication text, then we get the same positions. But not sure it matters in practice. > It's much more CPU and IO intensive to compute the positions from this, because we have to read (potentially streamed / encrypted) all the resources, compared to just looking at the ZIP metadata > Like you said, it might be quite complicated to match a position in the web view > It will likely give a poor experience for image-based resources > This algorithm feels nice and less dependent of technical stuff like HTML structure, but what does it bring compared to the current implementation for the users? We will still have a varying number of positions per screen and they won't better match physical books. > 5:26 <https://readium.slack.com/archives/C2JNGCN6R/p1631892401014900> > IMHO the TF could focus on improving on the downsides of the current implementations which often depend on the ZIP instead of the resources themselves. Meaning for the same publication, we won't get the same number of positions if the EPUB is exploded, if the compression status of resources change or if an EPUB is streamed from the web. > 5:27 <https://readium.slack.com/archives/C2JNGCN6R/p1631892476015700> > (You can probably read that I'm biased for the old Readium positions strategy which was using the "original length" of resources instead of the archive entry length ) > 5:29 <https://readium.slack.com/archives/C2JNGCN6R/p1631892542016500> > One redeeming quality of the strategy suggested by the TF is that if the HTML structure changes but not the actual publication text, then we get the same positions. But not sure it matters in practice. >
Received on Friday, 17 September 2021 20:38:11 UTC