Quote:
Originally Posted by jhowell
Parsing the HTML files that make up the book in spine sequence to generate a list of image names in order of reference would probably work best. Those would correspond between AZW3 and KFX formats in most cases, although there may be some circumstances where the lists might need to be tweaked a bit. Those lists could be used as the basis for substituting the proper files by renaming the ones from KFX to match the AZW3 names.
|
Ah, that's what I thought. So not as trivial as I envisioned, though I could probably construct a regexp easily enough given a single HTML input (perhaps through Calibre's htmlz file format), then remove all duplicated entries? It's almost not worth the hassle, even if it's for archival and future-proofing purposes -- I'll have to give it some further thought. I never realized things were so disparately different until I looked at Amazon's details section which provided the estimated file size. Seeing that the KFX output roughly matches what Amazon reported, I then started web searching which lead me to this forum thread.
As I care more about the semantics for reading purposes, assume that these low-resolution images are human-viewable for the majority of screen sizes and image types? I mean it's roughly half the size for each image, e.g., 22.8 vs. 10.4 MB for 206 images, hopefully the pictures, maps, hand-written material, etc are not too blurry? It would be nice to have the best of both worlds, but if it's not easily feasible then maybe I'd call it a day and switch to a different reading platform in future which others have also considered here...
Thanks for taking the time to answer. It is highly appreciated.