View Single Post
Old 11-08-2021, 11:40 PM   #11
Tex2002ans
Wizard
Tex2002ans ought to be getting tired of karma fortunes by now.Tex2002ans ought to be getting tired of karma fortunes by now.Tex2002ans ought to be getting tired of karma fortunes by now.Tex2002ans ought to be getting tired of karma fortunes by now.Tex2002ans ought to be getting tired of karma fortunes by now.Tex2002ans ought to be getting tired of karma fortunes by now.Tex2002ans ought to be getting tired of karma fortunes by now.Tex2002ans ought to be getting tired of karma fortunes by now.Tex2002ans ought to be getting tired of karma fortunes by now.Tex2002ans ought to be getting tired of karma fortunes by now.Tex2002ans ought to be getting tired of karma fortunes by now.
 
Posts: 2,306
Karma: 13057279
Join Date: Jul 2012
Device: Kobo Forma, Nook
Quote:
Originally Posted by anonlivros View Post
I'm glad the tutorial continues to help more people. Thanks for share, Tex!


Quote:
Originally Posted by anonlivros View Post
3) I also strongly recommend using Scantailor for dewarp and crop, generating B&W images. This step must be prior to abbyy finereader.
ScanTailor Advanced.

And the steps would go:
  • Images/Scans
    • Follow the anonlivros tutorial (and/or DIY Book Scanner forums)
  • ScanTailor Advanced
    • Crop + Dewarp + Grayscale/B&W
  • ABBYY Finereader (or Tesseract)
    • Optical Character Recognition (OCR)

This would get you digitized PDFs with text backend, which would allow you to search/copy/paste, etc.

From there, you can do the usual PDF->EPUB steps. (Which are their own in-depth workflows.)

Quote:
Originally Posted by anonlivros View Post
I know, I know... The tutorial needs to be updated.

I've learned some interesting things over the last few months, especially from sharing experience with Tex. (Which I reserve a gratitude that I couldn't properly express in this post)


I didn't check on the Github to see if it was expanded since.

And we definitely have to meetup and talk again before the end of the year!

Quote:
Originally Posted by Quoth View Post
BUT my 2002 vintage flatbed scanner with optional ADF is close to 30 M pixels at 600 dpi for a larger book. The advantage of a camera is a V shaped holder to avoid damaging the spine. Pirates and professional scanning of cheap common books cut off the spine and use a duplex ADF.
Any of these methods work, but I'd rank everything as a balance of these 4 categories:

Non-Destructive->Destructive
  • Camera
  • Scanner*
  • Feed Scanner

High->Low Quality
  • Scanner
  • Feed Scanner
  • Camera*

Fast->Slow Speed
  • Camera
  • Feed Scanner
  • Scanner

High->Low Labor
  • Scanner
  • Camera
  • Feed Scanner

Non-Destructive vs. Destructive

If the book must stay in-tact, then Scanner or Camera.

If you don't mind destroying the book, then cutting the spine off + feeding it into the scanner as a stack of paper saves tons of time+labor.

* Note: If you have very fragile/large books, the Scanner may be too rough on the spine, so your only non-destructive choice is Camera + V-shaped holder.

Quality

The most important, because every later stage depends on this.

Remember, digitizing books is an entire process, and getting pictures of pages is just Stage 1.

High quality input:
  • High DPI (300/600+) + great/even lighting + non-warped pages
  • = very accurate OCR + little image cleanup

Low quality images:
  • Low DPI + bad lighting (haloing) + warped pages
  • = OCR has more errors + more typos/formatting issues + fixing images takes longer

You may even have to redo a lot of your work when you stumble across a non-recoverable error later on. (Like a photograph/chart/graph being distorted beyond fixing... or horrible speckling that only appears when you try to B&W your image.)

* Note: 20+ years ago, most cameras were still too low DPI. The images may have been readable to a human, but feeding that lower-quality image through OCR, you'd get a much higher error rate.

Within the past 10 years though, the quality of your typical cellphone camera has dramatically improved. Now, everyone carries something in their pockets that may work "well enough".

This is where anonlivros's gooseneck+cellphone method comes into play. It's an extremely cheap (<$50) way of reusing materials you most likely already have (cellphone + lamp)... getting you 80%+ of the way there.

* * *

Recommendation: If doing conversion professionally though, I'd recommend a superb quality scanning company like Golden Images. He was recommended by Hitch+me quite a few times over the years.

The immaculate quality scans will save you all that time+labor in the long-run.

Higher quality input = better+faster conversion with less errors.

* * *

Speed

Cameras, when you get the workflow down, can take a few seconds per page.

Scanners, at higher DPI, are very slow.

Labor

Scanners+Cameras require you to turn pages, make sure everything is lined up, hold books down, no fingers in the way, etc.

Feed scanners, you stick in a stack of paper and can go off doing something else.

Quote:
Originally Posted by Quoth View Post
Google uses a camera and the V shaped book holder. However much of their output is only good enough for "Search" and thus poor quality.
It's quite good. And all the raw images are still there as JPEG2000. They are much higher quality than the auto-generated B&W/Color PDFs.

A lot of this was discussed a few months ago in:

2021: "Archive.org ePub"

I even showed the difference between the auto-generated Archive.org "EPUB"s vs. EPUBs generated out right out of Finereader with minimal intervention.

Note: Although again, you can rapidly get 80%+ of the way there in your PDF->EPUB, but that final quality push is what takes up the majority of time:
  • Headings + TOCs
  • Splitting chapters
  • Formatting
    • bold/italics + super-/sub-script
    • blockquotes/poetry/indexes/scenebreaks
    • Alignment (left/center/right)
  • Fixing tables/figures/images
  • Clickable footnotes
  • Typo corrections
  • [...]

99.99% text accuracy also seems fantastic... until you actually read a book with .01% errors in it. (At least a few errors on every page.)

Quote:
Originally Posted by Quoth View Post
A lot is on the Internet Archive and all of it as search results in Google Books. They don't bother with proofing. They also ought to have lost the court case because they are scanning and storing complete copyright works without the copyright holder's permission.
Fair Use. It's a library. And you don't "need permission".

There was even an article on that court case a few days ago:

Techdirt.com: "Internet Archive Would Like To Know What The Association Of American Publishers Is Hiding"

and a few days before that, on the ridiculousness of "library ebook licenses":

Techdirt.com: "Publishers Want To Make Ebooks More Expensive And Harder To Lend For Libraries; Ron Wyden And Anna Eshoo Have Questions"

Quote:
A UK campaign to fight that development in the world of academic publishing, called #ebookSOS, spells out the problems. Ebooks are frequently unavailable to institutions to license as ebooks. When they are on offer, they can be ten or more times the cost of the same paper book. [...] One title cost £29.99 as a physical book, and £1,306.32 for a single-user ebook license. As if those prices weren't high enough, it's common for publishers to raise the cost with no warning, and to withdraw ebook licenses already purchased. [...]

Quote:
Publishers are increasingly offering titles via an etextbook model, via third party companies, licensing content for use by specific, very restricted, cohorts of students on an annual basis. Quotes for these are usually hundreds, or sometimes thousands, times more than a print title, and this must be paid each year for new cohorts of students to gain access. This is exclusionary, restricts interdisciplinary research, and is unsustainable.
On Controlled Digital Lending (CDL), I'd also recommend checking out these fantastic webinars given by Internet Archive a few months ago:

Libraries already use these methods + Interlibrary Loans to loan scans of their own works (especially fragile/rare/non-public-facing material).

This allows you to consolidate it, instead of every individual library having to (poorly) re-scan + re-digitize.

Similarly with the Blind/Low-Vision/Disabilities associations + universities within each country all independently digitizing. Why? Why waste all that time and effort when you can do it once, in high-quality PDFs/EPUBs, then lend it from there?

On universities digitizing, see the great webinar:

On Copyright/Libraries + "Permission Culture" see:

Side Note: Luckily, Internet Archive has been archiving scholarly articles as well:

And speaking of digitizing + Link Rot... ~18% websites cited within scholarly articles are already dead:

This knowledge needs to be preserved + easily searchable/accessible.

Last edited by Tex2002ans; 11-09-2021 at 06:09 PM.
Tex2002ans is offline   Reply With Quote