So far I’ve been really impressed with the speed, quality and file size of OCR in DT3. However, that’s not maintained once I start merging files or manipulating pages, so I’d welcome any comments or suggestions. For the record, I’m working mainly with 300 ppi B&W scans.
Background: I’m a long-time user of DTPO2 to OCR scans from my Brother DCP-7045N. Because this is not a duplex scanner I do a lot of merging within DT, as well as scripted page manipulations in PDFpen. The thing that binds it all together is Apago’s excellent PDF Shrink app, which I have integrated into DT via AppleScript. Any bloat introduced by moving pages around is easily fixed with PDF Shrink. The end result in DTPO2 was good-quality files of sensible size.
Enter DT3 and OCR seems to yield smaller files than before, with good optical quality. The problem is that if I merge files or delete pages, the file size balloons. PDF Shrink can fix this up to a point, but this time with a significant loss in optical quality, to the point where it’s not a workable solution.
I know that OCR and PDF manipulation more generally must be very complicated under the hood, and everyone seems to have their own workflow. But DT3 is increasing in power, as witnessed by Smart Rules starting to replace Hazel and even simple things like the “reverse pages” command (which I used to have to do in PDFpen). It would be nice not to have to depend on so many helper apps. So is it reasonable to hope that a future release will do a better job of keeping file sizes down?