Why LLMs suck at OCR (Pulse AI Blog)
Original link: Why LLMs suck at OCR (Pulse AI Blog)
Related: Ingesting Millions of PDFs and why Gemini 2.0 Changes Everything
My notes
This week, there was a viral blog about Gemini 2.0 being used for complex PDF parsing, leading many to the same hypothesis we had nearly a year ago at this point. Data ingestion is a multistep pipeline, and maintaining confidence from these nondeterministic outputs over millions of pages is a problem.
When an LLM processes a document image, it first embeds it into a high-dimensional vector space through the attention mechanism. This transformation is lossy by design.
As amazing as LLM results appear to be, I can’t help but wonder why is this even a direction that we are headed in? Shouldn’t traditional computer vision be a better option given how spatially oriented PDFs are?
I have tried NanoNets before. I don’t know what their tech stack is but it seems to be the most reliable solution out there right now. I guess this Pulse AI might be just as good but I can’t see a free trial on their website so can’t try it.
A few months ago I investigated CSV cleaning with LLMs and good lord that is an unsolved problem. I thought it’d be an easy one to solve but I guess containing the LLM within the guardrails isn’t that easy.
I am sure someone somewhere will solve both these problems in another year or so. That would be fun. Data cleaning will finally be a solved problem.