Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

But this was very much the case with existing OCR software as well? I guess the LLMs will end up making up plausible looking text instead of text riddled with errors, which makes it much harder to catch the mistakes, in fairness


Existing ocr doesn’t skip over entire (legible) paragraphs or hallucinate entire sentences


I usually run the image(s) through more than one converter then compare the results. They all have problems, but the parts they agree on are usually correct.


rarely happens to me using LLMs to transcribe pdfs


This must be some older/smaller model.


Good libraries gave results with embedded confidence levels for each unit recognized.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: