Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

LLM/VLM-based OCR is highly prone to hallucination - the model does not know when it can’t read a text, it can’t estimate its own confidence, and it deals with fuzzy/unclear texts by simply making things up. I would be very nervous using it for anything critical.


There are really amazing products coming


I’ll believe it when I see it.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: