Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

[flagged]


They aren't wrong, though.

At best, the sources cited by an LLM system would be a listing of the items used for RAG, or other external data sources that were merged into the prompt for the LLM. These items would ideally be appended to the response by a governing system around the LLM itself. I don't know of any major providers that do this right now.

The median case is having the LLM itself generate the text for the citation section, in which case there really is no mechanism tying the content of a citation to the other content generated. IF you're lucky and within the bounds the LLM was trained on, then the citation may be relevant, but the links are generated by the same token prediction mechanism as the rest of the response.


> Can you please at least look at any of the major offerings of the past three years before being both snarky and wrong?

All of the examples on that website are from the last three years.

Can you clarify about how I’m wrong about LLMs not reliably citing sources? Are the 490 examples of made up sources appearing in court filings not valid? Is the link you posted where you asked chatgpt how many people there are (that included a broken link in the sources) valid?

Perplexity for example, kind of famous for citing sources, is currently in litigation over… inventing sources. https://www.wired.com/story/dow-jones-new-york-post-sue-perp...




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: