Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Doesn't this just mean that the LLM ingested training data where people talk about banning controversial propaganda type of newspapers while nobody talks about banning nyt or wapo?


I think if people took the time to understand how LLMs choose word weights based on training data, they would understand that these results are somewhat deterministic.

Instead, the preferred heuristic is to look for a bogeyman.


That's a scary proposition as it means the most vocal side wins the AI battle even if their position is bad or even if it is morally evil.

If one side is constantly vocalizing their position while the other side remains silent, then the vocal side wins.


It's almost like hoovering up the internet, feeding it into a computer, and then treating it like it's Deep Thought is a bad idea.


Garbage in, garbage out. This is why training data is important.


It's almost certainly the RLHF, not the base model.


But the base model, when its trained on the whole internet, will have some extreme biases on topics where there's a large and vocal group on one side and the other side is very silent. So RLHF is the attempt to correct for the biases on the internet.


> So RLHF is the attempt to correct for the biases on the internet.

...or it can be used to reinforce a specific ideology. Completely dependent on who does the RLHF and what their motivations are.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: