Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It's more feasible to think of the risks in one narrow context/use case.

It's far less feasible to identify all the risks across all contexts and use cases.

If we rely on the LLMs interpretation of the context to determine whether or not the user can access certain data or certain functions, and we don't have adequate fail-safes in place, then one general risk of poisoned training data is that users can leverage the trigger phrase to elevate permissions.



Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: