Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Careful with this idea, I had someone take a thread we were engaged in and feed it to an LLM, asking it to confirm his feelings about the conversation, only to post it back to the group thread. It was used to attack me personally in a public space.

Fortunately

1. The person was transparent about it, even posting a link to the chat session

2. They had to follow on prompt to really engage the sycophancy

3. The forum admins stepped in to speak to this individual even before I was aware of it

I actually did what you suggested, fed everything back into another LLM, but did so with various prompts to test things out. The responses where... interesting, the positive prompt did return something quite good. A (paraphrased) quote from it

"LLMs are a powerful rhetorical tool. Bringing one to a online discussion is like bringing a gun to a knife fight."

That being said, how you prompt will get you wildly different responses from the same (other) inputs. I was able to get it to sycophant my (not actually) hurt feelings.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: