July 22nd, 2024

The serious science of trolling LLMs

Trolling large language models manipulates responses for attention. Vendors address troll transcripts to maintain human-like illusions for commercial use. Trolling reveals LLM limitations, evolving into a scientific pursuit for model understanding.

Read original articleLink Icon
The serious science of trolling LLMs

The article discusses the practice of trolling large language models (LLMs) by manipulating prompts to generate nonsensical or outrageous responses for social media attention. While some view this as a pointless activity since LLMs do not learn from mistakes like humans, the author argues that vendors invest resources in addressing viral troll transcripts to maintain the illusion of human-like responses. The deceptive nature of making LLMs appear more human is crucial for commercial applications, where customers expect transparency about the models' limitations. Trolling is seen as a way to reveal the shortcomings of LLMs and differentiate reasoning from data recall. The author suggests that trolling is evolving into a legitimate scientific pursuit, shedding light on the true capabilities and limitations of these models. The article emphasizes the importance of understanding when and how LLMs fail to ensure their responsible use in various applications.

Link Icon 7 comments
By @talldayo - 9 months
> It’s when it fails at a simple task that we know what the limitations are — and trolls are the torch-bearers of this new englightenment.

I really don't agree. Watching AI fail spectacularly is a great bread and circus, but identifying something's deficiencies is not the same as fixing it. The thing these trolls get so damningly right is their denouncement of AI; we can't rely on shit text-generators that fly by the seat of their pants. The insinuated "but it's useful and does real work" shtick is over; I don't even see people talking about RAG value-add anymore.

The bubble popped, and now expectations are returning to normal. It's less of an enlightenment period and more of the "AI winter" that we've always known was going to occur once we exhausted the current scaling limits. The trolls are scratching beneath the surface and proving the prophecy right.

By @cptcobalt - 9 months
I wanted more examples of trolls and how the behavior degrades. The one scripted joke example doesn't land with a punch.

> In short, I believe that the revealed objectives of LLM vendors differ starkly from what they say. They pour considerable resources into responding to every trollish LLM transcript that goes viral on the internet — and it’s useful to ponder why.

I'd be pretty interested in reading at how people do poke at LLMs and what this actually reveals about the vendors priorities.

By @geon - 9 months
Generative AI is basically like playing the game where each participant says one word of a sentence, one after the other.

https://youtu.be/kNJNV0GXO-s?si=3TKUpFoFHV7FNSHR

The AI has no idea where it is going before it starts outputting. If it ends up with a sentence that can’t be completed sensibly, it will output nonsense.

I think a better model needs to be able to plan out the idea to convey in advance and possibly backtrack.

By @emilamlom - 9 months
I am curious what the author would think about LLM based AI like Neuro-sama (AI vtuber that streams on twitch). Different use-case than the llm's the author was "trolling", but much better at humor and more resilient to trolls than usual.
By @thih9 - 9 months
Summarizing, we’re dealing with closed source software coupled with an AI model trained on unspecified data. Vendors glorify the latter and never say how much work is handled by the former.

Trust seems a big factor here.

By @ysofunny - 9 months
> unlike humans, language models can’t be humiliated and do not learn from mistakes.

wut. I was pretty sure they do learn from mistakes. but they call it 'training'