ChatGPT Isn't 'Hallucinating'–It's Bullshitting – Scientific American
AI chatbots like ChatGPT can generate false information, termed as "bullshitting" by authors to clarify responsibility and prevent misconceptions. Accurate terminology is crucial for understanding AI technology's impact.
Read original articleThe article discusses the use of AI chatbots like ChatGPT and highlights the issue of these programs generating false information, which the authors argue should be termed as "bullshitting" rather than "hallucinating." They explain that these chatbots, based on large language models, aim to produce human-like text without necessarily caring about the truthfulness of the information. The authors emphasize that referring to AI-generated falsehoods as "hallucinations" is misleading and can impact public understanding of the technology. They suggest that using accurate terminology is crucial for clarifying responsibilities when such technology is used in critical areas like healthcare. By labeling ChatGPT as a "bullshit machine," the authors aim to prevent anthropomorphizing these chatbots and ensure accountability for the information they generate. The article underscores the importance of language in shaping perceptions of technology and the need to accurately describe AI capabilities to avoid misconceptions and misplaced trust.
Related
Lessons About the Human Mind from Artificial Intelligence
In 2022, a Google engineer claimed AI chatbot LaMDA was self-aware, but further scrutiny revealed it mimicked human-like responses without true understanding. This incident underscores AI limitations in comprehension and originality.
The Encyclopedia Project, or How to Know in the Age of AI
Artificial intelligence challenges information reliability online, blurring real and fake content. An anecdote underscores the necessity of trustworthy sources like encyclopedias. The piece advocates for critical thinking amid AI-driven misinformation.
ChatGPT is hallucinating fake links to its news partners' biggest investigations
ChatGPT by OpenAI generates fake URLs for major news sites, failing to link to correct articles despite promises. Journalists express concerns over reliability and demand transparency due to hallucinated URLs.
Google Researchers Publish Paper About How AI Is Ruining the Internet
Google researchers warn about generative AI's negative impact on the internet, creating fake content blurring authenticity. Misuse includes manipulating human likeness, falsifying evidence, and influencing public opinion for profit. AI integration raises concerns.
"AI", students, and epistemic crisis
An individual shares a concerning encounter where a student confidently presented false facts from an AI tool, leading to a clash over misinformation. Educators face challenges combating tailored misinformation in the digital era.
The term "hallucination" may be imprecise, but it is a lot closer to the truth than "bullshit".
The bullshit is only in the actions of marketers trying to engage technically illiterate wallets in funding a bubble out of false promises (such as "dreaming", "imagination", "hallucinations", "bullshitting", "superintelligence", "superalignment", "smarter than human", and so forth).
This is just wrong. Accurate modelling of language at the scale of modern LLMs requires these models to develop rich world models during pretraining, which also requires distinguishing facts from fiction. This is why bullshitting happens less with better, bigger models: the simple answer is that they just know more about the world, and can also fill in the gaps more efficiently.
We have empirical evidence here: it's even possible to peek into a model to check whether the model 'thinks' what it's saying is true or not. From “Discovering Latent Knowledge in Language Models Without Supervision” (2022) [1]:
> Specifically, we introduce a method for accurately answering yes-no questions given only unlabeled model activations. It works by finding a direction in activation space that satisfies logical consistency properties, such as that a statement and its negation have opposite truth values. (...) We also find that it cuts prompt sensitivity in half and continues to maintain high accuracy even when models are prompted to generate incorrect answers. Our results provide an initial step toward discovering what language models know, distinct from what they say, even when we don't have access to explicit ground truth labels.
So when a model is asked to generate an answer it knows is incorrect, it's internal state still tracks the truth value of the statements. This doesn't mean the model can't be wrong about what it thinks is true (or that it won't try to fill in the gaps incorrectly, essentially bullshitting), but it does mean that the world models are sensitive to truth.
More broadly, we do know these models have rich internal representations, and have started learning how to read them. See for example “Language Models Represent Space and Time” (Wes & Tegmark, 2023) [2]:
> We discover that LLMs learn linear representations of space and time across multiple scales. These representations are robust to prompting variations and unified across different entity types (e.g. cities and landmarks). In addition, we identify individual "space neurons" and "time neurons" that reliably encode spatial and temporal coordinates. While further investigation is needed, our results suggest modern LLMs learn rich spatiotemporal representations of the real world and possess basic ingredients of a world model.
For anyone curious, I can recommend the Othello-GPT paper as a good introduction to this problem (“Do Large Language Models learn world models or just surface statistics?”) [3].
[1]: https://arxiv.org/abs/2310.02207
Related
Lessons About the Human Mind from Artificial Intelligence
In 2022, a Google engineer claimed AI chatbot LaMDA was self-aware, but further scrutiny revealed it mimicked human-like responses without true understanding. This incident underscores AI limitations in comprehension and originality.
The Encyclopedia Project, or How to Know in the Age of AI
Artificial intelligence challenges information reliability online, blurring real and fake content. An anecdote underscores the necessity of trustworthy sources like encyclopedias. The piece advocates for critical thinking amid AI-driven misinformation.
ChatGPT is hallucinating fake links to its news partners' biggest investigations
ChatGPT by OpenAI generates fake URLs for major news sites, failing to link to correct articles despite promises. Journalists express concerns over reliability and demand transparency due to hallucinated URLs.
Google Researchers Publish Paper About How AI Is Ruining the Internet
Google researchers warn about generative AI's negative impact on the internet, creating fake content blurring authenticity. Misuse includes manipulating human likeness, falsifying evidence, and influencing public opinion for profit. AI integration raises concerns.
"AI", students, and epistemic crisis
An individual shares a concerning encounter where a student confidently presented false facts from an AI tool, leading to a clash over misinformation. Educators face challenges combating tailored misinformation in the digital era.