Have we stopped to think about what LLMs model?
Recent discussions critique claims that large language models understand language, emphasizing their limitations in capturing human linguistic complexities. The authors warn against deploying LLMs in critical sectors without proper regulation.
Read original articleRecent discussions surrounding large language models (LLMs) highlight a critical examination of what these technologies actually represent in terms of language and cognition. A peer-reviewed paper by Abeba Birhane and Marek McGann critiques the prevalent claims that LLMs can "understand" language, arguing that such assertions misinterpret the nature of human linguistic capabilities. The authors emphasize that LLMs, which are built on vast datasets and statistical techniques, do not replicate the complexities of human language, which is inherently social, contextual, and embodied. They argue that LLMs operate under flawed assumptions about language completeness and data representation, failing to capture the nuances of human interaction, such as emotional context and social participation. The paper warns that treating LLMs as language understanding machines can lead to misguided policies and social implications. Furthermore, the authors express concern over the deployment of LLMs in critical sectors like education and healthcare without adequate testing and regulation, highlighting the potential risks of misinformation and unreliability. While the AI industry continues to promote the economic benefits of LLMs, the authors call for a more cautious and skeptical approach to their development and application.
- The paper critiques the exaggerated claims about LLMs' understanding of language.
- LLMs are based on flawed assumptions about language and data representation.
- Human language is complex and cannot be fully captured by LLMs.
- There are significant risks in deploying LLMs in critical sectors without proper regulation.
- A more cautious approach to LLM development is advocated by researchers.
This is just not true. Too much BS and it risks getting shut down.
Like in the short story Lena, 2021-01-04 by qntm
In short story, the weights of the LLM are a brain scan.
But same situation. People could use multiple copies of the AI. But each time, they would have to 'talk it into' doing what they wanted
In addition I don't think it makes sense to compare this to building bridges or pharma.
I don't think ChatGPT is more likely to harm a person with misinformation than just plain Google or YouTube would.
In fact already existing search and recommendation algorithms I believe are more likely to lead you down the misinfo rabbit hole.
At least ChatGPT to an extent is biased to try and stay objective as opposed to any rabbit holes leading people to fringe content.
The stolen data used in the training sets are filled with online communities you would shudder to be forced to experience, and books you’d refuse to read.
It’s why they’ll suddenly suggest you put glue in pizza sauce, or why they read in a soulless overly verbose “m’lady” tone.
More data made them less useful but better at fooling people with a superficial interest in them, and that demographic is so large it affords these companies leverage in funding rounds.
Markets truly are irrational.