Delving into ChatGPT usage in academic writing through excess vocabulary
A study by Dmitry Kobak et al. examines ChatGPT's impact on academic writing, finding increased usage in PubMed abstracts. Concerns arise over accuracy and bias despite advanced text generation capabilities.
Read original articleThe study titled "Delving into ChatGPT usage in academic writing through excess vocabulary" by Dmitry Kobak and colleagues investigates the prevalence of large language model (LLM) usage in academic literature. Analyzing 14 million PubMed abstracts from 2010-2024, the research reveals a significant increase in the frequency of certain style words, indicating LLM usage in at least 10% of 2024 abstracts. This percentage varied across disciplines, countries, and journals, reaching as high as 30% in some PubMed sub-corpora. The impact of LLM-based writing assistants on scientific literature is deemed unprecedented, surpassing the influence of major world events like the Covid pandemic. The study highlights concerns regarding the potential inaccuracies, biases, and misuse associated with LLMs despite their human-level text generation capabilities.
Related
Optimizing AI Inference at Character.ai
Character.AI optimizes AI inference for LLMs, handling 20,000+ queries/sec globally. Innovations like Multi-Query Attention and int8 quantization reduced serving costs by 33x since late 2022, aiming to enhance AI capabilities worldwide.
We no longer use LangChain for building our AI agents
Octomind switched from LangChain due to its inflexibility and excessive abstractions, opting for modular building blocks instead. This change simplified their codebase, increased productivity, and emphasized the importance of well-designed abstractions in AI development.
GitHub – Karpathy/LLM101n: LLM101n: Let's Build a Storyteller
The GitHub repository "LLM101n: Let's build a Storyteller" offers a course on creating a Storyteller AI Large Language Model using Python, C, and CUDA. It caters to beginners, covering language modeling, deployment, programming, data types, deep learning, and neural nets. Additional chapters and appendices are available for further exploration.
Francois Chollet – LLMs won't lead to AGI – $1M Prize to find solution [video]
The video discusses limitations of large language models in AI, emphasizing genuine understanding and problem-solving skills. A prize incentivizes AI systems showcasing these abilities. Adaptability and knowledge acquisition are highlighted as crucial for true intelligence.
Detecting hallucinations in large language models using semantic entropy
Researchers devised a method to detect hallucinations in large language models like ChatGPT and Gemini by measuring semantic entropy. This approach enhances accuracy by filtering unreliable answers, improving model performance significantly.
I'm not quite sure this follows. At the very least, I think they should also consider the possibility of social contagion: if some of your colleagues start using a new word in work-related writing, you usually pick that up. The spread of "delve" was certainly bootstrapped by ChatGPT, but I'm not sure that the use of LLMs is the only possible explanation for its growing popularity.
Even in the pre-ChatGPT days, it was common for a new term to come out of nowhere and then spread like a wildfire in formal writing. "Utilize" for "use", etc.
delves
crucial
potential
these
significant
important
They're not the first to make this observation. Others have picked up that LLM's like the word "delves".LLMs are trained on texts which contain much marketing material. So they tend to use some marketing words when generating pseudo-academic content. No surprise there. I'm surprised it's not worse.
What happens if you use a prompt containing "Write in a style that maximizes marketing impact"?"
('You can't always use "Free", but you can always use "New"' - from a book on copywriting.)
While the word frequency stats are damning, there doesn’t seem to be any evidence presented that directly ties the changes to LLMs specifically.
"Delving".. Sounds like the authors might have used an LLM while writing this paper as well.
An old partner edited papers for a large publisher - largely written by non-native speakers and already heavily machine translated - and would almost always use ChatGPT for the first pass when extensive changes were needed.
She was paid by the word and also had a pretty intense daily minimum quota so it was practically required to get enough done to earn a liveable wage and avoid being replaced by another remote “contractor”.
It’s an issue I’ve noticed personally, as I’m seeing an increasing number of reviews that lack substance and are almost entirely made of filler content. Here’s an excerpt from a particularly egregious recent example I ran into, which had this to say on the subject of meaningful comparison to recent work:
> Additionally, while the bibliography appears to be comprehensive, there could be some minor improvements, such as including more recent or relevant references if applicable.
The whole review was written like this, with no specific suggestions for improvement, just vague “if applicable” filler. Infuriating.
https://trends.google.com/trends/explore?date=today%205-y&q=...
The leap from how the LLMs write now to how a professional sounding scientist might write their paper is probably not that big.
Its possible now to train them on your own writing style.
human and machine both should aim for brevity and clarity, and feel shame otherwise.
then we can read more and better in our lives.
We truly live in the informations age.
On a side note, do people still use ChatGPT to fill out their papers? I found Claude to be way better at spitting out more content. In my experience, ChatGPT has been in the middle while Gemini is the worst, it even cuts-off in the middle of the sentence.
Related
Optimizing AI Inference at Character.ai
Character.AI optimizes AI inference for LLMs, handling 20,000+ queries/sec globally. Innovations like Multi-Query Attention and int8 quantization reduced serving costs by 33x since late 2022, aiming to enhance AI capabilities worldwide.
We no longer use LangChain for building our AI agents
Octomind switched from LangChain due to its inflexibility and excessive abstractions, opting for modular building blocks instead. This change simplified their codebase, increased productivity, and emphasized the importance of well-designed abstractions in AI development.
GitHub – Karpathy/LLM101n: LLM101n: Let's Build a Storyteller
The GitHub repository "LLM101n: Let's build a Storyteller" offers a course on creating a Storyteller AI Large Language Model using Python, C, and CUDA. It caters to beginners, covering language modeling, deployment, programming, data types, deep learning, and neural nets. Additional chapters and appendices are available for further exploration.
Francois Chollet – LLMs won't lead to AGI – $1M Prize to find solution [video]
The video discusses limitations of large language models in AI, emphasizing genuine understanding and problem-solving skills. A prize incentivizes AI systems showcasing these abilities. Adaptability and knowledge acquisition are highlighted as crucial for true intelligence.
Detecting hallucinations in large language models using semantic entropy
Researchers devised a method to detect hallucinations in large language models like ChatGPT and Gemini by measuring semantic entropy. This approach enhances accuracy by filtering unreliable answers, improving model performance significantly.