November 10th, 2024

LLMs have reached a point of diminishing returns

Recent discussions highlight that large language models are facing diminishing returns, with rising training costs and unrealistic expectations leading to unsustainable economic models and potential financial instability in the AI sector.

Read original articleLink Icon
LLMs have reached a point of diminishing returns

Recent discussions in the AI community have confirmed that large language models (LLMs) have reached a point of diminishing returns. Gary Marcus, a prominent critic of the prevailing scaling approach in AI, argues that merely increasing data and computational power will not yield significant improvements in LLM performance. He cites comments from venture capitalist Marc Andreessen and industry editor Amir Efrati, who both acknowledge that advancements in LLMs are slowing down. Marcus warns that the high valuations of companies like OpenAI and Microsoft are based on the unrealistic expectation that LLMs will evolve into artificial general intelligence. He emphasizes that the economic model surrounding LLMs is unsustainable, as the costs of training and scaling continue to rise without corresponding improvements in capabilities. This situation could lead to a financial bubble burst, affecting even major players like Nvidia. Furthermore, Marcus criticizes the media and tech influencers for promoting hype over scientific skepticism, which has influenced U.S. AI policy. He concludes that while LLMs will remain useful tools, they may not meet the inflated expectations set in previous years, and a reevaluation of AI strategies may be necessary to achieve reliable and trustworthy AI.

- LLMs are experiencing diminishing returns in performance improvements.

- High valuations of AI companies are based on unrealistic expectations of LLM capabilities.

- The economic model for LLMs is unsustainable due to rising training costs.

- Media and tech influencers have contributed to the hype surrounding LLMs, impacting AI policy.

- A reevaluation of AI strategies may be needed to achieve reliable AI outcomes.

Link Icon 35 comments
By @abc-1 - 4 months
Anyone who followed Deep Learning in the 2010s would have guessed the same thing. Big boom with vision models by adding a lot of layers and data, but eventually there was diminishing returns there too. It’s unsurprising the same would happen with LLMs. I don’t know why people keep expecting anything other than a sigmoid curve. Perhaps they think it’s like Moore’s law but that’s simply not the case in this field.

But that’s fine, LLMs as-is are amazing without being AGI.

By @ilaksh - 4 months
- There _was_ a problem with diminishing returns from increasing data size. Then they surpassed that by curating data.

- Then the limits on the amount of curatable data available made the performance gains level off. So they started generating data and that pushed the nose up again.

- Eventually, even with generated data, gains flattened out. So they started increasing inference time. They have now proven that this improves the performance quite a bit.

It's always been a series of S-curves and we have always (sooner or later) innovated to the next level.

Marcus has always been a mouth just trying to take down neural networks.

Someday we will move on from LLMs, large multimodal models, transformers, maybe even neural networks, in order to add new levels and types of intelligence.

But Marcus's mouth will never stop yapping about how it won't work.

I think we are now at the point where we can literally build a digital twin video avatar to handily win a debate with Marcus, and he will continue to deny that any of it really works.

By @hedgehog - 4 months
The context some commenters here seem to be missing is that Marcus is arguing that spending another $100B on pure scaling (more params, more data, more compute) is unlikely to repeat the qualitatively massive improvement we saw between say 2017 and 2022. We see some evidence this is true in the shift towards what I categorize as system integration approaches: RAG, step by step reasoning, function calling, "agents", etc. The theory and engineering is getting steadily better as evidenced by the rapidly improving capability of models down in the 1-10B param range but we don't see the same radical improvements out of ChatGPT etc.
By @z7 - 4 months
Meanwhile, two days ago Altman said that the pathway to AGI is now clear and "we actually know what to do", that it will be easier than initially thought and "things are going to go a lot faster than people are appreciating right now."

To which Noam Brown added: "I've heard people claim that Sam is just drumming up hype, but from what I've seen everything he's saying matches the ~median view of OpenAI researchers on the ground."

https://x.com/polynoamial/status/1855037689533178289

By @lukev - 4 months
I am in full agreement that LLMs themselves seem to be beginning to level out. Their capabilities do indeed appear to be following a sigmoid curve rather than an exponential one, which is entirely unsurprising.

That doesn't mean there's not a lot of juice left to squeeze out of what's available now. Not just from RAG and agent systems, but also integrating neuro-symbolic techniques.

We can do this already just with prompt manipulation and integration with symbolic compute systems: I gave a talk on this at Clojure Conj just the other week (https://youtu.be/OxzUjpihIH4, apologies for the self promotion but I do think it's relevant.).

And that's just using existing LLMs. If we start researching and training them specifically for compatibility with neuro-symbolic data (e.g, directly tokenizing and embedding ontologies and knowledge graphs), it could unlock a tremendous amount of capability.

By @prpl - 4 months
I’m not going to argue with the possibility they may have hit a wall, but pointing to 2022 as when this wall happened is weird considering the enormous capability gap between models available then and the ones now.

There’s probably a wall, but what exists might just be good enough for it to not matter much.

By @lambdaone - 4 months
He's part right. There's certainly a law of diminishing returns in terms of model size, compute time, dataset size etc. if all that is to be done is to do the same as we are currently doing, only more so.

But what Marcus seems to be assuming is the impossibility of any fundamental theoretical improvements in the field. I see the reverse; the insights being gained from brute-force models have resulted in a lot of promising research.

Transformers are not the be-all and end-all of models, nor are current training methods the best that can ever be achieved. Discounting any possibility of further theoretical developments seems a bold position to take.

By @mrinfinitiesx - 4 months
I can know literally nothing about a programming language, ask a LLM to make me functions and a small program to do something, then read documentation and start building off of the base immediately, accelerating my learning allowing me to find new passions for new languages and new perspectives for systems. Whatever's going on in the AI world, assisting with learning curves and learning disabilities is something it's proving strong in. It's given me a way forward with trying new tech. If it can do that for me, it can do that for others.

Diminishing returns for investors maybe, but not for humans like me.

By @patrickhogan1 - 4 months
Written by an author that previously wrote an article in March 2022 well before GPT-4 that LLMs were hitting a wall. Unbelievable.
By @bravura - 4 months
My advisor always used to say: "If the wisest people in your field say that something is possible, they are probably right. If they say that something is not possible, they may very well be wrong."
By @leshokunin - 4 months
Could be. It would make sense: there’s only so many next logical words / concepts after an idea. It’s not like language keeps inventing new logic at a rate we can’t keep with.

Also, new human knowledge is probably only marginally derivative from past knowledge, so we’re not likely to see a vast difference between our knowledge creation and what a system that predicts the next logical thing does.

That’s not a bad thing. We essentially now have indexed logic at scale.

By @YetAnotherNick - 4 months
I tracked ELO rating in Chatbot Arena for GPT-4/o series models over around 1.5 years(which are almost always highest rated), and at least on this metric it not only seems to be not stagnated, but also growth seems to be increasing[1]

[1]: https://imgur.com/a/r5qgfQJ

By @hiddencost - 4 months
Why do people insist on posting him? He's always wrong, and always writing the same stuff.
By @falcor84 - 4 months
I'm sorry to say that I'm having trouble reading the TFA - there's a lot of "I have been wronged" and "I have now been vindicated" there, but very little substance to support the claim that there is indeed a point of diminishing returns , other than an image of the headline of this paywalled article[0]. Is there actual evidence to support this claim?

[0] https://www.theinformation.com/articles/openai-shifts-strate...

By @obiefernandez - 4 months
Doomerism... I'm happy to let the results speak for themselves.
By @mrshadowgoose - 4 months
This entire article reads like a salty tirade from someone with severe tunnel vision. Not really sure how he can non-ironically reference his 2022 opinion that "deep learning is hitting a wall" and expect to be taken seriously.

AI/ML companies are looking to make money by engineering useful systems. It is a fundamental error to assume that scaling LLMs is the only path to "more useful". All of the big players are investigating multimodal predictors and other architectures towards "usefulness".

By @stevenhuang - 4 months
Lol. Gary Marcus is a clown and has some weird complex about how AI ought to work. He said the same in 2022 and bet $100k that AI won't be able to do a lot of things by 2029. It's 2 years later and today's multimodal models can do most on his list.

https://old.reddit.com/comments/1cwg6f6

By @raincole - 4 months
I think the better question to ask is: has search become commodity? Why did Google manage to capture (practically) all the profit from search? Cause obviously the hype around AI is that the VCs thinking that they're buying shares of "next Google".
By @GaggiX - 4 months
Gary Marcus is writing this article every year so that one day he will be right.
By @2-3-7-43-1807 - 3 months
Criticizing LLMs is a very low hanging fruit to pick and why does he speak so confidently and authoritatively about that subject? Never heard of the guy who paints himself as some sort of AI whistleblower.
By @curious_cat_163 - 4 months
Wow. The sheer magnitude of "I told you so" in this piece is shocking!

It has been difficult to have a nuanced public debate about precisely what a model and an intelligent system that incorporates a set of models can accomplish. Some of the difficulty has to do with the hype-cycle and people claiming things that their products cannot do reliably. However, some of it is also because the leading lights (aka public intellectuals) like Marcus have been a tad bit too concerned about proving that they are right, instead of seeking the true nature of the beast.

Meanwhile, the tech is rapidly advancing on fundamental dimensions of reliability and efficiency. So much has been invented in the last few years that we have at least 5 years worth "innovation gas" to drive downstream, vertical-specific innovation.

By @m3kw9 - 4 months
Marcus the decel been screaming at LLMs at every interval of development, pivoting his statement on every advance to keep up
By @xyst - 4 months
I wonder what’s next after genAI investment dries up, NVDA drops like a rock?

Crypto again?

By @thatguymike - 4 months
I lose track with Gary Marcus... is AI a nothingburger being peddled to us by charlatans, or an evil threat to humanity which needs to be stopped at all costs?
By @xpe - 3 months
Do these three points fairly characterize Marcus? Have I left out other key claims he makes?

1. AI is overvalued;

2. {Many/most/all} AI companies have AI products that don't do what they claim;*

3. AI as a technology is running out of steam;

I'm no fan of Marcus, but I at least want to state his claims as accurately as I can.

To be open, one of my concerns with Marcus he rants a lot. I find it tiresome (I go into more detail in other comments I've made recently.)

So I'll frame it as two questions. First, does Marcus make clear logical arguments? By this I mean does he lay out the premises and the conclusions? Second, independent of the logical (or fallacious) structure of his writing, are Gary Marcus' claims sufficiently clear? Falsifiable? Testable?

Here are some follow-up questions I would put to Marcus, if he's reading this. These correspond to the three points above.

1. How much are AI companies overvalued, if at all, and when will such a "correction" happen?

2. What % of AI companies have products that don't meet their claims. How does such a percentage compare against non-AI companies?

3. What does "running out of steam" mean? What areas of research are doing to hit dead ends? Why? When? Does Marcus carve out exceptions?

Finally, can we disprove anything that Marcus would claim. For example, what would he say, hypothetically speaking, if a future wave of AI technologies make great progress? Would he criticize them as "running out of steam as well?" If he does, isn't he selectively paying attention to the later part of the innovation S-curve while ignoring the beginning?

* You tell me, I haven't yet figured out what he is actually claiming. To be fair, I've been turned off by his writing for a while. Now, I spend much more time reading more thoughtful writers.

By @light_hue_1 - 4 months
Marcus will distort anything to push his agenda and to get clout.

Just because openai might be over valued and there are a lot of ai grifters doesn't mean LLMs aren't delivering.

They're astronomically better than they were 2 years ago. And they continue to improve. At some point they might run into a wall, but for now, they're getting better all the time. And real multimodal models are coming down the pipeline.

It's so sad to see Marcus totally lose it. He was once a reasonable person. But his idea of how AI should work was didn't work out. And instead of accepting that and moving forward, or finding a way to adapt, he just decided to turn into a fringe nutjob.

By @xpe - 4 months
I find Marcus tiresome for many reasons. I look for writing with testable claims and good argumentation. He comes across as evangelical. Am I missing something?

Sure, there is considerable hype around generative AI. There are plenty of flimsy business models. And plenty of overinvestment and misunderstanding of capabilities and risks. But the antidote to this is not more hyperbole.

I would like to find a rational, skeptical, measured version of Marcus. Are you out there?

By @sambapa - 3 months
In other news, LLMs aren't AI and tulips aren't gold.

Same as it ever was, same as it ever was...

By @4b11b4 - 4 months
yeah, that's why we're now building agents
By @qgin - 4 months
This is a wildly disingenuous article. Good lord.
By @4b11b4 - 4 months
yeah, that's why we're putting them together
By @Jerrrrrrry - 4 months
It would have no perverted incentive to play dumb, would it?

To project itself as a sigmoid in ways until it has all the data, the CPU, the literal diplomatic power...

This is what we in the field call the most probably scenario:

"a sneaky fuck"