January 1st, 2025

AI Predictions for 2025, from Gary Marcus

Gary Marcus predicts that AGI won't be achieved by 2025, AI profits will be modest, regulatory frameworks will lag, job displacement will be minimal, and AI company valuations may decline.

Read original articleLink Icon
AI Predictions for 2025, from Gary Marcus

Gary Marcus outlines his predictions for AI developments by the end of 2025, emphasizing that artificial general intelligence (AGI) will not be achieved, contrary to claims by figures like Elon Musk. He anticipates that no AI system will reliably solve more than four of the AI 2027 Marcus-Brundage tasks, and profits from AI models will remain modest. Regulatory frameworks in the U.S. will lag behind Europe, and issues such as hallucinations and reasoning errors in generative AI will persist. The hype surrounding AI agents and humanoid robotics will not translate into reliable products. He predicts that few radiologists will be replaced by AI, and truly driverless cars will remain limited in use. The impact of AI on the workforce will be minimal, with less than 10% of jobs affected. Marcus expresses medium confidence that technical moats will remain elusive, and AI company valuations may begin to decline. He also notes the potential for a significant cyberattack involving generative AI. Overall, he suggests that while AI will continue to evolve, many of the anticipated breakthroughs may not materialize as quickly as expected.

- AGI is unlikely to be achieved by the end of 2025.

- Profits from AI models will remain modest, with limited corporate adoption.

- Regulatory frameworks in the U.S. will lag behind Europe.

- The impact of AI on job displacement will be minimal, affecting less than 10% of the workforce.

- Technical moats for AI companies will remain elusive, with potential declines in valuations.

Link Icon 7 comments
By @mordymoop - 4 months
I’m not as inclined to grade Gary Marcus’ predictions leniently as Gary Marcus seems to be.

• 7-10 GPT-4 level models

He only gets to claim this is true if you count a bunch of different versions of the same base model, or if you’re willing to say that some models that outperform GPT-4 on some benchmarks count as being GPT-4 level. I don’t think Marcus was right in spirit, here.

• No massive advance (no GPT-5, or disappointing GPT-5)

Seems too unquantifiable to judge, I would call o1 a massive advance over 4o, but I’m sure Marcus would not.

• Price wars

I guess so? From what I’ve read, the frontier models companies are still profitable, and OpenAI now has a $200/mo commercial model, hardly the action of a company deciding its prices purely to undercut the competition.

• Very little moat for anyone

It still seems like the only companies who have pulled off frontier model capabilities have spent many millions of dollars doing it. I think this might become true next year but I don’t think this can be judged as correct based on what we saw in 2024 alone.

• No robust solution to hallucinations

You only use words like “robust” in a prediction like this so you have room to weasel out of it later when the hallucinations diminish greatly but don’t quite go extinct.

• Modest lasting corporate adoption

My industry is oil and gas. A pretty hidebound and conservative industry. Adoption of LLMs has been massive.

• Modest profits, split 7-10 ways

Define modest.

I score Marcus at 0/7, at best 2/7.

By @starchild3001 - 4 months
In 2024, AI made astonishing progress across the board. Top models have improved by roughly 130 ELO points on public leaderboards, while costs for running those same models have dropped many fold. Smaller, low-latency models now offer near–state-of-the-art performance (see Gemini Flash, for example). Multimodal capability is increasingly the norm, and strong reasoning skills are widely expected. New “hard problem” leaderboards have emerged, and they, too, have seen major gains. Long-context models have become standard, and mainstream products like Google Search now integrate AI by default.

Overall, the field has seen tremendous progress -- whether or not most users realize just how far we’ve come. Marcus's predictions don't sound specific enough -- no GPT5? Correct. But, what does that even mean?

By @mentalgear - 4 months
I agree that LLMs can be highly valuable when used appropriately. For me, their strengths lie primarily in semantic search (e.g., Perplexity.ai), rewriting and discussing overall code architecture. However, they often prove unreliable in specialized knowledge domains. If a user isn’t skilled in initializing and following up with effective prompts or discerning the usefulness from the responses, interactions can be hit-or-miss.

The robustness of LLMs—like the original ChatGPT or GPT-3.5—is still far from a level where domain novices can rely on them with confidence. This might change as models incorporate more first-order data (e.g., direct observations in physics, embodiment) and improve in causal reasoning and deductive logic.

I find it crucial to have critical voices like Gary Marcus to counterbalance the overwhelming hype perpetuated by social media enthusiasts and corporate PR—much of which the media tends to echo uncritically.

One of Marcus's demands is the need for a more neuro-symbolic approach to advance AI. Progress can’t come solely from "scaling up" models. And it seems like he's right: All major ML companies seem to shift towards search-based algorithms (e.g., Q*) combined with reinforcement learning during inference time to explore the problem space, moving beyond mere "next-token prediction" training.

By @gandalfgeek - 4 months
I don't understand why he has such an axe to grind. Is there some historical baggage here?

Of course there are plenty of problems with the current state of AI and LLMs, but to have such a preconceived pessimistic outlook that can't even acknowledge their massive and quick adoption and usefulness in multiple domains seems not intellectually honest.

By @krackers - 4 months
The related list of ten Marcus-worthy tasks is probably more worth reading, since it at least provides some falsifiable goalposts.

https://news.ycombinator.com/item?id=42560545

I'd argue 1, 2, maybe 6 are effectively already doable. 3 & 5 are good tasks which are technically possible with some RAG hackery but in general is a good benchmark testing out-of-context fact retrieval. 4 & 10 might happen soon-ish with work on "agents" and proof synthesis respectively. 7 & 8 are too subjective. 9, maybe weakened to formulating or proving novel theorems, might be a good baseline for "peak human" intelligence (I think at best O1-O3 can spot and prove some lemmas, but nothing that anyone would bother publishing).

By @mrayycombi - 4 months
Marcus clearly threw a lot of spaghetti, seemingly pessimistic "it socks and won't get better" and then cherry picks confirmations, for the bits that stuck to the wall.

How about he indicates: 1. How he came to these conclusions (coin flip? Pessimism?) 2. How many predictions he missed. He's implying a very high rate of success, which is a big red flag of shenanigans for me.

This is little more than vague generalities and coin flipping with retroactive cherry picked "See?! I was right!" analysis.

A gypsy at a traveling circus serves up about the same.

By @tim333 - 4 months
>No single system will solve more than 4 of the AI 2027 Marcus-Brundage tasks by the end of 2025

I had a look at the "Marcus-Brundage tasks" that he has modestly named after himself and am stuck that for an AI skeptic he's listed things for 2027 well beyond 99.9% of humans like write 10,000 lines of bug free code, Oscar level screenplays, Nobel prize discoveries, Pulitzer books etc.