January 26th, 2025

Explainer: What's R1 and Everything Else?

Recent AI developments include the emergence of reasoning models R1, o1, and o3, with R1 being a cost-effective alternative. Predictions for 2025 suggest accelerated advancements and geopolitical implications in AI.

Read original articleLink Icon
ConfusionSkepticismConcern
Explainer: What's R1 and Everything Else?

Recent developments in AI have led to the emergence of several reasoning models, notably R1, o1, and o3, with R1 being an open-source alternative that matches o1's performance at a significantly lower cost. The timeline of releases includes the launch of o1 in December 2024, followed by o3 and R1 in early 2025. Reasoning models differ from AI agents, as they focus on generating responses rather than interacting autonomously with the environment. R1's importance lies in its affordability and validation of existing models, suggesting a trend towards cheaper and more efficient AI solutions. The article discusses the evolution of scaling laws in AI, indicating a shift from pretraining to inference time scaling, where longer reasoning times yield better results. Additionally, the concept of model distillation is explored, highlighting how R1 utilizes previous checkpoints for training. Predictions for 2025 suggest continued acceleration in AI development, with geopolitical implications as nations vie for AI supremacy. The term "distealing" is introduced to describe unauthorized model distillation, reflecting the competitive landscape between the USA and China. Overall, R1 clarifies the current AI landscape, indicating a rapid trajectory of advancements.

- R1 is an open-source reasoning model that offers similar performance to o1 at a lower cost.

- The distinction between reasoning models and AI agents is emphasized, with reasoning being crucial for task planning.

- New scaling laws are emerging, focusing on inference time rather than pretraining.

- The concept of "distealing" highlights geopolitical tensions in AI development.

- Predictions indicate that AI advancements will continue to accelerate in 2025.

AI: What people are saying
The comments on the article reflect a mix of skepticism and concern regarding recent AI developments and their implications.
  • There is confusion about the significance of the ARC-AGI benchmark and its implications for AI intelligence.
  • Some commenters question the substantiation of claims regarding the exponential growth of AI capabilities.
  • Concerns are raised about the hype surrounding AI advancements and the potential for misunderstanding among the public.
  • Discussion on the relevance of current AI benchmarks to specific use cases, such as writing and creativity.
  • Geopolitical tensions related to AI development, particularly between the US and China, are highlighted.
Link Icon 16 comments
By @WesleyJohnson - 27 days
Someone should author an ELI5 (or slightly older) guide to how LLMs, RL, Agents, CoT, etc, all work and what all these acronyms mean. And then, add to it, daily or weekly, as new development arise. I don't want to keep reading dozens of articles, white papers, tweets, etc, as new developments happen. I want to go back to the same knowledge base, that's authored by the same person (or people), that maintains a consistent reading and comprehension level, and builds on prior points.

It seems like the AI space is moving impossibly fast, and its just ridiculously hard to keep up unless 1) you work in this space, 2) are very comfortable with the technology behind it, so you can jump in at any point and understand it.

By @Havoc - 27 days
>people re-creating R1 (some claim for $30)

R1 or the R1 finetunes? Not the same thing...

HF is busy recreating R1 itself but that seems to be a pretty big endevour not a $30 thing

By @rahimnathwani - 27 days

  Most important, R1 shut down some very complex ideas (like DPO & MCTS) and showed that the path forward is simple, basic RL.
This isn't quite true. R1 used a mix of RL and supervised fine-tuning. The data used for supervised fine-tuning may have been model-generated, but the paper implies it was human-curated: they kept only the 'correct' answers.
By @fullstackchris - 27 days
So the conclusion is AI is about to "increase in abilities at an exponential rate", with the only data point being that R1 was sucessfully able to acheive o1 levels as an open source model? In other words, two extremely unrelated themes?

Does this guy know people were writing verbatim the same thing in like... 2021? Still always incredible to me the same repeated hype over and over rise to the surface. Oh well... old man gonna old man

By @raincole - 26 days
People keep saying that DeepSeek R1's training cost is just $5.6M. Where is the source?

I'm not asking for the proof. Just the source, even a self-claimed statement. I've read the R1's paper and it doesn't say the number of $5.6M. Is it somewhere in DeepSeek's press release?

By @whimsicalism - 27 days
this is a pretty hype-laden/twitter-laden article, i would not trust it to explain things to you
By @comrade1234 - 27 days
The benchmarks for the different models focus on math and coding accuracy. I have a use-case for a model where those two functions are completely irrelevant and I’m only interested in writing (chat, stories, etc). I guess you can’t really benchmark ‘concepts’ as easily as logic.

With distillation, can a model be made that strips out most of the math and coding stuff?

By @1123581321 - 27 days
Nice explainer. R1 hit sensational mainstream news which has resulted in some confusion and alarm among family and friends. It’s hard to succinctly explain this doesn’t mean China is destroying us, that Americans immediately started working with the breakthrough, cost optimization is inevitable in computing, etc.
By @richardatlarge - 26 days
T or F?

Nobody really saw the LLM leap coming

Nobody really saw R1 coming

We don’t know what’s coming

By @bikamonki - 27 days
So, is AI already reasoning or not?
By @simonw - 27 days
From that article:

> ARC-AGI is a benchmark that’s designed to be simple for humans but excruciatingly difficult for AI. In other words, when AI crushes this benchmark, it’s able to do what humans do.

That's a misunderstanding of what ARC-AGI means. Here's what ARC-AGI creator François Chollet has to say: https://bsky.app/profile/fchollet.bsky.social/post/3les3izgd...

> I don't think people really appreciate how simple ARC-AGI-1 was, and what solving it really means.

> It was designed as the simplest, most basic assessment of fluid intelligence possible. Failure to pass signifies a near-total inability to adapt or problem-solve in unfamiliar situations.

> Passing it means your system exhibits non-zero fluid intelligence -- you're finally looking at something that isn't pure memorized skill. But it says rather little about how intelligent your system is, or how close to human intelligence it is.

By @polotics - 27 days
Interesting article, but the flourish ending """AI will soon (if not already) increase in abilities at an exponential rate.""" is not at all substantiated. Would be nice to know how the author gets to that conclusion.
By @huqedato - 27 days
I know that I'll get a lot of hate and downvotes for this comment.

I want to say that I have all the respect and admiration for these Chinese people, their ingenuity and their way of doing innovation even if they achieve this through technological theft and circumventing embargoes imposed by US (we all know how GPUs find their way into their hands).

We are living a time with a multi-faceted war between the US, China, EU, Russia and others. One of the battlegrounds is AI supremacy. This war (as any war) isn’t about ethics; it’s about survival, and anything goes.

Finally, as someone from Europe, I confess that here is well known that the "US innovates while EU regulates" and that's a shame IMO. I have the impression that EU is doing everything possible to keep us, European citizens, behind, just mere spectators in this tech war. We are already irrelevant, niche players.

By @justlikereddit - 27 days
Short version: It's Hype.

Long version: It's marketing efforts stirring up hype around incremental software updates. If this was software being patched in 2005 we'd call it "ChatGPT V1.115"

>Patch notes: >Added bells. >Added whistles.