January 27th, 2025

A Boy Who Cried AGI

Mark Zuckerberg suggests AI will soon match mid-level engineers, sparking debate on AGI's timeline. The author stresses the need for clear definitions, cautious preparation, and public discourse on AGI ethics.

Read original articleLink Icon
A Boy Who Cried AGI

Mark Zuckerberg recently stated that AI will soon match the capabilities of mid-level engineers, igniting debates about the timeline for achieving artificial general intelligence (AGI). While some assert that AGI is already here, others believe it remains decades away. The author, an experienced machine learning practitioner, expresses concern over the lack of a clear definition for AGI and the challenges in evaluating large language models (LLMs). The ambiguity surrounding intelligence leads to teams prematurely declaring problems solved without delivering real value. The concept of technological singularity, where machines surpass human intelligence, is acknowledged as a serious concern, yet its timing remains uncertain. The author warns against overpromising advancements in AGI, likening it to the fable of "The Boy Who Cried Wolf," where repeated false alarms can lead to complacency. This overhyping can hinder genuine preparation for potential AGI developments. The author emphasizes the need for public discourse on regulations and the ethical implications of AGI, advocating for transparency in the machine learning community. Ultimately, the author calls for a balanced approach to AGI, recognizing both its potential benefits and risks, and the importance of preparing for an uncertain future.

- Zuckerberg claims AI will soon perform tasks of mid-level engineers.

- The definition of AGI remains unclear, complicating evaluation efforts.

- Overpromising on AGI advancements risks public complacency.

- The author advocates for public discussions on AGI regulations and ethics.

- The timeline for achieving AGI is uncertain, requiring cautious preparation.

Link Icon 2 comments
By @proc0 - 3 months
What's often left out of the discussion is that even with DeepSeek R1 the costs of leveraging human levels of expertise is very costly at the moment. The evidence for this is obvious. All the examples for how the latest LLM is somehow good at coding are simple. It's someone entering a prompt for a snake game, or a bouncing ball, or a simple landing page.

If the current state of the art had the capability of a mid-level engineer, it would be as simple as entering a prompt like "create a facebook clone, here are AWS credentials...", and it should create a fully functioning social media site including the backend and DB setup. After all we're talking about a mid-level engineer for ALL the different sub-fields, backend, frontend, embedded systems, etc.

Of course, it isn't anywhere near this level, and anything LLMs produce is full of errors even in the simplest of examples... and it's still took a gigantic effort to get to this point by brute-forcing the transformer architecture with the entire internet. The impact will be nowhere near the current hype, but I do think we will eventually get there.