We're in the brute force phase of AI – once it ends, demand for GPUs will too
Gartner highlights a transitional phase in AI development, emphasizing the limited use of generative AI and recommending a return to traditional methods and composite AI for more effective outcomes.
Read original articleAnalyst firm Gartner has indicated that the current reliance on GPUs for AI workloads signifies a "brute force" phase in AI development, where programming techniques are not yet refined. Erick Brethenoux, Gartner's chief of research for AI, noted that specialized hardware often becomes obsolete once standard machines can perform the same tasks. He emphasized that generative AI, which has dominated discussions, accounts for only a small fraction of actual use cases. Many organizations are returning to established AI methods, such as machine learning and rule-based systems, after exploring generative AI without significant business benefits. Brethenoux highlighted the potential of composite AI, which combines generative AI with traditional techniques, as a more effective approach. Gartner's vice president, Bern Elliot, echoed these sentiments, cautioning against using generative AI for tasks beyond content generation and knowledge discovery due to its unreliability. He recommended implementing guardrails to ensure the accuracy of generative outputs. Overall, the discussion at Gartner's Symposium suggests a shift back to more practical AI applications as organizations reassess their strategies.
- Gartner warns that the reliance on GPUs indicates a transitional phase in AI development.
- Generative AI is overhyped, representing only a small percentage of actual use cases.
- Organizations are returning to established AI techniques after exploring generative AI.
- Composite AI, which integrates generative and traditional AI methods, is recommended for better outcomes.
- Caution is advised when using generative AI due to its unreliability in various applications.
Related
US intelligence community is embracing generative AI
The US intelligence community integrates generative AI for tasks like content triage and analysis support. Concerns about accuracy and security are addressed through cautious adoption and collaboration with major cloud providers.
Pop Culture
Goldman Sachs report questions generative AI's productivity benefits, power demands, and industry hype. Economist Daron Acemoglu doubts AI's transformative potential, highlighting limitations in real-world applications and escalating training costs.
A third of AI projects will be abandoned by businesses by the end of next year
Gartner predicts 30% of generative AI projects will be abandoned by next year due to poor data quality, rising costs, and unclear value, despite potential applications across various sectors.
Why the collapse of the Generative AI bubble may be imminent
The Generative AI bubble is predicted to collapse soon, with declining investor enthusiasm and funding, potentially leading to failures of high-valued companies by the end of 2024.
Gartner AI Hype Cycle: Autonomous AI Is on the Way (Apparently)
Gartner's 2024 Hype Cycle shows generative AI entering disillusionment, while autonomous AI development accelerates. Critics argue the Hype Cycle misrepresents tech trends, warning of potential financial sinkholes in AI investments.
> In economics, the Jevons paradox occurs when technological progress increases the efficiency with which a resource is used (reducing the amount necessary for any one use), but the falling cost of use induces increases in demand enough that resource use is increased, rather than reduced.
I think it was literally lack of imagination. We were like "well, I automated most of the paper pushing we used to do in the office, guess my job is done!" and this occupied 0.001% of a computer's time. We invented all sorts of ways for people to only pay for that tiny slice of active time (serverless, async web frameworks, etc).
Now we're in an era where we can actually use the computers we've built. I don't think we're going back
- [X] Text
- [X] Images
- [X] Audio
- [ ] Videos (in progress)
- [ ] 3D Meshes and Textures (in progress)
- [ ] Genetics (in progress)
- [ ] Physics Simulation (in progress)
- [ ] Mathematics
- [ ] Logic and Algorithms aka Planning and Optimization
- [ ] Reasoning
- [ ] Emotion
- [ ] Consciousness
We still have a lot of data to crunch but it's not nearly enough so we're also going to have to collect and generate a lot more of it. Some of these items require data that we don't even know how to collect yet. Barring some kind of disastrous event, draconian regulation, or politically/culturally motivated demonization of ML I don't see GPU demand dropping any time soon.Two Reddit threads really highlight this.
- ~10 years ago: https://www.reddit.com/r/StableDiffusion/comments/y9zxj1/you...
- Today: https://www.reddit.com/r/StableDiffusion/comments/1f0b45f/fl...
The upgrade in throughput from GPT-4 to GPT-4o and GPT-4o Mini actually unlocked use cases for the startup I'm at.
People that think demand for GPU compute capacity is going to decrease are probably wrong in the same way that people who thought the demand for faster processors and more RAM would wane were wrong. We are just barely at the start of finding the use cases and how to eat those GPU cycles.
> The need for specialist hardware, he observed, is a sign of the "brute force" phase of AI, in which programming techniques are yet to be refined and powerful hardware is needed. "If you cannot find the elegant way of programming … it [the AI application] dies," he added.
The thing is that even if there is an elegant and efficient programmatic/algorithmic solution, having more and faster hardware only makes it better and pushes the limits even more.To be clear, I agree that LLMs are not anywhere close to AGI and I don't think they ever will be (just a component). But that doesn't mean they aren't useful enough to chew up a lot of compute for the foreseeable future.
Machine learning is a trade off between model size (training cost), model run time (inference cost), and quality.
When some task is solved (e.g., hot word detection or speech to text), it becomes a commodity and some harder task becomes the priority.
I know that as soon as I can output 100req/s on the cheap on a llama-level model I will put it EVERYWHERE. And my clients too.
DMCA handling? Content flagging alerts? Fuzzy categorization? Natural UI for end user complex queries?
All LLM baby.
And much, much more.
- if a company, say, AMD, found a way to produce GPUs at a fraction (say, x=10%) of the price of NVDA, would that increase aggregate demand, or keep it about the same (substitution for nVidia)? Would the price difference be enough to incentivize creation of a CUDA- alternative ecosystem? If not, what does the fraction x need to be?
- Very reductively, GPUs seem to be universally needed because they are better at manipulating matrices than alternatives (eg, inverting a matrix, finding dot product, cosine similarity, etc). Are there alternative approaches that could come to market in the next 2-3 years that could be better, or better-per-$, than the current approach of just building bigger GPUs?
And even LLM's and photo generation open up a bajillion usecases that would have taken years of research and development before. But nobody focuses on these - instead, they focus on S&P 500 companies and how they haven't earned much with GenAI.
Because these companies are usually known as the peak of human creativity, imagination and are ready to jump on a new technology without much red tape in it.
Honestly, it's been like two-three years tops. Even talking to tech startup CEO's I don't get the feeling they remotely understand the technology or application, as 90% of things I've heard them say is "oooh we could make a chatbot!" or "let's replace developers with it - oh it can't one shot generate my whole codebase? pft that sucks".
If these folks don't know how to use it, surely Jim VP of Engineering #62 at ACME & CO that hasn't used any tech except ERP's for the last 10 years will have an idea how to.
Really don’t see why he thinks brute force is going away. I mean human brains have billions of neurons too after all.
SOTA will remain at the edge of what compute can produce for a long time to come. SOTA is a moving frontier, and there will be demand for incrementally smarter models, because they save you time by making fewer mistakes.
No matter how much more efficient algorithmic innovations make ML model training, compute will make those algorithms smarter. It's a coefficient.
Related
US intelligence community is embracing generative AI
The US intelligence community integrates generative AI for tasks like content triage and analysis support. Concerns about accuracy and security are addressed through cautious adoption and collaboration with major cloud providers.
Pop Culture
Goldman Sachs report questions generative AI's productivity benefits, power demands, and industry hype. Economist Daron Acemoglu doubts AI's transformative potential, highlighting limitations in real-world applications and escalating training costs.
A third of AI projects will be abandoned by businesses by the end of next year
Gartner predicts 30% of generative AI projects will be abandoned by next year due to poor data quality, rising costs, and unclear value, despite potential applications across various sectors.
Why the collapse of the Generative AI bubble may be imminent
The Generative AI bubble is predicted to collapse soon, with declining investor enthusiasm and funding, potentially leading to failures of high-valued companies by the end of 2024.
Gartner AI Hype Cycle: Autonomous AI Is on the Way (Apparently)
Gartner's 2024 Hype Cycle shows generative AI entering disillusionment, while autonomous AI development accelerates. Critics argue the Hype Cycle misrepresents tech trends, warning of potential financial sinkholes in AI investments.