Nvidia's Project Digits is a 'personal AI supercomputer'
Nvidia's Project Digits, a compact personal AI supercomputer, features the GB10 Grace Blackwell Superchip, offers up to a petaflop performance, and targets AI researchers, priced at $3,000 from May 2025.
Read original articleNvidia introduced Project Digits at CES 2025, describing it as a "personal AI supercomputer" that utilizes the Grace Blackwell hardware platform in a compact design. CEO Jensen Huang highlighted that Project Digits runs the complete Nvidia AI software stack and functions as a cloud computing platform suitable for desktop use. Targeted at AI researchers, data scientists, and students, it features the GB10 Grace Blackwell Superchip, capable of delivering up to a petaflop of performance for AI model development. The system can handle models with up to 200 billion parameters, with the option to link two units for a combined capacity of 405 billion parameters. The hardware includes an Nvidia Blackwell GPU, a 20-core Nvidia Grace CPU, 128GB of memory, and up to 4TB of flash storage. Project Digits will be available starting in May for $3,000, making it a premium product aimed at a specific market segment. Huang believes that this technology will empower developers and researchers to significantly contribute to advancements in AI.
- Nvidia's Project Digits is a compact personal AI supercomputer unveiled at CES 2025.
- It features the GB10 Grace Blackwell Superchip, offering up to a petaflop of computing performance.
- The system can run AI models with up to 200 billion parameters and can be linked for greater capacity.
- Project Digits will be available for $3,000 starting in May 2025.
- The product is aimed at AI researchers, data scientists, and students, promoting accessibility to advanced AI tools.
Related
Intel's Gaudi 3 will cost half the price of Nvidia's H100
Intel's Gaudi 3 AI processor is priced at $15,650, half of Nvidia's H100. Intel aims to compete in the AI market dominated by Nvidia, facing challenges from cloud providers' custom AI processors.
XAI's Memphis Supercluster has gone live, with up to 100,000 Nvidia H100 GPUs
Elon Musk launches xAI's Memphis Supercluster with 100,000 Nvidia H100 GPUs for AI training, aiming for advancements by December. Online status unclear, SemiAnalysis estimates 32,000 GPUs operational. Plans for 150MW data center expansion pending utility agreements. xAI partners with Dell and Supermicro, targeting full operation by fall 2025. Musk's humorous launch time noted.
Nvidia and MediaTek Collaborate on 3nm AI PC CPU
Nvidia and MediaTek are collaborating on a 3nm AI CPU, entering production this month, with mass production expected in late 2025, potentially priced around $300 and enhancing graphical performance.
Nvidia's Christmas Present: GB300 and B300 – Reasoning Inference, Amazon, Memory
Nvidia launched the GB300 and B300 GPUs, enhancing reasoning model performance with a 50% increase in FLOPS, upgraded memory, and a restructured supply chain benefiting OEMs and hyperscalers.
Nvidia bets on robotics to drive future growth
Nvidia is focusing on robotics for growth, launching Jetson Thor in 2025. The global robotics market is expected to grow from $78 billion to $165 billion by 2029, despite safety challenges.
- Many users express enthusiasm for the performance and pricing, considering it a good value compared to existing options.
- Concerns about software support and Nvidia's history with previous products, like the Jetson Nano, are frequently mentioned.
- Some commenters highlight potential applications in fields like bioinformatics and personal AI, suggesting a broader market for the device.
- There are discussions about the implications for local computing versus cloud solutions, emphasizing privacy and data security.
- Several users question the actual specifications and capabilities, particularly regarding CPU performance and software compatibility.
Nvidia Jetson Nano, A SBC for "AI" debuted with already aging custom Ubuntu 18.04 and when 18.04 went EOL, Nvidia abandoned it completely without any further updates to its proprietary jet-pack or drivers and without them all of Machine Learning stack like CUDA, Pytorch etc. became useless.
I'll never buy a SBC from Nvidia unless all the SW support is up-streamed to Linux kernel.
The owner of the market, Illumina, already ships their own bespoke hardware chips in servers called DRAGEN for faster analysis of thousands of genomes. Their main market for this product is in personalised medicine, as genome sequencing in humans is becoming common.
Other companies like Oxford Nanopore use on-board GPUs to call bases (i.e., from raw electric signal coming off the sequencer to A, T, G, C) but it's not working as well as it could due to size and power constraints. I feel like this could be a huge game changer for someone like ONT, especially with cooler stuff like adaptive sequencing.
Other avenues of bioinformatics, such as most day-to-day analysis software, is still very CPU and RAM heavy.
"DGX OS 6 Features The following are the key features of DGX OS Release 6:
Based on Ubuntu 22.04 with the latest long-term Linux kernel version 5.15 for the recent hardware and security updates and updates to software packages, such as Python and GCC.
Includes the NVIDIA-optimized Linux kernel, which supports GPU Direct Storage (GDS) without additional patches.
Provides access to all NVIDIA GPU driver branches and CUDA toolkit versions.
Uses the Ubuntu OFED by default with the option to install NVIDIA OFED for additional features.
Supports Secure Boot (requires Ubuntu OFED).
Supports DGX H100/H200."
It's a garden hermit. Imagine a future where everyone has one of those(not exactly this version but some future version), it lives with you it learns with you and unlike the cloud based SaaS AI you can teach it things immediately and diverge from the average to your advantage.
$100M, 2.35MW, 6000 ft^2
>>Designed for AI researchers, data scientists, and students, Project Digits packs Nvidia’s new GB10 Grace Blackwell Superchip, which delivers up to a petaflop of computing performance for prototyping, fine-tuning, and running AI models.
$3000, 1kW, 0.5 ft^2
https://www.okdo.com/wp-content/uploads/2023/03/jetson-agx-o...
I wonder what the specifications are in terms of memory bandwidth and computational capability.
>This paper describes how the performance of AI machines tends to improve at the same pace that AI researchers get access to faster hardware. The processing power and memory capacity necessary to match general intellectual performance of the human brain are estimated. Based on extrapolation of past trends and on examination of technologies under development, it is predicted that the required hardware will be available in cheap machines in the 2020s.
and this is about the first personal unit that seems well ahead of his proposed specs. (He estimated 0.1 petaflops. The nvidia thing is "1 petaflop of AI performance at FP4 precision").
edit: While the title says "personal", Jensen did say this was aimed at startups and similar, so not your living room necessarily.
https://s3.amazonaws.com/cms.ipressroom.com/219/files/20250/...
Source: https://nvidianews.nvidia.com/news/nvidia-puts-grace-blackwe...
Ideally we can configure things like Apple Intelligence to use this instead of OpenAI and Apple's cloud.
tinybox red and green are for people looking for a quiet home/office machine. tinybox pro is for people looking for a loud compact rack machine.” [0]
This is really game changer.
They should make a deal with Valve to turn this into 'superconsole' that can run Half Life 3 (to be announced) :)
Just like Mac OS is free when you buy a Mac, having the latest high-quality LLM for free that just happens to run well on this box is a very interesting value-prop. And Nvidia definitely has the compute to make it happen.
Do I understand that right? It seems way to cheap.
At $3,000, it will be considerably cheaper than alternatives available today (except for SoC boards with extremely poor performance, obviously). I also expect that Nvidia will use its existing distribution channels for this, giving consumers a shot at buying the hardware (without first creating a company and losing consumer protections along the way).
Joking aside, personally will buy this workstation in a heartbeat if I have the budget to spare, one in the home and another in the office.
Currently I have an desktop/workstation for AI workloads with similar 128GB RAM that I bought few years back that cost around USD5K without the NVIDIA GPU that I bought earlier for about USD1.5K, with a total of about USD6.5K without a display monitor. This the same price of NeXT workstation (with a monitor) when it's sold back in 1988 without adjusting for inflations (now around USD18K) but it is more than 200 times faster in CPU speed and more than 1000 times RAM capacity than the original 25 MHz CPU and 4 MB RAM, respectively. The later updated version of NeXT has graphic accelerator with 8 MB VRAM, since the workstation has RTX 2080 it is about 1000 times more. I believe the updated NeXT with graphic accelerator is the one that used to develop original Doom software [1].
If NVIDIA can sell the Project Digits Linux desktop at USD3K with similar or more powerful setup configurations, it's going to be a winner and probably can sell by truckloads. It seems to has NeXT workstation vibe to it that used to develop the original WWW and Doom software. Hopefully it will be used to develop many innovative software but now using open source Linux software eco-system not proprietary one.
The latest Linux kernel now has real-time capability for more responsive desktop experience and as saying goes, good things come to those who wait.
[1] NeXT Computer:
Future versions will get more capable and smaller, portable.
Can be used to train new types models (not just LLMs).
I assume the GPU can do 3D graphics.
Several of these in a cluster could run multiple powerful models in real time (vision, llm, OCR, 3D navigation, etc).
If successful, millions of such units will be distributed around the world within 1-2 years.
A p2p network of millions of such devices would be a very powerful thing indeed.
One can only wish for this, but Nvidia would be going against the decades-long trend to emaciate local computing in favor of concentrating all compute on somebody else's linux (aka: cloud).
If one can skip buying gaming rig with a 5090 with its likely absurd price then this 3k becomes a lot easier for dual use hobbyists to swallow
Edit 5090 is 2k
But it's clear that everyone's favorite goal is keretsuification. If you're looking for abnormal profits, you can't do better than to add a letter to FAANG. Nvidia already got into the cloud business, and now it's making workstations.
The era of specialists doing specialist things is not really behind us. They're just not making automatic money, nor most of it. Nvidia excelled in that pool, but it too can't wait to leave it. It knows it can always fail as a specialist, but not as a kereitsu.
I'm bracing for a whole new era of unsufferable binary blobs for Linux users, and my condolences if you have a non-ultramainstream distro.
While I'm quite the "AI" sceptic I think it might be interesting to have a node in my home network capable of a bit of this and that in this area, some text-to-speech, speech-to-text, object identification, which to be decent needs a bit more than the usual IoT- and ESP-chips can manage.
https://s3.amazonaws.com/cms.ipressroom.com/219/files/20250/...
This goes against every definition of cloud that I know off. Again proving that 'cloud' means whatever you want it to mean.
First product that directly competes on price with Macs for local inferencing of large LLMs (higher RAM). And likely outperforms them substantially.
Definitely will upgrade my home LLM server if specs bear out.
Edit: Sorry fucked up my math. I wanted to do 40x52x4, $4/hr being the cloud compute price but that us actually $8300, so it is actually equivalent to about 4.5 months of cloud compute. 40 hours because I presume that this will only be used for prototyping and debugging, i.e during office hours.
Related
Intel's Gaudi 3 will cost half the price of Nvidia's H100
Intel's Gaudi 3 AI processor is priced at $15,650, half of Nvidia's H100. Intel aims to compete in the AI market dominated by Nvidia, facing challenges from cloud providers' custom AI processors.
XAI's Memphis Supercluster has gone live, with up to 100,000 Nvidia H100 GPUs
Elon Musk launches xAI's Memphis Supercluster with 100,000 Nvidia H100 GPUs for AI training, aiming for advancements by December. Online status unclear, SemiAnalysis estimates 32,000 GPUs operational. Plans for 150MW data center expansion pending utility agreements. xAI partners with Dell and Supermicro, targeting full operation by fall 2025. Musk's humorous launch time noted.
Nvidia and MediaTek Collaborate on 3nm AI PC CPU
Nvidia and MediaTek are collaborating on a 3nm AI CPU, entering production this month, with mass production expected in late 2025, potentially priced around $300 and enhancing graphical performance.
Nvidia's Christmas Present: GB300 and B300 – Reasoning Inference, Amazon, Memory
Nvidia launched the GB300 and B300 GPUs, enhancing reasoning model performance with a 50% increase in FLOPS, upgraded memory, and a restructured supply chain benefiting OEMs and hyperscalers.
Nvidia bets on robotics to drive future growth
Nvidia is focusing on robotics for growth, launching Jetson Thor in 2025. The global robotics market is expected to grow from $78 billion to $165 billion by 2029, despite safety challenges.