Nvidia announces next-gen RTX 5090 and RTX 5080 GPUs
Nvidia announced its RTX 50-series GPUs, including the RTX 5090, 5080, 5070 Ti, and 5070, featuring GDDR7 memory, improved performance, and DLSS 4 technology, launching between January and March 2025.
Read original articleNvidia has officially announced its next-generation RTX 50-series GPUs during a CES keynote, featuring four models: the RTX 5090 priced at $1,999, the RTX 5080 at $999, the RTX 5070 Ti at $749, and the RTX 5070 at $549. The RTX 5090 and RTX 5080 will be available on January 30, 2025, while the other two models will launch in February. The RTX 5090 is designed to be compact, featuring a new Founders Edition design with two double flow-through fans and GDDR7 memory. It boasts 32GB of GDDR7, 21,760 CUDA cores, and a memory bandwidth of 1,792GB/sec, claiming to be twice as fast as the RTX 4090. The RTX 5080 offers 16GB of GDDR7, 10,752 CUDA cores, and a memory bandwidth of 960GB/sec, promising significant performance improvements over the RTX 4080. Additionally, Nvidia is introducing the RTX 50-series for laptops, with models set to launch in March. The new GPUs will utilize DLSS 4 technology, enhancing performance and image quality through AI advancements. Nvidia's announcement marks a significant upgrade from the previous RTX 40-series, focusing on improved ray tracing and overall gaming performance.
- Nvidia's RTX 50-series includes four models: RTX 5090, 5080, 5070 Ti, and 5070.
- The RTX 5090 is claimed to be twice as fast as the RTX 4090, with a compact design.
- DLSS 4 technology will enhance performance and image quality across the new GPUs.
- RTX 50-series laptops will be available starting in March 2025.
- The GPUs feature GDDR7 memory and improved power efficiency compared to previous models.
Related
RTX 5080 rumored to have 400W TGP
Nvidia plans to finalize designs for the RTX 50-series, including the RTX 5080 with a 400-watt TGP, by September 2024, with a potential launch at CES 2025.
Leak claims RTX 5090 has 600W TGP, 32GB VRAM, 512-bit bus
Preliminary specs for Nvidia's RTX 5080 and RTX 5090 suggest significant performance advancements, with the 5090 featuring 600W TGP and 21,760 CUDA cores, potentially launching in early 2025.
Leak claims RTX 5090 has 600W TGP, RTX 5080 hits 400W
Preliminary specifications for Nvidia's RTX 5080 and RTX 5090 graphics cards suggest significant performance advancements, with the RTX 5090 featuring 600W TGP and 21,760 CUDA cores, potentially launching in early 2025.
Zotac confirms GeForce RTX 5090 with 32GB GDDR7 memory
ZOTAC has confirmed the NVIDIA GeForce RTX 5090 with 32GB GDDR7 memory, alongside the RTX 5080 and 5070 series, expected to launch in early 2025, focusing on high performance.
Nvidia GeForce RTX 5090 reportedly features TDP of 575W, RTX 5080 set at 360W
NVIDIA's GeForce RTX 5090 and RTX 5080 graphics cards will have TDPs of 575W and 360W, respectively, with significant CUDA cores and memory, supporting PCIe 5.0 and DisplayPort 2.1a.
- Many users express skepticism about the actual performance improvements, suggesting that gains are primarily due to DLSS 4 rather than hardware advancements.
- Concerns about power consumption and thermal output are prevalent, with the RTX 5090's TDP reaching 575W.
- Commenters highlight the ongoing issue of GPU availability and pricing, fearing that gamers will struggle to purchase these new cards.
- There is a notable discussion about the adequacy of VRAM, with many feeling that 32GB is insufficient for future gaming and AI workloads.
- Some users are critical of Nvidia's marketing strategies, perceiving a trend towards "fake" performance metrics rather than genuine improvements.
The 5090 just has way more CUDA cores and uses proportionally more power compared to the 4090, when going by CUDA core comparisons and clock speed alone.
All of the "massive gains" were comparing DLSS and other optimization strategies to standard hardware rendering.
Something tells me Nvidia made next to no gains for this generation.
I can understand lack of supply, but why can't I go on nvidia.com and buy something the same way I go on apple.com and buy hardware?
I'm looking for GPUs and navigating all these different resellers with wildly different prices and confusing names (on top of the already confusing set of available cards).
3090 - 350W
3090 Ti - 450W
4090 - 450W
5090 - 575W
3x3090 (1050W) is less than 2x5090 (1150W), plus you get 72GB of VRAM instead of 64GB, if you can find a motherboard that supports 3 massive cards or good enough risers (apparently near impossible?).
32 GB VRAM on the highest end GPU seems almost small after running LLMs with 128 GB RAM on the M3 Max, but the speed will most likely more than make up for it. I do wonder when we’ll see bigger jumps in VRAM though, now that the need for running multiple AI models at once seems like a realistic use case (their tech explainers also mentions they already do this for games).
Presumably the pro hardware based on the same silicon will have 64GB, they usually double whatever the gaming cards have.
https://www.nvidia.com/en-us/geforce/graphics-cards/50-serie...
When was the last time Nvidia made a high end GeForce card use only 2 slots?
Translation: No significant actual upgrade.
Sounds like we're continuing the trend of newer generations being beaten on fps/$ by the previous generations while hardly pushing the envelope at the top end.
A 3090 is $1000 right now.
2x faster in DLSS. If we look at the 1:1 resolution performance, the increase is likely 1.2x.
Seemingly NVIDIA is just playing number games, like wow 3352 is a huge leap compared to 1321 right? But how does it really help us in LLMs, diffusion models and so on?
Does anyone know what these might cost in the US after the rumored tariffs?
Going from 60 to 120fps is cool. Going from 120fps to 240fps is in the realm of diminishing returns, especially because the added latency makes it a non starter for fast paced multiplayer games.
12GB VRAM for over $500 is an absolute travesty. Even today cards with 12GB struggle in some games. 16GB is fine right now, but I'm pretty certain it's going to be an issue in a few years and is kind of insane at $1000. The amount of VRAM should really be double of what it is across the board.
You can try out pretty much all GPUs on a cloud provider these days. Do it.
VRAM is important for maxing out your batch size. It might make your training go faster, but other hardware matters too.
How much having more VRAM speeds things up also depends on your training code. If your next batch isn’t ready by the time one is finished training, fix that first.
Coil whine is noticeable on my machine. I can hear when the model is training/next batch is loading.
Don’t bother with the founder’s edition.
If that holds up in the benchmarks, this is a nice jump for a generation. I agree with others that more memory would've been nice, but it's clear Nvidia are trying to segment their SKUs into AI and non-AI models and using RAM to do it.
That might not be such a bad outcome if it means gamers can actually buy GPUs without them being instantly bought by robots like the peak crypto mining era.
I would expect something like the 5080 super will have something like 20/24Gb of VRAM. 16Gb just seems wrong for their "target" consumer GPU.
* Neural texture stuff - also super exciting, big advancement in rendering, I see this being used a lot (and helps to make up for the meh vram blackwell has)
* Neural material stuff - might be neat, Unreal strata materials will like this, but going to be a while until it gets a good amount of adoption
* Neural shader stuff in general - who knows, we'll see how it pans out
* DLSS upscaling/denoising improvements (all GPUs) - Great! More stable upscaling and denoising is very much welcome
* DLSS framegen and reflex improvements - bleh, ok I guess, reflex especially is going to be very niche
* Hardware itself - lower end a lot cheaper than I expected! Memory bandwidth and VRAM is meh, but the perf itself seems good, newer cores, better SER, good stuff for the most part!
Note that the material/texture/BVH/denoising stuff is all research papers nvidia and others have put out over the last few years, just finally getting production-ized. Neural textures and nanite-like RT is stuff I've been hyped for the past ~2 years.
I'm very tempted to upgrade my 3080 (that I bought used for $600 ~2 years ago) to a 5070 ti.
RTX 5090: 32 GB GDDR7, ~1.8 TB/s bandwidth. H100 (SXM5): 80 GB HBM3, ~3+ TB/s bandwidth.
RTX 5090: ~318 TFLOPS in ray tracing, ~3,352 AI TOPS. H100: Optimized for matrix and tensor computations, with ~1,000 TFLOPS for AI workloads (using Tensor Cores).
RTX 5090: 575W, higher for enthusiast-class performance. H100 (PCIe): 350W, efficient for data centers.
RTX 5090: Expected MSRP ~$2,000 (consumer pricing). H100: Pricing starts at ~$15,000–$30,000+ per unit.
It seems obvious to me that even NVIDIA knows that 5090s and 4090s are used more for AI Workloads than gaming. In my company, every PC has 2 4090s, and 48GB is not enough. 64GB is much better, though I would have preferred if NVIDIA went all in and gave us a 48GB GPU, so that we could have 96GB workstations at this price point without having to spend 6k on an A6000.
Overall I think 5090 is a good addition to the quick experimentation for deep learning market, where all serious training and inference will occur on cloud GPU clusters, but we can still do some experimentation on local compute with the 5090.
I always end up late to the party and the prices end up being massively inflated - even now I cant seem to buy a 4090 for anywhere close to the RRP.
it is easy to be carried away with vram size, but keeping in mind that most people with apple silicon (who can enjoy several times more memory) are stuck at inference, while training performance is off the charts through cuda hardware.
the jury is yet to be out on actual ai training performance, but i bet 4090, if sold at 1k or below, would be better value than lower tier 50 series. the "ai tops" of the 50 series is only impressive for the top model, while the rest are either similar or with lower memory bandwidth despite the newer architecture.
i think by now the training is best left on the cloud and overall i'd be happy rather owning a 5070 ti at this rate.
Gaming performance has been plateaued for some time now, maybe an 8k monitor wave can revive things
I miss when high-end GPUs were $300-400, and you could get something reasonable for $100-200. I guess that's just integrated graphics these days.
The most I've ever spent on a GPU is ~$300, and I don't really see that changing anytime soon, so it'll be a long time before I'll even consider one of these cards.
This is the same thing they did with the RTX 4000 series. More fake frames, less GPU horsepower, "Moore's Law is Dead", Jensen wrings his hands, "Nothing I can do! Moore's Law is Dead!" which is how Intel has been slacking since 2013.
I'm planning to upgrade (prob to a mid-end) as my 5 year old computer is starting to show it's age, and with the new GPUs releasing this might be a good time.
Related
RTX 5080 rumored to have 400W TGP
Nvidia plans to finalize designs for the RTX 50-series, including the RTX 5080 with a 400-watt TGP, by September 2024, with a potential launch at CES 2025.
Leak claims RTX 5090 has 600W TGP, 32GB VRAM, 512-bit bus
Preliminary specs for Nvidia's RTX 5080 and RTX 5090 suggest significant performance advancements, with the 5090 featuring 600W TGP and 21,760 CUDA cores, potentially launching in early 2025.
Leak claims RTX 5090 has 600W TGP, RTX 5080 hits 400W
Preliminary specifications for Nvidia's RTX 5080 and RTX 5090 graphics cards suggest significant performance advancements, with the RTX 5090 featuring 600W TGP and 21,760 CUDA cores, potentially launching in early 2025.
Zotac confirms GeForce RTX 5090 with 32GB GDDR7 memory
ZOTAC has confirmed the NVIDIA GeForce RTX 5090 with 32GB GDDR7 memory, alongside the RTX 5080 and 5070 series, expected to launch in early 2025, focusing on high performance.
Nvidia GeForce RTX 5090 reportedly features TDP of 575W, RTX 5080 set at 360W
NVIDIA's GeForce RTX 5090 and RTX 5080 graphics cards will have TDPs of 575W and 360W, respectively, with significant CUDA cores and memory, supporting PCIe 5.0 and DisplayPort 2.1a.