Nvidia Blackwell and GeForce RTX 50-Series GPUs: Specifications, release dates, pricing, and everything we know (updated)

Jump to:

The next-generation Nvidia Blackwell GPU architecture and RTX 50-series GPUs are coming, basically on schedule. Nvidia officially detailed the first four cards in the Blackwell RTX 50-series family at CES 2025, during CEO Jensen Huang’s keynote on January 6. We expect the various Blackwell GPUs will join the ranks of the best graphics cards, replacing their soon-to-be-prior-generation counterparts.

When we spoke with some people in early 2024, the expectation was that we’d see at least the RTX 5090 and RTX 5080 by the time the 2024 holiday season rolled around. But then came the delay of Blackwell B200 along with packaging problems, and that appears to have pushed things back. Now, we’re looking at a January 2025 announcement with at least one or two models coming before the end of the month, and perhaps as many as four different desktop cards — and the possibility of laptop RTX 50-series also exists.

Nvidia already provided many of the core details for its data center Blackwell B200 GPU. The AI and data center variants will inevitably differ from consumer parts, but there are some shared aspects between past consumer and data center Nvidia GPUs, and that should continue. That gives some good indications of certain aspects of the future RTX 50-series GPUs.

Things are starting to clear up now, with hard specifications and pricing details for the first four GPUs. There’s still no official word on the 5060-class GPUs, but those should arrive at some point in the coming months.

Let’s talk about specifications, technology, pricing, and other details. We’ve been updating this article for a while now, as information became available, and we’re now in the home stretch. Here’s everything we know about Nvidia Blackwell and the RTX 50-series GPUs.

Blackwell and RTX 50-series Release Dates

Despite what we personally heard in early 2024, the RTX 50-series didn’t make it out the door in 2024, but the first models will launch in January 2025. There were some delays, but not directly related to the consumer GPUs.

Nvidia’s data center Blackwell B100/B200 GPUs encountered packaging problems and were also delayed. Given how much money the data center segment raked in over the past year (see Nvidia’s latest earnings), putting more money and wafers into getting B200 ready and available makes sense. Gamers? Yeah, we’re no longer Nvidia’s top priority.

The consumer Blackwell GPUs are “late,” based on historical precedent. The Ada Lovelace RTX 40-series GPUs first appeared in October 2022. The Ampere RTX 30-series GPUs first appeared in September 2020. Prior to that, RTX 20-series launched two years earlier in September 2018, and the GTX 10-series was in May/June 2016, with the GTX 900-series arriving in September 2014. That’s a full decade of new Nvidia GPU architectures arriving approximately every two years. Even so, we’re still only a few months beyond the normal cadence.

And now we’re into 2025, and Nvidia spilled the beans on the RTX 5090, 5080, 5070 Ti, and 5070 — along with mobile variants — during the CES 2025 keynote. The top-tier RTX 5090 and 5080 will arrive first, in January 2025. The RTX 5070 Ti and RTX 5070 will come next, probably in February. The 5060-class hardware could come any time within the next six months after the first 50-series GPUs. As usual, we expect Blackwell GPUs to follow the typical staggered release schedule.

Still TSMC 4N, 4nm Nvidia

One of the surprising announcements at GTC 2024 was that Blackwell B200 will use the TSMC 4NP node — “4nm Nvidia Performance,” or basically a tuned/tweaked variation of the N4 node used on the RTX 40-series. Blackwell B200 also uses a dual-chip solution, with the two identical chips linked via a 10 TB/s NV-HBI (Nvidia High Bandwidth Interface) connection.

While it’s certainly true that process names have largely become detached from physical characteristics, many expected Nvidia to move to a variant of TSMC’s cutting-edge N3 process technology. Instead, it opted for a refinement of the existing 4N node that has already been used with Hopper and Ada Lovelace GPUs for the past two years.

For the consumer Blackwell chips, though, Nvidia will stick with TSMC 4N. Yes, the same node and process as Ada, with no changes that we’re aware of. Going this route certainly offers some cost savings, though TSMC doesn’t disclose the contract pricing agreements with its various partners. Perhaps Nvidia just didn’t think it needed to move to a 3nm-class node for this generation.

AMD will be moving to TSMC N4, while Intel will use TSMC N5 for Battlemage. So, even though Nvidia didn’t choose to pursue 3nm or 2nm this round, it’s still equal to or better than the competition. And the Ada architecture was already well ahead in terms of efficiency, performance, and features in many areas.

Next generation GDDR7 memory

Blackwell GPUs will move to GDDR7 memory, at least for the RTX 5070 and above. We don’t know for certain what the 5060-class will use, and we don’t have official clock speeds, but we do know that the 5070 through 5090 are all using GDDR7 memory.

The current generation RTX 40-series GPUs use GDDR6X and GDDR6 memory, clocked at anywhere from 17Gbps to 23Gbps. GDDR7 has target speeds of up to 36Gbps, 50% higher than GDDR6X and 80% higher than vanilla GDDR6. SK hynix says it will even have 40Gbps chips, though the exact timeline for when those might be available hasn’t been given. Regardless, GDDR7 will provide a much-needed boost to memory bandwidth at all levels.

Nvidia won’t actually ship cards with memory clocked at 36Gbps, though. In the past, it used 24Gbps GDDR6X chips but clocked them at 22.4Gbps or 23Gbps — and some 24Gbps Micron chips were apparently down-binned to 21Gbps in the various RTX 4090 graphics cards that we tested. The RTX 5090, 5070 Ti, and 5070 will clock their GDDR7 at 28Gbps, while the RTX 5080 opts for a higher memory speed of 30Gbps. Either way, that’s still a healthy bump to bandwidth.

At 28Gbps, GDDR7 memory provides a solid 33% increase in memory bandwidth compared to the 21Gbps GDDR6X used on the RTX 4090. The RTX 5080 opts for 30Gbps GDDR7, a 30% increase in bandwidth relative to the RTX 4080 Super, and 34% more bandwidth than the original 4080’s 22.4Gbps memory. As with so many other aspects of Blackwell, it remains to be seen just how far Nvidia and its partners will push things.

Nvidia will keep using a large L2 cache with Blackwell. This will provide even more effective memory bandwidth — every cache hit means a memory access that doesn’t need to happen. With a 50% cache hit rate as an example, that would double the effective memory bandwidth, though note that hit rates vary by game and settings, with higher resolutions in particular reducing the hit rate.

GDDR7 also potentially addresses the issue of memory capacity versus interface width. At GTC, we were told that 16Gb chips (2GB) are in production, with 24Gb (3GB) chips also coming. Are the larger chips with non-power-of-two capacity ready for upcoming Blackwell GPUs? Apparently not for the initial salvo, or at least not for the initial desktop cards.

There’s been at least one rumor suggesting Nvidia might have 16GB (2GB chips) and 24GB (3GB chips) variants of the RTX 5080. As long as the price difference isn’t too onerous and the other specs remain the same, that wouldn’t be a bad approach. The base models announced so far all come with 2GB chips, while upgraded variants could have 50% more VRAM capacity courtesy of the 3GB chips.

We’ve already seen this with the RTX 5090 Laptop GPU, which uses the same GB203 chip as the desktop RTX 5080/5070 Ti, with a 256-bit memory interface. But then it uses 3GB GDDR7 chips so that the resulting graphics chip has 24GB of VRAM. If Nvidia can do this for the mobile 5090, why isn’t it already doing it for the desktop 5080? Probably because it doesn’t feel that it needs to — and to leave room for a mid-cycle refresh using those chips in the future.

Right now, there’s no pressing need for consumer graphics cards to have more than 24GB of memory. But RTX 5090 has a 512-bit interface, meaning it will come with a default 32GB configuration and could offer a 48GB variant in the future. The higher capacity GDDR7 chips could be particularly beneficial for professional and AI focused graphics cards, where large 3D models and LLMs are becoming increasingly common. A 512-bit interface with 3GB chips on both sides of the PCB could yield a professional RTX 6000 Blackwell Generation as an example with 96GB of memory.

More importantly, the availability of 24Gb chips means Nvidia (along with AMD and Intel) could put 18GB of VRAM on a 192-bit interface, 12GB on a 128-bit interface, and 9GB on a 96-bit interface, all with the VRAM on one side of the PCB. We could also see 24GB cards with a 256-bit interface, and 36GB on a 384-bit interface — and double that capacity for professional cards. Pricing will certainly be a factor for VRAM capacity, but it’s more likely a case of “when” rather than “if” we’ll see 24Gb GDDR7 memory chips on consumer GPUs.

But there’s more going on than just raw VRAM capacity. Shown during the CES 2025 keynote, RTX Neural Materials could cut the VRAM requirements of textures by about one third. If that needs to be implemented on a per-game basis, it won’t help 8GB cards in all situations, but if it’s a driver-side enhancement, 8GB could actually be “enough” for most games again.

Blackwell architectural updates

The Blackwell architecture will have various updates and enhancements over the previous generation Ada Lovelace architecture. Nvidia hasn’t gone into a lot of detail, but one thing is clear: AI is a big part of Nvidia’s plans for Blackwell. We know that AI TOPS (teraops) performance per tensor core has been doubled, but Nvidia is pulling a fast one there. Technically, that’s for FP4 workloads (which should be classified as FP4 TFLOPS, not AI TOPS, but we digress).

The native support for FP4 and FP6 formats boosts raw compute with lower precision formats, but the per-tensor-core compute has remained the same otherwise. Every generation of Nvidia GPUs has contained other architectural upgrades as well, and we can expect the same to occur this round.

Nvidia has increased the potential ray tracing performance in every RTX generation, and Blackwell continues that trend, doubling the ray/triangle intersection rates for the RT cores. With more games like Alan Wake 2 and Cyberpunk 2077 pushing full path tracing — not to mention the potential for modders to use RTX Remix to enhance older DX10-era games with full path tracing — there’s even more need for higher ray tracing throughput.

What other architectural changes might Blackwell bring? Considering Nvidia is sticking with TSMC 4N for the consumer parts, we don’t anticipate massive alterations. There will still be a large L2 cache, and the tensor cores are now in charge of handling the optical flow calculations for frame generation and multi frame generation. DLSS 4 “neural rendering” is coming with various other enhancements including multi-frame generation on the RTX 50-series.

Raw compute, for both graphics and more general workloads, sees a modest bump on the RTX 5090 compared to the RTX 4090, though other GPUs don’t seem to be getting as much of an increase. Again, AI might make up for that, but a lot remains to be seen. The 5070 for example offers up 31 TFLOPS of compute, compared to 29 TFLOPS on the 4070. The 5090 has 107 TFLOPS compared to 83 TFLOPS on the 4090.