/r/hardware
/r/hardware is a place for quality computer hardware news, reviews, and intelligent discussion.
The goal of /r/hardware is a place for quality hardware news, reviews, and intelligent discussion.
/r/hardware IS NOT the place to come for help of any kind.
Techsupport and PC building questions should be posted to /r/techsupport or /r/buildapc instead.
The goal of /r/hardware is a place for quality hardware news, reviews, and intelligent discussion.
/r/hardware is not the place to come for help of any kind. This includes tech support and PC building questions.
Rule Description Follow the Reddit Content Policy You can find it here. TL;DR: If you can't say something respectfully, don't say it at all. Insults and personal attacks aren't welcome here. Post should be about hardware Posts should be about hardware news, reviews, technical discussion or how-tos and buyers guides. No editorializing titles Please use the "suggest title" button for link submissions, or copy the title of the original link. Do NOT editorialize the title of the submission; (minor) changes for clarity may be acceptable if the original title is clickbait, or failed to summarize its actual content. Original Source Policy Content submitted should be of original source, or at least contain partially original reporting on top of existing information. Exceptions can be made for content in foreign language or any other exceptional cases. Fully paywalled articles are not allowed. Please contact the moderators through modmail if you have questions. No memes, jokes, or direct links to images Memes, direct image links, and low effort content will be removed. This includes both posts and comments. Images submitted as self posts must include an informative description for context. Top level comments must be substantive and contain more than 20 characters. No tech support or PC building questions These belong in /r/techsupport, /r/buildapc or any relevant tech subreddits. If your post asks a question about a specific product or need, it probably belongs elsewhere. This includes 'What should I buy?', 'Does a ____ do what I need?', 'How much is _____ worth?' and 'How long until (something obvious) happens?' type questions. Serious and intelligent discussion We allow discussion posts on relevant topics to hardware; there should however be at least background information and relevancy to demonstrate such a discussion is worthwhile. Low effort posts such as "What is ______ and what does it do?" is not acceptable. Rumor Policy No unsubstantiated rumors - Rumors or other claims/information not directly from official sources must have evidence to support them. Any rumor or claim that is just a statement from an unknown source containing no supporting evidence will be removed. Misc. Rules No content behind paywalls. No submissions to streaming sites outside of event megathreads. No submissions to crowdfunding sites or ongoing campaigns. No bots.
Spam and self promotion are not allowed. Users who want to post links to their website should be aware of the reddit definition of spam and read this article concerning self promotion on reddit. The 10% rule is enforced on /r/hardware; repeat offenders and/or your website or channel may be banned.
/r/hardware
I'm aware that german computerbase claim to have did the benchmark test for rtx mega geometry but I would like to see more people testing rather than from 1 source. I'm interested to know if having rtx geometry having any performance improvement to rtx 5000 series since Nvidia many times emphasize that 5000 series gpu build for neural rendering
There has been this theory going around, that because of the 5080 being 50% of the performance of a 5090, that the 5080 is actually a 5070, renamed.
But this is blatantly untrue.
Here's the truth.
The 5080 is a 378 mm² die, from the GB203-400-A1 series. It's made on the TSMC n5 process.
It has a 256 bit GDDR7 memory bus with 960GB/s bandwidth, and 45.6 billion transistors.
The 4080 is a 379 mm² die, from the AD103-300-A1 series (notice 103 vs 203, both are 03 series).
It has a 256 bit GDDR6 memory bus with 716GB/s bandwidth, and 45.9 billion transistors.
Lets do a comparison:
Process: Same on both sides
Memory capacity: Same on both sides
Memory bit width: Same on both sides
Die size: 5080 is 0.7% smaller, nearly rounding error.
Admittedly, the 4080 is a 5% cutdown, but fun fact, THE 4090 IS A 11.2% CUTDOWN. The 4090 is a 8/9ths cutdown of the rtx ada 6000. The equivelent card for the 5090 hasn't come out yet.
So where do people get the conception that the 5080 should be the 5070?
It's because they're comparing it to the 5090. But there is a problem here. The 5090 and 4090 are not the same tier.
The 4090 is 609sqmm in size, with a 384 bit memory bus.
The 5090 is 750sqmm in size, with a 512 bit memory bus.
The 5080 isn't a 5070. It's that the 5090 is something much larger than anything we have ever had before.
Edit: Windows 11 automatically enables HAGS upon install and most reviewers used MFG that requires HAGS to be turned on to work. It's safe to assume HAGS was enabled in all reviews that used Windows 11.
(Why AMP and HAGS are related): In the Blackwell Whitepaper NVIDIA describes their AI Management Processor (AMP) as dedicated GPU context scheduler tapping directly into the HAGS functionality of Windows 10 and 11. The AMP is a key architectural highlight of Blackwell but to use it properly Hardware Accelerated GPU Scheduling (HAGS) should be enabled in Windows 10. It's always enabled on Windows 11 with a fresh install. HAGS off relegates scheduling the CPU, and on 50 series that could be problematic. That's because AMP sits right in front of the Gigathread Engine and could result in overhead and other problems if it's not possible to bypass it, but remains an unanswered question.
(Unanswered questions): HAGS isn't enabled by default in either Windows 10 or 11, but is required for Framegen and MFG. It was definitely turned on for the MFG and FG tests in the RTX 5090 and 5080 launch reviews, but reviewers didn't mention if it on or off for application and game testing. Without that info it's simply impossible to know when it was on or off.
(Request for testing + official statement): Established benchmarking channels like Gamers Nexus and Hardware Unboxed should test HAGS on vs off in Windows 11 and see if it affects Blackwell's gaming and application performance. NVIDIA should issue an official statement about whether or not HAGS should be turned on at all times for RTX 5080 or 5090 even when not using MFG or FG.
CLAS - Cluster Acceleration Structure (Up to 100x more triangles): https://github.com/KhronosGroup/Vulkan-Docs/blob/main/proposals/VK_NV_cluster_acceleration_structure.adoc
PTLAS - Partitioned Top Level Acceleration Structure (BVH reuse and caching): https://github.com/KhronosGroup/Vulkan-Docs/blob/main/proposals/VK_NV_partitioned_acceleration_structure.adoc
Blackwell Whitepaper (Explains RTX Mega Geometry): https://images.nvidia.com/aem-dam/Solutions/geforce/blackwell/nvidia-rtx-blackwell-gpu-architecture.pdf
Disclaimer (Please read this before up- or downvoting): I'm reporting this early based on the limited info available to spread awareness and encourage more testing by qualified testers and analysis by qualified experts. That's why I've marked as a rumour even though it's not a rumour nor remotely close to settled facts. We're nowhere near knowing the true culprit(s) for Blackwell's performance inconsistency and this only neccessitates additional work and analysis so we can draw valid conclusions.
Not trying to push an agenda here or arm copers, just highlighting a potentially serious issue with the Blackwell's AMP logic unit that warrants investigation. If the AMP issue is real and can be fixed in software, then it'll require an army of NVIDIA software engineers to help rewrite application and game specific code and/or rework the NVIDIA driver stack.
Blackwell's overall performance consistency and application support is extremely lackluster compared RTX 40 series and resembles an Intel ARC launch more than a usual rock solid NVIDIA launch where relatively uniform uplifts are observed across the board and applications and every usually just works. Application performance sees the wildest inconsistencies but they do extend to gaming as well. The lackluster performance at 1080p and 1440p is an issue plaguing both the 5090 and 5080, that gets somewhat resolved at 4K.
With 50 series Delta Force and Counter-Strike 2 experience FPS regression from 4080S to 5080 as shown in Hardware Unboxed's 5080 review. In the same review TLOU and Spiderman manage atrocious 0% gains for 5080 vs 4080S at 1440p. The result of that is that when using upscalers the performance gains of 50 vs 40 series tank. And remember upscaling is almost mandatory for heavy RT games a key selling for NVIDIA graphics cards. But the worst example yet is probably TechPowerUp's Elden Ring ray tracing performance at 1080p where the 5080 trails even the 4070 Super, available >here<. In TechPowerUp's review Elden Ring at native 1080p both 5090 and 5080 fail with 5-8 FPS regression vs the 4090 and 4080S.
But the most odd thing so far has been that RT performance uplift was consistently worse than raster uplift in nearly every single 5080 review. This is clearly shown in the TechPowerUp review, available >here< where in the majority of games 5080 and 5090 saw larger penalties from turning on RT than equivalent 40 series cards.
Then there's the obvious lack of support for a ton of professional and AI applications where reviewers had to wait for an update adding Blackwell support, but that obviously didn't happen, not even a week later when the 5080 launched. IDK if this is just me but I don't recall this level of incompatability with professional applications for any of the previous launches (20-40 series), isn't it unprecedented for an NVIDIA generation?
And when applications work their performance is sometimes broken resulting in a 5090 loosing to even a RTX 4080. Just watch some of the professional workload centric reviews and you'll see how bad it is.
The most insane performance degradation I've seen is outside of professional workloads is the testing by Guru3D with 3D Mark ray tracing, available >here<. When testing the 5080 in 3D Mark Hybrid ray tracing benchmark they observed a 21% lead over the 4080S. But when they ran the full path tracing benchmark the 5080 was now 31% slower than the 4080S and could only match a 4070 Super. The 5090 has the same issue although to a lesser degree showing a lead vs the 4090 of 45% (Hybrid RT) vs 24% (Full PT).
Thanks to NVIDIA's Blackwell GPU Architecture Whitepaper we might have a likely culprit, although it's likely it's not the only one, although probably still the most significant. The new AI Management Processor is in fact much more than just an AI workload scheduler:
"The AI Management Processor (AMP) is a fully programmable context scheduler on the GPU designed to offload scheduling of GPU contexts from the system CPU. AMP enhances the scheduling of GPU contexts in Windows to more efficiently manage different workloads running on the GPU. A GPU context encapsulates all the state information the GPU needs to execute one or more tasks."
"The AI Management Processor is implemented using a dedicated RISC-V processor located at the front of the GPU pipeline, and it provides faster scheduling of GPU contexts with lower latency than prior CPU-driven methods. The Blackwell AMP scheduling architecture matches the Microsoft architectural model that describes a configurable scheduling core on the GPU through Windows Hardware-Accelerated GPU Scheduling (HAGS), introduced in Windows 10 (May 2020 Update)."
AMP is a direct on-die RISC-V CPU context scheduler with extremely low latency and high bandwidth access to the Gigathread Engine. It sits in front of Gigathread engine and offloads context scheduling from the CPU and taps into Hardware Accelerated GPU Scheduling (HAGS) supported in Windows 10 and 11. This tight cointegration is crucial for MFG, neural rendering and LLM integration into video games, and beneficial to multitasking, content creation, and existing gaming experiences. It doesn't just magically work as intended and requires a proper code implementation and can be a double edged sword (more on that later).
Doing the largest redesign of the GPU wide frontend (not GPC level) since Fermi introduced the Gigathread Engine in 2010 without significant game and/or driver code rewrites is asking for trouble. On 40 series and prior the CPU communicated directly with the Gigathread Engine. But on 50 series, assuming no code rewrites, the CPU has to communicate to the Gigathread Engine through the AMP which adds significant latency and scheduling overhead or either partially or completely breaks scheduling. This results in severe performance degradation without code rewrites as seen with Elden Ring RT and 3D Mark Path tracing. It's also not surprising that when implementing a change this profound some applications just straight up refuse to work.
There's a Twitter post by "Osvaldo Pinali Doederlein" on AMP, where people are discussing why AMP could be causing Blackwell's inconsistent performance and how big the performance impact of Blackwell's new CPU -> AMP -> Gigathread scheduler paradigm is without code rewrites. Psst it's likely -10 to -20%.
The 5090 also seems to have worsened memory and SRAM latencies and L1 bandwidth regression vs the 4090 as reported by "harukaze5719" on Twitter. This is unrelated but could in some instances explain some of the 5090’s performance degradation (vs mean uplift).
(Conclusion): It's too early pass definitively judgment on Blackwell's performance issues, what's causing them, how likely the issues are to be either software and/or hardware related, and if they can even be fixed. With that said there's clearly a lot wrong and the issue spans across many games and many different types of compute and AI applications.
Getting 5080 gaming uplifts anywhere from sub -20% (Elden ring RT 1080p) to +40% (CB2077 HW Canucks 5080 Review) is unprecendented for an NVIDIA launch. NVIDIA Blackwell has severe performance inconsistencies and is reminiscent of ARC Battlemage. This shouldn't be dismissed as node stagnation or something to be expected. No there's clearly something wrong a fundamental level, which could be either a hardware flaw, broken software or a combination of both.
Hopefully NVIDIA can iron out the issues and improve Blackwell's overall performance and consistency over time, and improve it enough to deter AMD from doing something really stupid with RDNA 4's pricing.