cross-posted from: https://lemmy.world/post/2149490
In January I replaced my 5 year old GTX 1080 Ti with an Asus ROG TUF OC 4090. My old 1080 Ti never had an official waterblock made for it by anyone, so I was never able to incorporate it into my loop. I made sure that whatever model of 4090 I got, it had to have waterblock support from more than just one vendor. I’m finally done with my system. For now anyways. After years of tweaking, upgrades, and loop rebuilds, I’m happy with how it looks, and how it performs.
Specs
- Case: Lian Li O11 Dynamic XL
- CPU: AMD Ryzen 7 5800X3D
- Motherboard: Gigabyte Aorus X570 Master
- RAM: Corsair Vengeance Pro RGB (4 x 8GB)
- GPU: Asus ROG TUF OC 4090
- NVMe #1 - 1TB Western Digital Black SN850 (OS & Applications)
- NVMe #2 - 2TB Western Digital Black SN850X (Steam Library)
- Corsair Hydro X D5 Pump
- Alphacool Strix/TUF 4090 Block
- Optimus PC Foundation AM4 CPU Block
- 2 x Alphacool NexXxoS XT45 Full Copper 360mm Radiator
- Bitspower fittings
There’s a good chance however that next year when the Ryzen 8000 chips come out I’ll upgrade my platform to that. But for now I can finally enjoy the months of hard work and waiting for parts.
It’s always odd to see a custom loop setup with AMD parts. Like driving a Ferrari with a Honda engine. Nothing against AMD (or Honda for that matter) but it has been pretty consistently the budget-friendly option. Juxtaposed against the cost of all that a custom loop requires, it just seems odd. Was there a particular rationale for going with the 5800X3D? No offense intended; you’ve got a really great build!
If you’re shopping in the RTX4080 price range, the AMD equivalent has more performance in gaming so long as you don’t use ray tracing. They just don’t have an RTX4090 equivalent, so a more accurate analogy would be putting a corvette engine in a ferrari
Interesting! Thanks for sharing your experience.
I kind of hear what you’re saying, but other than the GPU, and it’s waterblock, the rest of the system is not new. I originally bought the motherboard in 2019 with a 3900X when the X570 AM4 platform and the Ryzen 3000 series CPU’s came out. Then upgraded the CPU to a 5800X3D just over a year ago.
Why the 5800X3D? 1) Because at the time I bought it, it was THE fastest gaming CPU. It’s still an incredible CPU that has a ton of performance. 2) I got the 5800X3D because I primarily play games on my PC. I don’t do any video editing, streaming, or productivity work. My rig is custom built for gaming, and only gaming. Even going as far as running a customized Windows 10 install with a bunch of background processes and analytics bloat stripped out. This is what the 5800X3D was made for. You look at benchmarks and reviews of the 5800X3D and it was beating or matching the 12900K, for less money, and most importantly, for less power draw. Performs better, runs cooler, costs less. I see that as a no brainer, not a compromise. There was nothing budget friendly about any of this platform when the parts were bought.
If I was buying the platform now, new, yes… then I could see why you’d think that. Honestly at this point, I’d have to go to the AM5 platform, and the 7800X3D to get any better performance than I have already. Since the AM5 platform is brand new, I’ll give it a year and wait till the Ryzen 8000 series chips launch and look at the X3D variants then. I went bleeding edge with my X570 and the Ryzen 3900X and there was definitely early adopter pains. I get more than enough performance from my system, in the games I play. I judge my happiness based on how the games feel to me. If the games load fast, feel smooth, and perform well then I’m happy. I don’t look too closely at framerates, it’s just about the experience for me.
Honestly, if Intel launched a platform and a CPU that had the better performance per $ with less power draw, and better thermals than AMD, then I’d probably buy that platform, but as it stands that hasn’t been the case, for a while now. If the 13900K beats my 5800X3D by 10% in a game, but the difference is 270FPS vs 300FPS, I honestly see that as a null win for Intel. Because those extra 30FPS will never ever be noticed or seen. Especially when I cap my FPS to my monitor’s refresh rate. Anything above that is just wasted electricity.