But the GTX 1080 is easily, hands-down, no-comparison the most powerful single-GPU graphics card ever released—especially when you overclock it.
Ignoring the auto-overclocking tools in the Precision X beta, I was able to manually boost the core clock speed by 250MHz and the memory clock speed by an additional 100MHz. Depending on what was going on in a given game’s given scene, and how Nvidia’s GPU Boost technology reacted to it, doing so resulted in clock speeds ranging from 1,873MHz to 2,088MHz. Yes, that’s clock speeds in excess of 2GHz on air, with no voltage tweaks.
In other words: Buckle up. This is going to be a wild ride.
First up: Ubisoft’s The Division, a third-person shooter/RPG set that mixes elements of Destiny and Gears of War. The game’s set in a gorgeous and gritty recreation of post-apocalyptic New York, running on Ubisoft’s new Snowdrop engine. Despite incorporating Nvidia Gameworks features—which we disabled during benchmarking to even the playing field—the game scales well across all hardware and isn’t part of Nvidia’s “The Way It’s Meant to be Played” lineup. In fact, it tends to perform better on Radeon cards.
Until the GTX 1080 enters the fray.
As you can see, the reference GTX 1080 offers a whopping 71-percent performance increase over the GTX 980 at 4K resolution and Ultra graphics settings. The GTX 1080 is designed as a generational replacement for the GTX 980, remember—not the Titan X. That said, the GTX 1080 outpunches the Titan X by 34.7 percent at 4K, and 24 percent at 1440p resolution, despite costing $400 less than Nvidia’s flagship.
This glorious murder-simulating sandbox’s Glacier engine is heavily optimized for AMD titles, with Radeon cards significantly outpunching their GeForce GTX 900-series counterparts, especially at higher resolutions. Because of that, while the GTX 1080 offers a significant performance leap over the GTX 980 (72.7 percent at 4K) and Titan X (33.8 percent), the performance gain over the Fury X is much more modest (8.8 percent) with all settings cranked to Ultra and FXAA enabled. That drives home how important in-engine support for a particular graphics architecture can be.
Note that these results are using DirectX 11. Hitman theoretically supports DirectX 12, but a recent update broke it, and the game refused to launch in DX12 on both PCWorld’s GPU testbed as well as my personal gaming rig despite ostensibly being fixed. Alas.
Rise of the Tomb Raider
Sign up for CIO Asia eNewsletters.