As we’ve done in the past for GTA V and Watch_Dogs 2, we’re now taking a look at Destiny 2’s texture resolution settings. Our other recent Destiny 2 content includes our GPU benchmark and CPU benchmark.
All settings other than texture resolution were loaded from the highest preset and left untouched for these screenshots. There are five degrees of quality, but only highest, medium, and lowest are shown here to make differences more obvious. The blanks between can easily be filled in.
Our Destiny 2 GPU benchmark was conducted alongside our CPU benchmark, using many of the same learnings from our research for the GPU bench. For GPU testing, we found Destiny 2 to be remarkably consistent between multiplayer and campaign performance, scaling all the way down to a 1050 Ti. This remained true across the campaign, which performed largely identically across all levels, aside from a single level with high geometric complexity and heavy combat. We’ll recap some of that below.
For CPU benchmarking, GN’s Patrick Lathan used this research (starting one hour after the GPU bench began) to begin CPU tests. We ultimately found more test variance between CPUs – particularly at the low-end – when switching between campaign and multiplayer, and so much of this content piece will be dedicated to the research portion behind our Destiny 2 CPU testing. We cannot yet publish this as a definitive “X vs. Y CPU” benchmark, as we don’t have full confidence in the comparative data given Destiny 2’s sometimes nebulous behaviors.
For one instance, Destiny 2 doesn’t utilize SMT with Ryzen, producing utilization charts like this:
The Destiny 2 beta’s arrival on PC provides a new benchmarking opportunity for GPUs and CPUs, and will allow us to plot performance uplift once the final game ships. Aside from being a popular beta, we also want to know if Bungie, AMD, and nVidia work to further improve performance in the final stretch of time prior to the official October 24 launch date. For now, we’re conducting an exploratory benchmark of multiplayer versus campaign test patterns for Destiny 2, quality settings, and multiple resolutions.
A few notes before beginning: This is beta, first off, and everything is subject to change. We’re ultimately testing this as it pertains to the beta, but using that experience to learn more about how Destiny 2 behaves so that we’re not surprised on its release. Some of this testing is to learn about settings impact to performance (including some unique behavior between “High” and “Highest”), multiplayer vs. campaign performance, and level performance. Note also that drivers will iterate and, although nVidia and AMD both recommended their respective drivers for this test (385.41, 17.8.2), likely change for final release. AMD in particular is in need of a more Destiny-specific driver, based on our testing, so keep in mind that performance metrics are in flux for the final launch.
Note also: Our Destiny 2 CPU benchmark will be up not long after this content piece. Keep an eye out for that one.
Blizzard announced in January that Overwatch had surpassed the 25 million player milestone, but despite being nearly a year old, there’s still no standardized way to benchmark the game. We’ve developed our own method instead, which we’re debuting with this GPU optimization guide.
Overwatch is an unusual title for us to benchmark. As a first person shooter, the priority for many players is on sustained high framerates rather than on overall graphical quality. Although Overwatch isn’t incredibly demanding (original recommended specs were a GTX 660 or a Radeon HD 7950), users with mid-range hardware might have a hard time staying above 60FPS at the highest presets. This Overwatch GPU optimization guide is for those users, with some graphics settings explanations straight from Blizzard to GN.
Benchmarking Mass Effect: Andromeda immediately revealed a few considerations for our finalized testing. Frametimes, for instance, were markedly lower on the first test pass. The game also prides itself in casting players into a variety of environs, including ship interiors, planet surfaces of varying geometric complexity (generally simpler), and space stations with high poly density. Given all these gameplay options, we prefaced our final benchmarking with an extensive study period to research the game’s performance in various areas, then determine which area best represented the whole experience.
Our Mass Effect: Andromeda benchmark starts with definitions of settings (like framebuffer format), then goes through research, then the final benchmarks at 4K, 1440p, and 1080p.
With Ryzen around the corner, we wanted to publish a full CPU benchmark of Watch Dogs 2 in our test course, as we’ve recently found the game to be heavily thread-intensive and responsive to CPU changes. The game even posts sizable gains for some overclocks, like on the i5-2500K, and establishes a real-world platform of when CPU choice matters. It’s easy to bottleneck GPUs with Watch Dogs 2, which is something of a unique characteristic for modern games.
Watch Dogs 2 is a familiar title by now at the GN test bench, and while we’ve published a GPU benchmark and a more recent CPU optimization guide, we never published a comprehensive CPU benchmark. We’ve gathered together all our results here, from the 2500K revisit all the way to Kaby Lake reviews (see: 7600K review & 7350K review), and analyzed what exactly makes a CPU work well with Watch Dogs 2 and why.
In this Watch Dogs 2 CPU benchmark, we’ll recap some graphics optimization tips for CPUs and test whether an i7 is worth it, alongside tests of the 7600K, 7700K, 6600K, 7350K, FX-8370, and more.
One interesting aspect of the Watch_Dogs 2 benchmarking we did for our 2500K revisit was the difference in performance between i5s and i7s. At stock speeds, the i7-2600K was easily outpacing the i5-2500K by roughly 15 FPS—and even more interestingly, the i7-6700K managed to hit our GTX 1080’s ceiling of 110-115 FPS, while the i5-6600K only managed 78.7 with the same settings. Watch_Dogs 2 is clearly a game where the additional threads are beneficial, making it an exciting test opportunity as that’s not a common occurrence. We decided to look into settings optimization for CPUs with Watch Dogs 2, and have tested a few of the most obvious graphics settings to see which ones can really help.
This Watch Dogs 2 graphics optimization guide focuses on CPU performance to try and figure out which settings can be increased (with GPU overhead) and decreased (with CPU limits).
Before even getting started here, let’s put out the obvious disclaimer. This GPU benchmark is for the beta version of For Honor, which means a few things: (1) the game’s not final yet and, despite being just two weeks away, there are still some graphics settings missing from the menu; (2) nVidia’s current drivers are optimized for the beta, but the company plans another update some point soon for further optimizations; (3) AMD has not yet released drivers for the game, though we did ask for early access and were told that the company won’t be ready until launch day. There are day-0 drivers planned from AMD.
Regardless, we tested anyway to see how the beta performs and get a baseline understanding of what we should expect overall from the new multiplayer brawler title. For Honor thus far has proven impressively detailed in geometry and texturing (especially texturing), and deserves high marks for the art department. Granted, that generally means more abuse on the video card or CPU (for the complex geometric draw calls), so we’ve got some For Honor graphics settings scaling tests as well.
This graphics card benchmark tests For Honor’s performance at 4K, 1440p, and 1080p with Extreme settings. We tested using a real, in-game benchmark rather than the built-in benchmark, which generally makes performance look a lot worse than it is in reality (we have a chart demonstrating this). Settings scaling was tested from low to extreme, as was multiplayer and ‘singleplayer’ (bot match). We primarily ran For Honor benchmarks with the AMD RX 480 8GB & 4GB, RX 470 4GB, RX 460 2GB, & 390X cards vs. the GTX 1080, 1070, 1060 6GB & 3GB, 1050 & Ti, and 970 AIB partner cards.
Ubisoft's newest dystopian efforts start strong with allusions to modern-day challenges pertaining to privacy and "cyber warfare," working to build-up our character as a counter-culture hacker. And, as with Ubisoft's other AAA titles, the game builds this world with high-resolution textures, geometrically complex and dense objects, taxing shadow/lighting systems, and an emphasis on graphics quality.
Watch Dogs 2 is a demanding title to run on modern hardware. We spent the first 1-2 hours of our time in Watch Dogs 2 simply studying the impact of various settings on performance, further studying locales and their performance hits. Areas with grass and foliage, we found, most heavily hit framerate. Nightfall or dark rain play a role in FPS hits, too, particularly when running high reflection qualities and headlight shadows.
We look at performance of 11 GPUs in this Watch Dogs 2 video card benchmark, including the RX 480 vs. GTX 1060, GTX 1070, GTX 1080, RX 470, R9 Fury X, and more.
We've been through Battlefield 1 a few times now. First were the GPU benchmarks, then the HBAO vs. SSAO benchmark, then the CPU benchmark. This time it's RAM, and the methodology remains mostly the same. Note that these results are not comparable to previous results because (1) the game has received updates, (2) memory spec has changed for this test, and (3) we have updated our graphics drivers. The test platforms and memory used are dynamic for this test, the rest remaining similar to what we've done in the past. That'll be defined in the methodology below.
Our CPU benchmark had us changing frequencies between test platforms as we tried to determine our test patterns and methodology / bench specs for the endeavor. During that exploratory process, we noticed that memory speeds of 3200MHz were measurably faster in heuristic testing than speeds of, say, 2400MHz. That was just done by eye, though; it wasn't an official benchmark, and we wanted to dedicate a separate piece to that.
This content benchmarks memory performance in Battlefield 1, focusing on RAM speed (e.g. 1600MHz, 1866, 2133, 2400, so forth) and capacity. We hope to answer whether 8GB is "enough" and find a sweet spot for price-performance in memory selection.
We moderate comments on a ~24~48 hour cycle. There will be some delay after submitting a comment.