Today we’re reviewing the RTX 2060, with additional tests on if an RTX 2060 has enough performance to really run games with ray-tracing – basically Battlefield, at this point – on the TU106 GPU. We have a separate tear-down going live showing the even more insane cooler assembly of the RTX 2060, besting the previous complexity of the RTX 2080 Ti, but today’s focus will be on performance in gaming, thermals, RTX performance, power consumption, and acoustics of the Founders Edition cooler.

The RTX 2060 Founders Edition card is priced at $350 and, unlike previous FE launches in this generation, it is also the price floor. Cards will start at $350 – no more special FE pricing – and scale based upon partner cost. We will primarily be judging price-to-performance based upon the $350 point, so more expensive cards would need to be judged independently.

Our content outline for this RTX 2060 review looks like this:

  • Games: DX12, DX11
  • RTX in BF V
  • Thermals
  • Noise
  • Power

We’re putting more effort into the written conclusion for this one than typically, so be sure to check that as well. Note that we have a separate video upload on the YouTube channel for a tear-down of the card. The PCB, for the record, is an RTX 2070 FE PCB. Same thing.

The XFX RX 590 Fatboy is a card we tore-down a few months ago, whereupon we complained about its thermal solution and noted inefficiencies in the design. These proved deficient in today’s testing, as expected, but the silicon itself – AMD’s GPU – remained a bit of a variable for us. The RX 590 GPU, ignoring XFX and its component of the review (momentarily), is potentially a stronger argument between the GTX 1060 and GTX 1070. It’s a pre-pre-overclocked RX 480 – or a pre-overclocked RX 580 – and, to AMD’s credit, it has pushed this silicon about as far is it can go.

Today, we’re benchmarking the RX 590 (the “Fatboy” model, specifically) against the GTX 1060, RX 580 overclocked, GTX 1070, and more.

Today’s benchmark is a case study by the truest definition of the phrase: We are benchmarking a single sample, overweight video card to test the performance impact of its severe sag. The Gigabyte GTX 1080 Ti Xtreme was poorly received by our outlet when we reviewed it in 2017, primarily for its needlessly large size that amounted to worse thermal and acoustic performance than smaller, cheaper competitors. The card is heavy and constructed using through-bolts and complicated assortments of hardware, whereas competition achieved smaller, more effective designs that didn’t sag.

As is tradition, we put the GTX 1080 Ti Xtreme in one of our production machines alongside all of the other worst hardware we worked with, and so the 1080 Ti Xtreme was in use in a “real” system for about a year. That amount of time has allowed nature – mostly gravity – to take its course, and so the passage of time has slowly pulled the 1080 Ti Xtreme apart. Now, after a year of forced labor in our oldest rendering rig, we get to see the real side-effects of a needlessly heavy card that’s poorly reinforced internally. We’ll be testing the impact of GPU sag in today’s content.

We already reviewed an individual NVIDIA Titan RTX over here, used first for gaming, overclocking, thermal, power, and acoustic testing. We may look at production workloads later, but that’ll wait. We’re primarily waiting for our go-to applications to add RT and Tensor Core support for 3D art. After replacing our bugged Titan RTX (the one that was clock-locked), we were able to proceed with SLI (NVLink) testing for the dual Titan RTX cards. Keep in mind that NVLink is no different from SLI when using these gaming bridges, aside from increased bandwidth, and so we still rely upon AFR and independent resources.

As a reminder, these cards really aren’t built for the way we’re testing them. You’d want a Titan RTX card as a cheaper alternative to Quadros, but with the memory capacity to handle heavy ML/DL or rendering workloads. For games, that extra (expensive) memory goes unused, thus demeaning the value of the Titan RTX cards in the face of a single 2080 Ti.

This is really just for fun, in all honesty. We’ll look at a theoretical “best” gaming GPU setup today, then talk about what you should buy instead.

Today, we’re reviewing the NVIDIA Titan RTX for overclocking, gaming, thermal, and acoustic performance, looking at the first of two cards in the lab. We have a third card arriving to trade for one defective unit, working around the 1350MHz clock lock we discovered, but that won’t be until after this review goes live. The Titan RTX costs $2500, outbidding the RTX 2080 Ti by about 2x, but only enables an additional 4 streaming multiprocessors. With 4 more SMs and 256 more lanes, there’s not much performance to be gained in gaming scenarios. The big gains are in memory-bound applications, as the Titan RTX has 24GB of GDDR6, a marked climb from the 11GB on an RTX 2080 Ti.

An example of a use case could be machine learning or deep learning, or more traditionally, 3D graphics rendering. Some of our in-house Blender project files use so much VRAM that we have to render instead with the slower CPU (rather than CUDA acceleration), as we’ll run out of the 11GB framebuffer too quickly. The same is true for some of our Adobe Premiere video editing projects, where our graph overlays become so complex and high-resolution that they exceed the memory allowance of a 1080 Ti. We are not testing either of these use cases today, though, and are instead focusing our efforts on the gaming and enthusiast market. We know that this is also a big market, and plenty of people want to buy these cards simply because “it’s the best,” or because “most expensive = most best.” We’ll be looking at how much the difference really gets you, with particular interest in thermal performance pursuant to the removal of the blower cooler.

Finally, note that we were stuck at 1350MHz with one of our two samples, something that we’ve worked with NVIDIA to research. The company now has our defective card and has traded us with a working one. We bought the defective Titan RTX, so it was a “real” retail sample. We just wanted to help NVIDIA troubleshoot the issue, and so the company is now working with it.

AMD launched its RX 580 2048 silently in China a few months ago, and in doing so damaged its brand credibility by rebranding the RX 570 as an RX 580. The point of having those two, distinct names is that they represent different products. The RX 580 2048 has 2048 FPUs (or streaming processors), which happens to be exactly what the RX 570 has. The RX 580 is also a few MHz higher in clock, which is fully attainable with an overclocked RX 570. Working with GamersNexus contacts in Taiwan, who then worked with contacts in China, we managed to obtain this China-only product launch so we could take a closer look at why, exactly, AMD thinks an RX 570 Ti deserves the name “RX 580.”

Taking an existing product with a relatively good reputation and rebuilding it as a worse product isn’t new. Don’t get us wrong: The RX 570, which is what the RX 580 2048 is, is a reasonably good card, especially with its new prices of roughly $150 (Newegg) to $180 elsewhere. That said, an RX 580 2048 is, by definition, not an RX 580. That’s lying. It is an RX 570, or maybe an RX 575, if AMD thinks that a 40MHz clock difference deserves a new SKU. AMD is pulling the same deceitful trick that NVIDIA pulled with its GT 1030 DDR4 card. It’s disgraceful, misleading, and predatory of consumers who may otherwise not understand the significance of the suffix “2048.” If they’re looking for an RX 580, they’re still finding one – except it isn’t one, and to brand the RX 580 2048 as an RX 580 is disgraceful.

We have a separate video scheduled to hit our channel with a tear-down of the card, in case you’re curious about build quality. Today, we’re using the DATALAND RX 580 2048 as our vessel for testing AMD’s new GPU. Keep in mind that, for all our scorn toward the GPU, DATALAND is somewhat unfortunately the host. DATALAND didn’t make the GPU – they just put it on the PCB and under the cooler (which is actually not bad). It also appears that DATALAND (迪兰) works alongside TUL, the parent company to PowerColor.

We paid about $180 USD for this card, which puts it around where some RX 570s sell for (though others are available for ~$150). Keep in mind that pricing in China will be a bit higher than the US, on average.

Finding the “best" workstation GPU isn't as straight-forward as finding the best case, best gaming CPU, or best gaming GPU. While games typically scale reliably from one to the next, applications can deliver wildly varying performance. Those gains and losses could be chalked up to architecture, drivers, and also whether or not we're dealing with a true workstation GPU versus a gaming GPU trying to fill-in for workstation purposes.

In this content, we're going to be taking a look at current workstation GPU performance across a range of tests to figure out if there is such thing as a champion among them all. Or, in the very least, we'll figure out how AMD differs from NVIDIA, and how the gaming cards differ from the workstation counterparts. Part of this will look at Quadro vs. RTX or GTX cards, for instance, and WX vs. RX cards for workstation applications. We have GPU benchmarks for video editing (Adobe Premiere), 3D modeling and rendering (Blender, V-Ray, 3ds Max, Maya), AutoCAD, SolidWorks, Redshift, Octane Bench, and more.

Though NVIDIA's Quadro RTX lineup has been available for a few months, review samples have been slow to escape the grasp of NVIDIA, and if we had to guess why, it's likely due to the fact that few software solutions are available that can take advantage of the features right now. That excludes deep-learning tests which can benefit from the Tensor cores, but for optimizations derived from the RT core, we're still waiting. It seems likely that Chaos Group's V-Ray is going to be one of the first plugins to hit the market that will support NVIDIA's RTX, though Redshift, Octane, Arnold, Renderman, and many others have planned support.

The great thing for those planning to go with a gaming GPU for workstation use is that where rendering is concerned, the performance between gaming and workstation cards is going to be largely equivalent. Where performance can improve on workstation cards is with viewport performance optimizations; ultimately, the smoother the viewport, the less tedious it is to manipulate a scene.

Across all of the results ahead, you'll see that there are many angles to view workstation GPUs from, and that there isn't really such thing as a one-size-fits all - not like there is on the gaming side. There is such thing as an ultimate choice though, so if you're not afraid of spending substantially above the gaming equivalents for the best performance, there are models vying for your attention.

The RTX 2080 Ti failures aren’t as widespread as they might have seemed from initial reddit threads, but they are absolutely real. When discussing internally whether we thought the issue of artifacting and dying RTX cards had been blown out of proportion by the internet, we had two frames of mind: On one side, the level of attention did seem disproportionate to the size of the issue, particularly as RMA rates are within the norm. Partners are still often under 1% and retailers are under 3.5%, which is standard. The other frame of mind is that, actually, nothing was blown out of proportion for people who spent $1250 and received a brick in return. For those affected buyers, the artifacting is absolutely a real issue, and it deserves real attention.

This content marks the closing of a storyline for us. We published previous videos detailing a few of the failures on our viewers’ cards (borrowed by GN on loan), including an unrelated issue of a 1350MHz lock and BSOD issue. We also tested cards in our livestream to show what the artifacting looks like, seen here. Today, we’re mostly looking at thermals, firmware, the OS, downclocking impact, and finding a conclusion of what the problem isn’t (rather than what it 100% is).

With over a dozen cards mailed in to us, we had a lot to sort through over the past week. This issue certainly exists in a very real way for those who spent $1200+ on an unusable video card, but it isn’t affecting everyone. It’s far from “widespread,” fortunately, and our present understanding is that RMA rates remain within reason for most of the industry. That said, NVIDIA’s response times to some RMA requests have been slow, from what our viewers have expressed, and replacements can take upwards of a month given supply constraints in some regions. That’s a problem.

This content stars our viewers and readers. We charted the most popular video cards over the launch period for NVIDIA’s RTX devices, as we were curious if GTX or RTX gained the most sales in this time. We’ve also got some AMD data toward the end, but the focus here is on a shifting momentum between Pascal and Turing architectures and what the consumers want.

We’re looking exclusively at what our viewers and readers have purchased over the two-month launch window since RTX was announced. This samples several hundred purchases, but is in no way at all a representative sample of the whole. Keep in mind that we have a lot of sampling biases here, the primary of which is that it’s our audience – that means these are people who are more enthusiast-leaning, likely buy higher end, and probably follow at least some of our suggestions. You can’t extrapolate this data market-wide, but it is an interesting cross-section for our audience.

Hardware news coverage has largely followed the RTX 2080 Ti story over the past week, and it's one of dying cards of unknown quantities online. We have been investigating the issue further and have a few leads on what's going on, but are awaiting for some of the dead cards to arrive at our office before proceeding further. We also learned about Pascal stock depletion, something that's been a curious topic when given the slow uptake on RTX.

Further news items include industry discussion on Intel's outsourcing to TSMC, its hiring of former AMD graphics staff, and dealings with 14nm shortages. Only one rumor is present this week, and that's of the nearly confirmed RX 590.

We moderate comments on a ~24~48 hour cycle. There will be some delay after submitting a comment.

Advertisement:

  VigLink badge