EA's Origin launcher has recently gained attention for hosting Apex Legends, one of the present top Battle Royale shooters, but is getting renewed focus as being an easy attack vector for malware. Fortunately, an update has already resolved this issue, and so the pertinent action would be to update Origin (especially if you haven't opened it in a while). Further news this week features the GTX 1650's rumored specs and price, due out allegedly on April 23. We also follow-up on Sony PlayStation 5 news, now officially confirmed to be working with a new AMD Ryzen APU and customized Navi GPU solution.
Show notes below the embedded video, for those preferring reading.
We’re still in China for our factory and lab tours, but we managed to coordinate with home base to get enough testing on the GTX 1660 done that a review became possible. Patrick ran the tests this time, then we just put the charts and script together from Dongguan, China.
This is a partner launch, so no NVIDIA direct sampling was done and, to our knowledge, no Founders Edition board will exist. Reference PCBs will exist, as always, but partners have control over most of the cooler design for this launch.
Our review will look at the EVGA GTX 1660 dual-fan model, which has an MSRP of $250 and lands $30 cheaper than the baseline GTX 1660 Ti pricing. The cheapest GTX 1660s will sell for about $220, but our $250 unit today has a higher power target allowance for overclocking and a better cooler. The higher power target is the most interesting, as overclocking performance can stretch upwards toward a GTX 1660 Ti at the $280 price-point.
We’ll get straight to the review today. Our focus will be on games, with some additional thermal and power tests toward the end. Again, as a reminder, we’re doing this remotely, so we don’t have as many non-gaming charts as normally, but we still have a complete review.
Our initial AMD Radeon VII liquid cooling mod was modified after the coverage went live. We ended up switching to a Thermaltake Floe 360 radiator (with different fans) due to uneven contact and manufacturing defects in the Alphacool GPX coldplate. Going with the Asetek cooler worked much better, dropping our thermals significantly and allowing increased overclocking and stock boosting headroom. The new drivers (19.2.3) also fixed most of the overclocking defects we originally found, making it possible to actually progress with this mod.
As an important foreword, note that overclocking with AMD’s drivers must be validated with performance at every step of the way. Configured frequencies are not the same as actual frequencies, so you might type “2030MHz” for core and get, for instance, 1950-2000MHz out. For this reason, and because frequency regularly misreports (e.g. “16000MHz”), it is critical that any overclock be validated with performance. Without validation, some “overclocks” can actually be bringing performance below stock while appearing to be boosted in frequency. This is very important for overclocking Radeon VII properly.
We recently revisited the AMD R9 290X from October of 2013, and now it’s time to look back at the GTX 780 Ti from November of 2013. The 780 Ti shipped for $700 MSRP and landed as NVIDIA’s flagship against AMD’s freshly-launched flagship. It was a different era: Memory capacity was limited to 3GB on the 780 Ti, memory frequency was a blazing 7Gbps, and core clock was 875MHz stock or 928MHz boost, using the old Boost 2.0 algorithm that kept a fixed clock in gaming. Overclocking was also more extensible, giving us a bigger upward punch than modern NVIDIA overclocking might permit. Our overclocks on the 780 Ti reference (with fan set to 93%) allowed it to exceed expected performance of the average partner model board, so we have a fairly full range of performance on the 780 Ti.
NVIDIA’s architecture has undergone significant changes since Kepler and the 780 Ti, one of which has been a change in CUDA core efficiency. When NVIDIA moved from Kepler to Maxwell, there was nearly a 40% efficiency gain when CUDA cores are processing input. A 1:1 Maxwell versus Kepler comparison, were such a thing possible, would position Maxwell as superior in efficiency and performance-per-watt, if not just outright performance. It is no surprise then that the 780 Ti’s 2880 CUDA cores, although high even by today’s standards (an RTX 2060 has 1920, but outperforms the 780 Ti), will underperform when compared to modern architectures. This is amplified by significant memory changes, capacity being the most notable, where the GTX 780 Ti’s standard configuration was limited to 3GB and ~7Gbps GDDR5.
Today, we’re reviewing the GTX 1660 Ti, whose name is going to trip us up for the entirety of its existence. The GTX 1660 Ti is NVIDIA’s mid-step between Pascal and Turing, keeping most of the Turing architectural changes to the SMs and memory subsystem, but dropping the official RTX support and RT cores in favor of a lower price. The EVGA GTX 1660 Ti XC that we’re reviewing today should have a list price of $280, sticking it between the $350 baseline of the RTX 2060 and the rough $200 price-point of modern 1060s, although sometimes that’s higher. For further reference, Vega 56 should now sell closer to $280, with the RX 590 still around the $260 range.
Apex Legends is one of the most-watched games right now and is among the top Battle Royale genre of games. Running on the Titanfall engine and with some revamped Titanfall assets, the game is a fast-paced FPS with relatively high poly count models and long view distances. For this reason, we’re benchmarking a series of GPUs to find the “best” video card for Apex Legends at each price category.
Our testing first included some discovery and research benchmarks, where we dug into various multiplayer zones and practice mode to try and find the most heavily loaded areas of the benchmark. We also unlocked FPS for this, so we aren’t going to bump against any 144FPS cap or limitation. This will help find which cards can play the game at max settings – or near-max, anyway.
Metro: Exodus is the next title to include NVIDIA RTX technology, leveraging Microsoft’s DXR. We already looked at the RTX implementation from a qualitative standpoint (in video), talking about the pros and cons of global illumination via RTX, and now we’re back to benchmark the performance from a quantitative standpoint.
The Metro series has long been used as a benchmarking standard. As always, with a built-in benchmark, one of the most important things to look at is the accuracy of that benchmark as it pertains to the “real” game. Being inconsistent with in-game performance doesn’t necessarily invalidate a benchmark’s usefulness, though, it’s just that the light in which that benchmark is viewed must be kept in mind. Without accuracy to in-game performance, the benchmark tools mostly become synthetic benchmarks: They’re good for relative performance measurements between cards, but not necessarily absolute performance. That’s completely fine, too, as that’s mostly what we look for in reviews. The only (really) important thing is that performance scaling is consistent between cards in both pre-built benchmarks and in-game benchmarks.
Finding something to actually leverage the increased memory bandwidth of Radeon VII is a challenge. Few games will genuinely use more memory than what’s found on an RTX 2080, let alone 16GB on the Radeon VII, and most VRAM capacity utilization reporting is wildly inaccurate as it only reports allocated memory and not necessarily used memory. To best benchmark the potential advantages of Radeon VII, which would primarily be relegated to memory bandwidth, we set up a targeted feature test to look at anti-aliasing and high-resolution benchmarks. Consider this an academic exercise on Radeon VII’s capabilities.
Our AMD Radeon VII review is one of our most in-depth in a while. The new $700 AMD flagship is a repurposed Instinct card, down-costed for gaming and some productivity tasks and positioned to battle the RTX 2080 head-to-head. In today’s benchmarks, we’ll look uniquely at Radeon VII cooler mounting pressure, graphite thermal pad versus paste performance, gaming benchmarks, overclocking, noise, power consumption, Luxmark OpenCL performance, and more.
We already took apart AMD’s Radeon VII card, remarking on its interesting Hitachi HM03 graphite thermal pad and vapor chamber. We also analyzed its VRM and PCB, showing impressive build quality from AMD. These are only part of the story, though – the more important aspect is the silicon, which we’re looking at today. At $700, Radeon VII is positioned against the RTX 2080 and now-discontinued GTX 1080 Ti (the two tested identically). Radeon VII has some interesting use cases in “content creation” (or Adobe Premiere, mostly) where GPU memory becomes a limiting factor. Due to time constraints following significant driver-related setbacks in testing, we will be revisiting the card with a heavier focus on these “content creator” tests. For now, we are focusing primarily on the following:
The AMD Radeon VII embargo for “unboxings” has lifted and, although we don’t participate in the marketing that is a content-filtered “unboxing,” a regular part of our box-opening process involves taking the product apart. For today, restrictions are placed on performance discussion and product review, but we are free to show the product and handle it physically. You’ll have to check back for the review, which should likely coincide with the release date of February 7.
This content is primarily video, as our tear-downs show the experience of taking the product apart (and discoveries as we go), but we’ll recap the main point of interest here. Text continues after the embedded video:
We moderate comments on a ~24~48 hour cycle. There will be some delay after submitting a comment.