NVidia Titan Xp Review vs. 1080 Ti: $200 Per Percentage Point Gained

By Published April 27, 2017 at 3:27 pm

Additional Info

  • Component: Video Card
  • Original MSRP: 1200
  • Manufacturer: NVIDIA

NVidia’s Titan Xp 2017 model video card was announced without any pre-briefing for us, marking it the second recent Titan X model card that took us by surprise on launch day. The Titan Xp, as it turns out, isn’t necessarily targeted at gaming – though it does still bear the GeForce GTX mark. NVidia’s Titan Xp followed the previous Titan X (that we called “Titan XP” to reduce confusion from the Titan X – Maxwell before that), and knocks the Titan X 2016 out of its $1200 price bracket.

The Titan Xp 2017 now firmly socketed into the $1200 category, we’ve got a gap between the GTX 1080 Ti at $700 MSRP ($750 common price) of $450-$500 to the TiXp. Even with that big of a gap, though, diminishing returns in gaming or consumer workloads are to be expected. Today, we’re benchmarking and reviewing the nVidia Titan Xp for gaming specifically, with additional thermal, power, and noise tests included. This card may be better deployed for neural net and deep learning applications, but that won’t stop enthusiasts from buying it simply to have “the best.” For them, we’d like to have some benchmarks online.

NVidia Titan Xp 2017 Specs vs. GTX 1080 Ti

  Tesla P100 Titan Xp 2017 GTX 1080 Ti GTX 1080 GTX 1070
GPU GP100 Cut-Down Pascal GP102-450 GP102-350-K1 GP104-400 Pascal GP104-200 Pascal
Transistor Count 15.3B 12B 12B 7.2B 7.2B
Fab Process 16nm FinFET 16nm FinFET 16nm FinFET 16nm FinFET 16nm FinFET
CUDA Cores 3584 3840 3584 2560 1920
GPCs 6 6 6 4 3
SMs 56 30 28 20 15
TPCs 28 TPCs 30 28 20 TPCs 15
TMUs 224 240 224 160 120
ROPs 96 (?) 96 88 64 64
Core Clock 1328MHz 1481MHz 1481MHz 1607MHz 1506MHz
Boost Clock 1480MHz 1582MHz 1582MHz 1733MHz 1683MHz
(1/32 FP64)
Memory Capacity 16GB 12GB 11GB 8GB 8GB
Memory Clock ? 11Gbps 11Gbps 10Gbps GDDR5X 4006MHz
Memory Interface 4096-bit 384-bit 352-bit 256-bit 256-bit
Memory Bandwidth ? ~547.2GB/s ~484GB/s 320.32GB/s 256GB/s
Total Power Budget ("TDP") 300W 250W 250W 180W 150W
Power Connectors ? 1x 8-pin
1x 6-pin
1x 8-pin
1x 6-pin
1x 8-pin 1x 8-pin
Release Date 4Q16-1Q17 04/06/17 3/10/17 5/27/2016 6/10/2016
Release Price - $1,200 $700 Reference: $700
MSRP: $600
Now: $500
Reference: $450
MSRP: $380

Above is the specs table for the Titan Xp and the GTX 1080 Ti, helping compare the differences between nVidia’s two FP32-focused flagships.

Clarifying Branding: GeForce GTX on Titan Xp Card

The initial renders of nVidia’s Titan Xp led us to believe that the iconic “GeForce GTX” green text wouldn’t be present on the card, a belief further reinforced by the lack of “GeForce GTX” in the actual name of the product. Turns out, it’s still marked with the LED-backlit green text. If you’re curious about whether a card is actually a Titan Xp card, the easiest way to tell would be to look at the outputs: TiXp (2017) does not have DVI, while Titan X (Pascal, 2016) does have DVI out. Further, the Titan Xp 2017 model uses a GP102-450 GPU, whereas Titan X (2016) uses a GP102-400 GPU.

A reader of ours, Grant, was kind enough to loan us his Titan Xp for review and inevitable conversion into a Hybrid mod (Part 1: Tear-Down is already live). Grant will be using the Titan Xp for neural net and machine learning work, two areas where we have admittedly near-0 experience; we’re focused on gaming, clearly. We took the opportunity to ask Grant why someone in his field might prefer the TiXp to a cheaper 1080 Ti, or perhaps to SLI 1080 Ti cards. Grant said:

“Data sizes vary and the GPU limits are based on data size and the applied algorithm. A simple linear regression can be done easily on most GPUs, but when it comes to convolutional neural networks, the amount of math is huge. I have a 4 gig data set that cannot run its CNN on the 1080 Ti, but it can do it on the Titan X.

Also, for us, CUDA cores matter a lot. And the good machine learning algorithms, even Google's tensorflow, need CUDAs and nVidia-specific drivers to use GPU.

SLI does not do anything for us. Multiple GPUs can be used to split-up data sets and run them in parallel, but it's tricky as hell with neural networks. With multiple cards, it's better for us to run one algorithm on one card and another algo on the next card.

Even Google created their own version of a GPU for deep learning that can be farmed much better than any nVidia option.”

According to this user, at least, the extra 1GB of VRAM on the TiXp is beneficial to the workload at hand, to the point that a 1080 Ti just wouldn’t even execute the task. This may explain part of why the 1080 Ti seemingly had such an odd memory pool: Yes, of course the GPU has a more limited bus, but there’s a reason for that. NVidia might have wanted to keep machine learning-class users in Titan-level hardware, rather than splitting supply of a 1080 Ti between audiences.

Whatever the reason for the Titan Xp, we’re testing it for gaming, because people will still buy it for gaming. The extra 1GB VRAM is irrelevant in our use case, but we’ve still got potential performance gains from other differences.

NVidia Titan Xp Tear-Down

Our tear-down for the Titan Xp is already live, seen here:

The card is basically a GTX 1080 Ti FE card. We’d recommend checking out our PCB & VRM component-level analysis of the GTX 1080 Ti FE card for more information on the Titan Xp, seeing as they’re mostly the same board.




The cooler leverages the same nVidia blower-style layout found on FE cards. There’s an aluminum heatsink with nickel-plated copper coldplate and vapor chamber for direct GPU contact, with the aluminum baseplate making up the bulk of the heat transfer needs for the power components. The blower fan takes care of both the VRM and GPU (and VRAM), and we measured several power+VRAM components during test to better understand thermal behavior of the product.

Overclock Stepping Table

Here’s our overclock progression table, including pass/fail for each step. The last numbers marked as “pass” will be used during benchmarking:

Peak Clock (MHz) AVG Clock (MHz) Core Offset (MHz) MEM CLK (MHz) MEM Offset (MHz) Power Target Voltage Fan TMP Pass/Fail
1810 1759 1425.6 100 0.93 2950 74 P
1835 1785 1425.6 120 0.98 2950 80 P
1873 1840 100 1425.6 120 0.98 2950 85 P
1924 1898 150 1425.6 120 0.98 3400 81 P
1962 1940 200 1425.6 120 0.98 3400 80 P
- - 250 1425.6 120 0.98 3400 80 F
2076 1974 225 1425.6 120 0.98 3400 70 F
2114 1974 225 1425.6 120 1.08 3400 80 F
1974 1949 200 1475 200 120 0.98 3400 80 P
1974 1949 200 1539 450 120 0.98 3400 80 P
1974 1949 200 1553 500 120 0.98 3400 80 P
1974 1936 200 1575 600 120 0.98 3400 80 F

All testing for this review was conducted prior to tear-down (other than where thermocouples were required, obviously). The Hybrid mod results will be posted separately.

Let’s get into the testing.

Continue to Page 2 for GPU test methodology.

Prev Next »

Last modified on April 27, 2017 at 3:27 pm
Steve Burke

Steve started GamersNexus back when it was just a cool name, and now it's grown into an expansive website with an overwhelming amount of features. He recalls his first difficult decision with GN's direction: "I didn't know whether or not I wanted 'Gamers' to have a possessive apostrophe -- I mean, grammatically it should, but I didn't like it in the name. It was ugly. I also had people who were typing apostrophes into the address bar - sigh. It made sense to just leave it as 'Gamers.'"

First world problems, Steve. First world problems.

  VigLink badge