Diablo III Mobile Performance Compared

So far we’ve determined that Diablo III isn’t a particularly taxing game, especially early on—at least not for your GPU; your mouse buttons might be a different story!—and that AMD, Intel, and NVIDIA graphics solutions deliver comparable image quality. The only question that remains is how quickly they can deliver that result to your display. We’ve used quite a few different laptops to see what sort of performance you can expect with Diablo III. Here’s the quick rundown.

First up, from AMD we have a Llano prototype with an A8-3500M APU and integrated HD 6620G graphics. There are faster clocked Llano APUs in terms of CPU performance, but by default all of the A8 GPUs run at 444MHz with 400 Radeon Cores. Second is our Trinity prototype laptop with and A10-4600M (HD 7660G graphics), running 384 Radeon Cores at a substantially higher 686MHz clock. A third option from AMD is the discrete Radeon HD 6630M, and we tested three laptops with that GPU; first is the Llano A8-3500M APU, second is a Sony VAIO C with a faster Intel i5-2410M CPU, and third is a Sony VAIO SE with an i7-2640M. This will at least give us some indication of whether or not CPU performance is a factor in Diablo III performance.

Unfortunately, we do have to make a note on the drivers for the HD 6630M laptops: all three laptops aren’t able to run the latest AMD reference drivers, as they all use some form of switchable graphics. The prototype Llano system (with drivers from June 2011) can be excused, as there’s not much point for AMD to invest a lot of time improving the drivers or end user experience on that laptop, but Sony’s laptops continue to be a concern with their often-over-six-months-old drivers. The VAIO C is using drivers that date back to June 2011 (released in October) while the VAIO SE actually is lucky as it had a driver update from Sony earlier this month; unfortunately, the driver build still appears to date back to December 2011. We didn’t notice any rendering issues with any of the 6630M laptops, but bear in mind that it’s possible performance is lower due to the outdated drivers.

From the Intel camp, we tested three different laptops. On the low end of the spectrum is a Dell Vostro V131 with i5-2410M CPU and HD 3000 graphics. We also tested with a quad-core i7-2820QM and HD 3000 graphics to see how much the slightly higher IGP clocks and significantly faster CPU matter with Diablo III. The third laptop is the ASUS N56VM Ivy Bridge prototype, with an i7-3720QM CPU and HD 4000 graphics. We do have a fourth Intel option on hand, an Intel Ultrabook with IVB ULV, but we can’t report the CPU model yet and I’m not sure about talking performance, so we’ll hold off discussing that for a few more days. Anand did test an ASUS UX21A in Diablo III and you can read his comments, but he used a different test sequence and again we can’t name the exact CPU he used, so stay tuned if you want to find out how dual-core (and potentially less expensive) Ivy Bridge matches up against Llano and Trinity.

Finally, from NVIDIA we’ve got the same ASUS N56VM with i7-3720QM, only this time we’ve enabled the GT 630M graphics. We also ran some tests with an Acer AS3830TG that has an i5-2410M CPU with GT 540M graphics. The Acer is known to have issues with CPU throttling in some games, but it does have higher clocks on the GPU than the N56VM, so this will give us some indication of how much—or how little—CPU performance matters with Diablo III. Finally, we also have in a second Clevo W110ER in for review, this time from AVADirect, with an i7-3610QM and GT 650M graphics. Overkill for Diablo III? Most likely, but it’s an awfully compact laptop for that much hardware!

Here are the benchmark results; again, keep in mind that the in-town comparisons are using an identical FRAPS run whereas the Old Ruins area is slightly randomized as far as monster locations and quantity and is more prone to variance between runs. Note that we didn’t bother running Sandy Bridge HD 3000 at our Enthusiast settings with the i7-2820QM; it was already struggling at our Mainstream settings, and the i5-2410M results will tell you everything you need to know about how well HD 3000 handles maxed out settings.

Update: As noted earlier, many are saying the later stages and higher difficulty levels can really start to drop frame rates. Take the following graphs as a reference point, and plan on dropping some detail settings and/or resolution later in the game on lower end hardware.

Diablo III - Value - New Tristram

Diablo III - Value - Old Ruins

Diablo III - Mainstream - New Tristram

Diablo III - Mainstream - Old Ruins

Diablo III - Enthusiast - New Tristram

Diablo III - Enthusiast - Old Ruins

There’s plenty of data to cover, so let’s just start at the top with the discrete NVIDIA GPUs. Not surprisingly, the GT 650M powers through Diablo III without any issues; even at maximum detail and 1080p resolution, it’s still pulling nearly 40 FPS. The second set of GPUs, the GT 630M in the N56VM and the GT 540M in the Acer AS3830TG, are in theory supposed to be roughly the same performance. However, we've seen in the past that the Acer sometimes has issues with throttling, so potentially the GT 540M is running with a thermally constrained CPU in the AS3830TG. The charts above clearly show that the Acer can’t keep up with the Ivy Bridge solution. Either Diablo III is very good at using multi-core CPUs (doubtful, given what we saw with Blizzard’s StarCraft II, not to mention a quick look at Perfmon with Diablo III), or the Acer is once again not hitting higher clock speeds.

Update #2: So it appears that the ASUS N56VM is not running a lower clocked GPU; in fact, the opposite is true. NVIDIA's control panel reports 475MHz on the GPU core, 950MHz on the shaders. I've been a bit confused about the performance since day one, but several other utilities reported 475MHz as well, including GPU-Z. Interestingly however, I just ran GPU-Z with the sensor logging option while doing a FRAPS run in Diablo III. Instead of 475/950MHz, the sensors tab is instead reporting 797.3/1594.7MHz. Mystery solved: the GT 630M in the N56VM is actually clocked almost 20% higher than the stock GT 540M. That would explain the differences seen above.

We did a quick check and found that the typical CPU clocks for the i5-2410M during our test sessions typically ranged from 800MHz to 1.7GHz range, which you can see in the above image. (Side note: we also tested with ThrottleStop active, which is what the above chart shows; it was set to a 21X multiplier, but clearly that didn't work as intended.) The average clock speeds of the two cores during our test sequance are a rather slow 1200MHz and 1085MHz, so clearly the CPU isn't really providing the sort of clocks we usually see on i5-2410M. However, Diablo III doesn’t appear to need a ton of CPU performance; given the new information we have on the GT 630M clocks (see update above), it appears that Diablo III simply doesn't push the Acer hard enough to activate higher CPU clocks most of the time.

The second grouping of scores is mostly in red/orange, representing the AMD GPUs/APUs. For the red bars, Trinity and Llano both provide acceptable performance at our Value settings, and they’re still fast enough for the Mainstream settings—remember as we mentioned in the intro that Diablo III is actually quite playable at anything above 20 FPS. Once we hit our Enthusiast settings, both drop quite a bit; Trinity remains tolerable, but Llano definitely can’t keep up and you’d need to drop the Shadow Quality to Low at the very least for 1080p. Another really interesting piece of information we discover is that Trinity with it’s integrated GPU is still faster across the board than the HD 6630M (though there’s a possibility HD 6630M is being hurt by the outdated drivers). As for the three way HD 6630M comparison, CPU performance does appear to help a bit—the i7-2640M is typically slightly faster than the i5-2410M and A8-3500M—but the largest spread is only 15% at our Value settings; at Mainstream the gap drops a bit to 10-12%, while at Enthusiast it’s under 10%. Given the frame rates, the extra 15% never really means the difference between unplayable and playable; all three laptops with HD 6630M tend to handle up to our Mainstream settings quite well.

The final three lines are the blue Intel IGP results. HD 4000 with quad-core Ivy Bridge trails Llano across all settings, though it’s often close enough. Performance at Mainstream is a bit questionable; sure, you can play Diablo III well enough in our experience at 20-25 FPS, but it’s not going to be the smoothest result. Llano may only be 3-4 FPS faster at Mainstream, but that 12% performance increase is just enough to make the result a bit smoother. Your best bet with HD 4000 is ultimately going to be turning the Shadow Quality down to Low/Off, and then running at 1600x900.

As for Sandy Bridge’s HD 3000 IGP, perhaps the less said the better. Even at our Value settings, it only qualifies as tolerable, and at Mainstream it’s quite choppy—you could still play Diablo III at 13-18 FPS in a pinch, but I wouldn’t recommend it, and I doubt it would work well in multiplayer. Once frame rates drop below 15 FPS, it appears the engine starts to slow down rather than just skipping animations. Our New Tristram run usually takes around 20 seconds to complete (even at 20.1 FPS on the HD 4000), but when frame rates are in the low teens the time for the town run increases to around 30 seconds. Single-player is still possible, but that’s as far as I’d go—and it will take longer for everything you do, thanks to the moderate slowdown. When the HD 3000 drops below 10 FPS, what was sluggish takes a major nosedive; the town run required just over 60 seconds to complete, and the Old Ruins run that usually requires about 100-110 seconds clocked in at 308 seconds. Yup, there’s a reason we didn’t try suffering through the Enthusiast benchmark a second time on HD 3000!

Other Performance Tests

We did a few other tests to round out our performance information, though we didn't repeat the tests multiple times or run them on all of the systems. For one test, we used our Enthusiast settings but with Shadows on Low/Off with the HD 4000; the result of the testing is scores that are slightly better than the Trinity scores with Shadows on High. With Low shadows at 1080p, New Tristram scored 20.1 FPS and the Old Ruins scored 18.5 FPS; drop the shadows to Off and New Tristram runs at 27.1 FPS with Old Ruins at 24.8 FPS. In total, the difference between High Shadow Quality and Low Shadow Quality is over 50%, and going from Low to Off is another 35%. The other test was to use our maxed out settings but at 1366x768, again on the HD 4000. The frame rates were 17.3/16.4, or around 35% faster than at 1080p.

Given those results, it appears that Shadow Quality is the single most demanding setting, trumping even resolution. On HD 4000, you can basically double your performance at 1080p by turning off the shadows. Without doing in-depth testing (remember, we're looking at about five minutes to set up and run each benchmark setting, so I've already spent around 10 hours just doing the basic set of results shown above, not to mention testing other settings!), I can't say for certain, but my general impression is that the results are similar with other IGPs/GPUs.

Diablo III Graphics Settings and Image Quality Detailed FRAPS Runs and Closing Thoughts
Comments Locked

87 Comments

View All Comments

  • DanNeely - Monday, May 28, 2012 - link

    All joking about account sharing not withstanding, would AT buying a new D3 account for testing and letting a volunteer (not me) level it up for late game/hell testing be a viable option?
  • JarredWalton - Monday, May 28, 2012 - link

    We do have a couple people playing the game, so at some point we'll be able to test later levels. Give me a chance to: A) have a holiday (today), B) write a few other articles, C) enjoy the game without more benchmarking of it (which totally kills the fun!). Probably in a week or so I can come back with results from Act II or later.

    Unless you can talk Anand into your idea? ;-)
  • DanNeely - Tuesday, May 29, 2012 - link

    My diplomacy skills are of the Europe 1914 level; the odds of my being able to sweet talk someone I don't know well into anything are slim to none.

    Better results in a week or so isn't that bad a delay. I'm just mildly frustrated since I've had a few people ask what sort of hardware they needed to play the game; and it seems that all the numbers I can find are from very early in the game and thus non-representative of what's really needed.
  • damianrobertjones - Saturday, May 26, 2012 - link

    I'm hoping that Windows 8 Metro games bring a stable platform and for once I'm glad that at we'll at least have HD4000 as a base platform.
  • dagamer34 - Saturday, May 26, 2012 - link

    I'm wondering how the HD 4000 compares to the GPUs in ARM SoCs as that will be the actual low mark if slower.
  • tipoo - Saturday, May 26, 2012 - link

    I'm curious about that as well, SoC GPUs like the SGX 543MP4 are getting pretty complex and Intel themselves used to use integrated PowerVR GPUs in their chipsets.
  • tipoo - Saturday, May 26, 2012 - link

    The GMA 3600 is based on the PowerVR SGX545.
  • JarredWalton - Saturday, May 26, 2012 - link

    And the Windows drivers for it are crap right now. I'm just saying....
  • Penti - Saturday, May 26, 2012 - link

    Actually Metro/WinRT won't be used for gaming, If you want a restricted environment there already is XNA so. Games will be too difficult and too less of an incentive or anything to gain to port to the WinRT framework. Or Windows Runtime as they call it. Game developers will never target Metro/WinRT if they don't have to and they don't on x86 machines, desktop is there, you can still build for Windows 7 etc where most users are and so on. Won't happen that much here until next gen consoles either. Plus macs have gotten a whole lot better in the department and plenty of game engines are available now. Taking those kind of engines and porting to C++/WinRT isn't something taken lightly it probably won't actually be possible without a rewrite which defeats the purpose. The performance wouldn't be good. The sandbox is probably too restrictive. It also means in practice it is a more restrictive environment then the mobile sandboxed OS's, several mobile OS's run Firefox for example. WinRT never will. WinRT never will run even IE.
  • oopyseohs - Saturday, May 26, 2012 - link

    Did I miss the part where you talk about using an external monitor, or how else were you able to run all of these GPUs at all three resolutions? I'm not saying the data isn't important, as it could be relevant to different notebooks that use the same or similar hardware just with higher-res screens.

    Also, I've played this game on an old desktop with with GTX 285 @ 1080p and everything turned up. While that is fairly smooth and playable, I still get quite a few moments of "stuttering" in Hell difficulty. I also play on basically the same Acer book with the GT 540M, and even at the lowest possible graphics settings and resolution in normal mode, it's hard for me to characterize that performance as anything other than horrible in comparison to the desktop.

Log in

Don't have an account? Sign up now