Shadow of Mordor

The final title in our testing is another battle of system performance with the open world action-adventure title, Shadow of Mordor. Produced by Monolith using the LithTech Jupiter EX engine and numerous detail add-ons, SoM goes for detail and complexity to a large extent, despite having to be cut down from the original plans. The main story itself was written by the same writer as Red Dead Redemption, and it received Zero Punctuation’s Game of The Year in 2014.

For testing purposes, SoM gives a dynamic screen resolution setting, allowing us to render at high resolutions that are then scaled down to the monitor. As a result, we get several tests using the in-game benchmark. For low-end graphics we examine at 720p with low settings, whereas mid and high-end graphics get 1080p Ultra. The top graphics test is also redone at 3840x2160, also with Ultra settings, and we also test two cards at 4K where possible.

Shadow of Mordor on ASUS GTX 980 Strix 4GB ($560)

Shadow of Mordor on ASUS GTX 980 Strix 4GB ($560)

Shadow of Mordor on MSI R9 290X Gaming LE 4GB ($380)

Shadow of Mordor on MSI R9 290X Gaming LE 4GB ($380)

Shadow of Mordor on MSI GTX 770 Lightning 2GB ($245)

Shadow of Mordor on MSI R9 285 Gaming 2GB ($240)

Shadow of Mordor on ASUS R7 240 DDR3 2GB ($70)

Shadow of Mordor on Integrated Graphics

The only real difference here between the newer Core i3-7350K and the older Core i7-2600K is with our mid-range cards (GTX 770 and R9 285), whereby the older CPU seems to have a deficit 'in general' to the other CPUs we've tested. This might be CPU instruction related, although these results aren't seen on the other cards.

Gaming: GRID Autosport Power and Overclocking
Comments Locked

186 Comments

View All Comments

  • JordanV - Tuesday, February 14, 2017 - link

    I think the sales argument for the big Intel chips as video encoders has been for x264 where the faster NVENC, VCE, and QuickSync technologies offer lower quality at a given bitrate for higher quality x264 settings. For most people, the hardware encoders are enough but for many others, the quality is not sufficient.

    The quality difference between hardware and software HEVC is smaller with higher quality software h265 encodes beating the quality of your Pascal x265 encodes but with a big performance penalty. It's not worth it for most people, but if you have limited bitrate/storage and want the best quality, it might be.
  • HerrKaLeun - Friday, February 3, 2017 - link

    Thanks for the great review, Ian.
    Considering one needs an expensive Z-board to OC, for most people buying an i5 makes more sense.
    I don't understand why so many people complain about Intel allegedly not making enough progress. Now you get a dual-core that comes close (or even exceeds in single threaded benches) to the former flagship quadcore. If you want to have a CPU that vastly exceeds the "old" quadcore, Intel also has newer quadcores. It is not like the i3 is the end of the lineup.... For the $317 that the 2600k used to cost you can get a KabyLake non-K i7, which sure vastly exceeds in performance (and much lower TDP). I assume someone who could afford an over $300 CPU 6 years ago can afford $300 now and upgrading to an i3 may not be what that person would do anyway. the trend goes to more cores.... most people here complain about Intel not offering mainstream hexa and octa cores... not sure why the same people allegedly are so eager to get dual-cores.
  • zodiacfml - Friday, February 3, 2017 - link

    Dual core is too weak for me. Web browsing can use more cores.
  • Hulk - Friday, February 3, 2017 - link

    Sorry to be dense.
    What does 2+2, 4+2, 4+3/e mean?
  • babysam - Saturday, February 4, 2017 - link

    The first number refers to the number of CPU cores. Te second number refers to the IGP configuration (the number of shaders, which may be a little bit different across generations, e.g. Haswell GT3 has 40 shaders, while Broadwell/Skylake GT3 have 48 shaders).
    The extra e means there is an extra eDRAM cache (Crystalwell) on the CPU package.
  • Hulk - Saturday, February 4, 2017 - link

    Thanks.
  • AndrewJacksonZA - Saturday, February 4, 2017 - link

    Thank you babysam.
  • babysam - Saturday, February 4, 2017 - link

    Thank you for you article (especially when many of us are waiting on the information of new CPU of both AMD and Intel). It is always good to have something to play (overclocked) with, but this is a little bit expensive.

    When I read the analysis of the first page, I see the lack of information on the CPU die size and transistor count disclosed by Intel recently. Also, I feel strange that the effect of the change of the 32nm to 22nm (from Haswell to Broadwell) have such a large difference on the 2C+GT2(which Intel claims there is a 37% reduction of the die, which can be seen in the table) and the 4C+GT3(which the difference are much smaller) CPU die. I feel even stranger when I see the Skylake 4C+GT3e die is a bit smaller than the Broadwell 4c+GT3e die. So I am quite curious on the sources of the die estimate.

    P.S. I found the origin of the 234mm^2 of the Skylake die size estimate.

    https://techreport.com/forums/viewtopic.php?t=1177...

    which based on the images of the following
    http://www.anandtech.com/show/10281/intel-adds-cry...

    It seems that the die described is the Skylake-H(which is a 4C+GT4e configuration). This makes the 241.5mm^2 estimate of the Broadwell 4C+GT3e a little bit unrealistic (Skylake GT4e have 72 shaders, while Broadwell GT3e have 48 only)
  • babysam - Saturday, February 4, 2017 - link

    Just find the die size of the Broadwell-H (4C+3e) in this document

    http://www.intel.com/content/www/us/en/embedded/pr...

    According to the document, the die size of Broadwell-H (4C+GT3e) should be 13.7mmx12.3mm = 168.51mm^2

    (Many thanks for the hints: https://forums.anandtech.com/threads/broadwell-cor... , which the got the answer two years ago.)
  • WoodyBL - Saturday, February 4, 2017 - link

    Am I the only one noticing that the i5-4690 was beating the i5-7600k in a lot of benchmarks? I'm having a hard time processing how that was even possible...

Log in

Don't have an account? Sign up now