Sleeping Dogs

Another Square Enix game, Sleeping Dogs is one of the few open world games to be released with any kind of benchmark, giving us a unique opportunity to benchmark an open world game. Like most console ports, Sleeping Dogs’ base assets are not extremely demanding, but it makes up for it with its interesting anti-aliasing implementation, a mix of FXAA and SSAA that at its highest settings does an impeccable job of removing jaggies. However by effectively rendering the game world multiple times over, it can also require a very powerful video card to drive these high AA modes.

Sleeping Dogs - 2560x1440 - Ultra Quality + High AA

Sleeping Dogs - Min. Frame Rate - 2560x1440 - Ultra Quality + High AA

Once more the tables turn with frame rates, and we see the 7990 hold on to a very small lead over the GTX 690.

Sleeping Dogs - Delta Percentages - 2560x1440 - Ultra Quality + High AA

Sleeping Dogs is another title where AMD has greatly improved on their frame consistency. With Catalyst 13.6 the 7990 had deltas over 100% (the variability was larger than the average frame time) which has since been heavily reduced with frame pacing. The end result eliminated AMD’s runt frames in this benchmark, and pushing them down to the 20% range. Still, this pales in comparison to the GTX 690. Though we do see the 6990 doing rather well for itself due to its lower frame rate, even with this metric compensating for that factor. It’s clearly a lot easier to schedule frames when you have more time to do it.

 

Sleeping Dogs - 95th Percentile FT - 2560x1440 - Ultra Quality + High AA

This is another case of where AMD’s massive improvements on frame pacing have led to an improvement in frame times at the 95th percentile. Here they shave off 7ms, and not unlike their frame rate are slightly ahead of the GTX 690.

Hitman: Absolution Battlefield 3
Comments Locked

102 Comments

View All Comments

  • chizow - Wednesday, August 7, 2013 - link

    There was discussions of microstutter on various forums associated with multi-GPU, but PCGH was the first site to publish it's findings in detail with both video evidence and hard data. From what I remember, they were the first to develop the methodology of using FRAPs frametimes and graphing the subsequent results to illustrate microstutter.
  • BrightCandle - Friday, August 2, 2013 - link

    One of the most shocking revelations to me is that AMDs quality assurance did not include checking the output of their cards frame by frame. I had always assumed that both NVidia and AMD had HDMI/DVI/VGA recorders that allowed them to capture the output of their cards so they could check them pixel by pixel, frame by frame and presumably check they were correct automatically.

    Such a technology would clearly have shown the problem immediately. I am stunned that these companies don't do that. Even FCAT is a blatantly blunt tool as it doesn't say anything about the contents of the frames. We still don't have any way to measure end to end latency for comparison either. All in all there is much to left to do and I am not confident that either company is testing these products well, its just I couldn't believe that AMD wasn't testing theirs for consistency (it was obvious when you played it something was wrong) at all.
  • krutou - Friday, August 2, 2013 - link

    AMD is in the business of being the best performance per price entry in every market segment. Technology and quality come second.

    How often does AMD introduce and/or develop technologies for their graphics cards? The only two that come to mind are Eyefinity and TressFX (100 times more overhyped than PhysX).
  • Death666Angel - Saturday, August 3, 2013 - link

    I think ATI had tessellation in their old DX8 chips. nVidia bought PhysX, so that shouldn't count. But I don't really see how having exclusive technology usable by a single GPU vendor is anything good. We need standardization and everybody having access to the same technologies (albeit with different performance deltas). Look at the gimmicky state of PhysX and imagine what it could be if nVidia would allow it to be fully utilized by CPUs and AMD GPUs?
  • krutou - Saturday, August 3, 2013 - link

    Because OpenCl and TressFX are doing so well right?
  • bigboxes - Sunday, August 4, 2013 - link

    March on, fanboi.
  • JamesWoods - Sunday, August 4, 2013 - link

    If you think that is all AMD/ATI has ever done for graphics then you sir, are ignorant. I was going to use a more degrading word there and thought better of it.
  • Will Robinson - Friday, August 2, 2013 - link

    LOL...what a load of tosh.
    "NVDA had to take them by the hand"?
    You and Wreckage ought to post in green text.
  • chizow - Friday, August 2, 2013 - link

    Agree with pretty much of all of this, although I would direct a lot of the blame on AMD's most loyal, enthusiastic supporters as well. Every time microstutter was mentioned and identified as being worst with AMD solutions, AMD's biggest fans would get hyperdefensive about it. If those most likely to have a problem were too busy denying any problem existed, it really should be no surprise it was never fixed.

    And this is the result. Years of denial and broken CF, finally fixed as a result of the scrutiny from the press and laughter of Nvidia fans which brought this to a head and forced AMD to take a closer look and formulate a solution.
  • EJS1980 - Friday, August 2, 2013 - link

    "Truth favors not one side."

Log in

Don't have an account? Sign up now