Age of Conan Analysis

The first thing we note here is that AMD leads the benchmark.




1680x1050    1920x1200    2560x1600


Both 3-way SLI and 3-way CrossFire perform very well in this game all the way up to 2560x1600. It looks like the GTX 280 and GTX 285 3-way tests are CPU limited here, but it is odd that the 4870 512MB 3-way solution scores so much lower than the 4850 3-way at 2560x1600 when they've got the same amount of RAM onboard.




1680x1050    1920x1200    2560x1600


Scaling from 1 to 3 GPUs comes in pretty well at 2560x1600 with the 9800 GTX+ and the GTX 260 netting around 140% improvement. the Radeon HD 4870 1GB does pretty well with about 128% improvement. Below the maximum resolution, we don't see scaling much better than 100% (though the GTX 260 and Radeon HD 4850 do manage it).




1680x1050    1920x1200    2560x1600


Moving from 2 to 3 GPUs shows a less impressive picture until we move to 2560x1600 where we see most of the hardware gain some good performance. The Radeon HD 4850 gains near its theoretical maximum at almost 45%. The oddity here is the GTX 260 which shows better performance scaling at lower resolutions peaking at 41% at 1920x1200.




1680x1050    1920x1200    2560x1600


As for value, the two most expensive solutions (the GTX 280 and 285) are pegged at the bottom of the chart. AMD's 4850 3-way shows good performance and good value, especially for a 3 card solution, in Age of Conan. The 4850 3-way doesn't quite reach the 4870 X2 which posts both better performance and better value.

We have 6 configurations that can't post 25 fps at 2560x1600, but some AMD single cards are able to squeak by and it looks as though two card solutions are a better balance of performance and cost than any 3-way option.

Prices, Stutter, and The Test Call of Duty World at War Analysis
Comments Locked

46 Comments

View All Comments

  • 7Enigma - Thursday, February 26, 2009 - link

    http://www.anandtech.com/mb/showdoc.aspx?i=3506&am...">http://www.anandtech.com/mb/showdoc.aspx?i=3506&am...

    Just a quick review (would have preferred to see the upper resolutions tested....wait did I just write that?), but shows that with CF you have a pretty competetive war. I would guess going tri-gpu (if the game supports it well) the i7 would pull ahead by quite a bit, especially when OC'd, but that remains to be seen.
  • Mr Roboto - Thursday, February 26, 2009 - link

    BORING! Not to sound like a total dick but seriously do we need 2 articles (with probably a third to come discussing quad GPU) to find out what a total waste multi GPU is? I mean until NV and ATI invest some serious research into hardware based SLI\XFire I'd never even consider buying two GPU's. The current software based multi GPU setups don't offer enough of a reward. Having to use a software based profile means you'll always be waiting for them to create one for the latest games, while some simply won't get one at all.

    Software in general is lagging so far behind current hardware that there is practically ZERO use for two GPU's let alone 3-4. Someone please name 10 games that can successfully use more than 2 cores at a time. OK now name 10 software applications that do the same (not counting video encoding apps). It's pathetic how bad it is. Look at all the XBox360 ports built on 5 year old hardware that are coming out in rapid succession and you have a pretty glib picture.

    Still waiting for some Lucid Hydra benchmarks showing 100% linear scaling. If it actually materializes on the desktop then I would consider multi GPU.
  • 7Enigma - Thursday, February 26, 2009 - link

    I find these articles well done and beneficial. The stereotype that tri/quad/sometimes even dual is a waste is just that a sterotype. For games that are coded well for multi-gpu and when you are on a large display I'm actually a bit impressed with the multi-gpu stats. A year ago you would likely not see even near the improvement from a single card. It's a testament to the drivers of the gpu manufacturers and as importantly the game programmers.

    What this also shows is games that are poorly suited for multi-gpu (either due to poor coding/drivers or bottlenecked at another point). Sure most of us don't have the 30" display that warrants the tri and upcoming quad articles, but for those that do it is not a waste, and look further down the rabbit hole to see where we will be in another 2 years or so with a single-gpu card.

    We have already pretty much approached a saturation point for mainstream graphics gaming (<$300) under 24" (1920X1200). That has to be dangerous for ATI/Nvidia because we are now becoming limited by our display size as opposed to the historical trend of ever demanding games keeping us at pretty much the same resolution year after year (with slow progress from 800X600 to 1024X768 to the now most common 1280X1024, to the quickly becoming gaming standard 22", and finally within a year or two to 24" gaming goodness).

    What I imagine you will see in the next year or two is heavy reliance on physics calculations on the gpu to burden them greatly bringing us back to the near always present gpu-limited status. If game makers fail to implement some serious gpu demands above and beyond what is being used today pretty soon we won't need to upgrade video cards near as frequently. And for the consumer that is actually a really good thing...
  • Pakman333 - Thursday, February 26, 2009 - link

    "BORING! Not to sound like a total dick"

    You do, and you probably are.

    If you don't care, then why read it? No one is forcing you to.
  • poohbear - Thursday, February 26, 2009 - link

    props for posting such a quick follow up! sometimes hardware sites promise follow-ups & never deliver.;)
  • Jynx980 - Thursday, February 26, 2009 - link

    Nice job on the resolution options for the graphs! This should alleviate frustration on both ends. I hope it's the standard from now on.
  • rcr - Thursday, February 26, 2009 - link

    I read some times ago, that the GTX285 will draw less power than a GTX280. But the graph of power consumption is here also made wih 3xGTX280OC.
  • ilkhan - Thursday, February 26, 2009 - link

    As a reader I hate going back and forth between pages. Id love to see the price for each option on the right hand side of the performance graphs. Would make it super easy to compare the value of each solution, rather than the seperate graphs you are using here.

    And on the plus side for 4 way scaling, at least theres only 4 cards to bench (9800GX2, 4850X2, 4870X2, and the GTX295). Maybe a couple more for memory options, but a lot less than the 3-way here.
  • LiLgerman - Wednesday, February 25, 2009 - link

    i currently have a 4870, 512mb TOP from asus and i was wondering if i can pair that with a 1gb model. Will i have to down clock my 4870 TOP to the speed of the 1gb model if i do crossfire?
    Thanks
  • DerekWilson - Thursday, February 26, 2009 - link

    AMD is good with asynchronous clocking (one can be faster than the other) ... but i believe memory sizes must match -- i know you could still put them in crossfire, but i /think/ that 512MB of the memory on the 1GB card will be disabled to make them work together.

    this is just what i remember off the top of my head though.

Log in

Don't have an account? Sign up now