Final Words

So ... wow. And this isn't even all the data. We've still got a four GPU update to bring out as soon as we finish aggregating and analyzing everything.

First things first though: let's sort through our 3-way tests and try to make some sense of all this.

The first thing to take away is that 3-way graphics systems are absolutely not something anyone without a 30" display needs or should even want. There are so many instances where 3-way fails to improve on 2-way, and because of the high price barrier to entry and the diminishing returns on adding more hardware, 3-way is just not for everyone and has questionable value for those who choose to go that route as well.

Honestly, NVIDIA and AMD trade blows, but the one clear thing to note is that 3-way GTX 280 and 285 are just way overpriced. You don't get what you pay for and you can still get huge performance from the GTX 260 which offers consistently better value. On the AMD front, 3-way 4870 1GB pairing a 4870 X2 and single 4870 1GB is the way to go if you want 3-way. The 4850 can scale in some games that don't crush it with too much data and it offers better value in some situations.

But even more than single or two card solutions, 3-way is a case by case basis. You've really got to look at the games you like and pick the solution that performs best there. This is in large part because it's just such an investment and these options are just not "worth" the cost unless you are really picking your graphics hardware for a purpose. There isn't an easy way to make a general recommendation.

Other than to stay away from 3-way GTX 280 and 285, we've got to split our recommendation between 3-way GTX 260 and 3-way 4870 1GB. If you've got to have a 3-way solution that is. And much of our data reinforces our recommendations from our look at 2-way multiGPU solutions which tend to offer a better balance of performance and value if more power than a single GPU can offer is required.

But really, 3-way, and especially with 3 cards, is just not for everyone and not something we put a lot of stake into as a good option. Saving the money to upgrade to a new architecture is going to get you a lot more for the money than trying to squeeze longevity out of hardware by stacking more than 2 of the same thing in one system.

Power Consumption
Comments Locked

46 Comments

View All Comments

  • 7Enigma - Thursday, February 26, 2009 - link

    http://www.anandtech.com/mb/showdoc.aspx?i=3506&am...">http://www.anandtech.com/mb/showdoc.aspx?i=3506&am...

    Just a quick review (would have preferred to see the upper resolutions tested....wait did I just write that?), but shows that with CF you have a pretty competetive war. I would guess going tri-gpu (if the game supports it well) the i7 would pull ahead by quite a bit, especially when OC'd, but that remains to be seen.
  • Mr Roboto - Thursday, February 26, 2009 - link

    BORING! Not to sound like a total dick but seriously do we need 2 articles (with probably a third to come discussing quad GPU) to find out what a total waste multi GPU is? I mean until NV and ATI invest some serious research into hardware based SLI\XFire I'd never even consider buying two GPU's. The current software based multi GPU setups don't offer enough of a reward. Having to use a software based profile means you'll always be waiting for them to create one for the latest games, while some simply won't get one at all.

    Software in general is lagging so far behind current hardware that there is practically ZERO use for two GPU's let alone 3-4. Someone please name 10 games that can successfully use more than 2 cores at a time. OK now name 10 software applications that do the same (not counting video encoding apps). It's pathetic how bad it is. Look at all the XBox360 ports built on 5 year old hardware that are coming out in rapid succession and you have a pretty glib picture.

    Still waiting for some Lucid Hydra benchmarks showing 100% linear scaling. If it actually materializes on the desktop then I would consider multi GPU.
  • 7Enigma - Thursday, February 26, 2009 - link

    I find these articles well done and beneficial. The stereotype that tri/quad/sometimes even dual is a waste is just that a sterotype. For games that are coded well for multi-gpu and when you are on a large display I'm actually a bit impressed with the multi-gpu stats. A year ago you would likely not see even near the improvement from a single card. It's a testament to the drivers of the gpu manufacturers and as importantly the game programmers.

    What this also shows is games that are poorly suited for multi-gpu (either due to poor coding/drivers or bottlenecked at another point). Sure most of us don't have the 30" display that warrants the tri and upcoming quad articles, but for those that do it is not a waste, and look further down the rabbit hole to see where we will be in another 2 years or so with a single-gpu card.

    We have already pretty much approached a saturation point for mainstream graphics gaming (<$300) under 24" (1920X1200). That has to be dangerous for ATI/Nvidia because we are now becoming limited by our display size as opposed to the historical trend of ever demanding games keeping us at pretty much the same resolution year after year (with slow progress from 800X600 to 1024X768 to the now most common 1280X1024, to the quickly becoming gaming standard 22", and finally within a year or two to 24" gaming goodness).

    What I imagine you will see in the next year or two is heavy reliance on physics calculations on the gpu to burden them greatly bringing us back to the near always present gpu-limited status. If game makers fail to implement some serious gpu demands above and beyond what is being used today pretty soon we won't need to upgrade video cards near as frequently. And for the consumer that is actually a really good thing...
  • Pakman333 - Thursday, February 26, 2009 - link

    "BORING! Not to sound like a total dick"

    You do, and you probably are.

    If you don't care, then why read it? No one is forcing you to.
  • poohbear - Thursday, February 26, 2009 - link

    props for posting such a quick follow up! sometimes hardware sites promise follow-ups & never deliver.;)
  • Jynx980 - Thursday, February 26, 2009 - link

    Nice job on the resolution options for the graphs! This should alleviate frustration on both ends. I hope it's the standard from now on.
  • rcr - Thursday, February 26, 2009 - link

    I read some times ago, that the GTX285 will draw less power than a GTX280. But the graph of power consumption is here also made wih 3xGTX280OC.
  • ilkhan - Thursday, February 26, 2009 - link

    As a reader I hate going back and forth between pages. Id love to see the price for each option on the right hand side of the performance graphs. Would make it super easy to compare the value of each solution, rather than the seperate graphs you are using here.

    And on the plus side for 4 way scaling, at least theres only 4 cards to bench (9800GX2, 4850X2, 4870X2, and the GTX295). Maybe a couple more for memory options, but a lot less than the 3-way here.
  • LiLgerman - Wednesday, February 25, 2009 - link

    i currently have a 4870, 512mb TOP from asus and i was wondering if i can pair that with a 1gb model. Will i have to down clock my 4870 TOP to the speed of the 1gb model if i do crossfire?
    Thanks
  • DerekWilson - Thursday, February 26, 2009 - link

    AMD is good with asynchronous clocking (one can be faster than the other) ... but i believe memory sizes must match -- i know you could still put them in crossfire, but i /think/ that 512MB of the memory on the 1GB card will be disabled to make them work together.

    this is just what i remember off the top of my head though.

Log in

Don't have an account? Sign up now