Power Consumption

A new and much needed addition to our GPU review is tracking power consumption. Here, we're using a simple meter to track the power consumption at the power supply level, so what we're left with is total system power consumption. But with the rest of the components in the test system remaining the same and the only variable being the video card, we're able to get a reasonably accurate estimate of peak power usage here.

At idle, all of the graphics cards pretty much perform the same, with the integrated graphics solution eating up the least amount of power at 106W for the entire system:

Peak System Power Consumption in Watts

Then, we fired up Doom 3 and ran it at 800x600 (High Quality) in order to get a peak power reading for the system.

Interestingly enough, other than the integrated graphics solution, the 6200 is the lowest power card here - drawing even less power than the X300, but that is to be expected given that the X300 runs at a 25MHz higher clock speed.

It's also interesting to note that there's no difference in power consumption between the 128-bit and 64-bit X300 cards. The performance comparison is a completely different story, however.

Peak System Power Consumption in Watts

In the end, none of these cards eat up too much power, with the X700 clearly leading the pack at 10% greater system power consumption than the 6600. It will be interesting to find out if there's any correlation here between power consumption and performance.

The Contenders The Test
Comments Locked

44 Comments

View All Comments

  • PrinceGaz - Tuesday, October 12, 2004 - link

    I'm assuming the 6200 you tested was a 128-bit version? You don't seem to mention it at all in the review, but I doubt nVidia would send you a 64-bit model unless they wanted to do badly in the benchmarks :)

    I don't think the X700 has appeared on an AT review before, only the X700 XT. Did you underclock your XT, or have you got hold of a standard X700? I trust those X700 results aren't from the X700 XT at full speed! :)

    As #11 and #12 mentioned, with the exception of Doom 3, the X600 Pro is faster than the 6200:

    Doom 3 - 39.3 60.1 (-35%)
    HL2 Stress Test - 91 76 (+20%)
    SW Battlefront - 45 33 (+36%)
    Sims 2 - 33.9 32.2 (+5%)
    UT2004 (1024x768) - 46.3 37 (+25%) [they were CPU limited at lower resolutions]
    BF Vietnam - 81 77 (+5%)
    Halo - 45.2 44 (+3%)
    Far Cry - 74.7 60.6 (+23%)

    So the X600 Pro is slower than the 6200 (128-bit) in Doom 3 by a significant amount, but its marginally faster than it in three games, and its significantly faster than the 6200 in the other three games and also the HL2 Stress Test. So that makes the X600 Pro the better card.

    The X700 absolutely thrashed even the 6600, let alone the 6200, in every game except of course Doom 3 where the 6600 was faster, and Halo where the X700 was a bit faster than the 6600 but not by such a large amount.

    Given the prices of the ATI cards, X300SE ($75), X300 ($100), X600 Pro ($130), X700 (MSRP $149); the 6600 is going to have to be priced at under its MSRP of $149 because of the far superior X700 at the same price point. Lets say a maximum of $130 for the 6600.

    If thats the case, I can't see how the 6200 could have a street-price of $149 (128-bit) and $129 (64-bit). How can the 6200 (128-bit) even have the same price as the faster 6600 anyway? Its also outperformed by the $130 X600 Pro which makes a $149 price ridiculous. I think the 6200 will have to be priced more like the X300 and X300SE-- $100 and $75 for the 128-bit and 64-bit versions respectively, if they are to be successful.

    Maybe most 6200's will end up being cheap 64-bit cards that are sold to people who aren't really bothered about gaming, or who mistakenly believe the amount of memory is the most important factor. You just have to look at how many 64-bit FX5200's are sold.
  • Shinei - Tuesday, October 12, 2004 - link

    The PT Barnum theory, wilburpan. There's a sucker born every minute, and if they're willing to drop $60 for a 64-bit version of a card when they could have had a 128-bit version, so much the better for profits. The FX5200 continues to be one of the best selling AGP cards on the market, despite the fact that it's worse than a Ti4200 at playing games, let alone DX9 games.
  • wilburpan - Tuesday, October 12, 2004 - link

    "The first thing to notice here is that the 6200 supports either a 64-bit or 128-bit memory bus, and as far as NVIDIA is concerned, they are not going to be distinguishing cards equipped with either a 64-bit or 128-bit memory configuration."

    This really bothers me a lot. If I knew there were two versions of this card, I definitely would want to know which version I was buying.

    What would be the rationale for such a policy?
  • wilburpan - Tuesday, October 12, 2004 - link

  • nserra - Tuesday, October 12, 2004 - link

    Why do you all keep talking about the Geforce 6600 cards (buying them) when the X700 was the clear winner?
    You all want to buy the worst card (less performing)? I dont understand.

    Why dont anantech use 3Dmark05?

    No doubt that mine 9700 was a magnificent buy almost 2 years ago. What a piece of cheat are the Geforce FX line of cards....
    Why didnt they use one (a 5600/5700) just to see...

    Even 4pipe line Ati cards can keep up with 8 pipe nvidia, gee what a mess... old tech yeah right.
  • coldpower27 - Tuesday, October 12, 2004 - link

    I am very happy you included Sims 2 into your benchmark suite:)

    I think this game like the amount of vertex processor on X700 plus it's advanatge in fillrate and memory bandwidth, could you please test the Sims 2 when you can on the high end cards from both vendors? :P
  • jediknight - Tuesday, October 12, 2004 - link

    What I'm wondering is.. how do previous generation top-of-the-line cards stack up to current gen mainstream cards?
  • AnonymouseUser - Tuesday, October 12, 2004 - link

    Saist, you are an idiot.

    "OpenGl was never really big on ATi's list of supported API's... However, adding in Doom3, and the requirement of OGL on non-Windows-based systems, and OGL is at least as important to ATi now as DirectX."

    Quake 3, RtCW, HL, CS, CoD, SW:KotOR, Serious Sam (1&2), Painkiller, etc, etc, etc, etc, are OpenGL games. Why would they ONLY NOW want to optimize for OpenGL?
  • Avalon - Monday, October 11, 2004 - link

    Nice review on the budget sector. It's good to see a review from you again, Anand :)
  • Bonesdad - Monday, October 11, 2004 - link

    Affordable gaming??? Not until the 6600GT AGP's come out...affordable is not replacing your mobo, cpu and video card...

Log in

Don't have an account? Sign up now