Introduction

Hardware is caught up in an eternal struggle to achieve the absolute best quality and performance in the latest games. With games like Doom 3 or Chronicles of Riddick, you don't need to have the best graphics card to enjoy them, but in order to play them (smoothly) with the highest quality settings, you need one that's top of the line. That's where NVIDIA's 7800 series comes in. This line of graphics cards is the most powerful that you can buy right now, and we intend ultimately to test and review as many as we can get our hands on.

Our last review covered EVGA's e-GeForce 7800 GTX, and this, the second in our series, will be on the MSI NX7800 GTX (VT2D256E). We are planning to review more and more variations of the 7800, and add them to the mix to see how they all stack up against each other. This will hopefully give you a good idea of which of these cards will best suit your needs.

Again, we'll be looking at heat, power draw and noise levels for the card. We'll also see how well the card performs with and without overclocking, and how it stands up to EVGA's 7800 and our reference G70. We've also added a set of tests with 4xAA enabled for comparison in the performance tests section.

Now that we have a second card in our series, we'll be looking closely at how these compare to each other. The most important factors in our decision will be performance and price. Before we get to the numbers, let's take a look at what we get out of the box.

The Card
Comments Locked

42 Comments

View All Comments

  • Fluppeteer - Friday, July 29, 2005 - link

    "Advertise" is perhaps a strong word, but the PDF data sheet on the eVGA web site
    does say that one output is dual link (even though the main specifications say
    the maximum digital resolution is 1600x1200, which is nonsense, like all resolution
    claims, even for most single link cards).

    I couldn't (last I looked) find anything about dual link support on the MSI site.
    But then, MSI have in the past ignored that the 6800GTo was dual link, and then
    claimed that their (real) 6800GT *was* dual link, and that the SiI transmitters
    were unnecessary... (Although I'm still mystified how the PNY AGP 6600GT seems to
    have dual dual link support without external transmitters.)

    I'm presuming both heads have analogue output, btw (I only ask because the GTo,
    for some astonishing reason, only has digital output on its single link head).

    Past experience (with the 6800) suggests that the reason none of the manufacturers
    mention it is that very few people actually know what dual link DVI *is*. A lot
    probably haven't tried it - there being, last I looked, only three monitors which
    can use it anyway, two of which are discontinued. nVidia caused a lot of confusion
    by claiming support in the chipset and putting an external transmitter on their
    reference card, which most manufacturers left off without updating their specs.
    Unfortunately, nVidia seem to fob off all their tech support to the manufacturers,
    who aren't always qualified to answer questions - I've not found anywhere to send
    driver feature requests, for example. Seeing the external transmitter make it to
    released boards is a vast relief to me.

    Now the Quadro 4500 has been announced, I'm hoping the 512MB boards will appear
    (and they might be DDL). Fingers crossed.
  • DerekWilson - Thursday, July 28, 2005 - link

    Yes. Again, the SI TMDS for dual-link is on the pcb. So far there are no 7800 cards that we have seen without dual-link on one port.

    NVIDIA didn't even make this clear at their initial launch. But it is there. If we see a board without dual-link we'll let you know.
  • Wulvor - Monday, July 25, 2005 - link

    For that extra $4 you are also paying for a longer Warranty. eVGA has a 1+1 warranty, so 1 year warranty out of the box, and another 1 year when you register online at eVGA. MSI on the other hand has a 3 year warranty, and BFG a lifetime warranty.
    It must be the corporate purchaser in me, $4 is well worth the extra year ( or 2 ), but I guess if you are going to be on the "bleeding" edge, then you are buying a new video card every 6 months anyways, so who cares?
  • smn198 - Monday, July 25, 2005 - link

    A suggestion:

    Regarding measuring the card's noise output and the way you measured the sound
    "We had to do this because we were unable to turn on the graphics card's fan without turning on the system."

    Would it be possible to try and measure the voltages going to the fan when the card is idle and under full load? Then supply the fan with these voltages when the system is off using a different power supply such as a battery (which is silent) and a variable resister.

    It would also be interesting to see a graph of how the noise increases when going from idle to full load over 10 minutes (or however long it takes to reach the maximum speed) on cards which have . Instead of trying to measure the noise with the system on, again measure the voltage over time and then using your battery, variable resistor and voltage meter recreate the voltages and use this in conjunction with the voltage/time data to produce noise/time data.

    Thanks
  • DerekWilson - Monday, July 25, 2005 - link

    We are definitely evaluating different methods for measuring sound. Thanks for the suggestions.

    Just to be clear, even after hours of looping tests on the 7800 GTX overclocked to 485/625, we never once heard an audible increase in the fan's speed.

    This is very much unlike our X850 parts that spin up and down frequenly during any given test.

    We have considered attempting to heat the environment to simulate a desert like climate (we've gotten plenty of email from military personel asking about heat tolerance on graphics cards), but it is more difficult than it would seem to heat the enviroment without causing other problems in our lab.

    Suggestions are welcome.

    Thanks,
    Derek Wilson
  • at80eighty - Tuesday, July 26, 2005 - link

    We have considered attempting to heat the environment to simulate a desert like climate [...] but it is more difficult than it would seem to heat the enviroment without causing other problems in our lab

    Derek, If you really wanna simulate desert like heat in the room, may i suggest inviting Monica Belluci to your lab ....should work like a charm :p
  • reactor - Monday, July 25, 2005 - link

    Ive been using MSI cards for a few years now, there fans always seem to run at top speed and ive found they usually run at higher RPM's(Slightly louder) than other manufacturers. I think that explains why the card is cooler while drawing more power, and why you didn't notice a difference in sound as the card was stressed. Im not entirely certain, but thats from my own expenriances with MSI cards.

    Good article, looking forward to the BFG.
  • yacoub - Monday, July 25, 2005 - link

    "As you can see, The EVGA slightly outperforms the MSI across the board at stock speeds."

    Either I'm reading it wrong or you mis-wrote that line, since I see the e-VGA normal and OC'd, the NVidia reference, and the MSI OC'd, but no MSI at stock speeds. Thus it's hard to compare th EVGA stock speeds vs the MSI stock speeds when one of them isn't on the charts.
  • DerekWilson - Monday, July 25, 2005 - link

    Check the bold print on the Performance page --

    MSI stock performance is the same as the NVIDIA reference performance at 430MHz ...

    To compare stock numbers compare the green bar to the EVGA @ 450/600

    Sorry for the confusion, but we actually tested all the games a second time and came up with the exact same numbers. Rather than add another bar, we thought it'd be easier to just reference the one.

    If you guys would rather see multipler bars for equivalent results across the board, we can certainly do that.

    Thanks,
    Derek Wilson
  • davecason - Monday, July 25, 2005 - link

    Since the MSI card drew a lot more power than expected but remained cooler than the eVGA card, I was thinking that some of the excess may be due to the cooling of the card itself. Maybe the fan on the MSI card works harder than the one on the eVGA card.

    The people at Anandtech could test the power usage of the stock video-card cooling fans independently to see what their effect is on power load. This may explain the 6 extra watts used by the MSI card. This information might be mildly useful to a person who was already stressing out their power supply with other things (such as several hard drives). Does anyone think that is worth doing?

Log in

Don't have an account? Sign up now