You’ve been living too perfect of a life if you’ve never used the phrase “it’s been a long day,” and for NVIDIA it has most definitely been a very long day. Just over two weeks ago the graphics industry was shook by some very hard hitting comments from Gabe Newell of Valve, primarily relating to the poor performance of NVIDIA cards under Half Life 2. All of the sudden ATI had finally done what they had worked feverishly for years to do, they were finally, seemingly overnight, crowned the king of graphics and more importantly – drivers. There were no comments on Half Life 2 day about ATI having poor drivers, compatibility problems or anything even remotely resembling discussions about ATI from the Radeon 8500 days.

Half Life 2 day was quickly followed up with all sorts of accusations against NVIDIA and their driver team; more and more articles were published with new discoveries, shedding light on other areas where ATI trounced NVIDIA. Everything seemed to all make sense now; even 3DMark was given the credibility of being the “I told you so” benchmark that predicted Half Life 2 performance several months in advance of September 12, 2003. At the end of the day and by the end of the week, NVIDIA had experienced the longest day they’ve had in recent history.

Some of the more powerful accusations went far beyond NVIDIA skimping on image quality to improve performance; these accusations included things like NVIDIA not really being capable of running DirectX 9 titles at their full potential, and one of the more interesting ones – that NVIDIA only optimizes for benchmarks that sites like AnandTech uses. Part of the explanation behind the Half Life 2 fiasco was that even if NVIDIA improves performance through later driver revisions, the performance improvements are only there because the game is used as a benchmark – and not as an attempt to improve the overall quality of their customers’ gaming experience. If that were true, then NVIDIA’s “the way it’s meant to be played” slogan would have to go under some serious rethinking; the way it’s meant to be benchmarked comes to mind.

But rewind a little bit; quite a few of these accusations being thrown at NVIDIA were the same ones thrown at ATI. I seem to remember the launch of the Radeon 9700 Pro being tainted with one accusation in particular – that ATI only made sure their drivers worked on popular benchmarking titles, with the rest of the top 20 games out there hardly working on the new R300. As new as what we’re hearing these days about NVIDIA may seem, let us not be victim to the near sightedness of the graphics industry – this has all happened before with ATI and even good ol’ 3dfx.

So who are you to believe? These days it seems like the clear purchase is ATI, but on what data are we basing that? I won’t try to build up suspense senselessly, the clear recommendation today is ATI (how’s that for hype-less journalism), but not because of Half Life 2 or any other conspiracies we’ve seen floating around the web these days.

For entirely too long we’ve been basing GPU purchases on a small subset of tests, encouraging the hardware vendors to spend the majority of their time and resources optimizing for those games. We’re not just talking about NVIDIA, ATI does it too, and you would as well if you were running either of those two companies. We’ve complained about the lack of games with built-in benchmarks and cited that as a reason to sticking with the suite that we’ve used – but honestly, doing what’s easy isn’t a principle I founded AnandTech on 6+ years ago.

So today we bring you quite a few new things, some may surprise you, some may not. ATI has released their Fall refresh product – the Radeon 9800XT and they are announcing their Radeon 9600XT. NVIDIA has counterattacked by letting us publish benchmarks from their forthcoming NV38 GPU (the successor to the NV35 based GeForce FX 5900 Ultra). But quite possibly more important than any of those announcements is the suite of benchmarks we’re testing these cards in; how does a total of 15 popular games sound? This is the first installment of a multipart series that will help you decide what video card is best for you, and hopefully it will do a better job than we have ever in the past.

The extensive benchmarking we’ve undertaken has forced us to split this into multiple parts, so expect to see more coverage on higher resolutions, image quality, anti-aliasing, CPU scaling and budget card comparisons in the coming weeks. We’re working feverishly to bring it all to you as soon as possible and I’m sure there’s some sort of proverb about patience that I should be reciting from memory to end this sentence but I’ll leave it at that.

Now that the long-winded introduction is done with, let’s talk hardware before we dive into a whole lot of software.

The Newcomers
Comments Locked

263 Comments

View All Comments

  • WooDaddy - Wednesday, October 1, 2003 - link

    Good eye #93. Evan actually mentioned it earlier too.

    Anand/Derek mentioned Nvidia being better at Doom 3. Y'all sneaky son-of-a-guns must be beta testing it in the background or sumthin. I know Carmack said it ran better but I betcha y'all got your hands on a copy. Go ahead. Admit it! Quit holding out on us. We wanna see the benchmark! I got a shiny nickel with your name on it if you put it out there...

    Overall great review. I sorta agree that 1024x768 is kinda like the 640x480 of yesteryear now, but most of us can gather what 1280 will run at. For the fanboys/girls, "You should've included counterstrike and hexen 2. waah!" Honestly, I know how long it can take to set up and benchmark those tests in a _controlled_ environment. Do you guys use automated software testers?

    Question though.Even though FFXI ran slow, is it still playable? I don't want to believe that it runs that slow all the time.
  • Anonymous User - Wednesday, October 1, 2003 - link

    k, I'm going to ignore everyone who's bitching because they didn't read it and thus haven't already twigged that IQ comparisons will be in part 2

    Re the PCI slot thing, doesn't that apply equally to two-slot cards? If putting a PCI card next to the AGP slot on a one-slot card is bad, surely putting a PCI card in the first slot after a two-slot card isn't exactly smart either? You still lose an extra PCI slot over what you would have with a one-slot card
  • Anonymous User - Wednesday, October 1, 2003 - link

    What about Max Payne 2 ? i like to see it in next benchs
  • Anonymous User - Wednesday, October 1, 2003 - link

    Please use Battlefield 1942 in benchmarking in the future! It's an awesome game and has some very nice and demanding mods like desert combat. Please use desert combat in benchmarking too. Try flying around, blowing up stuff and checking if framerate ever goes to unacceptable levels. Gamers rarely care about average or maximum fps, if game is running 50fps or 150 fps it doesn't matter, but if it ever runs as sluggishly as <10 fps in heat of the battle, it is very annoying.

    Just tell us with your own words, which graphics card brings playable framerates!
  • Anonymous User - Wednesday, October 1, 2003 - link

    I would like to see Battlefield 1942 added into the benchmarks. Especially since it is such a popular game and they have battlefield vietnam coming out before too long. Thanks.
  • Anonymous User - Wednesday, October 1, 2003 - link

    Thanks for benching so many games. Since I only play a few games I look for performance in those games in particular. My games were covererd and I really appreciate that.
  • Anonymous User - Wednesday, October 1, 2003 - link

    I see the fanbois are out in full force. :/
  • Anonymous User - Wednesday, October 1, 2003 - link

    over all a sad review, using drivers that are not out for every one to use, no IQ tests to see if the drivers are cheating at all, and then comments like "From these two graphs, it seems like NVIDIA is the clear winner, but in watching this demo run so many times, we noticed that the NVIDIA cards were running choppier than the ATI cards, and we again had some image quality questions we need to answer"

    so that pretty much does it for me. I won't take this with a grain of salt untill they rip apart the drivers, and make sure Nvidia is not up to any "optimazations" Ive lost all trust in Nvidia. I hope the Nv40 can turn this around.

  • Anonymous User - Wednesday, October 1, 2003 - link

    #69 just because you run games at 1280x1024, doesn't make you the majority representation of gamers. Most gamers run at 1024x768. Most computer users resolution is at 1024x768, like 55% or something like that.
  • Anonymous User - Wednesday, October 1, 2003 - link

    The "Prescott" string on Page 4 is white?! Just select the 3 last lines from "2.8 GHz ...". Has it been white all the time?

Log in

Don't have an account? Sign up now