DirectX 7 Performance

Below you can see our DirectX 7 based video processor chart:

GF2 GTS 200 333 4 2 0.5 128 1600 25 5081 100.0% 100.0% 100.0% 100.0%
DirectX 7
7500 290 460 2 3 0.5 128 1740 36 7019 108.8% 138.1% 145.0% 130.6%
GF4 MX460 300 550 2 2 0.5 128 1200 38 8392 75.0% 165.2% 150.0% 130.1%
GF2 Ultra 250 460 4 2 0.5 128 2000 31 7019 125.0% 138.1% 125.0% 129.4%
GF2 Ti 250 400 4 2 0.5 128 2000 31 6104 125.0% 120.1% 125.0% 123.4%
GF4 MX440 8X 275 500 2 2 0.5 128 1100 34 7629 68.8% 150.2% 137.5% 118.8%
7500 LE 250 360 2 3 0.5 128 1500 31 5493 93.8% 108.1% 125.0% 109.0%
GF4 MX440 275 400 2 2 0.5 128 1100 34 6104 68.8% 120.1% 137.5% 108.8%
GF2 Pro 200 400 4 2 0.5 128 1600 25 6104 100.0% 120.1% 100.0% 106.7%
7500 AIW 250 333 2 3 0.5 128 1500 31 5081 93.8% 100.0% 125.0% 106.3%
GF2 GTS 200 333 4 2 0.5 128 1600 25 5081 100.0% 100.0% 100.0% 100.0%
GF4 MX440 SE 250 333 2 2 0.5 128 1000 31 5081 62.5% 100.0% 125.0% 95.8%
Radeon DDR 183 366 2 3 0.5 128 1098 23 5585 68.6% 109.9% 91.5% 90.0%
GF4 MX4000 275 400 2 2 0.5 64 1100 34 3052 68.8% 60.1% 137.5% 88.8%
GF4 MX420 250 333 2 2 0.5 64 1000 31 2541 62.5% 50.0% 125.0% 79.2%
Radeon LE 148 296 2 3 0.5 128 888 19 4517 55.5% 88.9% 74.0% 72.8%
GF2 MX400 200 166 2 2 0.5 128 800 25 2541 50.0% 49.8% 100.0% 66.6%
Radeon SDR 166 166 2 3 0.5 128 996 21 2533 62.3% 49.8% 83.0% 65.0%
7200 183 183 2 3 0.5 64 1098 23 1396 68.6% 27.5% 91.5% 62.5%
GF2 MX 175 166 2 2 0.5 128 700 22 2541 43.8% 49.8% 87.5% 60.4%
GeForce 256 DDR 120 300 4 1 0.5 128 480 15 4578 30.0% 90.1% 60.0% 60.0%
GF2 MX200 175 166 2 2 0.5 64 700 22 1266 43.8% 24.9% 87.5% 52.1%
GeForce 256 SDR 120 166 4 1 0.5 128 480 15 2533 30.0% 49.8% 60.0% 46.6%
7000 AGP^ 183 366 1 3 0 64 549 0 2792 34.3% 55.0% 0.0% 29.8%
7000 PCI^ 166 333 1 3 0 64 498 0 2541 31.1% 50.0% 0.0% 27.0%
Radeon VE^ 183 183 1 3 0 64 549 0 1396 34.3% 27.5% 0.0% 20.6%
* RAM clock is the effective clock speed, so 250 MHz DDR is listed as 500 MHz.
** Textures/Pipeline is the maximum number of texture lookups per pipeline.
*** Nvidia says their GFFX cards have a "vertex array", but in practice it generally functions as indicated.
**** Single-texturing fill rate = core speed * pixel pipelines
+ Multi-texturing fill rate = core speed * maximum textures per pipe * pixel pipelines
++ Vertex rates can vary by implementation. The listed values reflect the manufacturers' advertised rates.
+++ Bandwidth is expressed in actual MB/s, where 1 MB = 1024 KB = 1048576 Bytes.
++++ Relative performance is normalized to the GF2 GTS, but these values are at best a rough estimate.
^ Radeon 7000 and VE had their T&L Engine removed, and cannot perform fixed function vertex processing.

Now we're talkin' old school. There are those people in the world that simply can't stand the thought of having less than the latest and greatest hardware on the planet in their PC, and then there are people that have social lives. Okay, it's not that bad, but not everyone needs a super powerful graphics card. In fact, there are plenty of businesses running computers with integrated graphics that would be thoroughly outclassed be even the five year old GeForce 256. If you're only playing older 3D games or just want to get the cheapest non-integrated card you can find, DX7 cards fit the bill. A Home Theater PC that plays movies has no need for anything more, for instance. Or maybe you have a friend that's willing to just give you his old graphics card, and you want to know if it will be better than the piece of junk you already have? Whatever the case, here are the relative performance figures for the DX7 era cards.

No special weighting was used, although with this generation of hardware you might want to pay closer attention to memory bandwidth than the other areas. Fill rate is still important as well, but vertex fill rate is almost a non-issue. In fact, these cards don't even advertise vertex rates - they were measured in triangle rates. Since they had a fixed-function Transform and Lighting (T&L) pipeline, triangles/sec was the standard unit of measurement. The vertex pipelines are listed as "0.5" for the DX7 cards, emphasizing that they are not programmable geometry processors. As luck would have it, 0.5 times clock speed divided by 4 also matches the advertised triangle rates, at least on the NVIDIA cards. Vertex rates are anywhere from two to four times this value, depending on whether or not edges are shared, but again these rates are not achievable with any known benchmark. One item worth pointing out is that the Radeon 7000 and VE parts have had their vertex pipeline deactivated or removed, so they are not true DX7 parts, but they are included as they bear the Radeon name.

Early adopters of the DX7 cards were generally disappointed, as geometry levels in games tended to remain relatively low. First, there was a demo called "Dagoth Moor Zoological Gardens" created for the launch of the original GeForce 256. It was created by a company called "The Whole Experience" and used upwards of 100,000 polygons. Unfortunately, they never released any commercial games using the engine (at least, none that we're aware of). Later, a different company at the launch of the GeForce 2 created a demo that had millions of polygons to show off the "future of gaming" - that company would eventually release a game based off of their engine that you might have hear of, Far Cry. Actually, Crytek Studios demoed for both the original GeForce 2 launch and the GeForce 3 launch. They used the same engine and the demo name "X-isle" was the same as well, but the GF3 version added support for some pixel shader and vertex shader effects. Four years after demonstrating the future, it finally arrived! Really, though, it wasn't that bad. Many games are in development for several years now, so you can't blame them too much for delaying. Besides, launching a game that only runs with the newest hardware is tantamount to financial suicide.

As far as performance is concerned, the GeForce2 was the king of this class of hardware for a long time. After the GeForce 3, NVIDIA revisited DX7 cards with the GF4MX line, which added additional hardware support for antialiasing and hardware bump mapping. While it only had two pixel pipelines in comparison to the 4 of the GF2, the higher core and RAM speeds generally allowed the GF4MX cards to match the GF2 cards, and in certain cases they beat it. The Radeon 7500 was also a decent performer in this class, although it generally trailed the GF2 slightly due to the 2x3 pixel pipeline, which could really only perform three texture operations if two of them came from the same texture. Worthy of mention is the Nforce2 IGP chipset, which included the GF4MX 440 core in place of the normally anemic integrated graphics most motherboards offer. Performance was actually more like the GF4MX420, due to the sharing of memory bandwidth with the CPU and other devices, but it remains one of the fastest performing integrated solutions to this day. Many cards were also crippled by the use of SDR memory or 64-bit buses - we still see such things with modern cards as well, of course. Caveat emptor, as they say. If you have any interest in gaming, stay away from 64-bit buses, and these days even 128-bit buses are becoming insufficient.

Bring on the Crazy Eighty Eight! Is it smaller than a bread box?
Comments Locked

43 Comments

View All Comments

  • suryad - Monday, September 6, 2004 - link

    What about the mobility x800 graphics card? I didnt see that thrown into the mix?
  • coldpower27 - Monday, September 6, 2004 - link

    Thank you Bloodshredder, yeh after reading a little about the Radeon LE, it's almost as good as a Radeon DDR, except with lower working frequencies.

    so if it's DDR then the correct no. are 148/296 and 32MB VRAM only.
  • Bloodshedder - Monday, September 6, 2004 - link

    For the Radeon LE, I noticed a question mark next to the amount of RAM. I own one of these cards, and can confirm that 32MB DDR is the only configuration it comes in.
  • Draven31 - Monday, September 6, 2004 - link

    You skipped which OpenGL version and features the various cards support... maybe add that when you add the various workstation cards to the listings...
  • coldpower27 - Monday, September 6, 2004 - link


    Yeh, Nvidia learned it's lesson, last gen, with the 0.13 micron new at the time process delaying the introduction of the NV30, thy learned to play it safe using a tried and tested process is a good idea for such high complexity chips initially, though they of course plan to shift these chips to the 110nm process when the process matures enough, possibly on the NV48 and R480 hopefully allowing higher clocks in the process:D, maybe not for R480 unless low-k is ready for 110nm by that time.

    It does make more sense to use the newer manufacturing process to help save costs on the volume shipping GPU, as the cost savings will beaccumulated much better in the mainstream and value arena's thanks to sheer volume.

    We also see this with Intel, when Intel yields on the 90nm were only so so, they introduced Prescott up to 3.2GHZ in quanitity, but introduced their Pentium 4 3.4GHZ on the northwood core on 0.13 micron. Though over time Intel is making all efforts to transfer everything to 90nm, with Prescott and Prescott 2M w/1066FSB for EE Edition.
  • JarredWalton - Monday, September 6, 2004 - link

    8 - Intel does this as well, testing a new process on their non-flagship parts. For example, after the launch of the P4, Intel piloted their 130 nm copper technology with the Tualatin CPU before releasing the Northwood. It probably has something to do with the amount of extra time a more complex design takes to test and verify.
  • stephenbrooks - Monday, September 6, 2004 - link

    Interesting how on the die sizes chart, I notice they're phasing in the 110nm process only for their mid-range-ish cards and sticking to the tried and tested 130nm for the high-end one. I suppose you can't blame them for that really, given it's their flagship product and all, but it could contribute to the huge die sizes.
  • JarredWalton - Monday, September 6, 2004 - link

    Thank, AtaStrumf - any errors in the numbers are ColdPower's fault. Heheheh. Really, he already caught a bunch of small mistakes, so hopefully the number of remaining errors is very small.

    For what it's worth, there are various versions of some of the chips that have different clock speeds and RAM speeds from what is listed. The models in the chart should reflect the most common configurations, though.

    BTW, the article text is now tweaked somewhat on the ATI and NVIDIA overview pages. Derek Wilson provided some additional insight on the subject of AA and AF that clarified things a little.
  • JarredWalton - Monday, September 6, 2004 - link

    Argon was the name for the .25 micron K7, while Pluto and Orion were .18 micron.

    #2 and #4: I realize you're kidding, but in all seriousness we did think about including other architectures. With the broken features on some of the more recent cards and the lack of T&L on 3dfx and older cards, we just decided to stick with the two major players. And hey - it's all fair, as we didn't include Cyrix/Via or Transmeta processors in the CPU cheatsheet! ;)
  • AtaStrumf - Monday, September 6, 2004 - link

    OMFG, this is awsome!!!! You really outdid youself this time! I have been collecting data on GPUs for quite a while and have been planing on making a spreadsheet just like the first two for my, so called, web site, but WAU, this rocks. Thanks for saving me a lot of work :)

    When I get the time, I'll check your munbers a bit, just to make sure there aren't any typos in there.

Log in

Don't have an account? Sign up now