The Elder Scrolls IV: Oblivion Performance

Version: 1.2.0416 Shivering Isles

Settings: Ultra High Quality settings defaults with vsync disabled

Our Oblivion test takes place in the south of the Shivering Isles, running through woods over rolling hills toward a lake. This is a straight line run that lasts around 20 seconds and uses FRAPS to record framerate. This benchmark is very repeatable, but the first run is the most consistent between cards, so we only run the benchmark once through and take that number.

Under Oblivion, AMD still shows some promise. The three way 3870 solution leads the way without AA enabled. At the same time, the 9800 GX2 clearly has the single card solution performance locked up. At not that much more money required for the 9800 GX2 over the three way 3870 solution, it’s good to see AMD being competitive in at least one benchmark.

Oblivion Performance


With AA enabled, the gap closes a bit. The three way AMD solution still has an advantage at the high end where it counts though. The very high end solutions are playable at 2560x1600 with everything turned up and AA/AF enabled as well.

Oblivion Performance


Crysis Performance Enemy Territory: Quake Wars Performance
Comments Locked

50 Comments

View All Comments

  • iceveiled - Tuesday, March 18, 2008 - link

    Probably the reason why a dual 8800 GT wasn't tested is because the mobo in the setup doesn't support SLi (it's an intel mobo)....
  • chizow - Tuesday, March 18, 2008 - link

    Its Skull Trail which does support SLI (they actually mention a GX2 SLI, ie Quad SLI, review upcoming). More likely NV put an embargo or warning on direct 8800GT/GTS comparisons so the spotlight didn't shift to artificial clock speed and driver discrepancies. After all, they do want to sell these abominations. ;)
  • madgonad - Tuesday, March 18, 2008 - link

    You clearly aren't paying attention to the market. A lot of people who cling to their PC gaming experience would also like to move their PC into the living room so that they can experience the big screen + 5/7.1 surround sound like their console brethren. The new Hybrid power and graphics solutions will allow a HTPC to have one of these Beasts as a partner for the onboard graphics. When watching movies or viewing the internet, this beast will be off and not making heat or noise. But once Crysis comes on, so does the discrete video card and it is off to the races. I have been waiting for the market to mature so that I can build a PC that games well, holds all my movies, and TiVos my shows - all in one box. All that I am waiting for is a Bitstream solution for the HD audio - which are due in Q2.
  • JarredWalton - Tuesday, March 18, 2008 - link

    That's true... but the HybridPower + SLI stuff isn't out yet, right? We need 790g or some such first. I also seem to recall NVIDIA saying that HybridPower would only work with *future* NVIDIA IGPs, not with current stuff. So until we have the necessary chipset, GPU, and drivers I for one would not even think of putting a 9800 GX2 into an HTPC. We also need better HDMI audio solutions.

    Anyway, we're not writing off HTPC... we're just saying that more the vast majority of HTPC users this isn't going to be the ideal GPU. As such, we focused on getting the gaming testing done for this article, and we can talk about the video aspects in a future article. Then again, there's not much to say: this card handles H.264 offload as well as the other G92 solutions, which is good enough for most HTPC users. Still need better HDMI audio, though.
  • casanova99 - Tuesday, March 18, 2008 - link

    While this is most likely a G92 variant, this isn't really akin to an SLI 8800GT setup, as the 8800GT has 112 shaders and 56 texture units. This card has 256 (128 * 2) shaders and 128 (64 * 2) texture units.

    It seems to match more with a 8800GTS 512MB, but with an underclocked core and shaders, paired with faster memory.
  • chizow - Tuesday, March 18, 2008 - link

    While this is true, it only perpetuates the performance myths Nvidia propagates with its misleading product differentiation. As has been shown time and time again, the differences in shaders/texture units with G92 have much less impact on performance compared to core clock and memory speeds. There's lots of relevant reviews with same-clocked 8800GT vs GTS performing nearly identically (FiringSquad has excellent comparisons), but you really need to look no further than the 9600GT to see how overstated specs like shaders are with current gen GPUs. If you dig enough you'll find the info you're looking for, like an 8800GT vs 8800GTS both at 650/1000 (shocker, 9800GTX is expected to weigh in at 675/1100). Problem is most reviews will take the artificial stock clock speeds of both and compare them, so 600/900 vs 650/1000 and then point to irrelevant core differences as the reason for the performance gap.
  • hooflung - Tuesday, March 18, 2008 - link

    Well, I am really disapointed in this review. It seems almost geared towards being a marketing tool for Nvidia. So it might be geared towards HD resolutions, what about the others resolutions? If AMD is competitive at 1680x1050 and 1900x1200 for ~200+ dollars less would the conclusion have been less favorable and start to nitpick the sheer awkwardness of this card? Also, I find it disturbing that 9600GTs can do nearly what this thing can do, probably at less power ( who knows you didn't review power consumption like every other card revier ) and cost half as much.

    To me, Nvidia is grasping at straws.
  • Genx87 - Tuesday, March 18, 2008 - link

    Eh? Grasping at straws with a solution that at times is clearly faster than the competition? This is a pretty solid single card offering if you ask me. Is it for everybody? Not at all. High end uber cards never are. But it definately took the crown back from AMD with authority.

  • hooflung - Tuesday, March 18, 2008 - link

    A single slot solution that isn't much better than a SLI 9600GT setup at those highest of high resolutions. Not for everyone is the understatement of the year. Yes I can see it is the single fastest on the block but at what cost? Another drop in the hat of a mislabeled 8 series product that is in a package about as illustrious as a cold sore.

    This card is a road bump. The article is written based on a conclusion that Nvidia cannot afford, and we cannot afford, to have a next generation graphics processor right now. To me, it smacks of laziness. Furthermore, gone are the times of 600 dollar graphics cards I am afraid. I guess Nvidia employees get company gas cards so the don't pay 3.39 a gallon for gasoline.

    How does this card flare for THE MAJORITY of users on 22" and 24" LCDs. I don't care about gaps at resolutions I need a 30" or HDTV to play on.
  • Genx87 - Tuesday, March 18, 2008 - link

    Sounds like you have a case of sour grapes. Dont get so hung up on AMD's failings. I know you guys wanted the x2 to trounce or remain competitive with this "bump". But you have to remember AMD has to undo years of mismanagement at the hands of ATI's management.

    600 dollar cards keep showing up because they sell. Nobody is forcing you buy one.

Log in

Don't have an account? Sign up now