Crysis: Warhead

Kicking things off as always is Crysis: Warhead. It’s no longer the toughest game in our benchmark suite, but it’s still a technically complex game that has proven to be a very consistent benchmark. Thus even four years since the release of the original Crysis, “but can it run Crysis?” is still an important question, and the answer when it comes to setups using a pair of high-end 28nm GPUs is “you better damn well believe it.”

Crysis was a game that Kepler didn’t improve upon by a great deal compared to the Fermi based GTX 580. NVIDIA sees some good SLI scaling here, but AMD’s performance lead with a single GPU translates into an equally impressive lead with multiple GPUs; in spite of all of its capabilities the GTX 690 trails the 7970CF by 18% here. So long as AMD gets good Crossfire scaling here, there’s just no opening for Kepler to win, allowing AMD to handily trounce the GTX 690 here.

As for the intra-NVIDIA comparisons, the GTX 690 does well for itself here. Performance relative to the GTX 680 SLI at 2560 is 98%, which represents a 77% lead over the GTX 680. Overall performance is quite solid; at 55.7fps we’re nearly to 60fps on Enthusiast quality at 2560 with 4x MSAA, which is the holy grail for a video card. Even 5760 is over 60fps, albeit at lower quality settings and without AA.

It’s taken nearly 4 years, but we’re almost there; Crysis at maximum on a single video card.

Our minimum framerates are much the same story for NVIDIA. The GTX 690 once again just trails the GTX 680 SLI, while interestingly enough the dual-GPU NVIDIA solutions manage to erode AMD’s lead at a single point: 2560. Here they only trail by 8%, versus 20%+ at 5760 and 1920. Though at 1920 we also see another interesting outcome: the GTX 580 SLI beats the GTX 680 SLI and GTX 690 in minimum framerates. This would further support our theory that the GTX 680 is memory bandwidth starved in Crysis, especially at the lowest performance points.

GeForce Experience & The Test Metro 2033
Comments Locked

200 Comments

View All Comments

  • von Krupp - Saturday, May 5, 2012 - link

    Not precisely. That $350 performance point? It used to be a $200 performance point. Similarly, that $350 point will turn into a $400 performance point. So, assuming I maintain the price tier, graphics returns for my dollar are gradually tapering off. I look at the performance I was getting out of my 7800 GT at 1280x1024, and it wasn't worth upgrading to a newer card, period, because of Windows XP, my single core CPU, and the fact that I was already maxing out every game I had and still getting decent frame rates. I think they key factor is that I do not care if I dip below 60 frames, as long as I'm above 30 and getting reasonable frame times.

    I also know that consoles extend the life of PC hardware. The 7800GT is a 20-pipe version of the GTX, which is in turn the GPU found in the PS3.Devs have gotten much better at optimization in titles that matter to me.
  • CeriseCogburn - Saturday, May 5, 2012 - link

    You spend well over $1,600 on a decent system.
    It makes no sense to spend all that money, then buy monitors the cards in question cannot successfully drive on 3 year old Crysis game, let alone well over half the benchmarks in this article set without turning DOWN the settings.
    You cannot turn up DX11 tesselation, keep it on medium.
    You cannot turn up MSAA past 4X, and better keep it at 2X.
    You had better turn down your visual distance in game.
    That in fact, with "all the console ports" moanings "holding us back".
    I get it, the obvious problem is none of you seem to, because you want to moan and pretend spending $1,000.00 on a monitor alone, or more, is "how it's done", because you whine you cannot even afford $500 for a single video card.
    These cards successfully drive 1920X1080 monitors in the benchmarks, but just barely - and if you turn the eye candy up, they cannot do it.
  • CeriseCogburn - Saturday, May 5, 2012 - link

    Thanks for telling everyone how correct I am by doing a pure 100% troll attack after you and yours could not avoid the facts.
    Your mommy, if you knew who she was, must be very disappointed.
  • geok1ng - Sunday, May 6, 2012 - link

    This card was not build for 2560x1600 gaming. a single 680 is more than enough for that.
    The 690 was built for 5760x1200 gaming.

    I would like to see triple 30" tests. Nothing like gaming at 7680x1600 to feel that you are spending well your VGA money.
  • CeriseCogburn - Sunday, May 6, 2012 - link

    You can use cards 2 generations back for that, but like these cards, you will be turning down most and near all of the eye candy, and be stuck rweaking and clocking, and jittering and wishing you had more power.
    These cards cannot handle 1920X at current "console port" games unless you turn them down, and that goes ESPECIALLY for the AMD cards that suck at extreme tesselation and have more issues with anything above 4XAA, and often 4XAA.
    The 5770 is an eyefinity card and runs 5760X1200 too.
    I guess none of you will ever know until you try it, and it appears none of you have spent the money and become disappointed turning down the eye candy settings - so blabbering about resolutions is all you have left.
  • _vor_ - Tuesday, May 8, 2012 - link

    "... blabbering..."

    Pot, meet kettle.
  • CeriseCogburn - Sunday, May 6, 2012 - link

    They cost $400 to $2,000 plus, not $150 like the 242 1080p.
    Thanks for playing.
  • hechacker1 - Monday, May 7, 2012 - link

    Nope, you can already get IPS, 27", 2560x1440 panels (the same that Apple uses) for $400.

    They're rare, but currently they are building them in batches of 1000 to see how strong demand is for them.

    Sure the 120Hz will sort of go to waste due to the slow IPS switching speed, but it will accept that signal with 0 input lag.

    The only problem is that only the 680 seems to have a ramdac fast enough to do 120Hz. Radeon's tend to cap out at 85Hz.
  • marine73 - Monday, May 7, 2012 - link

    After checking Newegg it would seem that, unfortunately for Nvidia, this will be another piece of vaporware. Perhaps they should scale the Kepler's to 22nm and contract Intel to fab them since TSMC has major issues with 28nm. Just a thought.
  • marine73 - Monday, May 7, 2012 - link

    I guess I should retract my comments about TSMC as other customers are not experiencing supply issues with 28nm parts. Apparently the issues are with Nvidia's design, which may require another redo. I'm guessing AMD will be out with their 8000 series before Nvidia gets their act together. Sad because I have used several generations of Nvidia cards and was always happy with them.

Log in

Don't have an account? Sign up now