Overclocking

NVIDIA provides software, nTune, which allows users to increase the clock speed of the memory and the GPU. This functionality used to be included in their driver but it's no significant inconvenience to have it provided as a separate download. We're still pleased with NVIDIA's support of user overclocking in general. For the 8800 series, we're still waiting on support for shader overclocking, as NVIDIA has made promises but has yet to deliver.

We were able to use nTune to find optimal core and memory clock speeds but in order to perform shader overclocking we had to rely on a tool called NiBiTor, an NVIDIA BIOS editor maintained by mvktech.net. Combined with a flash utility, we were able to download the BIOS from our video card, modify the shader clock speed, and flash it with our new settings. While we can't be sure that NVIDIA's driver doesn't alter this in any way while the card is running, if you buy a card with a higher than stock shader clock speed, this is what you'll get.

The first thing we did was find a maximum stable memory clock speed. We did this by increasing memory speed and running the fuzzy cube test within ATI Tool for about 10 minutes. Once we found the highest stable memory clock speed, we used this for all our tests in order to reduce the impact of memory on overall game performance. This also serves to reduce the number of combinations of different clock speeds we would have to test.

Maximum core and shader clock speeds were found independently and tested for stability using ATI Tool. The highest stable core clock we were able to run on our OCZ hardware is 640 MHz, while we found that 1600 MHz was about as high as we could go on the shader side. We test core clock scaling with memory fixed at 1020 MHz (2040 MHz effective data rate), and shader clock fixed at 1350 MHz (stock 8800 GTX speed). Our shader clock scaling tests are performed with the same memory speed at two different core clock speeds: 575 MHz, and 640 MHz.

To put this in perspective, here's how our OCZ overclocking compares to the other cards we've already tested.

Card Core Overclock Memory Overclock
ASUS GeForce EN8800 GTX 629MHz 1021MHz
BFG GeForce 8800 GTX 649MHz 973MHz
EVGA e-GeForce 8800 GTX w/ ACS3 659MHz 1013MHz
Leadtek Winfast GeForce 8800 GTX 627MHz 1033MHz
MSI GeForce NX8800 GTX 652MHz 1040MHz
OCZ GeForce 8800 GTX 640MHz 1020MHz
Sparkle Calibre 8800 GTX 631MHz 914MHz
Sparkle GeForce 8800 GTX 629MHz 1011MHz
XFX GeForce 8800 GTS 654MHz 866MHz


While OCZ doesn't have the highest showing in either core or memory clock speed department, they're position is quite high on both lists. The combination of a respectable memory and core clock speed is a good thing to hope for when buying a video card to overclock.

It is important to note here that every single card will overclock differently, even if the company has binned their parts to achieve high clock speeds. Your mileage may, of course, vary.

The OCZ GeForce 8800 GTX and The Test GeForce 8800 GTX Core Clock Scaling
Comments Locked

12 Comments

View All Comments

  • bigtoe36 - Wednesday, February 28, 2007 - link

    Derek

    I think you have the overclocking results a little wrong in the article.

    here is what I get your results to be. I removed the GTS card.

    Asus 621/1026
    BFG 648/972
    EVGA 648/1008
    Leadtech 621/1026
    MSI 648/1044
    OCZ 648/1026
    Sparkle C 621/918
    Sparkle 621/1008

    Shader domain will have been the major factor here holding some cards back, also did you try manipulating the PCIE bus as this will usually allow the cards to clock higher?

    Overall i get the MSI top and OCZ card second from your testing which is not at all bad.


  • Beenthere - Saturday, February 17, 2007 - link

    OCZ's PSUs and other products haven't been top shelf like their memory original was, so I'd take any claims with a large grain of salt.

    Many PC hardware companies have discovered how technically challenged most consumers are and thus have decided to profit from this. OCZ and other companies know that most hardware review sites will shill for them just to get free hardware to review. Few sites have the integrity to write an honest review lambasting a company for a crappy product. Instead the reviewer generally notes the positives and ignores the blemishes or outright malfuntions. When consumers purchase the product and find it's a piece of garbage, that's when the shit hits the fan.
  • DerekWilson - Monday, February 19, 2007 - link

    Personally, I really like OCZ PSUs ... I use them in my personal and test systems.

    There really was no reason to tear apart OCZ for their 8800 GTX. While there is nothing incredibly noteworthy about the part, there weren't any problems. While we can't say all OCZ 8800 GTX cards will overclock as well as the one we had, we were able to acheive higher clock speeds than a few others we've tested.

    We will certaily tear into hardware that deserves it.
  • lopri - Friday, February 16, 2007 - link

    I feel the pain of reviewing vendor cards. All of them essentially being reference cards, there is only as much a reviewer can explore. I'd have like some more comments regarding warranty and customer service experience, but the card being supplied by OCZ that would have also been somewhat pointless.

    However, the good part of the review where core clock / shader clock scaling was discussed in-dept is very informative. I've been aware of this behavior myself via RivaTuner monitoring and AT's review confirms it. For anyone hasn't understood this before, this information should be taken very seriously if s/he intends to overclock 8800 series cards. Basically, most (if not all) GTX are capable of 600+MHz core clock, and the performance jump occurs at 620~621MHz and 636~637MHz. This means an overclok of 630MHz will net you the same FPS as that of 620MHz.

    Most interesting bit for me was that after the last performance jump @649MHz(?), no ammount of overclocking will bring extra performance gain. I didn't know about this and thank Derek for this finding. Again, this is very interesting. I would like to know why this is the case, as well as the exact formula NV utilizes for this clock speed jump. Maybe in the next review? :)
  • adholmes - Friday, February 16, 2007 - link

    The idea that all of these cards are reference designs got me thinking. I wonder if the review was done with 5 different cards from the same manufacturer, would the results would be the same? Are the differences we observe due to vendor implementation, memory selection, etc., or just natural manufacturing variances?
  • DerekWilson - Friday, February 16, 2007 - link

    we'd love to get nvidia's formula for clocking their cards, but I don't think they'll give us the timing diagram :-)

    And I wouldn't say that there would be *no* ammount of overclocking that would bring more performance -- we just don't have a card that will go high enough ...

    similar to previous cards, it looks like the gap widens as clock speed increases -- in other words, the higher you go, the more you need to increase clock speed to hit the next performance bump.

    We might see something in the mid 670s ...

    I think aftermarket cooling would help -- which I might try. That would certainly make it interesting -- if water or phase change cooling could get you enough clock head room to reach the next plateau.
  • lopri - Friday, February 16, 2007 - link

    The reason why I thought there was some kind of 'ceiling' is that the touted rumor before the G80's debut. If I remember correctly, there was something along the line of :

    - Up to 1.5GHz Shader Core Clock

    Since the shader clock rises with core clock, if the shader clock's theoretical limit is 1.5GHz by design - again, this is a guesswork on my part - wouldn't there be some sort of ceiling?

    Anyway, I think NV should disclose this mystery regarding the whole clock domain thingy. :D
  • abhaxus - Friday, February 16, 2007 - link

    But 640mhz was 5th out of 9 in your chart, and 1020mhz was 4th out of 9. I just hate it when statistics are misused. The numbers LOOK high but in fact were close to median results in your small sample. In fact the core clock was below average... Nitpicky I know, but maybe it's flattering for you guys to know people read your articles closely enough to notice crap like that? I hope so.

    I liked the review and was very interested in the scaling of the GTX (although I am now leaning toward a 320mb GTS for my 1680x1050 LCD).
  • DerekWilson - Friday, February 16, 2007 - link

    640 MHz is actually 4th ... you're including the XFX 8800 GTS overclocked to 654 -- which shouldn't count in a comparison of 8800 GTX cards.

    1020 MHz more or less == 1021 MHz ... in fact, I'm absolutely positive that the OCZ 8800 GTX would run stable at 1021 -- I just didn't bother to bump the clock speed up that little bit. I suppose it's a slight difference in method between the way Josh and I determine clock speed. That extra MHz isn't going to do anything for anyone ... (it's cool to know that something happening an extra million times per second isn't going to have any impact on the outcome isn't it).

    Anyway, I do appreciate that you guys keep up with that sort of thing -- I know I did back when I was just reading the site. :-)

    If you're still interested in GTX scaling, please stay tuned -- I didn't originally include Vista because NVIDIA's driver didn't properly support overclocking, but with the BIOS editor I can just change the clock speeds at a low level. Combine that with the fact that NVIDIA just released a DX10 demo that uses all kinds of neat shader centered features and I'm really excited about doing some clock scaling with DX10 code.

    http://www.nzone.com/object/nzone_cascades_home.ht...">http://www.nzone.com/object/nzone_cascades_home.ht...

    Wish I'd had this before I finished testing for this article, but I think I can definitely find some other things to talk about with DX10, Vista, and overclocking.
  • abhaxus - Saturday, February 17, 2007 - link

    quote:

    (it's cool to know that something happening an extra million times per second isn't going to have any impact on the outcome isn't it).


    putting it that way... yes, it is actually very cool. At least cool in the way we nerds think about things.

    When you do that scaling with DX10 I would really really really love you guys if you ran some stuff on the GTS 320mb to see if there's a big disadvantage with the lower memory with dx10 shaders. As I said, I have a 22" WS LCD and so far it seems like the 320mb version is more than enough for what I do in current games, would be nice to know if it does well in the future also.

Log in

Don't have an account? Sign up now