Just One Small Problem: We Need a New Memory Technology

The R600 GPU had an incredibly wide 512-bit memory interface, the problem with such a large interface is that it artificially makes your die bigger as you’ve got to route those interface pads to the memory devices on the board. For RV770 to have the die size ATI wanted, it needed to have a 256-bit memory interface, but using (at the time) current memory technology that wouldn’t give the GPU enough memory bandwidth to hit the performance targets ATI wanted.

When the options were either make the chip too big or make the performance too low, ATI looked elsewhere: let’s use a new memory technology. Again, put yourself in ATI’s shoes, the time was 2005 and ATI had just decided to completely throw away the past few years of how-to-win-the-GPU-race and on top of that, even if the strategy were to succeed it would depend on a memory technology that hadn't even been prototyped yet.

The spec wasn’t finalized for GDDR5 at the time, there were no test devices, no interface design, nothing. Just an idea that at some point, there would be memory that could offer twice the bandwidth per pin of GDDR3, which would give ATI the bandwidth of a 512-bit bus, but with a physical 256-bit bus. It’s exactly what ATI needed, so it’s exactly what ATI decided to go with.

Unfortunately whether or not GDDR5 shipped by the summer of 2008 wasn’t all up to ATI, the memory manufacturers themselves had a lot of work to do. ATI committed a lot of resources both monetarily and engineering to working with its memory partners to make sure that not only was the spec ready, but that memory was ready, performing well and available by the summer of 2008. Note that the RV770 was going to be the only GPU that would use GDDR5, meaning that it was ATI and ATI alone driving the accelerated roadmap for this memory technology. It’s akin to you trying to single handedly bring 100Mbps internet to your city; it’ll happen eventually, but if you want it done on your time table you’re going to have to pickup a shovel and start burying a lot of your own cable.

ATI did much of the heavy lifting with the move to GDDR5, and it was risky because even if RV770 worked out perfectly but the memory wasn’t ready in time the GPU would get delayed. RV770 was married to GDDR5 memory, there was no other option, if in three years GDDR5 didn’t ship or had problems, then ATI would not only have no high end GPU, but it would have no performance GPU to sell into the market.

If GDDR5 did work out, then it meant that RV770 could succeed and that it would be another thing that NVIDIA didn’t have at launch. That is, of course, assuming that ATI’s smaller-die strategy would actually work...

The Power Paradigm Dave Baumann Saves the Radeon HD 4850
Comments Locked

116 Comments

View All Comments

  • nezuko - Thursday, December 4, 2008 - link

    I think those phrase is describe what Graphic-field is. Another year win, and another year lose. But from those situations, only hardworking and tough guy would be able to turn all upside down. And ATi team do make it. Now I relieved I make a decision to buy 4670, though not performance, it still does big bang for the buck. And with those Catalyst 8.12, I would be more grateful that I bought this video card. Has been downloaded it and now testing it.

    Would Anand make another article about those GP-GPU programming language to make a data parralel computing possible.

    Well, I considering to build my Leo Platform in the H2 of 2009 when the AM3 Deneb is out, Sata 3, and RD890.
  • JimiP - Thursday, December 4, 2008 - link

    Like many before me have said, this has to be one of the best articles I've ever read here at AT. It really puts things into perspective. We (the consumer) are always criticizing or praising everything that comes out and don't take into account the amount of hard work and time put into the release. I'm 4850 owner, and I couldn't be happier with the performance I've received. I would like to personally thank ATI/AMD and the entire team that put RV770 into play. Absolutely brilliant.

    I would also like to thank Anand for sharing this awesome experience with us.
  • zshift - Thursday, December 4, 2008 - link

    I have to say this was a great article. Great idea to write about the story behind these guys and the rv770. musta been a helluva relief when they realized how great the gpus were in the market, especially after taking such huge risks. For these guys to pull through the way they did, with the whole gddr5 issue and the die-shrink/physical limitations is amazing. I thought I was stressed in college. I can't imagine what its like to design something like this for 3 years not being even sure it'll work in the end. That's one hell of a resolve, makes me like ATI a bit more than I already do.

    Keep writing great articles here, this is my favorite site to read reviews on, and this is another reason why.

    go anand! :p
  • strikeback03 - Thursday, December 4, 2008 - link

    I agree with everyone else that the article is very well written. I am not sure if these would even be the right guys to ask, but did you bring up any of the driver issues your other recent articles have mentioned with them? As you have mentioned before, it is probably not the best business plan to assume nVidia will screw up again, and they should probably get their crossfire support in order for the good feelings about this strategy to continue.
  • Dyno1979 - Thursday, December 4, 2008 - link

    Definitely one of the best articles I've read lately. And I didn't even notice that typo, probably because I was reading the article instead of looking at it.

    5 stars
  • CarrellK - Thursday, December 4, 2008 - link

    The "sweet spot" strategy would have amounted to *nothing* without the efforts of many very talented engineers (and a little luck as Anand has noted). They made the 770 happen and deserve the lion's share of the credit.

    I didn't think Anand would use this for anything other than background here-and-there in future articles. I fully expected him to politely cut me off at some point and say "about those future architectures..." which would have lead to Eric, Mike, and Mark telling a different interesting story. Thanks to Anand et al for telling this part of the 770 story. Responding to a comment or two in the posts:

    * Sorry to quench the speculation - the AMD purchase had no effect on the 770's execution. Dirk Meyer and the other AMD executives supported Rick, a guy that they really didn't know, during some pretty tough times at AMD. They did their jobs so that we could do ours.

    * The price range for 770-based cards was determined back in 2005 - it was an essential factor limiting the GPU cost, one of the big gambles. We had no clue what nV's 2008 pricing would be, but we did know what the gamers wanted. At launch we were tempted oh so briefly to launch at a higher price given the competitor's product offerings. It took some will-power for the starving man (us) to pass up a banquet (profits). We had a sneaking suspicion there was a lot of unhappiness about the direction prices had gone, and didn't want to be a party to that for the sake of a few weeks better revenue. Greed never pays. Remembering your customers does.

    P.S. We don't keep any dart-board pictures of Anand around the office. However I *do* recall seeing his picture somewhere and thinking at the time that it *would* make a good dart target. Just a thought... :-)

  • lyeoh - Sunday, December 7, 2008 - link

    You guys got the sweet spot right as far as I'm concerned (I'm not sure if it's true for others - does it show up in the units sold?)

    Before the ATI 3800 (RV670), and Nvidia 8800GT, it seemed like after shelling out a few hundred US dollars, you'd only get low/medium quality at current games. And cheaper cards were pathetic to unusable for new games.

    So I stuck to playing old games with on my old video card (Ti4200) - which was decent in its time.

    After the beginning of the new "sweet spot" era, this year I bought a 9800GT (and a new PC). While the 9800GT is not as good as AMD/ATI's offerings in hardware performance terms, I was concerned about ATI's drivers/software. A colleague tried an ATI card on his office PC, but in the end he had to switch to Nvidia to get his multiscreen set up on Linux working the way he wanted, and I had seen a fair number of complaints from others. So far Nvidia's drivers have been OK for me whether in Windows or Linux.

    On the other hand I've seen too many Nvidia cards failing in hardware terms (bad caps, bad whatever). So pick your poison ;).

    But if the cards aren't totally crap, it often takes less time to just replace a faulty card, than to keep tinkering with drivers and software configs (sometimes to no avail).

    Anyway, many thanks for helping to make stuff affordable, even though I picked Nvidia again ;).

    In the end I'm still back to mostly playing old games though...
  • MrSpadge - Saturday, December 6, 2008 - link

    Thanks Andantech, ATI & AMD for this amazing article!

    And I'd like to add a point which has not been raised yet, at least in this discussion: the "small and fast enough" strategy only works because GPUs hit the realm where they're power limited!

    The point is, whenever you go multi-GPU you loose performance due to inefficiencies and communication delays and there are also some transistors lost to redundant logic. If you had the choice between one 100 Mio transistor chip or 2 50 Mio ones, then the 100 Mio one would certainly be faster; assuming both could run at the same clock speed, which previously was determined by chip design (basically identical in the example) and process (identical).

    But GT200 is too big, it can not fully fledge its clock speed wings because its power limited. Imagine GT200 at 1.5 - 1.8 GHz shader clock - it would be much more in line with performance expectations. RV770 on the other hand can be pushed quite a bit and on the 4870 it chews up lots of power for such a small chip - but that's OK because this power envelope has been accepted and the performance is there to justify it. And the 2 GPU versions are succesful because the power envelope on such "freak"-cards is larger.

    And another frequently overlooked aspect: not all of GT200s transistors contribute to game performance. The 30 shaders which are 64 bit capable must be large and don't help games at all (and probably won't for quite some time). This is a very forward looking feature for games and a feature of immediate benefit for GP-GPU.

    MrS
  • Frallan - Friday, December 5, 2008 - link

    Thank you m8!

    Not only for delivering good products but also for delivering good information and entertainment.

    Please convey to the other "Fellows" the heartfelt thanks of this community.
  • JimmiG - Thursday, December 4, 2008 - link

    Congratulations to Anandtech for one of the most interesting articles this year. Congratulations to ATI/AMD for putting out their best and most exciting product since R300/9700 Pro.

    The industry really needed something like RV770. When the 9700 Pro came out in 2002, it was at the cutting edge of technology and performance, far ahead of the previous champion, the Ti4600, yet it launched at only $399. Nvidia launched the 8800 Ultra and GTX280 at $800 and $600 respectively, even though neither GPU introduced any significant new features, only moderately higher framerates.

    I currently have a 4850 512MB which I bought in July and I love it... It runs all my favorite games at great framerates and with fantastic image quality at 1680x1050. Still, I wouldn't considering myself an "ATI fan". When it's time for me to upgrade again, I will buy the best card in the $200 range and won't care whether the sticker on the GPU fan is green or red.

Log in

Don't have an account? Sign up now