The New Video Decode Pipeline: UVD

UPDATE: Since the launch of the HD 2900 XT, we've learned that all HD 2000 series parts except the high end R600 based parts will have UVD. This means that the HD 2900 XT will feature the same video decode acceleration offered on the R5xx hardware. As higher end cards are usually paired with faster CPUs, AMD feels that UVD on R600 is unnecessary.

This comes as a bit of a surprise to us and to certain board partners who's box art claims UVD as a feature of the HD 2900 XT. We do apologize for any confusion we may have caused at launch.

NVIDIA beat AMD to the punch with their full H.264 decode acceleration on G86/G84. Of course, AMD is one upping NVIDIA this time around, as their UVD (Unified Video Decode) architecture is also capable of decoding VLC bitstreams on all three HD media codecs. This means that there are no cases where AMD hardware will not handle 100% of the video decode process (after the CPU has dealt with pulling the encrypted content off the disk and preparing it to send to the GPU that is).

Here's a diagram of the landscape as it stands now. Notice that G80 is not capable of the bitstream decode or the inverse transformation (either iDCT or otherwise), but G84/G86 come very close to matching AMD's capabilities.

At the same time, we should remember that bitstream decode is only really heavy under H.264. Certainly VC-1 and MPEG-2 will see some benefit, but they are already fairly manageable. NVIDIA stated that building the hardware to handle VLC bitstreams wouldn't have a high enough return on investment. AMD, however, indicated that their bitstream processors are at least a little bit adaptable and it wasn't that difficult to include VLC decode.

Either way, the best way to figure out what's going on is to take a look at performance and see if there really is any advantage to R600 over G86. Unfortunately, try as we might, we could not get UVD to work with the current drivers provided by AMD and the PowerDVD release that is supposed to enable the hardware acceleration on HD 2000 series parts. We will have to take a second look at hardware decode when AMD and CyberLink or Intervideo get their software in order.

For now, our information leads us to believe that performance won't be hugely improved over G84/G86 in MPEG-2 and VC-1 CPU offloading. Where we might start to see a difference is in AMD's 65nm HD 2000 and mobility series parts. These have the potential to decrease power consumption by large amounts and provide quiet running systems for HTPCs, or longer battery life for notebooks. We will have to wait to get our hands on the higher volume R6xx based parts though. Also worth nothing is that AMD's high-end hardware does something that NVIDIA's 8800 series cards currently don't, so NVIDIA users that want fast H.264 decoding support are stuck with slower 3D performance.

Tesselation and the Future AMD CFAA Performance and Image Quality
Comments Locked

86 Comments

View All Comments

  • imaheadcase - Tuesday, May 15, 2007 - link

    quote:

    Bad performance with AA turned on (everybody turns on AA), huge power consumption, late to the market.


    Says who? Most people I know don't care to turn on AA since they visually can't see a difference. Only people who are picky about everything they see do normally, the majority of people don't notice "jaggies" since the brain fixes it for you when you play.
  • Roy2001 - Tuesday, May 15, 2007 - link

    Says who? Most people I know don't care to turn on AA since they visually can't see a difference.
    ------------------------------------------
    Wow, I never turn it of once I am used to have AA. I cannot play games anymore without AA.
  • Amuro - Tuesday, May 15, 2007 - link

    quote:

    the majority of people don't notice "jaggies" since the brain fixes it for you when you play.

    Says who? No one spent $400 on a video card would turn off AA.
  • SiliconDoc - Wednesday, July 8, 2009 - link

    Boy we'd sure love to hear those red fans claiming they turn off AA nowadays and it doesn't matter.
    LOL
    It's just amazing how thick it gets.
  • imaheadcase - Tuesday, May 15, 2007 - link

    quote:

    Says who? No one spent $400 on a video card would turn off AA.


    Sure they do, because its a small "tweak" with a performance hit. I say who spends $400 on a video card to remove "jaggies" when they are not noticeable in the first place to most people. Same reason most people don't go for SLI or Crossfire, because it really in the end offers nothing substantial for most people who play games.

    Some might like it, but they would not miss it if they stopped using it for some time. Its not like its make or break feature of a video card.
  • motiv8 - Tuesday, May 15, 2007 - link

    Depends on the game or player tbh.

    I play within ladders without AA turned on, but for games like oblivion I would use AA. Depends on your needs at the time.

Log in

Don't have an account? Sign up now