Intel’s Gen 6 Graphics

All 2nd generation Core series processors that fit into an LGA-1155 motherboard will have one of two GPUs integrated on-die: Intel’s HD Graphics 3000 or HD Graphics 2000. Intel’s upcoming Sandy Bridge E for LGA-2011 will not have an on-die GPU. All mobile 2nd generation Core series processors feature HD Graphics 3000.

The 3000 vs. 2000 comparison is pretty simple. The former has 12 cores or EUs as Intel likes to call them, while the latter only has 6. Clock speeds are the same although the higher end parts can turbo up to higher frequencies. Each EU is 128-bits wide, which makes a single EU sound a lot like a single Cayman SP.

Unlike Clarkdale, all versions of HD Graphics on Sandy Bridge support Turbo. Any TDP that is freed up by the CPU running at a lower frequency or having some of its cores shut off can be used by the GPU to turbo up. The default clock speed for both HD 2000 and 3000 on the desktop is 850MHz; however, the GPU can turbo up to 1100MHz in everything but the Core i7-2600/2600K. The top-end Sandy Bridge can run its GPU at up to 1350MHz.

Processor Intel HD Graphics EUs Quick Sync Graphics Clock Graphics Max Turbo
Intel Core i7-2600K 3000 12 Y 850MHz 1350MHz
Intel Core i7-2600 2000 6 Y 850MHz 1350MHz
Intel Core i5-2500K 3000 12 Y 850MHz 1100MHz
Intel Core i5-2500 2000 6 Y 850MHz 1100MHz
Intel Core i5-2400 2000 6 Y 850MHz 1100MHz
Intel Core i5-2300 2000 6 Y 850MHz 1100MHz
Intel Core i3-2120 2000 6 Y 850MHz 1100MHz
Intel Core i3-2100 2000 6 Y 850MHz 1100MHz
Intel Pentium G850 Intel HD Graphics 6 N 850MHz 1100MHz
Intel Pentium G840 Intel HD Graphics 6 N 850MHz 1100MHz
Intel Pentium G620 Intel HD Graphics 6 N 850MHz 1100MHz

Mobile is a bit different. The base GPU clock in all mobile SNB chips is 650MHz but the max turbo is higher at 1300MHz. The LV/ULV parts also have different max clocks, which we cover in the mobile article.

As I mentioned before, all mobile 2nd gen Core processors get the 12 EU version—Intel HD Graphics 3000. The desktop side is a bit more confusing. In desktop, the unlocked K-series SKUs get the 3000 GPU while everything else gets the 2000 GPU. That’s right: the SKUs most likely to be paired with discrete graphics are given the most powerful integrated graphics. Of course those users don’t pay any penalty for the beefier on-die GPU; when not in use the GPU is fully power gated.

Despite the odd perk for the K-series SKUs, Intel’s reasoning behind the GPU split does makes sense. The HD Graphics 2000 GPU is faster than any desktop integrated GPU on the market today, and it’s easy to add discrete graphics to a desktop system if the integrated GPU is insufficient. The 3000 is simply another feature to justify the small price adder for K-series buyers.

On the mobile side going entirely with 3000 is simply because of the quality of integrated or low-end graphics in mobile. You can’t easily add in a discrete card so Intel has to put its best foot forward to appease OEMs like Apple. I suspect the top-to-bottom use of HD Graphics 3000 in mobile is directly responsible for Apple using Sandy Bridge without a discrete GPU in its entry level notebooks in early 2011.

I’ve been careful to mention the use of HD Graphics 2000/3000 in 2nd generation Core series CPUs, as Intel will eventually bring Sandy Bridge down to the Pentium brand with the G800 and G600 series processors. These chips will feature a version of HD Graphics 2000 that Intel will simply call HD Graphics. Performance will be similar to the HD Graphics 2000 GPU, however it won’t feature Quick Sync.

Image Quality and Experience

Perhaps the best way to start this section is with a list. Between Jarred and I, these are the games we’ve tested with Intel’s on-die HD 3000 GPU:

Assassin’s Creed
Batman: Arkham Asylum
Borderlands
Battlefield: Bad Company 2
BioShock 2
Call of Duty: Black Ops
Call of Duty: Modern Warfare 2
Chronicles of Riddick: Dark Athena
Civilization V
Crysis: Warhead
Dawn of War II
DiRT 2
Dragon Age Origins
Elder Scrolls IV: Oblivion
Empire: Total War
Far Cry 2
Fallout 3
Fallout: New Vegas
FEAR 2: Project Origin
HAWX
HAWX 2
Left 4 Dead 2
Mafia II
Mass Effect 2
Metro 2033
STALKER: Call of Pripyat
Starcraft II
World of Warcraft

This is over two dozen titles, both old and new, that for the most part worked on Intel’s integrated graphics. Now for a GPU maker, this is nothing to be proud of, but given Intel’s track record with game compatibility this is a huge step forward.

We did of course run into some issues. Fallout 3 (but not New Vegas) requires a DLL hack to even run on Intel integrated graphics, and we saw some shadow rendering issues in Mafia II, but for the most part the titles—both old and new—worked.


Modern Warfare 2 in High Quality

Now the bad news. Despite huge performance gains and much improved compatibility, even the Intel HD Graphics 3000 requires that you run at fairly low detail settings to get playable frame rates in most of these games. There are a couple of exceptions but for the most part the rule of integrated graphics hasn’t changed: turn everything down before you start playing.


Modern Warfare 2 the way you have to run it on Intel HD Graphics 3000

This reality has been true for more than just Intel integrated graphics however. Even IGPs from AMD and NVIDIA had the same limitations, as well as the lowest end discrete cards on the market. The only advantage those solutions had over Intel in the past was performance.

Realistically we need at least another doubling of graphics performance before we can even begin to talk about playing games smoothly at higher quality settings. Interestingly enough, I’ve heard the performance of Intel’s HD Graphics 3000 is roughly equal to the GPU in the Xbox 360 at this point. It only took six years for Intel to get there. If Intel wants to contribute positively to PC gaming, we need to see continued doubling of processor graphics performance for at least the next couple generations. Unfortunately I’m worried that Ivy Bridge won’t bring another doubling as it only adds 4 EUs to the array.

Quick Sync: The Best Way to Transcode Intel HD Graphics 2000/3000 Performance
Comments Locked

283 Comments

View All Comments

  • saikyan - Thursday, January 6, 2011 - link

    "perhaps we should return to just labeling these things with their clock speeds and core counts? After all, it’s what Apple does—and that’s a company that still refuses to put more than one button on its mice. Maybe it’s worth a try."

    I hate to sound like the resident Mac fanboy (I'm platform agnostic) but I want to point out:

    1. Apple sells by trim and display, they don't really make a big deal of the CPU (probably because they stick to low-end and midrange CPUs)

    2. They have been shipping multi-button mice for nearly six years now. Come on!
  • vol7ron - Friday, January 7, 2011 - link

    - gtx460 image quality definitely the worst
    - 6870 image quality next
    - quicksync/snb image quality are the best (marginally better than 6870); I did notice some color loss in the flowers behind the umbrella when I zoomed in on the quicksync picture, so I'd have to give SNB the title in terms of quality. QuickSync gets the title in terms of performance.
  • nitrousoxide - Monday, January 10, 2011 - link

    lmao
  • Burticus - Monday, January 10, 2011 - link

    My last Intel cpu was a prescott 2.4ghz P4 OC'd to over 3ghz... back in 2004? My last 3 main system builds all AMD.... I was thinking about going to an X6 in the near future, now I guess maybe not. My price point is pretty much $200 for the cpu + motherboard so maybe I'll have to wait a couple months.

    Suddenly my 2 year old Phenom II seems very, very slow...
  • magnusr - Tuesday, January 11, 2011 - link

    I just received my 2600K. It only had the normal fan. No special heatsink/fan for the 2600K. The same heatsink as the rest....

    This is a fraud since I placed my decision to take 2600K instead of the 2500K based on the better heatsink and the cache.
  • mmcnally - Tuesday, January 11, 2011 - link

    Same here.. Very disapointed as I would have purchased a better heatsink if I knew. I guess I'll just do the install with the standard crap HS and hold off on over-clocking until I get a better one.
  • swing848 - Tuesday, January 11, 2011 - link

    Many of us are using older equipment. And, for those of us with limited funds it would have been nice if you would have added the Intel Q9650 and run all game benchmarks at 3.4GHz [ the speed of the 2600K], except for the default 3.6GHz speed of the X4 975BE, leave it there.

    I have a QX9650 that I purchased from eBay and it does 4GHz+ with ease, in a Gigabyte P35-DS3R motherboard, even with my ancient cooler [Thermalright XP-90] that I pulled from a socket 478 motherboard [$5 adapter].

    Note: I lapped the XP-90 with a slight convex shape to better use with un-lapped CPUs.

    In any event, a "quick and dirty" or simple overclock would have yielded at least some usable information. To save time, no need to try to get the maximum speed from all components.

    As long as the CPUs were already overclocked, you could run all benchmarks at those speeds, not just games. Many of us overclock to get more for our money.

    You included the ancient Q6600 at it's slow default speed - in some of the benchmarks. Why didn't you include it in all benchmarks?

    Your normal benchmark page does not include a full, or nearly full, list of games and CPUs, so, comparisons are difficult to find, example here anandtech.com/bench/CPU/62

    Where does this leave those of us with older equipment that is still chugging along?
  • Kell_sw - Thursday, January 13, 2011 - link

    DRM inside the cpu? People is blind?. The sad thing, everybody is going to buy this.
  • Sweeo - Friday, January 14, 2011 - link

    I just bought a upgrade "kit" from an core 2 2.8 quad to i7 950 :(
    but I got 6 sata ports I noticed the new boards have 4+2
    will the more advanced boards have more ?
  • Ahumado - Sunday, January 16, 2011 - link

    I didn't see it discussed. Did I miss it?

Log in

Don't have an account? Sign up now