Intel’s Gen 6 Graphics

All 2nd generation Core series processors that fit into an LGA-1155 motherboard will have one of two GPUs integrated on-die: Intel’s HD Graphics 3000 or HD Graphics 2000. Intel’s upcoming Sandy Bridge E for LGA-2011 will not have an on-die GPU. All mobile 2nd generation Core series processors feature HD Graphics 3000.

The 3000 vs. 2000 comparison is pretty simple. The former has 12 cores or EUs as Intel likes to call them, while the latter only has 6. Clock speeds are the same although the higher end parts can turbo up to higher frequencies. Each EU is 128-bits wide, which makes a single EU sound a lot like a single Cayman SP.

Unlike Clarkdale, all versions of HD Graphics on Sandy Bridge support Turbo. Any TDP that is freed up by the CPU running at a lower frequency or having some of its cores shut off can be used by the GPU to turbo up. The default clock speed for both HD 2000 and 3000 on the desktop is 850MHz; however, the GPU can turbo up to 1100MHz in everything but the Core i7-2600/2600K. The top-end Sandy Bridge can run its GPU at up to 1350MHz.

Processor Intel HD Graphics EUs Quick Sync Graphics Clock Graphics Max Turbo
Intel Core i7-2600K 3000 12 Y 850MHz 1350MHz
Intel Core i7-2600 2000 6 Y 850MHz 1350MHz
Intel Core i5-2500K 3000 12 Y 850MHz 1100MHz
Intel Core i5-2500 2000 6 Y 850MHz 1100MHz
Intel Core i5-2400 2000 6 Y 850MHz 1100MHz
Intel Core i5-2300 2000 6 Y 850MHz 1100MHz
Intel Core i3-2120 2000 6 Y 850MHz 1100MHz
Intel Core i3-2100 2000 6 Y 850MHz 1100MHz
Intel Pentium G850 Intel HD Graphics 6 N 850MHz 1100MHz
Intel Pentium G840 Intel HD Graphics 6 N 850MHz 1100MHz
Intel Pentium G620 Intel HD Graphics 6 N 850MHz 1100MHz

Mobile is a bit different. The base GPU clock in all mobile SNB chips is 650MHz but the max turbo is higher at 1300MHz. The LV/ULV parts also have different max clocks, which we cover in the mobile article.

As I mentioned before, all mobile 2nd gen Core processors get the 12 EU version—Intel HD Graphics 3000. The desktop side is a bit more confusing. In desktop, the unlocked K-series SKUs get the 3000 GPU while everything else gets the 2000 GPU. That’s right: the SKUs most likely to be paired with discrete graphics are given the most powerful integrated graphics. Of course those users don’t pay any penalty for the beefier on-die GPU; when not in use the GPU is fully power gated.

Despite the odd perk for the K-series SKUs, Intel’s reasoning behind the GPU split does makes sense. The HD Graphics 2000 GPU is faster than any desktop integrated GPU on the market today, and it’s easy to add discrete graphics to a desktop system if the integrated GPU is insufficient. The 3000 is simply another feature to justify the small price adder for K-series buyers.

On the mobile side going entirely with 3000 is simply because of the quality of integrated or low-end graphics in mobile. You can’t easily add in a discrete card so Intel has to put its best foot forward to appease OEMs like Apple. I suspect the top-to-bottom use of HD Graphics 3000 in mobile is directly responsible for Apple using Sandy Bridge without a discrete GPU in its entry level notebooks in early 2011.

I’ve been careful to mention the use of HD Graphics 2000/3000 in 2nd generation Core series CPUs, as Intel will eventually bring Sandy Bridge down to the Pentium brand with the G800 and G600 series processors. These chips will feature a version of HD Graphics 2000 that Intel will simply call HD Graphics. Performance will be similar to the HD Graphics 2000 GPU, however it won’t feature Quick Sync.

Image Quality and Experience

Perhaps the best way to start this section is with a list. Between Jarred and I, these are the games we’ve tested with Intel’s on-die HD 3000 GPU:

Assassin’s Creed
Batman: Arkham Asylum
Borderlands
Battlefield: Bad Company 2
BioShock 2
Call of Duty: Black Ops
Call of Duty: Modern Warfare 2
Chronicles of Riddick: Dark Athena
Civilization V
Crysis: Warhead
Dawn of War II
DiRT 2
Dragon Age Origins
Elder Scrolls IV: Oblivion
Empire: Total War
Far Cry 2
Fallout 3
Fallout: New Vegas
FEAR 2: Project Origin
HAWX
HAWX 2
Left 4 Dead 2
Mafia II
Mass Effect 2
Metro 2033
STALKER: Call of Pripyat
Starcraft II
World of Warcraft

This is over two dozen titles, both old and new, that for the most part worked on Intel’s integrated graphics. Now for a GPU maker, this is nothing to be proud of, but given Intel’s track record with game compatibility this is a huge step forward.

We did of course run into some issues. Fallout 3 (but not New Vegas) requires a DLL hack to even run on Intel integrated graphics, and we saw some shadow rendering issues in Mafia II, but for the most part the titles—both old and new—worked.


Modern Warfare 2 in High Quality

Now the bad news. Despite huge performance gains and much improved compatibility, even the Intel HD Graphics 3000 requires that you run at fairly low detail settings to get playable frame rates in most of these games. There are a couple of exceptions but for the most part the rule of integrated graphics hasn’t changed: turn everything down before you start playing.


Modern Warfare 2 the way you have to run it on Intel HD Graphics 3000

This reality has been true for more than just Intel integrated graphics however. Even IGPs from AMD and NVIDIA had the same limitations, as well as the lowest end discrete cards on the market. The only advantage those solutions had over Intel in the past was performance.

Realistically we need at least another doubling of graphics performance before we can even begin to talk about playing games smoothly at higher quality settings. Interestingly enough, I’ve heard the performance of Intel’s HD Graphics 3000 is roughly equal to the GPU in the Xbox 360 at this point. It only took six years for Intel to get there. If Intel wants to contribute positively to PC gaming, we need to see continued doubling of processor graphics performance for at least the next couple generations. Unfortunately I’m worried that Ivy Bridge won’t bring another doubling as it only adds 4 EUs to the array.

Quick Sync: The Best Way to Transcode Intel HD Graphics 2000/3000 Performance
Comments Locked

283 Comments

View All Comments

  • hmcindie - Monday, January 3, 2011 - link

    Why is that Quick Sync has better scaling? Very evident in the Dark Knight police car image as all the other versions have definite scaling artifacts on the car.

    Scaling is something that should be very easy. Why is there so big a difference? Are these programs just made to market new stuff and no-one really uses them because they suck? So big scaling differences between codepaths make no sense.
  • JarredWalton - Monday, January 3, 2011 - link

    It looks to me like some of the encodes have a sharpening effect applied, which is either good (makes text legible) or bad (aliasing effects) depending on your perspective. I'm quite happy overall with the slightly blurrier QS encodes, especially considering the speed.
  • xxxxxl - Monday, January 3, 2011 - link

    I've been so looking forward to SB...only to hear that H67 cant overclock CPU?!?!?!?!
    Disappointed.
  • digarda - Monday, January 3, 2011 - link

    Who needs the IGP for a tuned-up desktop PC anyway? Some for sure, but I see the main advantages of the SB GPU for business laptop users. As the charts show, for desktop PC enthusiasts, the GPU is still woefully slow, being blown away even by the (low-end) Radeon 5570. For this reason, I can't help feeling that the vast majority of overclockers will still want to have discrete graphics.

    I would have preferred to dual core (4-thread) models to have (say) 32 shaders, instead of the 6 or 12 being currently offered. At 32nm, there's probably enough silicon real estate to do it. I guess Intel simply didn't want the quad core processors to have a lower graphics performance than the dual core ones (sigh).

    Pity that the socket 2011 processors (without a GPU) are apparently not going to arrive for nearly a year (Q4 2011). I had previously thought the schedule was Q3 2011. Hopefully, AMD's Bulldozer-based CPUs will be around (or at least imminent) by then, forcing intel to lower the prices for its high-end parts. On the other hand, time to go - looks like I'm starting to dream again...
  • Exodite - Monday, January 3, 2011 - link

    Using myself as an example showing the drawback of limiting overclocking on H67 would be the lack of a good selection of overclocking-friendly micro-ATX boards due to most, if not all, of those being H67.

    Granted, that's not Intel's fault.

    It's just that I have no need for more than one PCIe x16 slot and 3 SATA (DVD, HDD, SSD). I don't need PCI, FDD, PS2, SER, PAR or floppy connectors at all.

    Which ideally means I'd prefer a rather basic P67 design in micro-ATX format but those are, currently, in short supply.

    The perfect motherboard, for me, would probably be a P67 micro-ATX design with the mandatory x8/x8 Crossfire support, one x1 and one x4 slot, front panel connector for USB 3, dual gigabit LAN and the base audio and SATA port options.

    Gigabyte?

    Anyone? :)
  • geofelt - Monday, January 3, 2011 - link

    The only P67 based micro-ATX motherboard I have found to date is the
    Asus P8P67-M pro. (or evo?)

    Any others?
  • Rick83 - Monday, January 3, 2011 - link

    There's also a non-pro P8P67-M.

    Keep in mind though, that the over-clocking issue may not be as bad as pointed out. There are H67 boards being marketed for over-clocking ability and manuals showing how to adjust the multiplier for CPUs... I'm not yet convinced over-clocking will be disabled on H67.
  • smilingcrow - Monday, January 3, 2011 - link

    Major bummer as I was going to order a Gigabyte H67 board and an i5-2500K but am put off now. They seem to over-clock so well and with low power consumption that it seemed the perfect platform for me…
    I don’t mind paying the small premium for the K editions but being forced to use a P67 and lose the graphics and have difficulty finding a mATX P67 board seems crazy!

    I wonder if this limit is set in the chipset or it can be changed with a BIOS update?
  • DanNeely - Monday, January 3, 2011 - link

    Quick Sync only works if the IGP is in use (may be fixable via drivers later); for anyone who cares about video encoding performance that makes the IGP a major feature.
  • mariush - Monday, January 3, 2011 - link

    On the Dark Knight test...

    Looking at the Intel software encoding and the AMD encoding, it looks like the AMD is more washed out overall, which makes me think there's actually something related to colorspaces or color space conversion involved....

    Are you guys sure there's no PC/TV mixup there with the luminance or ATI using the color matrix for SD content on HD content or something like that?

Log in

Don't have an account? Sign up now