A Brief History of Switchable Graphics

So what is Optimus exactly? You could call it NVIDIA Switchable Graphics III, but that makes it sound like only a minor change. After our introduction, in some ways the actual Optimus technology is a bit of a letdown. This is not to say that Optimus is bad—far from it—but the technology only addresses a subset of our wish list. The area Optimus does address—specifically switching times—is dealt with in what can only be described as an ideal fashion. The switch between IGP and discrete graphics is essentially instantaneous and transparent to the end-user, and it's so slick that it becomes virtually a must-have feature for any laptop with a discrete GPU. We'll talk about how Optimus works in a minute, but first let's discuss how we got to our present state of affairs.

It turns out that the Optimus hardware has been ready for a while, but NVIDIA has been working hard on the software side of things. On the hardware front, all of the current G200M and G300M 40nm GPUs have the necessary internal hardware to support Optimus. Earlier laptop designs using those GPUs don't actually support the technology, but the potential was at least there. The software wasn't quite ready, as it appears to be quite complex—NVIDIA says that the GeForce driver base now has more lines of code than Windows NT, for example. NVIDIA was also keen to point out that they have more software engineers (over 1000) than hardware engineers, and some of those software engineers are housed in partner offices (e.g. there are NVIDIA employees working over at Adobe, helping with the Flash 10.1 software). Anyway, the Optimus hardware and software are both ready for public release, and you should be able to find first Optimus enabled laptops for sale starting today.


The original switchable graphics designs used a hardware switch to allow users to select either an IGP or discrete GPU. The first such system that we tested was the ASUS N10JC, but the very first implementation came from Sony in the form of the VAIO SZ-110B launched way back in April 2006. (The Alienware m15x was also gen1 hardware.) Generation one required a system reboot in order to switch between graphics adapters, with hardware multiplexers routing power to the appropriate GPU and more multiplexers to route the video signal from either GPU to the various display outputs. On the surface, the idea is pretty straightforward, but the actual implementation is much more involved. A typical laptop will have three separate video devices: the laptop LCD, a VGA port, and a DVI/HDMI port. Adding the necessary hardware requires six (possibly more) multiplexer ICs at a cost of around $1 each, plus more layers on the motherboard to route all of the signals. In short, it was expensive, and what's worse the required system reboot was highly disruptive.

Many users of the original switchable graphics laptops seldom switched, opting instead to use either the IGP or GPU all the time. I still liked the hardware and practically begged manufacturers to include switchable graphics in all future laptop designs. My wish wasn't granted, although given the cost it's not difficult to see why. As far as the rebooting, my personal take is that it was pretty easy to simply switch your laptop into IGP mode right before spending a day on the road. If you were going to spend most of the day seated at your desk, you'd switch to discrete mode. The problem is, as one of the highly technical folks on the planet, I'm not a good representation of a typical user. Try explaining to a Best Buy shopper exactly what switchable graphics is and how it works, and you're likely to cause more confusion than anything. Our readers grasp the concept, but the added cost for a feature many wouldn't use meant there was limited uptake.


Generation two involved a lot more work on the software side, as the hardware switch became a software controlled switch. NVIDIA also managed to eliminate the required system reboot (although certain laptop vendors continue to require a reboot or at least a logout, e.g. Apple). Again, that makes it sound relatively simple, but there are many hurdles to overcome. Now the operating system has to be able to manage two different sets of drivers, but Windows Vista in particular doesn't allow multiple display drivers to be active. The solution was to create a "Display Driver Interposer" that had knowledge of both driver sets. Launched in 2008, the first laptop we've reviewed with gen2 hardware and software was actually the ASUS UL80Vt, which took everything we loved about gen1 and made it a lot more useful. Now there was no need to reboot the system; you could switch between IGP and dGPU in about 5 to 10 seconds, theoretically allowing you the best of both worlds. We really liked the UL80Vt and gave it our Silver Editors' Choice award, but there was still room for improvement.

First, the use of a driver interposer meant that generic Verde drivers would not work with switchable graphics gen2. The interposer conforms to the standard graphics APIs, but then there's a custom API to talk to the IGP drivers. The result is that the display driver package contains both NVIDIA and Intel drivers (assuming it's a laptop with an Intel IGP), so driver updates are far more limited. If either NVIDIA or Intel release a new driver, there's an extra ~10 days of validation and testing that take place; if all goes well, the new driver is released, but any bugs reset the clock. Of course, that's only in a best-case situation where NVIDIA and Intel driver releases happen at the same time, which they rarely do. In practice, the only time you're likely to get a new driver is if there's a showstopper bug of some form and the laptop OEM asks NVIDIA for a new driver drop. This pretty much takes you back to the old way of doing mobile graphics drivers, which is not something we're fond of.

Another problem with gen2 is that there are still instances where switching from IGP to dGPU or vice versa will "block". Blocking occurs when an application is in memory that currently uses the graphics system. If blocking was limited to 3D games it wouldn't be a critical problem, but the fact is blocking can occur on many applications—including minesweeper and solitaire, web browsers where you're watching (or have watched) Flash videos, etc. If a blocking application is active, you need to close it in order to switch. The switch also results in a black screen as the hardware shifts from one graphics device to the other, which looks like potentially flaky hardware if you're not expecting the behavior. (This appears to be why Apple MacBook Pro systems require a reboot/logout to switch, even though they're technically gen2 hardware.) Finally, it's important to note that gen2 costs just as much as gen1 in terms of muxes and board layers, so it can still increase BOM and R&D costs substantially.

Incidentally, AMD switchable graphics is essentially equivalent to NVIDIA's generation two implementation. The HP Envy 13 is an example of ATI switchable graphics, with no reboot required and about 5-10 seconds required to switch between IGP and discrete graphics (an HD 4330 in this case—and why is it they always seem to use the slowest GPUs; why no HD 4670?).

For technically inclined users, gen2 was a big step forward and the above problems aren't a big deal; for your typical Best Buy shopper, though, it's a different story. NVIDIA showed a quote from Roger Kay, President of Endpoint Technology Associates, that highlights the problem for such users.

"Switchable graphics is a great idea in theory, but in practice people rarely switch. The process is just too cumbersome and confusing. Some buyers wonder why their performance is so poor when they think the discrete GPU is active, but, unknown to them, it isn't."

The research from NVIDIA indicates that only 1% of users ever switched between IGP and dGPU, which frankly seems far too low. Personally, if a laptop is plugged in then there's no real reason to switch off the discrete graphics, and if you're running on battery power there's little reason to enable the discrete graphics most of the time. It could be that only 1% of users actually recognize that there's a switch taking place when they unplug their laptop; it could also be that MacBook Pro users with switchable graphics represented a large percentage of the surveyed users. Personally, I am quite happy with gen2 and my only complaint is that not enough companies use the technology in their laptops.

Index NVIDIA Optimus Unveiled
Comments Locked

49 Comments

View All Comments

  • MasterTactician - Wednesday, February 10, 2010 - link

    But doesn't this solve the problem of an external graphics solution not being able to use the laptop's display? If the external GPU can pass the rendered frames back to the IGP's buffer via PCI-E than its problem solved, isn't it? So the real question is: Will nVidia capitalize on this?
  • Pessimism - Tuesday, February 9, 2010 - link

    Did Nvidia dip into the clearance bin again for chip packaging materials? Will the laptop survive its warranty period unscathed? What of the day after that?
  • HighTech4US - Tuesday, February 9, 2010 - link

    You char-lie droid ditto-heads are really something.

    You live in the past and swallow everything char-lie spews.

    Now go back to your church's web site semi-inaccurate and wait for your next gospel from char-lie.
  • Pessimism - Tuesday, February 9, 2010 - link

    So I suppose the $300+ million dollars in charges Nvidia took were merely a good faith gesture for the tech community with no basis in fact regarding an actual defect.
  • HighTech4US - Tuesday, February 9, 2010 - link

    The $300+ million was in fact a good faith measure to the vendors who bought the products that had the defect. nVidia backed extended warranties and picked up the total cost of repairs.

    The defect has been fixed long ago.

    So your char-lie comment as if it still exists deserves to be called what it is. A char-lie.
  • Pessimism - Tuesday, February 9, 2010 - link

    So you admit that a defect existed. That's more than can be said for several large OEM manufacturers.


  • Visual - Tuesday, February 9, 2010 - link

    Don't get me wrong - I really like the ability to have a long battery life when not doing anything and also have great performance when desired. And if switchable graphics is the way to achieve this, I'm all for it.

    But it seems counter-productive in some ways. If the external GPU was properly designed in the first place, able to shut down power to the unused parts of the processor, supporting low-power profiles, then we'd never have needed switching between two distinct GPUs. Why did that never happen?

    Now that Intel, and eventually AMD too, are integrating a low-power GPU inside the CPU itself, I guess there is no escaping from switchable graphics any more. But I just fail to see why NVidia or ATI couldn't have done it the proper way before.
  • AmdInside - Tuesday, February 9, 2010 - link

    Because it's hard to compete with a company that is giving away integrated graphics for free (Intel) in order to move higher priced CPUs. In a way, AMD is doing the same with ATI. Giving us great motherboards with ATI graphics and cheap cheap prices (which in many ways are much better than Intel's much higher priced offerings) in order to get you to buy AMD CPUs.
  • JarredWalton - Tuesday, February 9, 2010 - link

    Don't forget that no matter how pieces of a GPU go into a deep sleep state (i.e. via power gate transistors), you would still have some extra stuff receiving power. VRAM for example, plus any transistors/resistors. At idle the CULV laptops are down around 8 to 10W; even the WiFi card can suck up .5 to 1W and make a pretty substantial difference in battery life. I'd say the best you're likely to see from a discrete GPU is idle power draw that's around 3W over and above what an IGP might need, so a savings of 3W could be a 30% power use reduction.
  • maler23 - Tuesday, February 9, 2010 - link

    I've been waiting for this article ever since it was hinted in the last CULV roundup. The ASUS laptop is a little disappointing, especially the graphic card situation(the Alienware M11X kind of sucked up a lot of excitement there). Frankly, I'd just take a discounted UL-30VT and deal with manual graphics switching.

    Couple of questions:

    -Any chance for a review of the aforementioned Alienware M11X soon?

    -I've seen a couple of reviews with display quality comparisons including this one. How do to the Macbook and Macbook Pros fit into the rankings?

    cheers!

    -J

Log in

Don't have an account? Sign up now