Why Combine Quadro & Tesla?

So far we’ve covered how NVIDIA will be combining Quadro, but the more interesting question is “why?” In the NVIDIA hierarchy, Quadro is NVIDIA’s leading product. On the graphics side it’s fully unlocked, supporting quad-buffers, uncapped geometry performance, and uncapped viewport performance, while on the compute side it offers full speed FP64 support. Furthermore it’s available in the same configurations as a Tesla card, featuring the same number of CUDA cores and memory. Short of TCC, if you can compute it on a Tesla, you can compute it on a Quadro.

This actually creates some complexities for both NVIDIA and users. On a technical level, Fermi’s context switching is relatively fast for a GPU, but on an absolute level it’s still slow. CPUs can context switch in a fraction of the time, giving the impression of a concurrent thread execution even when we know that’s not the case. Furthermore for some reason context switching between rendering and compute on Fermi is particularly expensive, which means the number of context switches needs to be minimized in order to keep from wasting too much time just on context switching.

As a result of the time needed to context switch, Quadro products are not well suited to doing rendering and compute at the same time. They certainly can, but depending on what applications are being used and what they’re trying to do the result can be that compute eats up a great deal of GPU time, leaving the GUI to only update at a few frames per second with significant lag. On the consumer side NVIDIA’s ray-tracing Design Garage tech demo is a great example of this problem, and we took a quick video on a GTX 580 showcasing how running the CUDA based ray-tracer severely impacts GUI performance.

Alternatively, a highly responsive GUI means that the compute tasks aren’t getting a lot of time, and are only executing at a fraction of the performance that the hardware is capable of. As part of their product literature NVIDIA put together a few performance charts, and while they should be taken with a grain of salt, they do quantify the performance advantage of moving compute over to a dedicated GPU.

For these reasons if an application needs to do both compute and rendering at the same time then it’s best served by sending the compute task to a dedicated GPU. This is the allocation work developers previously had to take into account and that NVIDIA wants to eliminate. At the end of the day the purpose of Maximus is to efficiently allow applications to do both rendering and compute by throwing their compute workload on another GPU, because no one wants to spend $3500 on a Quadro 6000 only for it to get bogged down.

It’s worth noting that this situation closely mirrors the situation for software developers. For debug purposes NVIDIA recommends programmers have two GPUs, so that one GPU can be locked down debugging a compute or rendering task as necessary, while the other GPU is available to display the results. So NVIDIA encouraging users to have two GPUs for technical reasons is not new, but it is expanded. It also means there’s an obvious avenue for further development as NVIDIA wants to move GPU multitasking closer and closer to where CPU multitasking is today.

Moving on, the second reason NVIDIA is pursing Maximus is a result of their own actions. Because Quadro is NVIDIA’s leading product, it commands a leading price: a Quadro 6000 card is $3500 or more. This is a product of NVIDIA’s well engineered market segmentation – a GF110 GPU can be in a $500 GTX 580, a $2500 Tesla C2075, or a $3500 Quadro 6000. By disabling a few critical features on other products (e.g. geometry performance or FP64 performance) NVIDIA can push customers into buying a product at a price NVIDIA believes is best for the target market.

So what’s the problem? The Quadro 6000 is both a highly capable rendering product at a highly capable compute product, but not every professional user needs that much rendering power even if they need the compute power. Those users still need a Quadro card for its uncapped rendering performance, but they don’t necessarily need features such as Quadro 6000’s massive geometry throughput. The result is that NVIDIA was pricing themselves right out of their own market.

The solution to that is combining Quadro and Tesla. Maximus allows a Tesla C2075 to be used with any Fermi based Quadro (600/2K/4K/5K/6K), which allows NVIDIA to more appropriately tap the overlapping market of Quadro users that need top-tier compute performance. The end result for those users is that they not only pay less – a Quadro 2000 and a Tesla C2075 is $3000 versus over $3500 for a single Quadro 6000 – but they gain the aforementioned advantages of not having conflicting tasks slowing down the performance of a single Quadro card. Admittedly this is a lot of effort on NVIDIA’s part to tap a very specific market, but at the end of the day the professional market is a highly profitable market, making it worth NVIDIA’s time.

Final Words

Wrapping things up, NVIDIA has made it clear that they’re going to be pushing Maximus hard right out of the gate. Today of course was also the launch of Intel’s new Sandy Bridge E platform for high-end desktops and workstations, and in this industry there’s very little coincidence. It’s in NVIDIA’s interest to latch into workstation upgrade sales, and this is how you do it. They’ve already lined up HP, Lenovo, Fujitsu, and Dell to offer workstations pre-configured for Maximus, and we’re told those workstations will be made available for purchase today.

As to whether Maximus will be successful or not, this is going to depend both on software and marketing. On the software side NVIDIA needs to deliver on the transparency Maximus promises to developers and users – the concept is simple, but for the professional market the execution must be precise. Optimus graphics switching misbehaves now and then, but professional users will not be as willing to put up with any undesired behavior out of Maximus.

Marketing on the other hand is equally about promoting Maximus and promoting CUDA. A lot of NVIDIA’s promotional material for Maximus could easily be confused for CUDA promotional material, and this is because the videos and case studies are largely about how CUDA improved a process or a product while Maximus was the icing on the cake. Though we consider CUDA old, the fact of the matter is that much of the professional market NVIDIA is targeting has still not heard of CUDA, or has a limited understanding at best. As such NVIDIA will be using the launch of Maximus to promote the benefits of CUDA to certain targeted markets such as manufacturing, design, and broadcasting, just as much as they will be promoting the benefits of having multiple GPUs.

NVIDIA’s Maximus Technology: Quadro + Tesla, Launching Today Press Release
Comments Locked

29 Comments

View All Comments

  • Lonbjerg - Monday, November 14, 2011 - link

    When other OS have a market penetration that makes for a viable profit, I am sure NVIDIA will look at them.
    But until then...you get what you pay for.
  • Filiprino - Monday, November 14, 2011 - link

    LOL. GNU/Linux is the most used OS on supercomputers. You get what you pay for.
  • Solidstate89 - Monday, November 14, 2011 - link

    And?

    What does Optimus or Maximus have to do with large Super Computers? You don't need a graphical output for Super Computers, and last I checked a laptop with switchable graphics sure as hell isn't remotely comparable to a Super Computer.
  • Filiprino - Thursday, November 17, 2011 - link

    But you as a worker don't have always access to your rendering farms, supercomputers or workstations. Having a laptop to program on the go is a must for increased productivity.
  • Ryan Smith - Monday, November 14, 2011 - link

    On the compute side of things I'd say NVIDIA already has strong support for Linux. Tesla is well supported, as Linux is the natural OS to pair them up with for a compute cluster.

    As for Optimus or improved Quadro support, realistically I wouldn't hold my breath. It's still a very small market; if NVIDIA was going to chase it, I'd expect they'd have done so by now.
  • iwod - Monday, November 14, 2011 - link

    Hopefully all these technology lays the ground work for next generation GPU based on 28nm.

    We have been stuck with the current generation for too long... i think.
  • beginner99 - Monday, November 14, 2011 - link

    Since this is artifical, is it possible to turn a 580 into a tesla and quadro? bascially getting the same performance for a lot less?
  • Stuka87 - Monday, November 14, 2011 - link

    A GTX580 and a Fermi based Quadro may share a lot in common, they are not the same thing. There are many differences between them.

    I seriously doubt you could make a single 580 act as both, much less one or the other.
  • jecs - Monday, November 14, 2011 - link

    I am an artist working on CG and this is what I know.

    There is one software I know that can take advantage of both, Quadro's and GeForce's cards at their full potential and that is the Octane render. And as Ryan Smith explain there are software designed to use 2 graphic cards for different purposes and this rendering engine is optimized to use let's say a basic Nvidia Quadro for display or for the viewport, and a set of one or more 580-590 or any GeForce for rendering. This is great for your economy but Octane is a particular solution and not something Nvidia is pushing directly. The engineers at Refractive Software are the ones responsible to support the implementation and Nvidia could do any change at anytime that can disrupt any functionality without any compromise.

    With Maximus I can see Nvidia is heading in the same direction but endorsing Tesla as the workforce. The problem for smaller studios and professionals on their own is that Tesla cards are still in the 2K level whether a GeForce's is in the $200-700.

    So Maximus is a welcome as even high end studios are asking for these features that are cost effective and Nvidia is responding to their needs but Maximus is still to expensive for students or smaller studios working on CG.

    Well, Nvidia may not be willing yet to give you a "cheap" card to do high end work on it as they spend a lot on R&D on their drivers, lets be fair. So ask either to your favorite software producer to implement and fully support commercial gaming cards as an alternate hardware for rendering or compute but I can guess they wont be willing to support mayor features and develop and support their own drivers.
  • erple2 - Wednesday, November 16, 2011 - link

    Just how much money is your time worth? How many "wasted" hours of your time waiting for a render to finish adds up to 1800-1300 dollar price difference?

    This is an argument that I've had with several customers that seem to not quite fully realize that time wasted for things to finish working (on the CPU, GPU, RAM, Disk etc) side is time that your Engineer or Developer is very expensively twiddling their thumbs. It gets even more crazy when you realize that everyone downstream is also twiddling their thumbs waiting for work to finish.

    I wonder if it's a holdover from the old "bucket" style of accounting - having separate buckets for Hardware vs. Software vs. Engineering vs. etc. It's ultimately all the same bucket for a given project - saving $15,000 on hardware costs means that your development staff (and therefore everyone else that does work after Development) is delayed hours each week just to wait for things to finish working. So while the Hardware Bucket looks great ("We saved $15,000!"), your Development bucket winds up looking terrible ("We're now $35,000 behind schedule just from development, so everything downstream is in backlog").

    The flip side of that is that you can work on more projects over the long haul if you reduce "machine" time - more projects means more money (or accomplishments) which is good for the future.

Log in

Don't have an account? Sign up now