The Architecture

We'll start, logically, at the front end of a Bulldozer module. The fetch and decode logic in each module is shared by both integer cores. The role this logic plays is to fetch the next instruction in the thread being executed, decode the x86 instruction into AMD's own internal format, and pass the decoded instruction onto the scheduling hardware for execution.

AMD widened the K8 front end with Bulldozer. Each module is now able to fetch and decode up to four x86 instructions from a single thread in parallel. Each of the four decoders are equally capable. Remembering that each Bulldozer module appears as two cores, the front end can only pick 4 instructions to fetch and decode from a single thread at a time. A single Bulldozer module can switch between threads as often as every clock.

Decode hardware isn't very expensive on its own, but duplicating it four times across multiple cores quickly adds up. Although decode width has increased for a single core, multi-core Bulldozer configurations can actually be at a disadvantage compared to previous AMD architectures. Let's look at the table below to understand why:

Front End Comparison
  AMD Phenom II AMD FX Intel Core i7
Instruction Decode Width 3-wide 4-wide 4-wide
Single Core Peak Decode Rate 3 instructions 4 instructions 4 instructions
Dual Core Peak Decode Rate 6 instructions 4 instructions 8 instructions
Quad Core Peak Decode Rate 12 instructions 8 instructions 16 instructions
Six/Eight Core Peak Decode Rate 18 instructions (6C) 16 instructions 24 instructions (6C)

For a single instruction thread, Bulldozer offers more front end bandwidth than its predecessor. The front end is wider and just as capable so this makes sense. But note what happens when we scale up core count.

Since fetch and decode hardware is shared per module, and AMD counts each module as two cores, given an equivalent number of cores the old Phenom II actually offers a higher peak instruction fetch/decode rate than the FX. The theory is obviously that the situations where you're fetch/decode bound are infrequent enough to justify the sharing of hardware. AMD is correct for the most part. Many instructions can take multiple cycles to decode, and by switching between threads each cycle the pipelined front end hardware can be more efficiently utilized. It's only in unusually bursty situations where the front end can become a limit.

Compared to Intel's Core architecture however, AMD is at a disadvantage here. In the high-end offerings where Intel enables Hyper Threading, AMD has zero advantage as Intel can weave in instructions from two threads every clock. It's compared to the non-HT enabled Core CPUs that the advantage isn't so clear. Intel maintains a higher instantaneous decode bandwidth per clock, however overall decoder utilization could go down as a result of only being able to fill each fetch queue from a single thread.

After the decoders AMD enables certain operations to be fused together and treated as a single operation down the rest of the pipeline. This is similar to what Intel calls micro-ops fusion, a technology first introduced in its Banias CPU in 2003. Compare + branch, test + branch and some other operations can be fused together after decode in Bulldozer—effectively widening the execution back end of the CPU. This wasn't previously possible in Phenom II and obviously helps increase IPC.

A Decoupled Branch Predictor

AMD didn't disclose too much about the configuration of the branch predictor hardware in Bulldozer, but it is quick to point out one significant improvement: the branch predictor is now significantly decoupled from the processor's front end.

The role of the branch predictor is to intercept branch instructions and predict their target address, rather than allowing for tons of cycles to go by until the branch target is known for sure. Branches are predicted based on historical data. The more data you have, and the better your branch predictors are tuned to your workload, the more accurate your predictions can be. Accurate branch prediction is particularly important in architectures with deep pipelines as a mispredict causes more instructions to be flushed out of the pipe. Bulldozer introduces a significantly deeper pipeline than its predecessor (more on this later), and thus branch prediction improvements are necessary.

In both Phenom II and Bulldozer, branches are predicted in the front end of the pipe alongside the fetch hardware. In Phenom II however, any stall in the fetch pipeline (e.g. fetching an instruction that wasn't in cache) would stop the whole pipeline including future branch predictions. Bulldozer decouples the branch prediction hardware from the fetch pipeline by way of a prediction queue. If there's a stall in the fetch pipeline, Bulldozer's branch prediction hardware is allowed to run ahead and continue making future predictions until the prediction queue is full.

We'll get to the effectiveness of this approach shortly.

Scheduling and Execution Improvements

As with Sandy Bridge, AMD migrated to a physical register file architecture with Bulldozer. Data is now only stored in one location (the physical register file) and is tracked via pointers back to the PRF as operations make their way through the execution engine. This is a move to save power as copying data around a chip is hardly power efficient.

The buffers and queues that feed into the execution engines of the chip are all larger on Bulldozer than they were on Phenom II. Larger data structures allows for better instruction level parallelism when trying to execute operations out of order. In other words, the issue hardware in Bulldozer is beefier than its predecessor.

Unfortunately where AMD took one step forward in issue hardware, it does a bit of a shuffle when it comes to execution resources themselves. Let's start with the positive: Bulldozer's integer execution cores.

Integer Execution

Each Bulldozer module features two fully independent integer cores. Each core has its own integer scheduler, register file and 16KB L1 data cache. The integer schedulers are both larger than their counterparts in the Phenom II.

The biggest change here is each integer core now has two ports instead of three. A single integer core features two AGU/ALU ports, compared to three in the previous design. AMD claims the third ALU/AGU pair went mostly unused in Phenom II, and as a result it's been removed from Bulldozer.

With larger structures feeding into the integer cores, AMD should be able to have an easier time of making use of the integer units than in previous designs. AMD could, in theory, execute more integer operations per core in Phenom II however AMD claims the architecture was typically bound elsewhere.

The Shared FP Core

A single Bulldozer module has a single shared FP core for use by up to two threads. If there's only a single FP thread available, it is given full access to the FP execution hardware, otherwise the resources are shared between the two threads.

Compared to a quad-core Phenom II, AMD's eight-core (quad-module) FX sees no drop in floating point execution resources. AMD's architecture has always had independent scheduling for integer and floating point instructions, and we see the same number of execution ports between Phenom II cores and FX modules. Just as is the case with the integer cores, the shared FP core in a Bulldozer module has larger scheduling hardware in front of it than the FPU in Phenom II.

The problem is AMD had to increase the functionality of its FPU with the move to Bulldozer. The Phenom II architecture lacks SSE4 and AVX support, both of which were added in Bulldozer. Furthermore, AMD chose Bulldozer as the architecture to include support for fused multiply-add instructions (FMA). Enabling FMA support also increases the relative die area of the FPU. So while the throughput of Bulldozer's FPU hasn't increased over K8, its capabilities have. Unfortunately this means that peak FP throughput running x87/SSE2/3 workloads remains unchanged compared to the previous generation. Bulldozer will only be faster if newer SSE, AVX or FMA instructions are used, or if its clock speed is significantly higher than Phenom II.

Looking at our Cinebench 11.5 multithreaded workload we see the perfect example of this performance shuffle:

Cinebench 11.5—Multi-Threaded

Despite a 9% higher base clock speed (more if you include turbo core), a 3.6GHz 8-core Bulldozer is only able to outperform a 3.3GHz 6-core Phenom II by less than 2%. Heavily threaded floating point workloads may not see huge gains on Bulldozer compared to their 6-core predecessors.

There's another issue. Bulldozer, at least at launch, won't have to simply outperform its quad-core predecessor. It will need to do better than a six-core Phenom II. In this comparison unfortunately, the Phenom II has the definite throughput advantage. The Phenom II X6 can execute 50% more SSE2/3 and x87 FP instructions than a Bulldozer based FX.

Since the release of the Phenom II X6, AMD's major advantage has been in heavily threaded workloads—particularly floating point workloads thanks to the sheer number of resources available per chip. Bulldozer actually takes a step back in this regard and as a result, you will see some of those same workloads perform worse, if not the same as the outgoing Phenom II X6.

Compared to Sandy Bridge, Bulldozer only has two advantages in FP performance: FMA support and higher 128-bit AVX throughput. There's very little code available today that uses AMD's FMA instruction, while the 128-bit AVX advantage is tangible.

Cache Hierarchy and Memory Subsystem

Each integer core features its own dedicated L1 data cache. The shared FP core sends loads/stores through either of the integer cores, similar to how it works in Phenom II although there are two integer cores to deal with now instead of just one. Bulldozer enables fully out-of-order loads and stores, an improvement over Phenom II putting it on parity with current Intel architectures. The L1 instruction cache is shared by the entire bulldozer module, as is the L2 cache.

The instruction cache is a large 64KB 2-way set associative cache, similar in size to the Phenom II's L1 cache but obviously shared by more "cores". A four-core Phenom II would have 256KB of total L1 I-Cache, while a four core Bulldozer will have half of that. The L1 data caches are also significantly smaller than Bulldozer's predecessor. While Phenom II offered a 64KB L1 D-Cache per core, Bulldozer only offers 16KB per integer core.

The L2 cache is much larger than what we saw in multi-core Phenom II designs however. Each Bulldozer module has a private 2MB L2 cache.

There's a single 8MB L3 cache that's shared among all Bulldozer modules on a chip. In its first incarnation, AMD has no plans to offer a desktop part without an L3 cache. However AMD indicated that the L3 cache was only really useful in server workloads and we might expect future Bulldozer derivatives (ahem, Trinity?) to forgo the L3 cache entirely as a result.

Cache accesses require more clocks in Bulldozer, due to a combination of size and AMD's desire to make Bulldozer a very high clock speed part...

Introduction The Pursuit of Clock Speed
Comments Locked

430 Comments

View All Comments

  • Hector2 - Wednesday, October 12, 2011 - link

    Ivy Bridge on 22nm (Sandy Bridge on steroids) is just around the corner. It's not going to be pretty
  • superccs - Wednesday, October 12, 2011 - link

    If Ghz is BD's salvation, then why did you post some OC'd benchmarks? Cmon, really you are going to make us do it?
  • chizow - Wednesday, October 12, 2011 - link

    Anand, you said it yourself. AMD hasn't been competitive since 2006 and Core 2. Has there been a real choice in the CPU space for the last 5 years? Do we even need choice? The fact of the matter is Intel is only competing with *THEMSELVES* in the CPU space and this has been the case for the last 5 years. They need to keep dropping prices and releasing faster processors because that's the only way they are going to sell more CPUs, irrelevant of whatever drek AMD is releasing in their wake.

    I really wish tech writers and pseudo-economists would stop repeating the lie of a meme that "we need competition" in order for innovation to continue, when that's simply not the case.
  • bji - Wednesday, October 12, 2011 - link

    Your comments are a gross oversimplification. There are many segments of the x86 market and AMD has been competitive in some of them over the past 5 years. Just because they haven't been competitive in your favorite segment (which I am guessing is high end) doesn't mean that they haven't been competitive at all.

    I bought an Atom board and a Zacate board and the Zacate was clearly and thoroughly superior. So AMD wins at the very low end.

    At the low midrange I believe that AMD has always, except until maybe very recently, been the better choice; of course AMD has achieved this by basically making no money on their parts which obviously is not sustainable.

    At the mid-midrange AMD has been sometimes competitive and sometimes not over the past 5 years. I personally bought a 6 core Phenom II last year around this time because the workload I care about - massively parallel software compiles - was heads and shoulders better on AMD than similarly priced Intel at the time.

    Once you get into the high midrange and all segments of the high end, then I finally agree with your comments - AMD has not been competitive there for 5 years.

    On the laptop front, AMD has mostly been uncompetitive in the time frame in question except where they have reduced price to an unsustainable (for them) level and stolen the occasional win. The only win for AMD in that space is, once again, their Atom equivalents which win in netbooks.

    Your post is an oversimplification. However, if AMD doesn't pull a rabbit out of their hat, then for the next 5 years your oversimplification will end up being true.

    Competition driving innovation is not a lie, and your appeal to people to stop repeating this truth will go unheeded so don't waste your breath.
  • chizow - Wednesday, October 12, 2011 - link

    No, its not really a gross oversimplication, its just an accurate generalization born out repeatedly over the last 5 years.

    Intel has AMD beat in every single price-performance segment in the x86 market except in the rare cases where mGPUs come into play, like your Zacate example. Which is ironic in and of itself, given some of the asinine remarks AMD leadership has made in the past with regard to ULP products.

    Just look at this low to mid-range line-up, the pricing segments have been steady for the last 4-5 years during Intel's dominance, the ONLY things that change are the SKU names and clockspeeds.

    http://www.anandtech.com/show/4524/the-sandy-bridg...

    This all occurs independent of any meaningful competition from AMD, as every single Intel option offers better performance for the same or slightly more money.

    Even in multi-threaded apps Intel often achieves performance parity with FEWER cores and less TDP. Its embarassing really how Intel can even compete with AMD's 6 core processors with their *2* core processors even in the threaded apps you're referring to.

    http://www.anandtech.com/show/3674/amds-sixcore-ph...

    Tick Tock doesn't stop when AMD stumbles, nor does Intel's commitment to process fabrication leadership. Why? Because their investors and consumers demand innovation INDEPENDENT of what AMD does or does not do to compete. Competition does not drive innovation in the x86 desktop market and has not for the last 5 years, the bottom line does.
  • Al1458 - Wednesday, October 12, 2011 - link

    AMD is the only viable alternative to Intel. If AMD should not get out of this slump, we could well see AMD leave the CPU market which would leave one CPU manufacturer, and we all know how restrained Intel is with its pricing...
  • hoohoo - Wednesday, October 12, 2011 - link

    I really want to see some server benchmarks: LAMP & IIS stack, DBMSs, HPC benchmarks.

    I do not think the real potential is being explored, (for FPU intensive code especially) a recompile is needed with a compiler that knows what a BD is.
  • Kyanzes - Wednesday, October 12, 2011 - link

    Simply horrible. Surely there has to be some mistake.
  • mapesdhs - Wednesday, October 12, 2011 - link


    Sometimes I think it's a pity a company like IBM doesn't make a consumer chip, ie. a rival to Intel
    which has enough cash and expertise to do the job properly, come up with something that can be
    a good fit for both the consumer & server spaces. IBM certainly has plenty of CPU design experience
    for servers, eg. the 4.25GHz 8-core Power7.

    Ian.
  • FunBunny2 - Wednesday, October 12, 2011 - link

    They did. Called the PPC. Nobody bought it.

Log in

Don't have an account? Sign up now