AMD’s Industry Problem

A significant number of small form factor and portable devices have been sold since the start of the century - this includes smartphones, tablets, laptops, mini-PCs and custom embedded designs. Each of these markets is separated by numerous facets: price, performance, mobility, industrial design, application, power consumption, battery life, style, marketing and regional influences. At the heart of all these applications is the CPU that takes input, performs logic, and provides output dependent on both the nature of the device and the interactions made. Both the markets for the devices, and the effort placed into manufacturing the processors, is large and complicated. As a result we have several multi-national and worldwide companies hiring hundreds or thousands of engineers and investing billions of dollars each year into processor development, design, fabrication and implementation. These companies, either by developing their own intellectual property (IP) or licensing then modifying other IP, aim to make their own unique products with elements that differentiate them from everyone else. The goal is to then distribute and sell, so their products end up in billions of devices worldwide.

The market for these devices is several hundreds of billions of dollars every year, and thus to say competition is fierce is somewhat of an understatement. There are several layers between designing a processor and the final product, namely marketing the processor, integrating a relationship with an original equipment manufacturer (OEM) to create a platform in which the processor is applicable, finding an entity that will sell the platform under their name, and then having the resources (distribution, marketing) to the end of the chain in order to get the devices into the hands of the end user (or enterprise client). This level of chain complexity is not unique to the technology industry and is a fairly well established route for many industries, although some take a more direct approach and keep each stage in house, designing the IP and device before distribution (Samsung smartphones) or handling distribution internally (Tesla motors).

In all the industries that use semiconductors however, the fate of the processor, especially in terms of perception and integration, is often a result of what happens at the end of the line. If a user, in this case either an end user or a corporate client investing millions into a platform, tries multiple products with the same processor but has a bad experience, they will typically relate the negativity and ultimately their purchase decision towards both the device manufacturer and the manufacturer of the processor. Thus it tends to be in the best interest of all parties concerned that they develop devices suitable for the end user in question and avoid negative feedback in order to develop market share, recoup investment in research and design, and then generate a profit for the company, the shareholders, and potential future platforms. Unfortunately, with many industries suffering a race-to-the-bottom, cheap designs often win due to budgetary constraints, which then provides a bad user experience, giving a negative feedback loop until the technology moves from ‘bearable’ to ‘suitable’.

Enter Carrizo

One such platform that was released in 2015 is that of AMDs Carrizo APU (accelerated processor unit). The Carrizo design is the fourth generation of the Bulldozer architecture, originally released in 2011. The base design of the microarchitecture is different to the classical design of a processor - at a high level, rather than one core having one logic pipeline sharing one scheduler, one integer calculation port and one floating point calculation port resulting in one thread per core, we get a compute module with two logic pipelines sharing two schedulers, two integer calculation ports and only one floating point pipeline for two threads per module (although the concept of a module has been migrated to that of a dual core segment). With the idea that the floating point pipeline is being used infrequently in modern software and compilers, sharing one between two aims to save die area, cost, and additional optimizations therein.

The deeper reasons for this design lie in typical operating system dynamics - the majority of logic operations involving non-mathematical interpretations are integer based, and thus an optimization of the classical core design can result in the resources and die area that would normally be used for a standard core design to be focused on other more critical operations. This is not new, as we have had IP blocks in both the desktop and mobile space that have shared silicon resources, such as video decode codecs sharing pipelines, or hybrid memory controllers covering two memory types, to save die area but enable both features in the market at once.

While interesting in the initial concept, the launch of Bulldozer was muted due to its single threaded performance compared to that of AMD’s previous generation product as well as AMD’s direct competitor, Intel, whose products could ultimately process a higher number of instructions per clock per thread. This was countered by AMD offering more cores for the same die area, improving multithreaded performance for high workload throughput, but other issues plagued the launch. AMD also ran at higher frequencies to narrow the performance deficit, and at higher frequencies, the voltage required to maintain those frequencies related in a higher power consumption compared to the competition. This was a problem for AMD as Intel started to pull ahead on processor manufacturing technology taking advantage of lower operating voltages, especially in mobile devices.

Also, AMD had an issue with operating system support. Due to the shared resource module design of the processor, Microsoft Windows 7 (the latest at the time) had trouble distinguishing between modules and threads, often failing to allocate resources to the most suitable module at runtime. In some situations, it would cause two threads would run on a single core, with the other cores being idle. This latter issue was fixed via an optional update and in future versions of Microsoft Windows but still resulted in multiple modules being on 'active duty', affecting power consumption.

As a result, despite the innovative design, AMDs level of success was determined by the ecosystem, which was rather unforgiving in both the short and long term. The obvious example is in platforms where power consumption is directly related to battery life, and maintaining a level of performance required for those platforms is always a balance in managing battery concerns. Ultimately the price of the platform is also a consideration, and along with historical trends from AMD, in order to function this space as a viable alternative, AMD had to use aggressive pricing and adjust the platforms focus, potentially reducing profit margins, affecting future developments and shareholder return, and subsequently investment.

The Devices: #5 The Lenovo Y700 (Carrizo, FX-8800P + R9 385MX) How to Iterate Through Design
Comments Locked

175 Comments

View All Comments

  • basicmath - Tuesday, February 9, 2016 - link

    No it's really not, this laptop came from the factory with dual channel capability but that capability was not utilised because that would have shown the platform in a much better light, he even states that he checked the chips in the G2 to confirm that it was single channel. Upgrading the RAM on a laptop is a simple process that any end user can perform. The only discernible difference between the APU in the G2 & G3 is the number of GPU cores so why did he even bother testing the G3 without using dual channel configuration?
  • Intel999 - Sunday, February 7, 2016 - link

    @Ian

    I look forward to that R-series test as it will provide a sneak peek at how much DDR4 relieves the bottleneck on integrated graphics when Bristol Ridge comes out.

    That $70 Athlon X4 845 is intriguing as well.
  • AS118 - Saturday, February 6, 2016 - link

    Good article (although to be fair, I mostly skimmed it), and I agree with the conclusions. AMD should try harder to make sure their high-end products are paired with good components. Single-channel ram, bad screens, and slow hard drives with an A10 or mobile FX defeats the purpose of having those higher end APU's.

    Plus, people will get a bad impression of AMD if a lot of them have poor trackpads, etc. I wish they'd make their own "signature" brand of laptops, and find someone to help make them a thing.
  • TheinsanegamerN - Thursday, February 11, 2016 - link

    Both clevo and MSI have treated AMD well before, Im sure either would love to have exclusive rights to the high end AMD notebook.

    That being said, I doubt AMD has the intelligence to pull it off. They seem to be run by monkeys 90% of the time.
  • Cryio - Saturday, February 6, 2016 - link

    I'm really sorry for AMD. Kaveri and Carizzo on mobile, when configured to the proper ram, cooling and when using the highest performing part ... would've provided awesome performance, compared to Haswell and Broadwell. But no one bothered.

    Bristol Ridge will basically be Carizzo with DDR4 support and since it will be even better binned 28 nm CPUs, maybe we'll get even higher frequency out of Excavator. As for GCN 2.0 GPUs ... it will be interesting to see.

    I love my Surface Pro 4, even given the disaster that is Skylake drivers and Windows 10 horrible efficiency compared to W8.1. But MAN. I would've loved a proper Carrizo based Surface Pro/Book.
  • Gadgety - Saturday, February 6, 2016 - link

    A confirmation, with in depth detail. Nice write up.
  • Khenglish - Sunday, February 7, 2016 - link

    I would have really liked to see some dual channel results, or at least pulling a memory stick from the Kaveri and Intel systems to get a fair comparison. AMD says Zen brings a 40% IPC improvement. It'd be great to have a baseline to see if that 40% improvement is enough. In the dual channel intel to single channel AMD comparisons it does not appear to be enough, but we don't know how big of a factor memory was.
  • Jon Irenicus - Sunday, February 7, 2016 - link

    I want to buy an amd part for my next notebook but as was mentioned in the article, oems only choose bargain basement platforms to put the amd inside. The elitebook is the one exception, along with the lenovo if you don't mind the bulk.

    But the elitebooks are super overpriced for what you get. They need to release an hp spectre version of a notebook with a zen apu, a dell xps notebook variant with an apu. Ideally, the models that include a discreet gpu should allow the apu to work in tandem.

    In 2017 dx12 will be in full effect with games, and having two gpus working together by default could give a lot of amd equipped systems a larger edge, especially if the oversized ipc deficits between excavator and intel parts is minimized with zen.

    The future really does rest on Zen, amd needs to laser focus on performance per watt and ipc, and equip the 2017 apus with polaris gpu parts or vega or whatever the first iterations will be called. That has to be the minimum. Put those in nice chassis with solid battery life and that is all they need.
  • Intel999 - Sunday, February 7, 2016 - link

    In DX12 dual graphics will be automatic. Even an Intel Igpu combined with a discrete Nvidia or AMD GPU will "merge" the graphic capabilities in the laptop.

    Theoretically, an AMD APU combined with an AMD GPU might have an advantage as all graphics would be from the same underlying graphic architecture. Time will tell if this bares out.
  • Ryan Smith - Monday, February 8, 2016 - link

    Note that it's only going to be as automatic as the game developer makes it, as devs will be responsible for implementing it. For the moment game devs are going to be the wildcard.

Log in

Don't have an account? Sign up now