Kabini: Competing in the Evolving Marketplace

Prior to the onslaught of Android tablets and the launch of Windows 8, I was pretty much done with Atom and Brazos—they simply didn’t provide enough performance to make them pleasant devices for me to use. What was rather anemic hardware for Windows Vista/7 starts to look much more palatable with the new OS, and Android on such a chip can run extremely well. AMD has Temash for that audience, but I suspect we’ll at least see some 15W Kabini tablets/hybrids at some point, and they should stack up quite well against the competition in terms of performance and features. That’s the real takeaway from today’s launch, and it shows that AMD is keen to carve out a market niche separate from the traditional desktop and laptop PCs.

In terms of the normal Windows experience, Kabini doesn’t make too many waves. Yes, it’s faster than any current Atom or Brazos laptop, and Windows runs reasonably well all things considered, but there are still applications where Kabini falls short. That's the problem with competing in the good enough part of the computing spectrum—everyone has a different definition of what's good enough. 

CPU performance is appreciably better than anything Bobcat or Atom based at this point. If you're ok with Clover Trail, then Kabini will feel really quick. The big Ivy Bridge cores still maintain a significant performance advantage, but presumably Kabini's offer is that you can find it in an Ultrabook-like system but at a much more reasonable pricetag. If that ends up being the case, I suspect many would choose form factor over extra CPU performance. So much of Kabini's success will be tied to what OEMs do with the parts. It's unfortunate that AMD doesn't have any Kabini APUs with Turbo Core working as I suspect that could do wonders for further driving single threaded performance. 

On the graphics front, Kabini's Radeon HD 8330 is stuck between a rock and a hard place. It's significantly faster than what we had with Brazos (and lightyears ahead of what you get with Clover Trail), but not faster than Intel's HD 4000 which we viewed as the minimum acceptable level of performance for processor graphics upon its introduction. However, if you're playing older titles, or when faced with more tablet-like 3D workloads, Kabini's GPU should do very well.

On the power front, Kabini is great. In our tests we found much better battery life than Brazos and even better battery life than 17W Ultrabook-class Ivy Bridge parts. The days of AMD being associated with poor battery life are long gone. Kabini manages compelling battery life and better performance than Brazos, which is exactly what AMD needs. Given how successful Brazos was (almost 50M units shipped), Kabini seems to have the right recipe.

The real question for me is what sort of laptops and devices we’ll see when manufacturers release Kabini into the wild. The prototype laptop is really weak on some areas I care about—the keyboard and touchpad just don’t impress, and build quality is flimsy at best—but what it does have is a great LCD for what will hopefully be a budget-friendly laptop. Give me a reasonable Ultrabook-style chassis (or maybe a dockable tablet) with Kabini and a decent quality 1080p touchscreen and do it at the right price and there are plenty of people that will jump at the offer.

Kabini Gaming and Battery Life
Comments Locked

130 Comments

View All Comments

  • HisDivineOrder - Thursday, May 23, 2013 - link

    Given AMD's traditional design wins and how those systems end up, I suspect this is not going to matter much. I have more hope of Bay Trail providing a solid deal for once than I do this.

    It's a shame because this really should be AMD's niche to dominate, but I doubt any OEM'll give them a serious try.
  • Desperad@ - Thursday, May 23, 2013 - link

    On competitive positioning, is it even near IB Pentium?
  • brainee - Saturday, May 25, 2013 - link

    I think so, yes. IB Pentium 2117U (17 Watt TDP) should be around 33 % faster in legacy Intel-optimised CPU benchmarks doing the math and according to say Techspot. I would think ULV Pentiums are more expensive for OEMs, notebooks is a different story. Not to mention Kabini should cost a fraction to make for AMD compared with even crippled 2C Ivybridges aka Celeron / Pentium. Kabini wins in games and Open CL, and in AVX-enabled applications it should eat the Pentium alive since the latter doesn't support AVX extensions (should be mentioned at least). I'd prefer AVX extensions to Cinebench but this site seems to suggest I am a minority...
  • yhselp - Saturday, May 25, 2013 - link

    Comparing a 3W SoC (Z2760) to a 15W SoC (A4-5000), and calling the former laughable... not really fair.

    Sure, Kabini is definitely faster than the old Atom architecture and, yes, I understand this is not a definitive comparison; nevertheless - it seems misleading.

    What would happen if we compare a 3W Kabini to a 15W Haswell? Laughable wouldn't even begin to describe the performance difference.
  • silverblue - Saturday, May 25, 2013 - link

    But... an A4-5000 doesn't use anywhere near 15W, as far as I've heard. Still, let's consider the evidence - the Z2760 is a 32-bit, dual core, hyperthreaded CPU at 1.8GHz with a low powered graphics unit and 1MB of L2. The A4-5000 is a 64-bit, quad core CPU at 1.5GHz with a far stronger graphics unit and 2MB of dynamic L2. Temash would be a different proposition I expect as the A4-1200 is only clocked at 1GHz.
  • yhselp - Saturday, May 25, 2013 - link

    Yes, absolutely, I agree - it's just that the direct comparisons and conclusions made are a bit stark.

    There's always another side to an argument; in your case, I could argue that comparing the brand new Jaguar to a terribly old Atom architecture isn't the way to go. Consider the following evidence - Silvermont is 64-bit, quad-core, 2MB L2 cache, OoO, 2GHz+, 22nm, far more energy efficient, supports 1st gen Core instructions and Turbo Boost; it would decimate Jaguar.

    In the article, I also discovered that the 2020M is referred to as a 1.8GHz 35W part, when it's actually 2.4GHz. Are the benchmarks done on a underclocked 2020M or was that simply a typo?

    That's the kind of stuff I'm talking about, not AMD vs. Intel.
  • jcompagner - Sunday, May 26, 2013 - link

    So this is the core that will be in the next 2 big consoles?
    Am i the only one that think that these are quite weak, even if you have 8 of them?

    That does mean now that if one of those 2 consoles are the lead in the development that the games will be forced to be really good multi threaded. (So i guess the next games for the pc will also be using multiply cores way more)

    Why did they go for the jaguar core thats really targetted for ver low end or mobile stuff?

    Why didn't they just go for a Richland 8 core system with a very good gpu that lets say is a 100W part?

    What is the guess that the TDP is of the xbox one or ps4? A console can take 100W just easily that doesn't matter, so why choose for a core that is dedicated for mobile?
  • yhselp - Sunday, May 26, 2013 - link

    Yes, the Jaguar core is 'weak', but what does 'weak' mean? That is such a vague definition. For one usage scenario Jaguar might be unacceptable, for another it might be overkill. Remember, Sony/MS are not building a contemporary PC. Jaguar might seem slow to us, and in a gaming desktop it would be, but that's not the point. Think of consoles, in this case the PS4 and the Xbox One, as non-PC devices such as tablets. Would you say the latest Samsung/Apple running on a Cortex A15 is slow? No, you would say it's super fast. Well, Jaguar is even faster. Yes, a console has to deal with different workloads than a tablet, but that's why it has very different hardware.

    Why did Sony/MS choose Jaguar? Jaguar is easier to integrated, more power efficient and most importantly cheaper than Richland. It's a far simpler architecture than Richland, and probably easier to work with in a console's life. Also, it's very important to note that Sony/MS wanted an integrated solution - they weren't going to build a system with a dedicated video card like a gaming PC.

    Cost, cost, cost - everything is about the cost. A console cannot be expensive (the way a gaming PC is) - it has to sell very well in order to establish an install base to sell games to. Sony/MS will probably sell their 8th gen consoles at a loss initially - AMD's Jaguar/GCN was their best/only choice. What else could they do at the same price or even at all? Silvermont isn't ready yet and NVIDIA probably wouldn't be willing to integrate a GPU of theirs the way AMD did, and both of those would be more expensive than Jaguar/GCN. Not to mention, MS has had a ton of trouble with NVIDIA in the original Xbox - they are probably not willing to go down that road again.

    It's not really an 8-core solution - it's two quad-core modules and communication between the two might be problematic; so games on the new PS/Xbox would probably run on four Jaguar cores at 1.6 GHz. However, don't forget that neither of the two consoles has a ton of raw graphics power under the hood - the Xbox GPU is roughly equivalent to an HD 7770 (but with better memory bandwidth), and the PS to an HD 7850. Games would be specifically developed for this kind of hardware (unlike PC games) and would most probably be GPU limited so the Jaguar cores would really be sufficient.

    I hope this answers your questions.
  • Kevin G - Monday, May 27, 2013 - link

    A Pile Driver module is much larger than a Jaguar core. For die size concerns, it going with Jaguar made sense if core counts are the same. Steam Roller cores are due out in 2014 which are expected to bring higher IPC and a slight clock speed increase compared to Pile Driver.

    Power consumption is also an issue. The bulk of the power consumption from the XBox One and PS4 SoC's will come from their GPU's. Adding a high power CPU core like Pile Driver would have ballooned power consumption close to 200W which makes cooling impractical and expensive. Jaguar still adds power but it is far more manageable in comparison.

    In addition, Steam Roller is tied to processes from Global Foundries (though IBM could likely manufacture them if need be). TSMC is the preferred foundry for bulk processes due to cost and a slight edge in density. Jaguar has been prepared to be manufactured at TSMC from the start. AMD could have stuck with GF but it would have had to port GCN functional units to that same process. Such efforts are currently underway for Kaveri that is looking to be a 2014 part. So for any type of 2013 launch, going that route was not an option.
  • aikyucenter - Sunday, June 30, 2013 - link

    Great OpenCL performance ... love it ... just make it faster launch and decrease TDP too = PERFECT :D

Log in

Don't have an account? Sign up now