The A6 SoC

Section by Anand Shimpi

All great tech companies have their showing up to the fight moment. I borrow this phrasing from former ATI/AMDer, current Qualcomm-er Eric Demers. While at ATI/AMD, Eric came to the conclusion that the best way to lose market share was by simply not showing up to the fight. Customers tend to spend their money at key points throughout the year (holidays, back to school, etc...). If you don't have something shiny and new when those upticks in spending happen, you're not going to win. Eric called it showing up to the fight. By showing up to the fight every year, you at least had a chance of selling whatever it is that you're trying to hock.

Intel came to a similar realization after the Pentium 4, which eventually resulted in its famous tick-tock cadence. Every year you get evolutionary improvements, either in power or performance (sometimes in both). Over the course of several years, especially if your competition isn't as aggressive, you end up with a series of products that look downright revolutionary.

Apple learned from the best and quickly adopted a similar approach after it released the iPhone in 2007. With the exception of last year's 4S launch, like clockwork, Apple brought out a new iPhone every year at around the same time. The summer launch cycle was pushed back to fall last year but, since then Apple continued its roughly 12 month cadence for the iPhone.

The smartphone SoC space is still operating on this hyper Moore's Law curve which allows for significant innovation on a yearly cadence rather than a big update every 18 - 24 months. Even Intel recognized this fact as it will shift Atom to a yearly update cadence starting towards the end of next year.

The fast pace of changes on the smartphone side combined with the similarly aggressive release schedules from its competitors explain the difference in Apple's approach to iPhone/iPad vs. new Mac releases. The former are launched with much more pomp and circumstance, and are on a 2-year chassis redesign cadence. There's also the fact that devices running iOS make up the largest portion of Apple's yearly revenue. At some point I would expect the innovation/release cadence to slow down, but definitely not for the next few years.

The first few iPhones largely leveraged Samsung designed and manufactured silicon. Back then I heard Samsung was paying close attention to Apple's requirements and fed that experience into its own SoC and smartphone design.

With a couple of successful iPhone generations under its belt, Apple set its sights much higher. Steve Jobs hired some of the brightest minds in CPU and GPU design and kept them close by. They would influence silicon supplier roadmaps as well as help ensure Apple was on the forefront of performance. Remember that CPU and GPU makers don't just set their own roadmaps, they ask their biggest customers and software vendors what they would like to see. As Apple grew in size, Apple's demands carried more weight.

Unlike the desktop/notebook CPU space, there was no truly aggressive SoC provider. The why is easy to understand. Mobile SoCs sell for $14 - $30, while the desktop and notebook CPUs that Intel invests so heavily in sell for around 10x that, despite being 1 - 4x the physical die size of their cheaper mobile counterparts. In short, most SoC providers felt that no one would be willing to pay for a big, high performance chip, so no one made them. Ultimately this led to a lot of embarassment, with companies like NVIDIA being known for their graphics prowess losing when it came to SoC GPU performance.

Realizing the lack of an Intel-like player in the mobile SoC space, Apple took it upon itself to build the silicon it needed to power the iPhone and iPad. By controlling its own SoC destiny it could achieve a level of vertical integration that no OEM has enjoyed in recent history. Apple would be able to define the experience it wanted, then work with the device, OS, application and SoC teams to deliver that experience. It's a very tempting thing to strive for, the risks are plentiful but the upside is tremendous.

The A4 SoC was Apple's first branded solution, although internally it still leveraged licensed IP blocks from ARM (Cortex A8) and Imagination Technologies (PowerVR SGX 535). Its replacement, the A5, moved to a dual-core Cortex A9 setup with a much beefier GPU from Imagination (PowerVR SGX 543MP2). For the 3rd generation iPad, Apple doubled up GPU core count and built the largest ARM based mobile SoC we've seen deployed.

When I first looked at the A4, I wrote the following:

Apple is not a microprocessor company, nor does Apple want to toss its hat in with the likes of Intel, NVIDIA, Qualcomm and TI as an SoC maker. History has shown us that the only way to be a successful microprocessor company is to be able to subsidize the high cost of designing a powerful architecture over an extremely large install base. That's why x86 survived, and it's why the ARM business model works.

Designing high performance SoCs just for use in the iPad and iPhone just doesn't make sense. In the short term, perhaps, but in the long run it would mean that Apple would have to grow the microprocessor side of its business considerably. That means tons of engineers, more resources that aren't product focused, and honestly re-inventing the wheel a lot.

The fact that the A4 appears to be little more than a 45nm, 1GHz Cortex A8 paired with a PowerVR SGX GPU tells me that Apple isn't off its rocker. I don't exactly know what Apple is doing with all of these CPU and GPU engineers in house, but licensing tech from the companies who have experience in building the architectures is still on the menu.

While I still believe that, long term, Apple will either have to commit to being a full blown chip company or buy processors from whoever ends up dominating the mobile SoC industry it's clear that for the foreseeable future Apple will be a device company that also makes mobile SoCs. Given the state of the mobile SoC space at this point, I can't blame Apple for wanting to build its own chips.

Apple SoC Evolution
  Apple A4 Apple A5 Apple A5r2 Apple A5X Apple A6
Intro Date 2010 2011 2012 2012 2012
Intro Product iPad iPad 2 iPad 2 iPad 3 iPhone 5
Product Targets iPad/iPhone 4 iPad 2/iPhone 4S iPad 2/iPhone 4S iPad 3 ?
CPU ARM Cortex A8 2 x ARM Cortex A9 2 x ARM Cortex A9 2 x ARM Cortex A9 2 x Apple Swift
CPU Frequency 1GHz/800MHz (iPad/iPhone) 1GHz/800MHz (iPad/iPhone) 1GHz/800MHz (iPad/iPhone) 1GHz 1.3GHz
GPU PowerVR SGX 535 PowerVR SGX 543MP2 PowerVR SGX 543MP2 PowerVR SGX 543MP4 PowerVR SGX 543MP3
Memory Interface 32-bit LPDDR2 2 x 32-bit LPDDR2 2 x 32-bit LPDDR2 4 x 32-bit LPDDR2 2 x 32-bit LPDDR2
Manufacturing Process Samsung 45nm LP Samsung 45nm LP Samsung 32nm LP HK+MG Samsung 45nm LP Samsung 32nm LP HK+MG

Apple's A6 is the next step in the company's evolution. Although it continues to license graphics IP from Imagination Technologies (PowerVR SGX 543MP3) and it licenses the ARMv7 instruction set from ARM, it is the first SoC to feature Apple designed CPU cores. The A6 is also the second Apple SoC to be built using Samsung's 32nm LP High-K + Metal Gate transistors. Thanks to UBM Tech Insights and Chipworks we have some great die shots of A6 as well as an accurate die size.

I've updated our die size comparison to put the A6 in perspective:

The new SoC is smaller than the A5 used in the iPhone 4S, but it's built on a newer process which will have some added costs associated with it (at least initially). Over time I would expect A6 pricing to drop below that of the A5, although initially there may not be much (if any at all) cost savings. Note that Apple's 32nm A5r2 is very close in size to the A6, which made it a great test part for Samsung's 32nm process. Apple likely caught the bulk of its process issues on A5r2, making an aggressive ramp for A6 on 32nm much easier than it would have been previously. It's clear that the Apple SoC team benefitted from the practical experience of its members.

Putting the A6 in perspective, we have the usual table we throw in our CPU reviews:

CPU Specification Comparison
CPU Manufacturing Process Cores Transistor Count Die Size
Apple A6 32nm 2 ? 97mm2
Apple A5X 45nm 2 ? 163mm2
Apple A5r2 32nm 2 ? 71mm2
Apple A5 45nm 2 ? 122mm2
Intel Ivy Bridge HE-4 (GT2) 22nm 4 1.4B 160mm2
Intel Ivy Bridge HM-4 (GT1) 22nm 4 ? 133mm2
Intel Ivy Bridge H-2 (GT2) 22nm 2 ? 118mm2
Intel Ivy Bridge M-2 (GT1) 22nm 2 ? 94mm2
Intel Sandy Bridge 4C 32nm 4 995M 216mm2
Intel Sandy Bridge 2C (GT1) 32nm 2 504M 131mm2
Intel Sandy Bridge 2C (GT2) 32nm 2 624M 149mm2
NVIDIA Tegra 3 40nm 4+1 ? ~80mm2
NVIDIA Tegra 2 40nm 2 ? 49mm2

Although the A6 is significantly smaller than the mammoth A5X, it's still quite large by mobile SoC standards. At 97mm2 Apple's A6 is slightly larger than a dual-core Ivy Bridge with GT1 graphics. Granted that's not a very impressive part, but it's still a modern chip that Intel sells for over $100. I'm still not sure what the die size sweet spot is for a smartphone/tablet SoC, perhaps something around 120mm2? I just can't see the 200mm2 chips we love on the desktop being the right fit for ultra mobile.


A6 die photo courtesy UBM Tech Insights

Looking at the A6 die we clearly see the two CPU cores, three GPU cores and 2 x 32-bit LPDDR2 memory interfaces. The Chipworks photo shows the GPU cores a bit better:


Apple A6 die photo courtesy Chipworks

Chipworks was first to point out that Apple's custom CPU cores appeared to be largely laid out by hand vs. using automated tools. Not using automated layout for all parts of a CPU isn't unusual (Intel does it all the time), but it is unusual to see in an ARM based mobile SoC. Shortly after the iPhone 5's launch we confirmed that the A6 SoC featured Apple's first internally designed ARM CPU cores. As a recap there are two types of ARM licensees: architecture and processor. A processor license gives you the right to take an ARM designed CPU core and integrate it into your SoC. Apple licensed ARM's Cortex A9 design in the A5/A5X SoCs for example. An architecture license gives you the right to design your own core that implements an ARM instruction set. Marvell and Qualcomm are both examples of ARM architecture licensees.

For years it's been rumored that Apple has held an ARM architecture license. With the A6 we now have conclusive proof. The question is, what does Apple's first custom ARM CPU core look like? Based on Apple's performance claims we know it's more than a Cortex A9. But to find out what the architecture looks like at a high level we had to do a lot of digging.

 

Build Quality Issues, Scuffgate Decoding Swift
Comments Locked

276 Comments

View All Comments

  • Calista - Sunday, October 21, 2012 - link

    English is not my native language (as I'm sure you have noticed) and so the flow in the language is far from flawless. But I still believe my opinions are valid and that the review was too long-winded.
  • Teknobug - Wednesday, October 17, 2012 - link

    I live in a big city and I don't know a single person that went and got the iPhone 5, most are happy with the iPhone 4 or whatever phone they're using, I don't see what's so great about the iPhone 5 other than it being built better than the iPhone 4's double sided glass structure (I've seen people drop their's on the train or sidewalk and it shattering on both sides!).

    And what now? iPad mini? I thought Apple wasn't interested in the 6-7" tablet market, Steve Jobs said 9" is small enough. I know Apple tried a 6" tablet a decade ago but the market wasn't read for it back then.
  • name99 - Wednesday, October 17, 2012 - link

    You know what AnandTech REALLY needs now?
    A comment moderation system like Ars Technica, so that low-content comments and commenters (like the above) can be suppressed.

    Teknobug is a PERFECT example of Ars' Troll Type #1: "Son of the "I don't even own a TV" guy: "

    This is the poster who thinks other people will find it interesting that he cares nothing about their discussion or their interests, and in fact judges himself as somehow morally superior as a result. The morphology of this on Ars Technica includes people popping into threads about Windows 8 to proclaim how they will never use Windows, people popping into threads about iOS 6 to proclaim that they never have and never will buy an Apple product, and people popping into Android related threads and claiming that they will never purchase "crappy plastic phones." In these cases, the posters have failed to understand that no one really cares what their personal disposition is on something, if they have nothing to add to the discussion.
  • ratte - Wednesday, October 17, 2012 - link

    yeah, my thoughts exactly.
  • worldbfree4me - Wednesday, October 17, 2012 - link

    I finished reading the review a few moments ago. Kudos again for a very thorough review, however I do a have a few questions and points that I would like to ask and make.

    Am I wrong to say, Great Job on Apple finally catching up to the Android Pack in terms of overall performance? The GS3, HTC X debuted about 6 months ago yes?

    Have these benchmark scores from the competing phones been updated to reflect the latest OS updates from GOOG such as OS 4.1.X aka Jelly Bean?

    Clearly the LG Optimus G is a preview of the Nexus 4,complete with a modern GPU In Adreno 320 and 2GB ram. I think based on history, the Nexus 4 will again serve as a foundation for all future Androids to follow. But again, good Job on Apple finally catching up to Android with the caveat being, iOS only has to push its performance to a 4inch screen akin to a 1080p LCD monitor verses a true gamers 1440p LCD Home PC setup. Ciao
  • Zinthar - Thursday, October 18, 2012 - link

    Caught up and passed, actually (if you were actually reading the review). As far as graphics are concerned, no smartphone has yet to eclipse the 4S's 543MP2 other than, of course, the iPhone 5.

    I have no idea what you're going on about with the Adreno 320, because that only gets graphics performance up to about the level of the PowerVR SGX 543MP2. Please see Anand's preview: http://www.anandtech.com/show/6112/qualcomms-quadc...
  • yottabit - Wednesday, October 17, 2012 - link

    Anand, as a Mech-E, I think somewhere the anodization facts in this article got very wonky

    I didn't have time to read thoroughly but I saw something about the anodized layer equaling half the material thickness? The idea of having half a millimeter anodized is way off the mark

    Typically there are two types of anodizing I use: regular, and "hard coat anodize" which is much more expensive

    If the iPhone is scuffing then it's definitely using regular anodizing, and the thickness of that layer is likely much less than .001" or one thousandth of an inch. More on the order of a ten-thousandth of an inch, actually. The thickness of traditional anodizing is so negligible that in fact most engineers don't even need to compensate for it when designing parts.

    Hard-coat anodize is a much more expensive process and can only result in a few darker colors, whereas normal anodizing has a pretty wide spectrum. Hard-coat thicknesses can be substantial, in the range of .001" to .003". This usually must be compensated for in the design process. Hard coat anodize results in a much flatter looking finish than typical anodize, and is also pretty much immune to scratches of any sort.

    Aluminum oxide is actually a ceramic which is harder than steel. So having a sufficient thickness of anodize can pretty much guarantee it won't be scratched under normal operating conditions. However it's much cheaper and allows more colors to do a "regular" anodize

    When I heard about scuffgate I immediately thought one solution would be to have a hardcoat anodize, but it would probably be cost prohibitive, and would alter the appearance significantly
  • guy007 - Wednesday, October 17, 2012 - link

    A little late to the party with the review, the iPhone 6 is almost out now...
  • jameskatt - Wednesday, October 17, 2012 - link

    Anand is pessimistic about Apple's ability to keep creating its own CPUs every year. But realize that the top two smartphone manufacturers (Apple and Samsung) are CRUSHING the competition. And BOTH create their own CPUs.

    Apple has ALWAYS created custom chips for its computers - except for a few years when Steve Jobs accidentally let their chip engineers go when they switched to Intel and Intel's motherboard designs.

    Apple SAVES a lot of money by designing its own chips because it doesn't have to pay the 3rd party profit on each chip.

    Apple PREVENTS Samsung from spying on its chip designs and giving the data to its own chip division to add to its own designs. This is a HUGE win given Samsung's copycat mentality.

    Apple can now always be a step ahead of the competition by designing its own chips. Realize that others will create copies of the ARM A15. But only Apple can greatly improve on the design. Apple, for example, greatly improved the memory subsystem on its own ARM chips. This is a huge weakness on otherARM chips. Apple can now custom design the power control as well - prolonging battery life even more. Etc. etc.
  • phillyry - Sunday, October 21, 2012 - link

    Good points re: copycat and profit margin savings.

    I've always been baffled by the fact that Apple outsources their part manufacturing to the competition. I know that Samsung is a huge OEM player but they are stealing Apple's ideas. They are doing a very good job of it and now improving on those ideas and techs, which is good for the consumer but still seems completely illogical to me from Apple's perspective. Must be the 20/20 hindsight kicking in again.

Log in

Don't have an account? Sign up now