Decoding Swift

Section by Anand Shimpi

Apple's A6 provided a unique challenge. Typically we learn about a new CPU through an architecture disclosure by its manufacturer, tons of testing on our part and circling back with the manufacturer to better understand our findings. With the A6 there was no chance Apple was going to give us a pretty block diagram or an ISSCC paper on its architecture. And on the benchmarking side, our capabilities there are frustratingly limited as there are almost no good smartphone benchmarks. Understanding the A6 however is key to understanding the iPhone 5, and it also gives us a lot of insight into where Apple may go from here. A review of the iPhone 5 without a deep investigation into the A6 just wasn't an option.

The first task was to know its name. There's the old fantasy that knowing something's name gives you power over it, and the reality that it's just cool to know a secret code name. A tip from a reader across the globe pointed us in the right direction (thanks R!). Working backwards through some other iOS 6 code on the iPhone 5 confirmed the name. Apple's first fully custom ARM CPU core is called Swift.

Next we needed to confirm clock speed. Swift's operating frequency would give us an idea of how much IPC has improved over the Cortex A9 architecture. Geekbench was updated after our original iPhone 5 performance preview to more accurately report clock speed (previously we had to get one thread running in the background then launch Geekbench to get a somewhat accurate frequency reading). At 1.3GHz, Swift clearly ran at a higher frequency than the 800MHz Cortex A9 in Apple's A5 but not nearly as high as solutions from Qualcomm, NVIDIA, Samsung or TI. Despite the only 62.5% increase in frequency, Apple was promising up to a 2x increase in performance. It's clear Swift would have to be more than just a clock bumped Cortex A9. Also, as Swift must remain relevant through the end of 2013 before the next iPhone comes out, it had to be somewhat competitive with Qualcomm's Krait and ARM's Cortex A15. Although Apple is often talked about as not being concerned with performance and specs, the truth couldn't be farther from it. Shipping a Cortex A9 based SoC in its flagship smartphone through the end of 2013 just wouldn't cut it. Similarly, none of the SoC vendors would have something A15-based ready in time for volume production in Q3 2012 which helped force Apple's hand in designing its own core.

With a codename and clock speed in our hands, we went about filling in the blanks.

Some great work on behalf of Chipworks gave us a look at the cores themselves, which Chipworks estimated to be around 50% larger than the Cortex A9 cores used in the A5.


Two Apple Swift CPU cores, photo courtesy Chipworks, annotations ours


Two ARM Cortex A9 cores , photo courtesy Chipworks, annotations ours

Looking at the die shots you see a much greater logic to cache ratio in Swift compared to ARM's Cortex A9. We know that L1/L2 cache sizes haven't changed (32KB/1MB, respectively) so it's everything else that has grown in size and complexity.

The first thing I wanted to understand was how much low level compute performance has changed. Thankfully we have a number of microbenchmarks available that show us just this. There are two variables that make comparisons to ARM's Cortex A9 difficult: Swift presumably has a new architecture, and it runs at a much higher clock speed than the Cortex A9 in Apple's A5 SoC. For the tables below you'll see me compare directly to the 800MHz Cortex A9 used in the iPhone 4S, as well as a hypothetical 1300MHz Cortex A9 (1300/800 * iPhone 4S result). The point here is to give me an indication of how much performance has improved if we take clock speed out of the equation. Granted the Cortex A9 won't see perfect scaling going from 800MHz to 1300MHz, however most of the benchmarks we're looking at here are small enough to fit in processor caches and should scale relatively well with frequency.

Our investigation begins with Geekbench 2, which ends up being a great tool for looking at low level math performance. The suite is broken up into integer, floating point and memory workloads. We'll start with the integer tests. I don't have access to Geekbench source but I did my best to map the benchmarks to the type of instructions and parts of the CPU core they'd be stressing in the descriptions below.

Geekbench 2
Integer Tests Apple A5 (2 x Cortex A9 @ 800MHz Apple A5 Scaled (2 x Cortex A9 @ 1300MHz Apple A6 (2 x Swift @ 1300MHz Swift / A9 Perf Advantage @ 1300MHz
Blowfish 10.7 MB/s 17.4 MB/s 23.4 MB/s 34.6%
Blowfish MT 20.7 MB/s 33.6 MB/s 45.6 MB/s 35.6%
Text Compression 1.21 MB/s 1.97 MB/s 2.79 MB/s 41.9%
Text Compression MP 2.28 MB/s 3.71 MB/s 5.19 MB/s 40.1%
Text Decompression 1.71 MB/s 2.78 MB/s 3.82 MB/s 37.5%
Text Decompression MP 2.84 MB/s 4.62 MB/s 5.88 MB/s 27.3%
Image Compression 3.32 Mpixels/s 5.40 Mpixels/s 7.31 Mpixels/s 35.5%
Image Compression MP 6.59 Mpixels/s 10.7 Mpixels/s 14.2 Mpixels/s 32.6%
Image Decompression 5.32 Mpixels/s 8.65 Mpixels/s 12.4 Mpixels/s 43.4%
Image Decompression MP 10.5 Mpixels/s 17.1 Mpixels/s 23.0 Mpixels/s 34.8%
LUA 215.4 Knodes/s 350.0 Knodes/s 455.0 Knodes/s 30.0%
LUA MP 425.6 Knodes/s 691.6 Knodes/s 887.0 Knodes/s 28.3%
Average - - - 37.2%

The Blowfish test is an encryption/decryption test that implements the Blowfish algorithm. The algorithm itself is fairly cache intensive and features a good amount of integer math and bitwise logical operations. Here we see the hypothetical 1.3GHz Cortex A9 would be outpaced by Swift by around 35%. In fact you'll see this similar ~30% increase in integer performance across the board.

The text compression/decompression tests use bzip2 to compress/decompress text files. As text files compress very well, these tests become great low level CPU benchmarks. The bzip2 front end does a lot of sorting, and is thus very branch as well as heavy on logical operations (integer ALUs used here). We don't know much about the size of the data set here but I think it's safe to assume that given the short run times we're not talking about compressing/decompressing all of the text in Wikipedia. It's safe to assume that these tests run mostly out of cache. Here we see a 38 - 40% advantage over a perfectly scaled Cortex A9. The MP text compression test shows the worst scaling out of the group at only 27.3% for Swift over a hypothetical 1.3GHz Cortex A9. It is entirely possible we're hitting some upper bound to simultaneous L2 cache accesses or some other memory limitation here.

The image compression/decompression tests are particularly useful as they just show JPEG compression/decompression performance, a very real world use case that's often seen in many applications (web browsing, photo viewer, etc...). The code here is once again very integer math heavy (adds, divs and muls), with some light branching. Performance gains in these tests, once again, span the 33 - 43% range compared to a perfectly scaled Cortex A9.

The final set of integer tests are scripted LUA benchmarks that find all of the prime numbers below 200,000. As with most primality tests, the LUA benchmarks here are heavy on adds/muls with a fair amount of branching. Performance gains are around 30% for the LUA tests.

On average, we see gains of around 37% over a hypothetical 1.3GHz Cortex A9. The Cortex A9 has two integer ALUs already, so it's possible (albeit unlikely) that Apple added a third integer ALU to see these gains. Another potential explanation is that the 3-wide front end allowed for better utilization of the existing two ALUs, although it's also unlikely that we see better than perfect scaling simply due to the addition of an extra decoder. If it's not more data being worked on in parallel, it's entirely possible that the data is simply getting to the execution units faster.

Let's keep digging.

MP

Geekbench 2
FP Tests Apple A5 (2 x Cortex A9 @ 800MHz Apple A5 Scaled (2 x Cortex A9 @ 1300MHz Apple A6 (2 x Swift @ 1300MHz Swift / A9 Perf Advantage @ 1300MHz
Mandlebrot 223 MFLOPS 362 MFLOPS 397 MFLOPS 9.6%
Mandlebrot MP 438 MFLOPS 712 MFLOPS 766 MFLOPS 7.6%
Dot Product 177 MFLOPS 288 MFLOPS 322 MFLOPS 12.0%
Dot Product MP 353 MFLOPS 574 MFLOPS 627 MFLOPS 9.3%
LU Decomposition 171 MFLOPS 278 MFLOPS 387 MFLOPS 39.3%
LU Decomposition MP 348 MFLOPS 566 MFLOPS 767 MFLOPS 35.6%
Primality 142 MFLOPS 231 MFLOPS 370 MFLOPS 60.3%
Primality MP 260 MFLOPS 423 MFLOPS 676 MFLOPS 60.0%
Sharpen Image 1.35 Mpixels/s 2.19 Mpixels/s 4.85 Mpixels/s 121%
Sharpen Image MP 2.67 Mpixels/s 4.34 Mpixels/s 9.28 Mpixels/s 114%
Blur Image 0.53 Mpixels/s 0.86 Mpixels/s 1.96 Mpixels/s 128%
Blur Image MP 1.06 Mpixels/s 1.72 Mpixels/s 3.78 Mpixels/s 119%
Average - - - 61.6%

The FP tests for Geekbench 2 provide some very interesting data. While we saw consistent gains of 30 - 40% over our hypothetical 1.3GHz Cortex A9, Swift behaves much more unpredictably here. Let's see if we can make sense of it.

The Mandlebrot benchmark simply renders iterations of the Mandlebrot set. Here there's a lot of floating point math (adds/muls) combined with a fair amount of branching as the algorithm determines whether or not values are contained within the Mandlebrot set. It's curious that we don't see huge performance scaling here. Obviously Swift is faster than the 800MHz Cortex A9 in Apple's A5, but if the A5 were clocked at the same 1.3GHz and scaled perfectly we only see a 9.6% increase in performance from the new architecture. The Cortex A9 only has a single issue port to its floating point hardware that's also shared by its load/store hardware - this data alone would normally indicate that nothing has changed here when it comes to Swift. That would be a bit premature though...

The Dot Product test is simple enough, it computes the dot product of two FP vectors. Once again there are a lot of FP adds and muls here as the dot product is calculated. Overall performance gains are similarly timid if we scale up the Cortex A9's performance: 9 - 12% increase at the same frequency doesn't sound like a whole lot for a brand new architecture.

The LU Decomposition tests factorize a 128 x 128 matrix into a product of two matrices. The sheer size of the source matrix guarantees that this test has to hit the 1MB L2 cache in both of the architectures that we're talking about. The math involved are once again FP adds/muls, but the big change here appears to be the size of the dataset. The performance scales up comparatively. The LU Decomposition tests show 35 - 40% gains over our hypothetical 1.3GHz Cortex A9.

The Primality benchmarks perform the first few iterations of the Lucas-Lehmar test on a specific Mersenne number to determine whether or not it's prime. The math here is very heavy on FP adds, multiplies and sqrt functions. The data set shouldn't be large enough to require trips out to main memory, but we're seeing scaling that's even better than what we saw in the LU Decomposition tests. The Cortex A9 only has a single port for FP operations, it's very possible that Apple has added a second here in Swift. Why we wouldn't see similar speedups in the Mandlebrot and Dot Product tests however could boil down to the particular instruction mix used in the Primarily benchmark. The Geekbench folks also don't specify whether we're looking at FP32 or FP64 values, which could also be handled at different performance levels by the Swift architecture vs. Cortex A9.

The next two tests show the biggest gains of the FP suite. Both the sharpen and blur tests apply a convolution filter to an image stored in memory. The application of the filter itself is a combination of matrix multiplies, adds, divides and branches. The size of the data set likely hits the data cache a good amount.

We still haven't gained too much at this point. Simple FP operations don't see a huge improvement in performance over a perfectly scaled Cortex A9, while some others show tremendous gains. There seems to be a correlation with memory accesses which makes sense given what we know about Swift's memory performance. Improved memory performance also lends some credibility to the earlier theory about why integer performance goes up by so much: data cache access latency could be significantly improved.

The A6 SoC Custom Code to Understand a Custom Core
Comments Locked

276 Comments

View All Comments

  • darwinosx - Tuesday, October 16, 2012 - link

    The iPhone 5 display is better than any current Android display.
    But Motorola and Android if you want a company that is dying and being sold and a copycat cheap phone with no service and support.
  • V-Money - Tuesday, October 16, 2012 - link

    Your wisdom and informative argument adds tremendous value to this post. For the record though, the OP said specifically battery life and 720p display, so the response was relevant.

    The rest of your post is petty, get over yourself. If you are going to play the copycat card you should have done it before Apple decided to go with a bigger screen and use a (eerily similar) notification bar to what Android phones have had for years.

    As for quality (of display or otherwise), that is subjective analysis and considering that Apple only releases one phone at a time and Android manufacturers many, its a stupid argument for anyone to make. Case and point, I can find many android phones that are much more terrible than the iPhone, but I can also find many that are better. The iPhone is a decent phone, but its not for everyone. Every consumer has their preference.

    My point being there is not one-size-fits-all phone, so quit acting high and mighty with your close mindedness. You are not better than those around you because you bought into Apple's marketing, you are just a fool dealing with the first world problem of living such a meaningless existence that you have to hold on to the imaginary power an inanimate object pretends to give to you.
  • Alucard291 - Tuesday, October 16, 2012 - link

    I feel that your argument may be too good for him to reply to :)

    He seems awfully angry :D
  • crankerchick - Tuesday, October 16, 2012 - link

    Great reply. If there's one place I just want to exchange comments without playing the "my toy is better than yours" game, it's here on AnandTech.
  • Gradly - Wednesday, October 17, 2012 - link

    I'm sick of ppl comparing iPhone to other devices. I'm sick of those telling you iPhone borrowed the notifications slider form android and skipping the myriad of things that other borrowed form iPhone. Apple has always said that "we are not the first but we do it the best". I'm sick of those who still don't realize that before iPhone ppl were living in caves actually.

    I'm an Apple lover not an Apple fanboy. I just adore the design, aesthetics, and GUI of Apple devices.
  • Penti - Wednesday, October 17, 2012 - link

    It's sadly Apple that goes and patent UI-elements to use against their competitors that is why it's always brought up. It would be totally unnecessary otherwise. You might look at who's the inspiration otherwise and it's often not Apple. In reality we had capacitive touch screens (it's not Apples tech of course) before, app store before, Android even had an SDK out before Apple. Competitors like Symbian/Nokia, HP WebOS, and Blackberry are even allowed to use stuff like bounce back effect even without (or before) any agreement with Apple. They should have credit but they didn't all the sudden bring out their device with what we now call smartphone features, it lacked most functions at first and slowly iterated, it did a lot poorer in many areas then it's competitors was doing even before iPhone and the first few years it also showed in sales numbers which were not high at the first 2-3 years. It did show us how important a good platform was. Guys like Rubin had already figured that out though. So I'm not sure what they would borrow. Full WebKit-browsers on mobile is a good example of stuff they are co-developing but it was out in Nokia devices in 2006, netfront and Opera was never good alternatives to build into your platform. Stock Android don't have the bounce back effect, UI's looking like Apples and so on. Not even TouchWiz on Samsung's tablets looks like or infringes anything (design-wise) by Apple. They clearly have their own ideas. They are not the "me too", others might try to emulate them more in a business sense though. But they will be punished by the market by their execution instead of by Apple. It's not like any of the major players are fruit ninja-clones though.

    iPhone was desperately rudimentary at first. It didn't do applications and the web, messaging, photos etc better then anybody. What they did good was to iterate and improve. They take enterprise / corporate customers more seriously then Microsoft and so on in this field. Even if it took some time for them to get there. So they do plenty of good. It's a good platform, but it's not like they gave their competitors their blueprints for their devices / os of today back in 2007 and both have made many improvements. Well maybe not Microsoft but it takes a few years to start over. Apple has even got into hardware (components) a bit. Commoditization and convergence has reached far beyond the mobile field. That's great even if Apple won't enter them. Still don't know why any competitor would like to turn themselves into a retail giant and employ mostly store staff as Apple does – Microsoft should start doing what they are good at instead. Google would be the most evil company in the world if they had started to patent and sue based on UI-features and methods. Or if they really tried to stop Bing and Bing Maps (and getting it banned in some markets) for example. It doesn't really matter who was first and who invented what if you take it to court were that doesn't really count and that creates a lot of BS surrounding the whole issue and companies involved that is largely unnecessary. But the real silly thing is why they fight. It's not based on IPR, it's basically that they want to be alone in doing whatever, even if they can't really make that claim to have sole rights to something. But ultimately courts do get that under control even when corporate leaders turn to fighting outside of releasing product.
  • slickr - Tuesday, October 16, 2012 - link

    LOL. Don't make me laugh. It has still the worst display and has had the worst display for at least 3 years.
  • A5 - Tuesday, October 16, 2012 - link

    Your response is just as dumb as his. The iPhones have excellent displays.
  • medi01 - Wednesday, October 17, 2012 - link

    None of the iPhones have anything to compare with AMOLEDs, on top of having idiotic resolution.

    On tablet space, only iPad 3 matched color gamut of THE FIRST Samsung Galaxy Tab.
  • thunng8 - Wednesday, October 17, 2012 - link

    How does 67.5% of sRGB on the galaxy tab 10.1 match the 94.4% on the ipad 3?

    http://www.anandtech.com/show/5688/apple-ipad-2012...

Log in

Don't have an account? Sign up now