Decoding Swift

Section by Anand Shimpi

Apple's A6 provided a unique challenge. Typically we learn about a new CPU through an architecture disclosure by its manufacturer, tons of testing on our part and circling back with the manufacturer to better understand our findings. With the A6 there was no chance Apple was going to give us a pretty block diagram or an ISSCC paper on its architecture. And on the benchmarking side, our capabilities there are frustratingly limited as there are almost no good smartphone benchmarks. Understanding the A6 however is key to understanding the iPhone 5, and it also gives us a lot of insight into where Apple may go from here. A review of the iPhone 5 without a deep investigation into the A6 just wasn't an option.

The first task was to know its name. There's the old fantasy that knowing something's name gives you power over it, and the reality that it's just cool to know a secret code name. A tip from a reader across the globe pointed us in the right direction (thanks R!). Working backwards through some other iOS 6 code on the iPhone 5 confirmed the name. Apple's first fully custom ARM CPU core is called Swift.

Next we needed to confirm clock speed. Swift's operating frequency would give us an idea of how much IPC has improved over the Cortex A9 architecture. Geekbench was updated after our original iPhone 5 performance preview to more accurately report clock speed (previously we had to get one thread running in the background then launch Geekbench to get a somewhat accurate frequency reading). At 1.3GHz, Swift clearly ran at a higher frequency than the 800MHz Cortex A9 in Apple's A5 but not nearly as high as solutions from Qualcomm, NVIDIA, Samsung or TI. Despite the only 62.5% increase in frequency, Apple was promising up to a 2x increase in performance. It's clear Swift would have to be more than just a clock bumped Cortex A9. Also, as Swift must remain relevant through the end of 2013 before the next iPhone comes out, it had to be somewhat competitive with Qualcomm's Krait and ARM's Cortex A15. Although Apple is often talked about as not being concerned with performance and specs, the truth couldn't be farther from it. Shipping a Cortex A9 based SoC in its flagship smartphone through the end of 2013 just wouldn't cut it. Similarly, none of the SoC vendors would have something A15-based ready in time for volume production in Q3 2012 which helped force Apple's hand in designing its own core.

With a codename and clock speed in our hands, we went about filling in the blanks.

Some great work on behalf of Chipworks gave us a look at the cores themselves, which Chipworks estimated to be around 50% larger than the Cortex A9 cores used in the A5.


Two Apple Swift CPU cores, photo courtesy Chipworks, annotations ours


Two ARM Cortex A9 cores , photo courtesy Chipworks, annotations ours

Looking at the die shots you see a much greater logic to cache ratio in Swift compared to ARM's Cortex A9. We know that L1/L2 cache sizes haven't changed (32KB/1MB, respectively) so it's everything else that has grown in size and complexity.

The first thing I wanted to understand was how much low level compute performance has changed. Thankfully we have a number of microbenchmarks available that show us just this. There are two variables that make comparisons to ARM's Cortex A9 difficult: Swift presumably has a new architecture, and it runs at a much higher clock speed than the Cortex A9 in Apple's A5 SoC. For the tables below you'll see me compare directly to the 800MHz Cortex A9 used in the iPhone 4S, as well as a hypothetical 1300MHz Cortex A9 (1300/800 * iPhone 4S result). The point here is to give me an indication of how much performance has improved if we take clock speed out of the equation. Granted the Cortex A9 won't see perfect scaling going from 800MHz to 1300MHz, however most of the benchmarks we're looking at here are small enough to fit in processor caches and should scale relatively well with frequency.

Our investigation begins with Geekbench 2, which ends up being a great tool for looking at low level math performance. The suite is broken up into integer, floating point and memory workloads. We'll start with the integer tests. I don't have access to Geekbench source but I did my best to map the benchmarks to the type of instructions and parts of the CPU core they'd be stressing in the descriptions below.

Geekbench 2
Integer Tests Apple A5 (2 x Cortex A9 @ 800MHz Apple A5 Scaled (2 x Cortex A9 @ 1300MHz Apple A6 (2 x Swift @ 1300MHz Swift / A9 Perf Advantage @ 1300MHz
Blowfish 10.7 MB/s 17.4 MB/s 23.4 MB/s 34.6%
Blowfish MT 20.7 MB/s 33.6 MB/s 45.6 MB/s 35.6%
Text Compression 1.21 MB/s 1.97 MB/s 2.79 MB/s 41.9%
Text Compression MP 2.28 MB/s 3.71 MB/s 5.19 MB/s 40.1%
Text Decompression 1.71 MB/s 2.78 MB/s 3.82 MB/s 37.5%
Text Decompression MP 2.84 MB/s 4.62 MB/s 5.88 MB/s 27.3%
Image Compression 3.32 Mpixels/s 5.40 Mpixels/s 7.31 Mpixels/s 35.5%
Image Compression MP 6.59 Mpixels/s 10.7 Mpixels/s 14.2 Mpixels/s 32.6%
Image Decompression 5.32 Mpixels/s 8.65 Mpixels/s 12.4 Mpixels/s 43.4%
Image Decompression MP 10.5 Mpixels/s 17.1 Mpixels/s 23.0 Mpixels/s 34.8%
LUA 215.4 Knodes/s 350.0 Knodes/s 455.0 Knodes/s 30.0%
LUA MP 425.6 Knodes/s 691.6 Knodes/s 887.0 Knodes/s 28.3%
Average - - - 37.2%

The Blowfish test is an encryption/decryption test that implements the Blowfish algorithm. The algorithm itself is fairly cache intensive and features a good amount of integer math and bitwise logical operations. Here we see the hypothetical 1.3GHz Cortex A9 would be outpaced by Swift by around 35%. In fact you'll see this similar ~30% increase in integer performance across the board.

The text compression/decompression tests use bzip2 to compress/decompress text files. As text files compress very well, these tests become great low level CPU benchmarks. The bzip2 front end does a lot of sorting, and is thus very branch as well as heavy on logical operations (integer ALUs used here). We don't know much about the size of the data set here but I think it's safe to assume that given the short run times we're not talking about compressing/decompressing all of the text in Wikipedia. It's safe to assume that these tests run mostly out of cache. Here we see a 38 - 40% advantage over a perfectly scaled Cortex A9. The MP text compression test shows the worst scaling out of the group at only 27.3% for Swift over a hypothetical 1.3GHz Cortex A9. It is entirely possible we're hitting some upper bound to simultaneous L2 cache accesses or some other memory limitation here.

The image compression/decompression tests are particularly useful as they just show JPEG compression/decompression performance, a very real world use case that's often seen in many applications (web browsing, photo viewer, etc...). The code here is once again very integer math heavy (adds, divs and muls), with some light branching. Performance gains in these tests, once again, span the 33 - 43% range compared to a perfectly scaled Cortex A9.

The final set of integer tests are scripted LUA benchmarks that find all of the prime numbers below 200,000. As with most primality tests, the LUA benchmarks here are heavy on adds/muls with a fair amount of branching. Performance gains are around 30% for the LUA tests.

On average, we see gains of around 37% over a hypothetical 1.3GHz Cortex A9. The Cortex A9 has two integer ALUs already, so it's possible (albeit unlikely) that Apple added a third integer ALU to see these gains. Another potential explanation is that the 3-wide front end allowed for better utilization of the existing two ALUs, although it's also unlikely that we see better than perfect scaling simply due to the addition of an extra decoder. If it's not more data being worked on in parallel, it's entirely possible that the data is simply getting to the execution units faster.

Let's keep digging.

MP

Geekbench 2
FP Tests Apple A5 (2 x Cortex A9 @ 800MHz Apple A5 Scaled (2 x Cortex A9 @ 1300MHz Apple A6 (2 x Swift @ 1300MHz Swift / A9 Perf Advantage @ 1300MHz
Mandlebrot 223 MFLOPS 362 MFLOPS 397 MFLOPS 9.6%
Mandlebrot MP 438 MFLOPS 712 MFLOPS 766 MFLOPS 7.6%
Dot Product 177 MFLOPS 288 MFLOPS 322 MFLOPS 12.0%
Dot Product MP 353 MFLOPS 574 MFLOPS 627 MFLOPS 9.3%
LU Decomposition 171 MFLOPS 278 MFLOPS 387 MFLOPS 39.3%
LU Decomposition MP 348 MFLOPS 566 MFLOPS 767 MFLOPS 35.6%
Primality 142 MFLOPS 231 MFLOPS 370 MFLOPS 60.3%
Primality MP 260 MFLOPS 423 MFLOPS 676 MFLOPS 60.0%
Sharpen Image 1.35 Mpixels/s 2.19 Mpixels/s 4.85 Mpixels/s 121%
Sharpen Image MP 2.67 Mpixels/s 4.34 Mpixels/s 9.28 Mpixels/s 114%
Blur Image 0.53 Mpixels/s 0.86 Mpixels/s 1.96 Mpixels/s 128%
Blur Image MP 1.06 Mpixels/s 1.72 Mpixels/s 3.78 Mpixels/s 119%
Average - - - 61.6%

The FP tests for Geekbench 2 provide some very interesting data. While we saw consistent gains of 30 - 40% over our hypothetical 1.3GHz Cortex A9, Swift behaves much more unpredictably here. Let's see if we can make sense of it.

The Mandlebrot benchmark simply renders iterations of the Mandlebrot set. Here there's a lot of floating point math (adds/muls) combined with a fair amount of branching as the algorithm determines whether or not values are contained within the Mandlebrot set. It's curious that we don't see huge performance scaling here. Obviously Swift is faster than the 800MHz Cortex A9 in Apple's A5, but if the A5 were clocked at the same 1.3GHz and scaled perfectly we only see a 9.6% increase in performance from the new architecture. The Cortex A9 only has a single issue port to its floating point hardware that's also shared by its load/store hardware - this data alone would normally indicate that nothing has changed here when it comes to Swift. That would be a bit premature though...

The Dot Product test is simple enough, it computes the dot product of two FP vectors. Once again there are a lot of FP adds and muls here as the dot product is calculated. Overall performance gains are similarly timid if we scale up the Cortex A9's performance: 9 - 12% increase at the same frequency doesn't sound like a whole lot for a brand new architecture.

The LU Decomposition tests factorize a 128 x 128 matrix into a product of two matrices. The sheer size of the source matrix guarantees that this test has to hit the 1MB L2 cache in both of the architectures that we're talking about. The math involved are once again FP adds/muls, but the big change here appears to be the size of the dataset. The performance scales up comparatively. The LU Decomposition tests show 35 - 40% gains over our hypothetical 1.3GHz Cortex A9.

The Primality benchmarks perform the first few iterations of the Lucas-Lehmar test on a specific Mersenne number to determine whether or not it's prime. The math here is very heavy on FP adds, multiplies and sqrt functions. The data set shouldn't be large enough to require trips out to main memory, but we're seeing scaling that's even better than what we saw in the LU Decomposition tests. The Cortex A9 only has a single port for FP operations, it's very possible that Apple has added a second here in Swift. Why we wouldn't see similar speedups in the Mandlebrot and Dot Product tests however could boil down to the particular instruction mix used in the Primarily benchmark. The Geekbench folks also don't specify whether we're looking at FP32 or FP64 values, which could also be handled at different performance levels by the Swift architecture vs. Cortex A9.

The next two tests show the biggest gains of the FP suite. Both the sharpen and blur tests apply a convolution filter to an image stored in memory. The application of the filter itself is a combination of matrix multiplies, adds, divides and branches. The size of the data set likely hits the data cache a good amount.

We still haven't gained too much at this point. Simple FP operations don't see a huge improvement in performance over a perfectly scaled Cortex A9, while some others show tremendous gains. There seems to be a correlation with memory accesses which makes sense given what we know about Swift's memory performance. Improved memory performance also lends some credibility to the earlier theory about why integer performance goes up by so much: data cache access latency could be significantly improved.

The A6 SoC Custom Code to Understand a Custom Core
Comments Locked

276 Comments

View All Comments

  • Sufo - Tuesday, October 16, 2012 - link

    Agreed. If his goal is to fly the flag for apple (who clearly need no flag flying - look at their stock prices, but i digress...), and discredit its detractors, he's doing an awful job. But then again, I do detect a whiff of troll.
  • Spunjji - Friday, October 19, 2012 - link

    Word.
  • doobydoo - Saturday, October 20, 2012 - link

    Bragging? About being an engineer?

    LOL
  • dagamer34 - Tuesday, October 16, 2012 - link

    If you wanted a "should I upgrade to this phone" review, there are hundreds of those reviews online. But AnandTech is pretty much the only place where you get a definitive review worth reading 5 years from now. They leave no stone unturned.
  • Arbee - Tuesday, October 16, 2012 - link

    Agreed. "Should I upgrade" is covered by literally dozens of newspapers, TV shows, and websites (Engadget, The Verge, Gizmodo, All Things D just to name 4). AT is the home of the 15+ page deep dive, and they do it just as well for Androids and Windows Phones.

    Also, I'm completely positive that if you sent Brian a GS3 with the iPhone 5's camera he'd write about it in exactly the same way. 2 weeks ago DPReview covered the iPhone 5's camera in a very similar way (including the same suggestions on how to avoid the problem, and a demonstration of inducing similar artifacts on the iPhone 4S and a couple of Android handsets). Optics is not a soft science, there is no room for fanboyism.
  • rarson - Wednesday, October 17, 2012 - link

    I totally disagree. He brings up a completely valid point because Anandtech usually separates the reviews from the in-depth tech examinations. There's absolutely no need for the review to be 20 pages when most people are looking for benchmarks and hands-on impressions. Considering the fact that going this in-depth made the review late, it makes no sense at all.

    At least half of this information in this article doesn't even fall under the category of a review.
  • darkcrayon - Tuesday, October 16, 2012 - link

    I think this type of review (hell, the site in general) is directed at people that want the maximum amount of compiled nitty gritty techy details... Notice his review was weeks after the larger more general consumer oriented sites. I think anyone wanting to know whether they should upgrade, that isn't interested in the technical details of the A6, would be better served reading those reviews anyway.

    Anand has said in previous reviews that he felt that iOS was intended to be more of an "appliance" OS. It's a pretty apt comparison of the two actually. That focus is why you can side load and more easily put custom software on Android, and also why you'll need anti-malware software for it before long as well. The point of an appliance is to have a reliable, consistent device that you spend more time using than tweaking.
  • daar - Wednesday, October 17, 2012 - link

    Point taken, darkcrayon.

    I prefer AT's reviews because they do a thorough and unbiased job at detailing/benchmarking and comparing different products. The suggestion was that the info about the SoC be split on it's own. If Intel released a new chip, call it i9, and the first sample was from an Alienware notebook, I would simply be suggesting that the technical info about the chip have it's own post and not be combined with the review of the notebook is all.

    I find it a bit strange that people are suggesting to go to other websites when I made the comment of comparisons to other products, and quite unlike most posts in AT reviews. If I make a comment about a few ATI features not being compared with Nvidia's, I would have been surprised to have people to tell me to go visit Tom's Hardware or the like.

    Not to say there wasn't any comparisons, but rather in contrast to say, for example, the One X review where Brian made the comment of how the construction of the device felt better than the GS3. It felt like punches were being pulled in this review is all.
  • phillyry - Sunday, October 21, 2012 - link

    Anand,

    I would like to know, however, how an Android device serves more as an all purpose device than an iPhone.

    Did you mean because of its customisable skins or because it can do some things that an iPhone cannot - presumably because of Apple's strong hold ('death grip') on the OS?

    This is pretty important to me because I am near the end of the term of my agreement and am in the market for a new 'phone'. I've considered W8P for precisely this reason but am waiting to see if they flop or not. I've always thought of Android as pretty darn similar to iOS but with slightly different interfaces and less user restrictions.

    Is there some other factor that makes an Android any more like a pocket computer, like the future x86 W8P phones will presumably eventually be, and less like an iPhone than I have imagined?
  • phillyry - Sunday, October 21, 2012 - link

    I also took notice of it when Anand referred to the iPhone as an appliance. Your remark saying, "The point of an appliance is to have a reliable, consistent device that you spend more time using than tweaking" would be comforting but I don't think that that's quite how Anand meant it. I was actually quite put off by the term because I think that he meant that the iPhone is made to be more of a tag along device that goes with your other Macs and plays a support role rather than a stand alone device. He pretty much says as much.

    Like I said, I found this a bit off putting but I think he's just saying how he sees it in terms of the respective companies' product lines and agendas. It actually makes a fair bit of sense. I found that when I got an iPhone it made me want an iPad. And then when I got an iPad it made me want a MacBook. Call it what you will but I remember thinking that they should be able to make it so that I can do everything I need to on an iPad but distinctly felt like I really needed a MacBook to really do all that I wanted. It could be argued, along the lines of Anand's original comment, that this is Apple's approach / business model.

    It also points to a distinction between Apple and the other big player that no one in this forum is talking about - Microsoft. Windows 8 appears to be meant to be the exact opposite of this approach. Instead of one device for each purpose it's one device for all purposes. It will be interesting to see if Microsoft's approach with Windows 8 will turn things around or simply flop, at least on the handheld device side of things.

Log in

Don't have an account? Sign up now