Modifying a Krait Platform: More Complicated

Modifying the Dell XPS 10 is a little more difficult than Acer's W510 and Surface RT. In both of those products there was only a single inductor in the path from the battery to the CPU block of the SoC. The XPS 10 uses a dual-core Qualcomm solution however. Ever since Qualcomm started doing multi-core designs it has opted to use independent frequency and voltage planes for each core. While all of the A9s in Tegra 3 and both of the Atom cores used in the Z2760 run at the same frequency/voltage, each Krait core in the APQ8060A can run at its own voltage and frequency. As a result, there are two power delivery circuits that are needed to feed the CPU cores. I've highlighted the two inductors Intel lifted in orange:

Each inductor was lifted and wired with a 20 mΩ resistor in series. The voltage drop across the 20 mΩ resistor was measured and used to calculate CPU core power consumption in real time. Unless otherwise stated, the graphs here represent the total power drawn by both CPU cores.

Unfortunately, that's not all that's necessary to accurately measure Qualcomm CPU power. If you remember back to our original Krait architecture article you'll know that Qualcomm puts its L2 cache on a separate voltage and frequency plane. While the CPU cores in this case can run at up to 1.5GHz, the L2 cache tops out at 1.3GHz. I remembered this little fact late in the testing process, and we haven't yet found the power delivery circuit responsible for Krait's L2 cache. As a result, the CPU specific numbers for Qualcomm exclude any power consumed by the L2 cache. The total platform power numbers do include it however as they are measured at the battery.

The larger inductor in yellow feeds the GPU and it's instrumented using another 20 mΩ resistor.

Visualizing Krait's Multiple Power/Frequency Domains

Qualcomm remains adament about its asynchronous clocking with multiple voltage planes. The graph below shows power draw broken down by each core while running SunSpider:

SunSpider is a great benchmark to showcase exactly why Qualcomm has each core running on its own power/frequency plane. For a mixed workload like this, the second core isn't totally idle/power gated but it isn't exactly super active either. If both cores were tied to the same voltage/frequency, the second core would have higher leakage current than in this case. The counter argument would be that if you ran the second core at its max frequency as well it would be able to complete its task quicker and go to sleep, drawing little to no power. The second approach would require a very fast microcontroller to switch between v/f modes and it's unclear which of the two would offer better power savings. It's just nice to be able to visualize exactly why Qualcomm does what it does here.

On the other end of the spectrum however is a benchmark like Kraken, where both cores are fairly active and the workload is balanced across both cores:

 

Here there's no real benefit to having two independent voltage/frequency planes, both cores would be served fine by running at the same voltage and frequency. Qualcomm would argue that the Kraken case is rare (single threaded performance still dominates most user experience), and the power savings in situations like SunSpider are what make asynchronous clocking worth it. This is a much bigger philosophical debate that would require far more than a couple of graphs to support and it's not one that I want to get into here. I suspect that given its current power management architecture, Qualcomm likely picked the best solution possible for delivering the best possible power consumption. It's more effort to manage multiple power/frequency domains, effort that I doubt Qualcomm would put in without seeing some benefit over the alternative. That being said, what works best for a Qualcomm SoC isn't necessarily what's best for a different architecture.

Introduction Krait: Idle Power
Comments Locked

140 Comments

View All Comments

  • extide - Friday, January 4, 2013 - link

    When will you post an article about Bay Trail / Valley View?? Usually you guys are pretty fast to post stuff about topics like this yet I have seen some info on other sites already...
  • jpcy - Friday, January 4, 2013 - link

    ...which I bet CISC users thought had ended about 18 years ago...

    It's good to see a resurgence of this highly useful, extremely low-power and very hardy British CPU platform.

    I remember back in the day when ARMs were used in the Acorn computers (possibly too long ago for most to remember, now - I still have an A7000 and a RISC PC with both a StrongARM and a DX2-66 lol) was at war with Intel's Pentium CPU range and AMD's K6's, boasting an almost 1:1 ration of MIPS:MHz - Horsepower for your money (something Intel and AMD were severely lacking in, if I remember correctly.)

    And now, well, who'dve thought it... These ARM CPUs are now in nearly everything we use... Phones, smartphones, tablets, notebooks...

    Suppose I was right in the argument with my mate in school afterall... RISC, superior technology (IMHO) may well take over, yet!
  • nofumble62 - Friday, January 4, 2013 - link

    No performance advantage, no battery life advantage. Why anyone would bother with incompatible software?
  • sseemaku - Friday, January 4, 2013 - link

    Looks like people have changed religion from AMD to ARM. Thats what I see from some comments.
  • mugiebahar - Saturday, January 5, 2013 - link

    Yeah n no. They wanted a no paid opinions to screw with the outcome. But Intel hype won over real life .

    Intel better and will get better - yes
    Any chance they will compete (performance and PRICE) and legacy. Support to phone apps - Never in the near future which is the only time for them.
  • tuxRoller - Saturday, January 5, 2013 - link

    Also, any chance for an actual performance comparison between the platforms?
    Apple's performance and power use look awesome. Better than I had imagined.
    I'd love to see how they compare on the same tests, however.
  • Kogies - Saturday, January 5, 2013 - link

    It appears the war has begun, well two wars in fact. The one you have articulately described, and the oft ensuing war-of-words...

    Thanks Anand, I appreciate the analysis you have given. It is excellent to get to see the level of granularity you have been able to achieve with your balance of art and science, and knowing where to hook into! I am very interested to see how the L2 cache power draw effects the comparison, just a little jitter in my mind. If nothing else, it looks as if the delicate balance of process tech., and desired performance/power may have a greater bearing on this "war" than mere ISA.

    With Krait 300, Haswell, and more A15's this is going to be a tremendous year. Keep up the good work.
  • Torrijos - Saturday, January 5, 2013 - link

    Any chance we could see the same tests run on the latest Apple iPad?
    That way we could have a chance to see what Apple tried to improve compared to the A15 generation.
  • urielshun - Saturday, January 5, 2013 - link

    The whole discussion about ARM and x86 is not important when you go for the ecomonics of each platform. ARM is dirty cheap and works well. It's 1/10th of the price of any current Atom with decent perfomace (talking about RK3066).

    Don't underestimate the Chinese they are having a field day with ARM's pricing model and and have shown amazing chips.

    In 8 years from now all SoC's would have reached the usuable performace and the only thing that will matter will be power and cost of integration.
  • iwod - Saturday, January 5, 2013 - link

    Where are you getting 1/10 of a price from? Unless they are produced on good old 40nm LP Node with Nothing else, or crap included, otherwise there just aren't any Chinese SoC selling for $4

Log in

Don't have an account? Sign up now