The SoC: TI's OMAP 4460

The launch platform for Ice Cream Sandwich was TI's OMAP 4460. Unlike previous Android releases however, it seems that other SoCs will see their ICS ports done in a much quicker manner. It took a very long time for Honeycomb to be ported to other SoCs, whereas a number of companies have already demonstrated ICS running on their hardware (e.g. Intel, NVIDIA). If this is the case going forward, the launch vehicle for a new Android version may not mean what it used to.

The OMAP 4460 is a fairly standard, yet full featured dual-core ARM Cortex A9 SoC. You get two A9 cores complete with MPE (NEON support), behind a shared 1MB L2 cache. The SoC features two 32-bit LPDDR2 memory channels as well. The GPU is provided by Imagination Technologies in the form of a PowerVR SGX 540.

Max clocks for the OMAP 4460 are 1.5GHz for the CPUs and 384MHz for the GPU. As with all SoCs, all final clocks are OEM customizable to hit their desired point on the performance/battery life curve. Google and Samsung settled on 1.2GHz for the cores and 307MHz for the GPU, both exactly 80% of the OMAP 4460's max frequencies. Sprint recently announced its Galaxy Nexus would run at 1.5GHz. It's quite possible that we'll see a jump in GPU clocks there as well since the two may run in lockstep.

From a CPU standpoint the 4460 is competitive with pretty much everything else on the market (A5, Exynos, Tegra 2, Snapdragon S3). The 4460 does have more memory bandwidth than Tegra 2, Tegra 3 and Snapdragon, but it's comparable to Apple's A5 and Samsung's Exynos 4210. It's the GPU that's a bit dated at this point; the PowerVR SGX 540 typically delivers Tegra 2-class performance. A quick look at GLBenchmark and Basemark results echoes our findings:

GLBenchmark 2.1 - Egypt - Offscreen (720p)
 
GLBenchmark 2.1 - Pro - Offscreen (720p)
 
RightWare Basemark ES 2.0 V1 - Taiji
 
RightWare Basemark ES 2.0 V1 - Hoverjet

At 720p, which happens to be the GN's native resolution, the OMAP 4460 is much faster than Tegra 2. It's also important to note just how much faster Tegra 3's GPU is by comparison.

I understand why Google didn't wait for a Krait based SoC, however I don't believe the OMAP 4460 was the best bet given the launch timeframe of the Galaxy Nexus. Based on performance alone, Google should have picked Tegra 3 as the launch platform for ICS. GPU performance is much better than the SGX 540 and there's comparable CPU performance. It's possible that Google needed the memory bandwidth offered by OMAP 4, but we'll find out for sure soon enough as the first Tegra 3 device (ASUS' TF Prime) is slated to get ICS this week.

I'm also less concerned about power consumption being an issue since NVIDIA added full power gating to all of the cores in Tegra 3. With a conservative enough power profile Google could have guaranteed battery life similar to OMAP 4460 out of Tegra 3.

I get the feeling that Google wasn't very pleased with NVIDIA after Honeycomb and chose to work with TI this time around for reasons other than absolute performance. If it weren't for the fact that Tegra 3 and other SoCs appear to be getting ICS in fairly short form I'd be more upset over this decision. To be honest, the choice of SoC simply hurts the Galaxy Nexus as a phone. If I were you, I'd wait for a Krait based device.

The Galaxy Nexus - Hardware and Aesthetics Camera - Stills and Video
Comments Locked

185 Comments

View All Comments

  • CoryS - Thursday, January 19, 2012 - link

    Guys, this is a NEXUS it is a dev device. That primary reason I got it was because of this...better hardware will be right around the corner...but we won't see another Nexus..especially on Verizon for some time.

    It is refreshing to have a community to fix issues OEMS ignore (yes even Apple) for a change. This is my first unlocked device, and i can't see myself ever going back to anything else.
  • medi01 - Friday, January 20, 2012 - link

    Wake up, Smartphone market (worldwide):
    1. Samsung 24%
    2. Apple 18%

    Android vs Apple = 3 vs 1 and gap is raising.

    Most people turn to apple due to FUD, like this article. Google "steppit out of the shade of its competitor" having three times Apple's market share and much more usable interface (try to quickly access settings like wlan/bluetooth/gps on ios)
  • steven75 - Friday, February 10, 2012 - link

    LOL dont you get it? You don't *need* to fiddle with those settings on iOS necause the battery life is so dramatically better.

    Also, funny reading this comment after Apple's Q4 report where they dominated.
  • Omid.M - Wednesday, January 18, 2012 - link

    I hope Samsung puts out this phone based on GN aesthetics but Exynos 5250 (plus MDM9xxx multi-mode/LTE modem) and blows away the competition.

    @moids
  • Chumster - Wednesday, January 18, 2012 - link

    Could someone clarify on what GPU/CPU he was talking about coming in Q2 devices? Cray? Crate? It was hard to pick up on my headphones.
  • mmp121 - Wednesday, January 18, 2012 - link

    Krait

    Read below:

    http://www.anandtech.com/show/4170/qualcomms-annou...

    Enjoy!
  • Conficio - Wednesday, January 18, 2012 - link

    Really, Google can't survive once Walled Garden platforms like iOS gain traction.

    While it is nice to control the OS (Chrome OS) on PC like devices and nice to stick it to Microsoft, it is essential in the world of smart phones. Google clearly saw that Apple did the unthinkable, wrestle control of the phone's apps away from the networks. That is an existential thread for Google. If there is a billion PC users world wide, there is a multitude of smart phone users, sooner or later.

    If a hardware manufacturer and OS provider like Apple (or Microsoft) controls the apps that can be provided to the phone and features, move from browser to apps on phones, then this is the end of (a profitable) google sooner or later.

    From anther point of view, Google is a huge data center that provides you with data services on their computing power (and you pay for it with advertisement somehow). Apple is a hardware manufacturer that sees it necessary to control the software to deliver a good user experience. Sure, two different approaches to a smart phone OS.
  • hackbod - Tuesday, January 24, 2012 - link

    "Google clearly saw that Apple did the unthinkable, wrestle control of the phone's apps away from the networks."

    There is this weird thing I see expressed a lot, as if Android is a reaction to the iPhone.

    It is not.

    In this particular case, it is obvious: Android's SDK was made available a few months after the original iPhone was on sale, well before there was *any* native SDK for the iPhone. At that time Apple's very clear official policy was that web-based apps was the One True Way to create applications for their phone. There was no concept of an App Store, no phone apps except what Apple shipped built in to the iPhone, nothing wrestled away from the networks in that department.

    If Android was a reaction to anything, it was to the current situation on desktop PCs, with one company controlling that platform, and being able to quite strongly dictate and control its ecosystem and thus large parts of the computer industry.

    One of the goals of Android was to try to keep that from happening in the upcoming mobile industry, by creating an open platform so that everybody in the industry can compete as equally as possible.

    (And an aside -- this also makes it funny to see the recent stuff going around about Google "losing control" of Android. Android was very much set up so that no one company, not even Google, could have anything like the control that Microsoft does over Windows. This should be pretty obvious to anyone who wants to actually write thoughtful articles on the topic and not just link bait.)
  • bjacobson - Wednesday, January 18, 2012 - link

    Can you talk more about this? From Diane Hackborne's post here (https://plus.google.com/u/0/105051985738280261832/... it sounds like the "limitation" is memory bandwidth in that hardwares that are "laggy" are laggy because they can't render to the entire screen 2 and 3x per frame for all the overlays. Which wouldn't seem like so much of a Tegra2 limitation in my opinion considering it has the power to play games like Quake 3 at 1600x1200 @ 60fps (I think...right?). What are your thoughts?
  • hackbod - Tuesday, January 24, 2012 - link

    I don't know about the performance of Tegra 2 playing Quake, but you need to be very careful when comparing the traditional 3d workload that GPUs are highly optimized to support (as exemplified by Quake) vs. the performance rendering 2d graphics.

    Traditional 3d games tend to rely, for example, on triangle rendering as much if not more than raw pixel fill rate, and GPUs are designed to be able to do that fast. When drawing 2d scenes, there are very few triangles but those triangles cover very large parts of the screen and are rendered as overlapping layers.

    On all of the hardware I have seen, for 2d rendering raw memory bandwidth (determining the number of times every pixel can be touched per frame) is the #1 impact on performance.

    Look back at that post -- for a typical scroll of all apps in launcher, without using overlay tricks (which aren't available on Tegra when the screen is rotated), you are looking at touching every pixels about 4 times to render all the layers and composite them to the screen. This is just not a typical 3D game workload.

Log in

Don't have an account? Sign up now