Conclusion: Less Lakes, More Coves Please

One thing that Intel has learned through the successive years of the reiterating the Skylake microarchitecture on the same process but with more cores has been optimization – the ability to squeeze as many drops out of a given manufacturing node and architecture as is physically possible, and still come out with a high-performing product when the main competitor is offering similar performance at a much lower power.

Intel has pushed Comet Lake and its 14nm process to new heights, and in many cases, achieving top results in a lot of our benchmarks, at the expense of power. There’s something to be said for having the best gaming CPU on the market, something which Intel seems to have readily achieved here when considering gaming in isolation, though now Intel has to deal with the messaging around the power consumption, similar how AMD had to do in the Vishera days.

One of the common phrases that companies in this position like to use is that ‘when people use a system, they don’t care about how much power it’s using at any specific given time’. It’s the same argument borne out of ‘some people will pay for the best, regardless of price or power consumption’. Both of these arguments are ones that we’ve taken onboard over the years, and sometimes we agree with them – if all you want is the best, then yes these other metrics do not matter.

In this review we tested the Core i9-10900K with ten cores, the Core i7-10700K with eight cores, and the Core i5-10600K with six cores. On the face of it, the Core i9-10900K with its ability to boost all ten cores to 4.9 GHz sustained (in the right motherboard) as well as offering 5.3 GHz turbo will be a welcome recommendation to some users. It offers some of the best frame rates in our more CPU-limited gaming tests, and it competes at a similar price against an AMD processor that offers two more cores at a lower frequency. This is really going to be a case of ‘how many things do you do at once’ type of recommendation, with the caveat of power.

The ability of Intel to pull 254 W on a regular 10 core retail processor is impressive. After we take out the system agent and DRAM, this is around 20 W per core. With its core designs, Intel has often quoted its goal to scale these Core designs from fractions of a Watt to dozens of Watts. The only downside now of championing something in the 250 W range is going to be that there are competitive offerings that do 24, 32, and 64 cores in this power range. Sure, those processors are more expensive, but they do a lot more work, especially for large multi-core scenarios.

If we remember back to AMD’s Vishera days, the company launched a product with eight cores, 5.0 GHz, with a 220 W TDP. At peak power consumption, the AMD FX-9590 was nearer 270 W, and while AMD offered caveats listed above (‘people don’t’ really care about the power of the system while in use’) Intel chastised the processor for being super power hungry and not high performance enough to be competitive. Fast forward to 2020, we have the reverse situation, where Intel is pumping out the high power processor, but this time, there’s a lot of performance.

The one issue that Intel won’t escape from is that all this extra power requires extra money to be put into cooling the chip. While the Core i9 processor is around the same price as the Ryzen 9 3900X, the AMD processor comes with a 125 W cooler which will do the job – Intel customers will have to go forth and source expensive cooling in order to keep this cool. Speaking with a colleague, he had issues cooling his 10900K test chip with a Corsair H115i, indicating that users should look to spending $150+ on a cooling setup. That’s going to be a critical balancing element here when it comes to recommendations.

For recommendations, Intel’s Core i9 is currently performing the best in a lot of tests, and that’s hard to ignore. But will the end-user want that extra percent of performance, for the sake of spending more on cooling and more in power? Even in the $500 CPU market, that’s a hard one to ask. Then add in the fact that the Core i9 doesn’t have PCIe 4.0 support, we get to a situation that it’s the best offering if you want Intel, and you want the best processor, but AMD has almost the same ST performance, better MT performance in a lot of cases, and a much better power efficiency and future PCIe support – it comes across as the better package overall.

Agisoft Photoscan 1.3.3, Complex Test

Throughout this piece I’ve mainly focused on the Core i9, with it being the flagship part. The Core i7 and the Core i5 are also in our benchmark results, and their results are a little mixed.

The Core i5 doesn’t actually draw too much power, but it doesn’t handle the high performance aspect in multi-threaded scenarios. At $262 it comes in at more expensive than the Ryzen 5 3600 which is $200, and the trade-off here is the Ryzen’s better IPC against the Core i5’s frequency. Most of the time the Core i5-10600K wins, as it should do given that it costs an extra $60, but again doesn’t have the PCIe 4.0 support that the AMD chip offers. When it comes to building that $1500 PC, this will be an interesting trade-off to consider.

For the Core i7-10700K, at 8C/16T, we really have a repeat of the Core i9-9900K from the previous generation. It offers a similar sort of performance and power, except with Turbo Boost Max 3.0 giving it a bit more frequency. There’s also the price – at $374 it is certainly a lot more attractive than it was at $488. Anand once said that there are no bad products, only bad prices, and a $114 price drop on the 8c overclocked part is highly welcome. At $374 it fits between the Ryzen 9 3900X ($432) and the Ryzen 7 3800X ($340), so we’re still going to see some tradeoffs against the Ryzen 7 in performance vs power vs cost.

Final Thoughts

Overall, I think the market is getting a little bit of fatigue from continuous recycling of the Skylake microarchitecture on the desktop. The upside is that a lot of programs are optimized for it, and Intel has optimized the process and the platform so much that we’re getting a lot higher frequencies out of it, at the expense of power. Intel is still missing key elements in its desktop portfolio, such as PCIe 4.0 and DDR4-3200, as well as anything resembling a decent integrated graphics processor, but ultimately one could argue that the product team are at the whims of the manufacturing arm, and 10nm isn’t ready yet for the desktop prime time.

When 10nm will be ready for the desktop we don’t know - Intel is set to showcase 10nm Ice Lake for servers at Hot Chips in August, as well as 10nm Tiger Lake for notebooks. If we’re stuck on 14nm on the desktop for another generation, then we really need a new microarchitecture that scales as well. Intel might be digging itself a hole, optimizing Skylake so much, especially if it can’t at least match what might be coming next – assuming it even has enough space on the manufacturing line to build it. Please Intel, bring a Cove our way soon.

For the meantime, we get a power hungry Comet Lake, but at least the best chip in the stack performs well enough to top a lot of our charts for the price point.

These processors should be available from today. As far as we understand, the overclockable parts will be coming to market first, with the rest coming to shelves depending on region and Intel's manufacturing plans.

Gaming: F1 2018
Comments Locked

220 Comments

View All Comments

  • catavalon21 - Wednesday, May 20, 2020 - link

    The ability to edit (or ^Z) would be most welcome, trust me.
  • eastcoast_pete - Wednesday, May 20, 2020 - link

    Isn't that Skylake running a bit dry by now? But, seriously, Intel really risks losing a lot of market share in future years by selling these "classics" at high prices, and that is if one can get one in the first place.
    Curious: how many commercial customers buy Intel desktops just because they have iGPUs, but want more CPU oomph than the 3200G has? Is that why Intel still dominates the OEM desktop market?
  • AnarchoPrimitiv - Wednesday, May 20, 2020 - link

    Intel dominates the OEM market through intimidation and threats of retribution... They were literally convicted of bribing OEMs to NOT use AMD CPUs all throughout the 2000s in several courts around the world. The trials uncovered emails between Intel executives that stated, and I quote, "Dell is the best friend money can buy".... The proof is in the fact that currently, the Ryzen 4000 mobile CPUs are the best mobile chips offered right now, but Dell only puts them in the low end laptops. Why? Because Intel is probably giving huge financial incentives to bar AMD from premium designs to perpetuate the myth that AMD isn't a premium brand
  • Retycint - Wednesday, May 20, 2020 - link

    Do keep in mind that these are baseless speculations, based on something that happened 2 decades ago. Both Intel and AMD have changed since then (new engineering team, new management etc) and there has been no evidence of Intel providing incentives to cripple AMD systems. Go take your conspiracy elsewhere.

    And before you inevitably accuse me of being an Intel shill, this isn't about Intel or AMD, it's about facts to support your claim, of which there have been none
  • Irata - Wednesday, May 20, 2020 - link

    Baseless speculation? Financial horsepower, MDF and meet the comp funds are current and no secret.

    Why do you think there are no Ryzen 4000 laptops with GPU above a 2060?
  • Spunjji - Tuesday, May 26, 2020 - link

    Not entirely baseless, as they made two distinct claims. I've been a party to how Intel's "Marketing Development Funds" work - and work it does, at all levels from OEM to reseller to retailer. These days they don't explicitly punish anyone for not buying AMD - they simply tie rebates that will improve the profit margins on a product to specific quantities of those products being sold. It's "nobody's fault" if those quantities happen to make the sale of an AMD product by a given retailer or reseller distinctly unlikely.

    As for incentivizing bad *builds* of AMD systems, though, I'm not so sure. Intel clearly do a lot of work building reference platforms, and the economics of doing integration testing for a new vendor is not trivial. Honestly though, it's hard to tell how we *would* know if this were going on, because it would absolutely be made to look innocent - just like last time.
  • brantron - Wednesday, May 20, 2020 - link

    "literally convicted of bribing"

    1) No. That's not what "literally" means.
    2) No. No one was even *charged* with a crime, much less convicted.
    3) No. It wasn't about bribery.

    The reason Athlon 64s weren't ubiquitous way back when is the same reason the 4000 APUs aren't today - there aren't enough to go around.

    If your post were to be rephrased without hyperbole, baseless accusations, and whataboutism unrelated to the topic of this article, it would read something like this:

    "6 months after AMD's announcement of Renoir, the number of 4000 APUs sold for desktops is literally zero (see how that works?) because TSMC is still slammed."
  • WaWaThreeFIVbroS - Thursday, May 21, 2020 - link

    Your ignorance is amusing
    It is technically a bribery

    https://www.extremetech.com/computing/184323-intel...
  • Spunjji - Tuesday, May 26, 2020 - link

    First 3 points: accurate, if not entirely on-topic. Nobody was charged with a crime, but Intel sure were fined a lot for collusion.

    Which gets to the 4th point: again, accurate, but not entirely relevant. AMD were definitely not able to match Intel for manufacturing, which is why they couldn't have beaten Intel out of the market entirely, but that was barely related to why they weren't getting into Dell systems. See the aforementioned proven-and-fined-for collusion.
  • drothgery - Friday, May 22, 2020 - link

    Or because premium designs take longer when the new chip isn't just another respin of the same thing, and AMD hadn't produced a viable high-end notebook chip in well over a decade so it made sense to wait and see if Ryzen 4000 was any good rather than designing in advance?

Log in

Don't have an account? Sign up now