Power Consumption

The nature of reporting processor power consumption has become, in part, a dystopian nightmare. Historically the peak power consumption of a processor, as purchased, is given by its Thermal Design Power (TDP, or PL1). For many markets, such as embedded processors, that value of TDP still signifies the peak power consumption. For the processors we test at AnandTech, either desktop, notebook, or enterprise, this is not always the case.

Modern high performance processors implement a feature called Turbo. This allows, usually for a limited time, a processor to go beyond its rated frequency. Exactly how far the processor goes depends on a few factors, such as the Turbo Power Limit (PL2), whether the peak frequency is hard coded, the thermals, and the power delivery. Turbo can sometimes be very aggressive, allowing power values 2.5x above the rated TDP.

AMD and Intel have different definitions for TDP, but are broadly speaking applied the same. The difference comes to turbo modes, turbo limits, turbo budgets, and how the processors manage that power balance. These topics are 10000-12000 word articles in their own right, and we’ve got a few articles worth reading on the topic.

In simple terms, processor manufacturers only ever guarantee two values which are tied together - when all cores are running at base frequency, the processor should be running at or below the TDP rating. All turbo modes and power modes above that are not covered by warranty. Intel kind of screwed this up with the Tiger Lake launch in September 2020, by refusing to define a TDP rating for its new processors, instead going for a range. Obfuscation like this is a frustrating endeavor for press and end-users alike.

However, for our tests in this review, we measure the power consumption of the processor in a variety of different scenarios. These include full workflows, real-world image-model construction, and others as appropriate. These tests are done as comparative models. We also note the peak power recorded in any of our tests.

I’m here plotting the 10900K against the 10850K as we load the threads with AIDA’s stress test. Peak values are being reported.

On the front page, I stated that one of the metrics on where those quality lines were drawn, aside from frequency, is power and voltage response. Moving the needle for binning by 100 MHz is relatively easy, but binning for power is more difficult beast to control. Our tests show that for any full-threaded workload, despite being a lower frequency than the 10900K, our 10850K actually uses more power. At the extreme, this is +15-20W more, or up to 2 W per core, showcasing just how strict the metrics on the 10900K had to be (and perhaps why Intel has had difficulty manufacturing enough). However, one could argue that it was Intel’s decision to draw the line that aggressive.

In more lightly threaded workloads, the 10850K actually seems to use less power, which might indicate that this could be a current density issue being the prime factor in binning.

For a real workload, we’re using our Agisoft Photoscan benchmark. This test has a number of different areas that involve single thread, multi-thread, or memory limited algorithms.

At first glance, it looks as if the Core i9-10850K consumes more power at any loading, but it is worth noting the power levels in the 80-100% region of the test, when we dip below 50 W. This is when we’re likely using 1 or 2 threads, and the power of the Core i9-10900K is much higher as a percentage here, likely because of the 5300 MHz setting.

After getting these results, it caused me to look more at the data underneath. In terms of power per core, when testing POV-Ray at full load the difference is about a watt per core or just under. What surprised me more was the frequency response as well as the core loading temperature.

Starting with the 10900K:

In the initial loading, we get 5300 MHz and temperatures up into the 85-90ºC bracket. It’s worth noting that at these temperatures the CPU shouldn’t be in Thermal Velocity Boost, which should have a hard ceiling of 70ºC, but most modern motherboards will ignore that ‘Intel recommendation’. Also, when we look at watts per core, on the 10900K we’re looking at 26 W on a single core, just to get 5300 MHz, so no wonder it drops down to 15-19W per core very quickly.

The processor runs down to 5000 MHz at 3 cores loaded, sitting at 81ºC. Then as we go beyond three cores, the frequency dips only slightly, and the temperature of the whole package increases steadily up and up, until quite toasty 98ºC. This is even with our 2 kg copper cooler, indicating that at this point it’s more about thermal transfer inside the silicon itself rather than radiating away from the cooler.

When we do the same comparison for the Core i9-10850K however, the results are a bit more alarming.

This graph comes in two phases.

The first phase is the light loading, and because we’re not grasping for 5300 MHz, the temperature doesn’t go into the 90ºC segment at light loading like the 10900K does. The frequency profile is a bit more stair shaped than the 10900K, but as we ramp up the cores, even at a lower frequency, the power and the thermals increase. At full loading, with the same cooler and the same benchmarks in the same board, we’re seeing reports of 102ºC all-package temperature. The cooler is warm, but not excessively so, again showcasing that this is more an issue for thermal migration inside the silicon rather than cooling capacity.

To a certain degree, silicon is already designed with thermal migration in mind. It’s what we call ‘dark’ silicon, essentially silicon that is disabled/not anything that acts as a thermal (or power/electrical) barrier between different parts of the CPU. Modern processors already have copious amounts of dark silicon, and as we move to denser process node technologies, it will require even more. The knock on effect on this is die size, which could also affect yields for a given defect density.

Despite these thermals, none of our benchmarks (either gaming or high-performance compute) seemed to be out of line based on expectations – if anything the 10850K outperforms what we expected. The only gripe is going to be cooling, as we used an open test bed and arguably the best air cooler on the market, and users building into a case will need something similarly substantial, probably of the liquid cooling variety.

Intel Core i9-10850K Review CPU Tests: Microbenchmarks
Comments Locked


View All Comments

  • Kevin G - Monday, January 4, 2021 - link

    Inconsistency error on the second page: the 10850K and 10900K swap colors in the graphs.

    With regards to the intercore latency, what is the internal bus topology like in the 10 core Comet Lake? Last I heard, Intel was still using the ring topology for client processors but did they finally add a second ring similar to what they did for the medium and high core core count Xeons pre-Sky Lake? That'd explain why two of the cores seem off from the rest (and I would presume the GPU and PCIe controller would also sit on that same ring).
  • Spunjji - Monday, January 4, 2021 - link

    I genuinely appreciate the detail put into this article, but it's amusing that it was prompted by what amounts to little more than a bait-and-switch on their top-end product. 😬 The way I'm looking at it, they either used the early reviews to lie about the price of their *theoretical* flagship CPU, or they used the early reviews to lie about the power characteristics of their *actually available* flagship CPU.

    Either way, the end result remains the same: never trust Intel PR, and avoid their products unless you like to use your PC as a small household water heater.
  • JessNarmo - Monday, January 4, 2021 - link

    All companies lie and do questionable stuff, all that matters is what you getting for the money.

    And in winter small house heater in much appreciated when running script overnight allows you to avoid using house air heater in the room ;D

    I think 10850k could be a pretty good deal if $400 or cheaper and you can't wait for AMD availability.

    But I think it's only viable option for those who still rocking DDR3 or those who have no PC at all for the rest of us waiting is better. CPU upgrade is just not worth it as long as you have anything Skylake based on zen2 based.
  • Qasar - Tuesday, January 5, 2021 - link

    " All companies lie and do questionable stuff " not compared to intel, they have been pretty much lieing for how many years now about how its 10nm tech is coming along.
  • GeoffreyA - Wednesday, January 6, 2021 - link

    Maybe Intel should go into the geyser business ;)
  • zamroni - Monday, January 4, 2021 - link

    Intel should remove the low spec on die integrated gpu and replace it with more powerful gpu chiplet.
  • JessNarmo - Monday, January 4, 2021 - link

    Intel should remove integrated GPU from all high end CPU's. It's completely pointless. No one is using $500 CPU and using integrated graphics. Silicone is better spent on better CPU.
  • heftig - Monday, January 4, 2021 - link

    I am, actually. For a soho+media server and distcc host which only needs a GPU for whatever can't be done over SSH.
  • temps - Monday, January 4, 2021 - link

    good to see anandtech still has the biggest gap between quality of articles and quality of comment section

    running a 9900K with integrated here, the computer is for audio production and has no use for a graphics card. intel outperforms ryzen for this workload too.
  • powerarmour - Tuesday, January 5, 2021 - link

    Sorry to disappoint if you've been hiding under a rock for a couple of years, but this is no longer the case, check any recent Scan dawbench articles...

Log in

Don't have an account? Sign up now