General Performance: SYSMark 2007

Our journey starts with SYSMark 2007, the only all-encompassing performance suite in our review today. The idea here is simple: one benchmark to indicate the overall performance of your machine. SYSMark 2007 ends up being more of a dual-core benchmark as the applications/workload show minimal use of more than two threads.

SYSMark 2007

The 2600K is our new champion, the $317 chip is faster than Intel's Core i7 980X here as SYSMark 2007 doesn't really do much with the latter's extra 2 cores. Even the 2500K is a hair faster than the 980X. Compared to the Core i5 750, the upgrade is a no brainer - Sandy Bridge is around 20% faster at the same price point as Lynnfield.

Compared to Clarkdale, the Core i3 2100 only manages a 5% advantage howeer.

Adobe Photoshop CS4 Performance

To measure performance under Photoshop CS4 we turn to the Retouch Artists’ Speed Test. The test does basic photo editing; there are a couple of color space conversions, many layer creations, color curve adjustment, image and canvas size adjustment, unsharp mask, and finally a gaussian blur performed on the entire image.

The whole process is timed and thanks to the use of Intel's X25-M SSD as our test bed hard drive, performance is far more predictable than back when we used to test on mechanical disks.

Time is reported in seconds and the lower numbers mean better performance. The test is multithreaded and can hit all four cores in a quad-core machine.

Adobe Photoshop CS4 - Retouch Artists Benchmark

Once again, we have a new king - the 2600K is 9.7% faster than the 980X in our Photoshop CS4 test and the 2500K is just about equal to it. The Core i3 2100 does much better compared to the i3 540, outpacing it by around 30% and nearly equaling the performance of AMD's Phenom II X6 1100T.

The Test Video Encoding Performance
Comments Locked

283 Comments

View All Comments

  • aviat72 - Tuesday, January 4, 2011 - link

    Though SB will be great for some applications, there are still rough edges in terms of the overall platform. I think it will be best to wait for SNB-E or at least the Z68. SNB-E seems to be the best future-proofing bet.

    I also wonder how a part rated for 95W TDP was drawing 111W in the 4.4GHz OC (the Power Consumption Page). SB's power budget controller must be really smart to allow the higher performance without throttling down, assuming your cooling system can manage the thermals.
  • marraco - Tuesday, January 4, 2011 - link

    I wish to know more about this Sandy Bridge "feature":

    http://www.theinquirer.net/inquirer/news/1934536/i...
  • PeterO - Tuesday, January 4, 2011 - link

    Anand, Thanks for the great schooling and deep test results -- something surely representing an enormous amount of time to write, produce, and massage within Intel's bumped-forward official announcement date.

    Here's a crazy work-around question:

    Can I have my Quick Synch cake and eat my Single-monitor-with-Discrete-Graphics-card too if I, say:

    1). set my discrete card output to mirror Sandy Bridge's IGP display output;

    2). and, (should something exist), add some kind of signal loopback adapter to the IGP port to spoof the presence of a monitor? A null modem, of sorts?

    -- I have absolutely no mobo/video signaling background, so my idea may be laugh in my face funny to anybody who does but I figure it's worth a post, if only for your entertainment. :)
  • Hrel - Wednesday, January 5, 2011 - link

    It makes me SO angry when Intel does stupid shit like disable HT on most of their CPU's even though the damn CPU already has it on it, they already paid for. It literally wouldn't cost them ANYTHING to turn HT on those CPU's yet the greedy bastards don't do it.
  • Moizy - Wednesday, January 5, 2011 - link

    The HD Graphics 3000 performance is pretty impressive, but won't be utilized by most. Most who utilize Intel desktop graphics will be using the HD Graphics 2000, which is okay, but I ran back to the AMD Brazos performance review to get some comparisons.

    In Modern Warfare 2, at 1024 x 768, the new Intel HD Graphics 2000 in the Core i3 2100 barely bests the E-350. Hmm--that's when it's coupled with a full-powered, hyper-threaded desktop compute core that would run circles around the compute side of the Brazos E-350, an 18w, ultra-thin chip.

    This either makes Intel's graphics less impressive, or AMD's more impressive. For me, I'm more impressed with the graphics power in the 18w Brazos chip, and I'm very excited by what mainstream Llano desktop chips (65w - 95w) will bring, graphics-wise. Should be the perfect HTPC solution, all on the CPU (ahem, APU, I mean).

    I'm very impressed with Intel's video transcoding, however. Makes CUDA seem...less impressive, like a bunch of whoop-la. Scary what Intel can do when it decides that it cares about doing it.
  • andywuwei - Wednesday, January 5, 2011 - link

    not sure if anybody else noticed. CPU temp of the i5@3.2GHz is ~140 degrees. any idea why it is so high?
  • SantaAna12 - Wednesday, January 5, 2011 - link

    Did I miss the part where you tell of about the DRM built into this chip?
  • Cb422 - Wednesday, January 5, 2011 - link

    When will Sandy Bridge be available on Newegg or Amazon for me to purchase?
  • DesktopMan - Thursday, January 6, 2011 - link

    Very disappointed in the lack of vt-d and txt on k-variants. They are after all the high end products. I also find the fact that only the k-variants having the faster GPU very peculiar, as those are the CPUs most likely to be paired with a discrete GPU.
  • RagingDragon - Thursday, January 6, 2011 - link

    Agreed. I find the exclusion of VT-d particularly irritating: many of the overclockers and enthusiasts to whom the K chips are marketed also use virtualization. Though I don't expect many enthusiasts, if any, to miss TXT (it's more for locked down corporate systems, media appliances, game consoles, etc.).

    With the Z68 chipset coming in the indeterminate near future, the faster GPU on K chips would have made sense if the K chips came with every other feature enabled (i.e. if they were the "do eveything chips").

    Also, I'd like to have the Sandy Bridge video encode/decode features separate from the GPU functionality - i.e. I'd like to choose between Intel and Nvidia/AMD video decode/encode when using a discrete GPU.

Log in

Don't have an account? Sign up now