AnandTech Storage Bench 2013

When Anand built the AnandTech Heavy and Light Storage Bench suites in 2011 he did so because we didn't have any good tools at the time that would begin to stress a drive's garbage collection routines. Once all blocks have a sufficient number of used pages, all further writes will inevitably trigger some sort of garbage collection/block recycling algorithm. Our Heavy 2011 test in particular was designed to do just this. By hitting the test SSD with a large enough and write intensive enough workload, we could ensure that some amount of GC would happen.

There were a couple of issues with our 2011 tests that we've been wanting to rectify however. First off, all of our 2011 tests were built using Windows 7 x64 pre-SP1, which meant there were potentially some 4K alignment issues that wouldn't exist had we built the trace on a system with SP1. This didn't really impact most SSDs but it proved to be a problem with some hard drives. Secondly, and more recently, we've shifted focus from simply triggering GC routines to really looking at worst case scenario performance after prolonged random IO.

For years we'd felt the negative impacts of inconsistent IO performance with all SSDs, but until the S3700 showed up we didn't think to actually measure and visualize IO consistency. The problem with our IO consistency tests is that they are very focused on 4KB random writes at high queue depths and full LBA spans--not exactly a real world client usage model. The aspects of SSD architecture that those tests stress however are very important, and none of our existing tests were doing a good job of quantifying that.

We needed an updated heavy test, one that dealt with an even larger set of data and one that somehow incorporated IO consistency into its metrics. We think we have that test. The new benchmark doesn't even have a name, we've just been calling it The Destroyer (although AnandTech Storage Bench 2013 is likely a better fit for PR reasons).

Everything about this new test is bigger and better. The test platform moves to Windows 8 Pro x64. The workload is far more realistic. Just as before, this is an application trace based test--we record all IO requests made to a test system, then play them back on the drive we're measuring and run statistical analysis on the drive's responses.

Imitating most modern benchmarks Anand crafted the Destroyer out of a series of scenarios. For this benchmark we focused heavily on Photo editing, Gaming, Virtualization, General Productivity, Video Playback and Application Development. Rough descriptions of the various scenarios are in the table below:

AnandTech Storage Bench 2013 Preview - The Destroyer
Workload Description Applications Used
Photo Sync/Editing Import images, edit, export Adobe Photoshop CS6, Adobe Lightroom 4, Dropbox
Gaming Download/install games, play games Steam, Deus Ex, Skyrim, Starcraft 2, BioShock Infinite
Virtualization Run/manage VM, use general apps inside VM VirtualBox
General Productivity Browse the web, manage local email, copy files, encrypt/decrypt files, backup system, download content, virus/malware scan Chrome, IE10, Outlook, Windows 8, AxCrypt, uTorrent, AdAware
Video Playback Copy and watch movies Windows 8
Application Development Compile projects, check out code, download code samples Visual Studio 2012

While some tasks remained independent, many were stitched together (e.g. system backups would take place while other scenarios were taking place). The overall stats give some justification to what we've been calling this test internally:

AnandTech Storage Bench 2013 Preview - The Destroyer, Specs
  The Destroyer (2013) Heavy 2011
Reads 38.83 million 2.17 million
Writes 10.98 million 1.78 million
Total IO Operations 49.8 million 3.99 million
Total GB Read 1583.02 GB 48.63 GB
Total GB Written 875.62 GB 106.32 GB
Average Queue Depth ~5.5 ~4.6
Focus Worst case multitasking, IO consistency Peak IO, basic GC routines

SSDs have grown in their performance abilities over the years, so we wanted a new test that could really push high queue depths at times. The average queue depth is still realistic for a client workload, but the Destroyer has some very demanding peaks. When we first introduced the Heavy 2011 test, some drives would take multiple hours to complete it; today most high performance SSDs can finish the test in under 90 minutes. The Destroyer? So far the fastest we've seen it go is 10 hours. Most high performance SSDs we've tested seem to need around 12 - 13 hours per run, with mainstream drives taking closer to 24 hours. The read/write balance is also a lot more realistic than in the Heavy 2011 test. Back in 2011 we just needed something that had a ton of writes so we could start separating the good from the bad. Now that the drives have matured, we felt a test that was a bit more balanced would be a better idea.

Despite the balance recalibration, there's just a ton of data moving around in this test. Ultimately the sheer volume of data here and the fact that there's a good amount of random IO courtesy of all of the multitasking (e.g. background VM work, background photo exports/syncs, etc...) makes the Destroyer do a far better job of giving credit for performance consistency than the old Heavy 2011 test. Both tests are valid; they just stress/showcase different things. As the days of begging for better random IO performance and basic GC intelligence are over, we wanted a test that would give us a bit more of what we're interested in these days. As Anand mentioned in the S3700 review, having good worst case IO performance and consistency matters just as much to client users as it does to enterprise users.

We're reporting two primary metrics with the Destroyer: average data rate in MB/s and average service time in microseconds. The former gives you an idea of the throughput of the drive during the time that it was running the Destroyer workload. This can be a very good indication of overall performance. What average data rate doesn't do a good job of is taking into account response time of very bursty (read: high queue depth) IO. By reporting average service time we heavily weigh latency for queued IOs. You'll note that this is a metric we've been reporting in our enterprise benchmarks for a while now. With the client tests maturing, the time was right for a little convergence.

AnandTech Storage Bench 2013 - The Destroyer

While IO consistency was lacking, the Hawk does surprisingly well in our new Storage Bench 2013. It's definitely no challenger to SanDisk Extreme II or Seagate 600, but the excellent sequential performance makes up for the lack of random write performance and consistency. A look at the average service time shows that the IO consistency is what is holding the Hawk back. 

AnandTech Storage Bench 2013 - The Destroyer

Performance Consistency Random & Sequential Performance
Comments Locked

26 Comments

View All Comments

  • ssj3gohan - Wednesday, June 26, 2013 - link

    Wait, what?! HIPM and DIPM have been standard features that are always enabled on every intel desktop system since the intel 5 series at least (and AM2+ on the AMD side, which had other reasons). The only reason not to have HIPM and DIPM is your choice of OS. It's a driver-enabled feature, it doesn't work without explicitly enabling it.
  • Kristian Vättö - Wednesday, June 26, 2013 - link

    That's not true. You can do the registry hack and get the option to enable HIPM+DIPM but that will have zero impact on power consumption on a desktop system. We've tried four different motherboards including Intel-branded ones and also both Windows 7 and 8 without success. That's why it took us so long to start doing HIPM+DIPM tests as Anand had to take an Ultrabook apart just for power testing and obviously I don't have access to that system due to geographical issues.
  • JDG1980 - Tuesday, June 25, 2013 - link

    If you want a Toshiba drive, why not just buy a Toshiba drive instead of one rebranded by a no-name company? I don't see the point of this.
  • JarredWalton - Tuesday, June 25, 2013 - link

    It's cheaper and uses the same hardware and firmware?
  • Kristian Vättö - Tuesday, June 25, 2013 - link

    Toshiba has quite poor retail presence, so Strontium may be targeting the markets where the availability of Toshiba SSDs is either very poor or non-existent.
  • HisDivineOrder - Tuesday, June 25, 2013 - link

    My concern would be what happens when they change products and firmwares are... nonexistent. SSD's are still far too temperamental to trust just some nobody who likes to throw caution to the wind.

    OCZ can't even be trusted in this regard.

    Samsung, Intel, Micron/Crucial, maaaaybe Sandisk.
  • bji - Tuesday, June 25, 2013 - link

    Can OCZ be trusted in *any* regard?
  • zanon - Tuesday, June 25, 2013 - link

    I wouldn't have a problem if it was Intel or some other major SSD OEM doing this as they could change the way GB is defined in the storage industry

    Kristian no they couldn't, seriously what the hell? We *still* have people arguing wrongly in freaking 2013? "Giga" is an SI prefix, it's base-10. That's what it is, anything else is wrong. This is particularly surprisingly from a technology enthusiast, as tech depends on precision, not random overloaded definitions. Maybe you'd like to suggest that they come up with their own definitions of "kilogram" and "pound" too with whatever mass they feel like? That sounds like a recipe for success!

    If they want to show base-2, then they can use KiB/MiB/GiB/TiB/PiB. I don't think those sound the greatest but whatever, they're precise, correct, and the international standard (IEC 60027, published in *1999*). Storage makers were never wrong to use the SI prefixes and have base-10 numbers of sectors.

    If anything the problem is Microsoft, who should long since have switched to displaying GB instead of GiB. Even Apple finally updated ages ago.
  • Kristian Vättö - Tuesday, June 25, 2013 - link

    The problem is that Giga is not only an SI prefix, it's also a JEDEC prefix where Giga is defined as 1024^3. I do agree that Giga should only be used as an SI prefix (i.e. meaning 1000^3) but I guess someone at JEDEC thought that since 1000^3 and 1024^3 are "close enough", lets just call them the same.

    I also agree that it's Microsoft who should just change to the SI standard and define GB as 1000^3 bytes like Apple did. My point about Intel was that it's useless for a small OEM like Strontium to differ from the norm and I don't see Intel or anyone else switching away from the SI standard because the advertised capacities would end up being smaller than they are now.
  • KAlmquist - Wednesday, June 26, 2013 - link

    "Storage makers were never wrong to use the SI prefixes"

    This is, I believe, inconsistent with your preceding paragraph. When disk drive manufacturers came up with their own definition of megabyte to make their disk drives seem larger, that created the same problems you envision if someone came up with their own definitions for "kilogram" and "pound".

    As you note, that the IEC has decided to endorse the disk drive manufacture's definition of "megabyte" and has invented the term "mebibyte" to refer to a megabyte. That's a response to the linguistic mess created by the disk drive manufacturers; it doesn't justify creating the mess in the first place.

    Meanwhile, as Kristian notes, JEDEC is sticking with the original meaning of the word. Words can acquire new meanings fairly quickly, but it takes a long time (on the order of 100 years) for a word to lose a well-established meaning.

Log in

Don't have an account? Sign up now