AnandTech Storage Bench 2013

When Anand built the AnandTech Heavy and Light Storage Bench suites in 2011 he did so because we didn't have any good tools at the time that would begin to stress a drive's garbage collection routines. Once all blocks have a sufficient number of used pages, all further writes will inevitably trigger some sort of garbage collection/block recycling algorithm. Our Heavy 2011 test in particular was designed to do just this. By hitting the test SSD with a large enough and write intensive enough workload, we could ensure that some amount of GC would happen.

There were a couple of issues with our 2011 tests that we've been wanting to rectify however. First off, all of our 2011 tests were built using Windows 7 x64 pre-SP1, which meant there were potentially some 4K alignment issues that wouldn't exist had we built the trace on a system with SP1. This didn't really impact most SSDs but it proved to be a problem with some hard drives. Secondly, and more recently, we've shifted focus from simply triggering GC routines to really looking at worst-case scenario performance after prolonged random IO.

For years we'd felt the negative impacts of inconsistent IO performance with all SSDs, but until the S3700 showed up we didn't think to actually measure and visualize IO consistency. The problem with our IO consistency tests is that they are very focused on 4KB random writes at high queue depths and full LBA spans–not exactly a real world client usage model. The aspects of SSD architecture that those tests stress however are very important, and none of our existing tests were doing a good job of quantifying that.

We needed an updated heavy test, one that dealt with an even larger set of data and one that somehow incorporated IO consistency into its metrics. We think we have that test. The new benchmark doesn't even have a name, we've just been calling it The Destroyer (although AnandTech Storage Bench 2013 is likely a better fit for PR reasons).

Everything about this new test is bigger and better. The test platform moves to Windows 8 Pro x64. The workload is far more realistic. Just as before, this is an application trace based test–we record all IO requests made to a test system, then play them back on the drive we're measuring and run statistical analysis on the drive's responses.

Imitating most modern benchmarks Anand crafted the Destroyer out of a series of scenarios. For this benchmark we focused heavily on Photo editing, Gaming, Virtualization, General Productivity, Video Playback and Application Development. Rough descriptions of the various scenarios are in the table below:

AnandTech Storage Bench 2013 Preview - The Destroyer
Workload Description Applications Used
Photo Sync/Editing Import images, edit, export Adobe Photoshop CS6, Adobe Lightroom 4, Dropbox
Gaming Download/install games, play games Steam, Deus Ex, Skyrim, Starcraft 2, BioShock Infinite
Virtualization Run/manage VM, use general apps inside VM VirtualBox
General Productivity Browse the web, manage local email, copy files, encrypt/decrypt files, backup system, download content, virus/malware scan Chrome, IE10, Outlook, Windows 8, AxCrypt, uTorrent, AdAware
Video Playback Copy and watch movies Windows 8
Application Development Compile projects, check out code, download code samples Visual Studio 2012

While some tasks remained independent, many were stitched together (e.g. system backups would take place while other scenarios were taking place). The overall stats give some justification to what we've been calling this test internally:

AnandTech Storage Bench 2013 Preview - The Destroyer, Specs
  The Destroyer (2013) Heavy 2011
Reads 38.83 million 2.17 million
Writes 10.98 million 1.78 million
Total IO Operations 49.8 million 3.99 million
Total GB Read 1583.02 GB 48.63 GB
Total GB Written 875.62 GB 106.32 GB
Average Queue Depth ~5.5 ~4.6
Focus Worst-case multitasking, IO consistency Peak IO, basic GC routines

SSDs have grown in their performance abilities over the years, so we wanted a new test that could really push high queue depths at times. The average queue depth is still realistic for a client workload, but the Destroyer has some very demanding peaks. When we first introduced the Heavy 2011 test, some drives would take multiple hours to complete it; today most high performance SSDs can finish the test in under 90 minutes. The Destroyer? So far the fastest we've seen it go is 10 hours. Most high performance SSDs we've tested seem to need around 12–13 hours per run, with mainstream drives taking closer to 24 hours. The read/write balance is also a lot more realistic than in the Heavy 2011 test. Back in 2011 we just needed something that had a ton of writes so we could start separating the good from the bad. Now that the drives have matured, we felt a test that was a bit more balanced would be a better idea.

Despite the balance recalibration, there's just a ton of data moving around in this test. Ultimately the sheer volume of data here and the fact that there's a good amount of random IO courtesy of all of the multitasking (e.g. background VM work, background photo exports/syncs, etc...) makes the Destroyer do a far better job of giving credit for performance consistency than the old Heavy 2011 test. Both tests are valid; they just stress/showcase different things. As the days of begging for better random IO performance and basic GC intelligence are over, we wanted a test that would give us a bit more of what we're interested in these days. As Anand mentioned in the S3700 review, having good worst-case IO performance and consistency matters just as much to client users as it does to enterprise users.

We're reporting two primary metrics with the Destroyer: average data rate in MB/s and average service time in microseconds. The former gives you an idea of the throughput of the drive during the time that it was running the Destroyer workload. This can be a very good indication of overall performance. What average data rate doesn't do a good job of is taking into account response time of very bursty (read: high queue depth) IO. By reporting average service time we heavily weigh latency for queued IOs. You'll note that this is a metric we've been reporting in our enterprise benchmarks for a while now. With the client tests maturing, the time was right for a little convergence.

AT Storage Bench 2013 - The Destroyer (Data Rate)

The SSD 530 does okay in our new Storage Bench 2013. The improvement from SSD 335 is again quite significant, which is mostly thanks to the improved performance consistency. However, the SF-2281 simply can't challenge the more modern designs and for ultimate performance the SanDisk Extreme II is still the best pick. 

AT Storage Bench 2013 - The Destroyer (Service Time)

Performance Consistency & TRIM Validation Random & Sequential Performance
Comments Locked

60 Comments

View All Comments

  • Duncan Macdonald - Saturday, November 16, 2013 - link

    Would it be possible for you to do an additional SSD test - how much does the write performance recover after a 30 minute idle period. Most consumer PCs (and even many servers) tend to have idle periods every day and if the garbage collection and free space erasure algorithms on the drive can get it back to a near new condition then this would be significent.
  • 'nar - Monday, November 18, 2013 - link

    I agree, GC and idle time makes TRIM unnecessary, and even works better anyway. These benchmarks are a gross exaggeration of anything done in real-world usage. On the one hand everyone recognizes that fact, but on the other they keep hammering consistency and incompressible data like everyone does video editing all day every day.
  • thefoodaddy - Saturday, November 16, 2013 - link

    The prices in that table for the Seagate and Crucial 240GB are, sadly, not $150 ($220 and $180, respectively)--way to get my hopes up, Dyntamitedata.com!
  • purerice - Sunday, November 17, 2013 - link

    On google shopping I just typed in "Seagate SSD 600" and selected 240GB. 50+stores had them and 1 store has them for $149.99 with $0 tax and $0 shipping
  • Kristian Vättö - Sunday, November 17, 2013 - link

    The prices were taken on November 12th and both drives were $150 back then (probably a temporary sale).
  • slickr - Saturday, November 16, 2013 - link

    Man these SSD's seem like a lot of hard work to me. I mean with all the firmware updates that need to be flushed, with all the failures that seem to be happening, with the inconsistent performances, with the fairly still high prices even after 4 years of SSD drives.

    I mean 4 years ago I though we would have at least 250GB for $150 by around this time, by around year 2014, but we we are still way off, I though 500GB SSD's would have started becoming more mainstream in 4 years, but now that we are here, now that I'm in the future it hasn't been done.

    In fact some of the drives are still plagued by the same problems some of the first SSD's had. I mean I agree that the average SSD is more reliable and generally faster, but this is not by much and the prices have been slow to come down.

    So I hope to see $150 250GB SSD's and more in the next several months, maybe 2014 will be the year, but I think if you just want reliability and security its best to go with normal hard drives that have huge capacity at cheap prices, I can get 1TB for $70 that is super cheap.
  • 'nar - Monday, November 18, 2013 - link

    I think you have been mislead by the benchmarks. They do not compare SSD's to hard drives, so you have no perspective. I recommend SSD's for everyone. They are faster and more reliable. Get a hard drive if you want your 1 TB of storage, which will be pictures, music, and video anyway, all things that would not benefit from SSD speeds.

    The only concern is that many people that complain about reliability fail to mention the model SSD that failed on them. I use Intel/Sandforce drive for systems I build for others, and OCZ/SandForce/Bigfoot on all of my own and never had a problem. I suspect that those looking for cheap, get cheap. If you want reliability don't look for the cheapest drive. As in all things, you get what you pay for. Find yourself a good drive, THEN look for a good price on it. Don't assume that any SSD made by a particular manufacturer is good.
  • name99 - Saturday, November 16, 2013 - link

    "The problem now is that every significant segment from a performance angle has been covered."

    Unfortunately no. If *I* were an SSD manufacturer, I'd try to differentiate myself by putting together a hybrid drive that isn't crap. It is insane that, with 2013 almost over, there is, as far as I can tell, precisely one HD available that is a hybrid drive --- and that HD is available in one form factor+size, only as a bare drive, and with a minuscule pool of flash.
    Complain about Apple all you like, but at least they have done (within the scope of what they control) something about this --- unlike freaking WD, Seagate, SanDisk and everyone else.

    WTF have SanDisk (or Sandforce, or Samsung, or Toshiba, or ...) down something about this? Put together a decent package of some RAM, some flash, a controller, firmware that does the caching properly, and sell it to WD or Seagate to glue onto 1TB+ size drives? Apple's solution is expensive, probably too expensive, because it's using pretty good quality flash and a lot of it. Cut down to 48 or 32GB of flash that's slightly slower and I think you could still give a heck of a kick to a drive at an additional cost of $30 to $50. I'd certainly be willing to pay this.

    I do not understand WD and Seagate. You go to Best Buy or Frys today, and they're each trying to reach out at you with a huge collection of basically identical drives --- they'll sell you a 2TB 2.5" in a green version, a black version, a red version, a blue version. (And those are not case colors, they are supposedly different models.)
    The one thing they won't sell is the thing that would actually make a difference, that I'd be willing to pay for, a freaking HYBRID version that consists of more than adding 8GB of crappy bargain bin flash and lame caching software that won't even capture writes.
  • Bob Todd - Sunday, November 17, 2013 - link

    Indeed. While it would be great if every laptop with a 2.5" drive had a mSATA or M.2 slot available, they are still the minority. I have SSDs as the boot drives of every machine sans one laptop that still has one of the 7200rpm 750GB Seagate SSHDs. I want at least 500GB of capacity for that machine, but I don't really want to drop the money for an SSD that big. A 7mm 500+ GB drive with 32+ GB of NAND needs to happen.
  • emvonline - Sunday, November 17, 2013 - link

    100% agree. A 32G SSD+1TB HDD would cover all storage needs and be very fast for 90% of all work. On the once a month timing that you load a rarely accessed 1GB video it would take 3 seconds more than a SSD. All this assumes the Cache software works correctly :-)

Log in

Don't have an account? Sign up now