AnandTech Storage Bench 2013

When Anand built the AnandTech Heavy and Light Storage Bench suites in 2011 he did so because we did not have any good tools at the time that would begin to stress a drive's garbage collection routines. Once all blocks have a sufficient number of used pages, all further writes will inevitably trigger some sort of garbage collection/block recycling algorithm. Our Heavy 2011 test in particular was designed to do just this. By hitting the test SSD with a large enough and write intensive enough workload, we could ensure that some amount of GC would happen.

There were a couple of issues with our 2011 tests that we've been wanting to rectify however. First off, all of our 2011 tests were built using Windows 7 x64 pre-SP1, which meant there were potentially some 4K alignment issues that wouldn't exist had we built the trace on a system with SP1. This didn't really impact most SSDs but it proved to be a problem with some hard drives. Secondly, and more recently, we've shifted focus from simply triggering GC routines to really looking at worst-case scenario performance after prolonged random IO.

For years we'd felt the negative impacts of inconsistent IO performance with all SSDs, but until the S3700 showed up we didn't think to actually measure and visualize IO consistency. The problem with our IO consistency tests is that they are very focused on 4KB random writes at high queue depths and full LBA spans–not exactly a real world client usage model. The aspects of SSD architecture that those tests stress however are very important, and none of our existing tests were doing a good job of quantifying that.

We needed an updated heavy test, one that dealt with an even larger set of data and one that somehow incorporated IO consistency into its metrics. We think we have that test. The new benchmark doesn't even have a name, we've just been calling it The Destroyer (although AnandTech Storage Bench 2013 is likely a better fit for PR reasons).

Everything about this new test is bigger and better. The test platform moves to Windows 8 Pro x64. The workload is far more realistic. Just as before, this is an application trace based test–we record all IO requests made to a test system, then play them back on the drive we're measuring and run statistical analysis on the drive's responses.

Imitating most modern benchmarks Anand crafted the Destroyer out of a series of scenarios. For this benchmark we focused heavily on Photo editing, Gaming, Virtualization, General Productivity, Video Playback and Application Development. Rough descriptions of the various scenarios are in the table below:

AnandTech Storage Bench 2013 Preview - The Destroyer
Workload Description Applications Used
Photo Sync/Editing Import images, edit, export Adobe Photoshop CS6, Adobe Lightroom 4, Dropbox
Gaming Download/install games, play games Steam, Deus Ex, Skyrim, Starcraft 2, BioShock Infinite
Virtualization Run/manage VM, use general apps inside VM VirtualBox
General Productivity Browse the web, manage local email, copy files, encrypt/decrypt files, backup system, download content, virus/malware scan Chrome, IE10, Outlook, Windows 8, AxCrypt, uTorrent, AdAware
Video Playback Copy and watch movies Windows 8
Application Development Compile projects, check out code, download code samples Visual Studio 2012

While some tasks remained independent, many were stitched together (e.g. system backups would take place while other scenarios were taking place). The overall stats give some justification to what we've been calling this test internally:

AnandTech Storage Bench 2013 Preview - The Destroyer, Specs
  The Destroyer (2013) Heavy 2011
Reads 38.83 million 2.17 million
Writes 10.98 million 1.78 million
Total IO Operations 49.8 million 3.99 million
Total GB Read 1583.02 GB 48.63 GB
Total GB Written 875.62 GB 106.32 GB
Average Queue Depth ~5.5 ~4.6
Focus Worst-case multitasking, IO consistency Peak IO, basic GC routines

SSDs have grown in their performance abilities over the years, so we wanted a new test that could really push high queue depths at times. The average queue depth is still realistic for a client workload, but the Destroyer has some very demanding peaks. When we first introduced the Heavy 2011 test, some drives would take multiple hours to complete it; today most high performance SSDs can finish the test in under 90 minutes. The Destroyer? So far the fastest we've seen it go is 10 hours. Most high performance SSDs we've tested seem to need around 12–13 hours per run, with mainstream drives taking closer to 24 hours. The read/write balance is also a lot more realistic than in the Heavy 2011 test. Back in 2011 we just needed something that had a ton of writes so we could start separating the good from the bad. Now that the drives have matured, we felt a test that was a bit more balanced would be a better idea.

Despite the balance recalibration, there is just a ton of data moving around in this test. Ultimately the sheer volume of data here and the fact that there's a good amount of random IO courtesy of all of the multitasking (e.g. background VM work, background photo exports/syncs, etc...) makes the Destroyer do a far better job of giving credit for performance consistency than the old Heavy 2011 test. Both tests are valid; they just stress/showcase different things. As the days of begging for better random IO performance and basic GC intelligence are over, we wanted a test that would give us a bit more of what we're interested in these days. As Anand mentioned in the S3700 review, having good worst-case IO performance and consistency matters just as much to client users as it does to enterprise users.

We are reporting two primary metrics with the Destroyer: average data rate in MB/s and average service time in microseconds. The former gives you an idea of the throughput of the drive during the time that it was running the Destroyer workload. This can be a very good indication of overall performance. What average data rate doesn't do a good job of is taking into account response time of very bursty (read: high queue depth) IO. By reporting average service time we heavily weigh latency for queued IOs. You'll note that this is a metric we've been reporting in our enterprise benchmarks for a while now. With the client tests maturing, the time was right for a little convergence.

AT Storage Bench 2013 - The Destroyer (Data Rate)

Update: It appears that something was off in the first run as the 1TB scored 261.52MB/s when I retested it.

I'm not sure if I'm comfortable with the score above. There are no other benchmarks that would indicate the EVO mSATA to be over 20% faster than the 2.5" EVO, so I'm thinking there has been some kind of an error in the test. Unfortunately I didn't have time to rerun the test because The Destroyer takes roughly 12 hours to run and another eight or so hours to be analyzed. However, I managed to run it on the 500GB EVO mSATA and as the graph above shows, its performance is on-par with the 2.5" EVO. I'll rerun the test on the 1TB sample once I get back and will update this based on its output.

AT Storage Bench 2013 - The Destroyer (Service Time)

Performance Consistency & TRIM Validation Random & Sequential Performance
Comments Locked

65 Comments

View All Comments

  • ahar - Thursday, January 9, 2014 - link

    Can we also have one for the article? ;)
    "...so the number's you're seeing here..."
  • Unit Igor - Saturday, January 11, 2014 - link

    Tell me Kristian please would EVO 120GB msata have any advantage over EVO 250gb msata in longer battery life when you compare power consuptipon vs. disk busy times and mb/s.I use my ultrabook only for mails ,somtimes wathing movies and surfing.I dont need more then 120GB SSD but i am willing to buy 250Gb if it would give me more battery life.What i wanted to see in your benchmark is MobileMark 2012 because msata is for laptops and that is where battery life play big role
  • guidryp - Thursday, January 9, 2014 - link

    "endurance is fine for consumer usage"

    Thanks for your opinion, but I'll stick with MLC.

    Do you also think Multi-TB HDDs are fine for consumer use? Since HDDs went over 1TB, they have been failing/wearing out for me regularly. I am sure you can find some theoretical numbers that say these are "fine for consumer usage" as well.

    There is a big trend to bigger sizes but lower reliability. That trend can get stuffed.

    Samsungs advantage of Being the only TLC player strikes me as a reason to avoid Samsung, so I can avoid TLC and decreasing endurance that goes with it.
  • Kristian Vättö - Thursday, January 9, 2014 - link

    That's just your experience, it's not a proof that over 1TB hard drives are less reliable. We can't go out and start claiming that they are less reliable unless we have some concrete proof of that (failures on our end, statistics etc).

    The same applies for TLC. All we have is the P/E cycle number and frankly it gives us a pretty good estimation of the drive's lifespan and those numbers suggest that the endurance of TLC is completely fine for consumer usage. Or do you think our calculations are incorrect?
  • MrSpadge - Thursday, January 9, 2014 - link

    And add to that that the P/E cycles are usually conservatively estimated by manufacturers. The SSD-burn-tests at XS sometimes exceed the ratings significantly.
  • guidryp - Thursday, January 9, 2014 - link

    I think if you examine any aggregate source of reviews like Newegg you will see a significant drop in drive satisfaction do to early failures, since drives went over 1TB. So it isn't just some personal fluke that half of my >1TB drives have failed worn out, so far.

    I am really sick of this trend of declining reliability being sold as good enough. If TLC is "good enough" I will take MLC with 3X "good enough" unless the we are talking about 1/3 the price for TLC.

    Weren't the Samsung 840s failing in days for Anand last year?

    Unlike reviewers, I use my products until they fail, so reliability matters a LOT, and is something that is going in the wrong direction IMO.
  • Kristian Vättö - Thursday, January 9, 2014 - link

    Reliability is not the same as endurance. TLC has lower endurance, that's a fact, but it's not less reliable. Endurance is something you can predict (in the end all electronics have a finite lifespan) but reliability you cannot since there's a lot else than just NAND that can fail. I would claim that today's SSDs are much more reliable than the SSDs we had two years ago -- there haven't been any widespread issues with current drives (compared to e.g. early SandForce drives).

    Yes, we had a total of three 840 and 840 Pros that failed but that was on pre-production firmware. The retail units shipped with a fixed firmware.

    This isn't a new trend. Historically we can go back all the way to 1920s when light bulb companies started rigging their products so the lifespan would be shorter, which would in turn increase sales. Is it fair? Of course not. Do all companies do it? Yes.

    I do see your point but I think you're exaggerating. Even TLC SSDs will easily outlive the computer as a whole since the system will become obsolete in in a matter of years anyway if it's not updated.
  • gandergray - Saturday, January 25, 2014 - link

    For information concerning hard drive failure rates that is more objective, please see the following article: http://www.extremetech.com/extreme/175089-who-make... .
  • althaz - Thursday, January 9, 2014 - link

    TLC is NOT a trade off in reliability, but a tradeoff in longevity.

    Longevity is measured in write-cycles and with heavy consumer loads TLC drives will still last for many years.
  • bsd228 - Thursday, January 9, 2014 - link

    Other than the fact that they both store data, SSDs and HDDs have nothing in common, so it's silly to presume a problem that isn't really what you think it is in the first place. HDDs got dirt cheap as we cross the TB threshold and with it went diligent QA. You want 2TB for $80, you're going to get a higher defect rate. And going to 4 or 5 platters just increases the failure points, but the razor thin margins are the culprit here.

    In contrast, a bigger SSD just means either more chips, or higher density ones. But 16 chips is nothing new, and since there are no mechanical parts, nothing to worry about. Aside from OCZ, the SSD track record for reliability has been pretty solid, and Samsung (and Intel) far better than that. If you want to stick to 256G in your laptop out of a silly fear of TLC, you're just hurting yourself. The Anand guys have already shown how overstated the wear issue has become.

Log in

Don't have an account? Sign up now