Random Read/Write Speed

The four corners of SSD performance are as follows: random read, random write, sequential read and sequential write speed. Random accesses are generally small in size, while sequential accesses tend to be larger and thus we have the four Iometer tests we use in all of our reviews.

Our first test writes 4KB in a completely random pattern over an 8GB space of the drive to simulate the sort of random access that you'd see on an OS drive (even this is more stressful than a normal desktop user would see). I perform three concurrent IOs and run the test for 3 minutes. The results reported are in average MB/s over the entire time. We use both standard pseudo randomly generated data (data is random within a write, but duplicated between writes) for each write as well as fully random data (data is random within a write and random across most writes) to show you both the maximum and minimum performance offered by SandForce based drives in these tests. The average performance of SF drives will likely be somewhere in between the two values for each drive you see in the graphs. For an understanding of why the type of data you're writing matters, read our original SandForce article.

Iometer - 4KB Random Write, 8GB LBA Space, QD=3

Random write performance has always been the weak spot of Toshiba’s controllers, this latest combination of controller and firmware is no different. Compared to all other SSDs, the Toshiba based SSDNow V+ 100 doesn’t look very good. It’s even slower than the old Indilinx based Corsair Nova. It’s still over 2x the speed of the fastest desktop 3.5” hard drive however, and enough to give you the feel of an SSD for the most part.

Crucial loses a decent amount of performance when going from 128GB to 64GB. The RealSSD C300 drops below even the worst case scenario performance of the Corsair Force F40.

Note that not all SandForce drives are created equal here. If a manufacturer doesn’t meet SandForce’s sales requirements, their drives are capped at a maximum of 50MB/s here. This is the case with the Patriot Inferno, although OCZ’s Agility 2 voluntarily enforces the limit.

Many of you have asked for random write performance at higher queue depths. What I have below is our 4KB random write test performed at a queue depth of 32 instead of 3. While the vast majority of desktop usage models experience queue depths of 0 - 5, higher depths are possible in heavy I/O (and multi-user) workloads:

Iometer - 4KB Random Write, 8GB LBA Space, QD=32

Kingston's performance doesn't change regardless of queue depth. The SandForce drives either stay the same or get a little faster as they have more requests to parallelize and write sequentially.

Our random read test is similar to the random write test however we lift the 8GB LBA space restriction:

Iometer - 4KB Random Read, QD=3

Random read speed is much better, but still not as good as the competition. At 19.7MB/s the SSDNow V+ 100 is a couple of orders of magnitude faster than the 600GB VelociRaptor. The Indilinx based Corsair Nova is even faster however, and nothing can top the RealSSD C300.

Prices and New Competitors Sequential Read/Write Speed
Comments Locked

96 Comments

View All Comments

  • Gonemad - Thursday, November 11, 2010 - link

    Yes, a longevity test. Put it on grind mode until it pops. P51 Mustangs benefited from engines tested that way.

    Now, this one, should it be fully synthetic or more life like? Just place the drive in "write 0, write 1" until failure, record how many times it can be used, or create some random workload scripted in such a manner that it behaves pretty much like real usage, overusing the first few bytes of every 4k sector... if it affects any results. What am I asking is, will it wear only the used bytes, or the entire 4k rewritten sector will be worn evenly, if I am expressing myself correctly here.

    On another comment, I always thought SSD drives were like overpriced undersized Raptors, since they came to be, but damn... I hope fierce competition drive the prices down. Way down. "Neck and neck to mag drives" down.

    And what about defragmenting utilities? Don't they lose their sense of purpose on a SSD? Are they blocked from usage, since the best situation you have on a SSD is actually randomly sprayed data, because there is no "needle" running over it at 7200rpm in a forcibly better sequential read? Should they be renamed to "optimization tools" when concerning SSD's? Should anybody ever consider manually giving permission to a system to run garbage collection, TRIM, whatever, while blocking it until strictly necessary, in order to increase life span?
  • Iketh - Thursday, November 11, 2010 - link

    win7 "automatically disables" their built-in defrag for SSDs, though if you go in and manually tell it to defrag an SSD, it will do it without question

    prefetch and superfetch are supposedly also disabled automatically when the OS is installed on an SSD, though I don't feel comfortable until i change the values myself in the registry
  • cwebersd - Thursday, November 11, 2010 - link

    I have torture-tested a 50 GB Sandforce-based drive (OWC Mercury Extreme Pro RE) with the goal to destroy it. I stopped writing our semi-random data after 21 days because I grew tired.
    16.5 TB/d, 360 disk fills/day, 21 days more or less 24/7 duty cycle (we stopped a few times for an hour or two to make adjustments)
    ~7500 disk fills total, 350 TB written
    The drive still performs as good as new, and SMART parameters look reasonably good - to the extent that current tools can interpret them anyway.

    If I normally write 20 GB/d this drive is going to outlast me. Actually, I expect it to die from "normal" (for electronics) age-related causes, not flash cells becoming unwritable.
  • Anand Lal Shimpi - Thursday, November 11, 2010 - link

    This is something I've been working on for the past few months. Physically wearing down a drive as quickly as possible is one way to go about it (all of the manufacturers do this) but it's basically impossible to do for real world workloads (like the AT Storage Bench). It would take months on the worst drives, and years on the best ones.

    There is another way however. Remember NAND should fail predictably, we just need to fill in some of the variables of the equation...

    I'm still a month or two away from publishing but if you're buying for longevity, SandForce seems to last the longest, followed by Crucial and then Intel. There's a *sharp* fall off after Intel however. The Indilinx and JMicron stuff, depending on workload, could fail within 3 - 5 years. Again it's entirely dependent on workload, if all you're doing is browsing the web even a JMF618 drive can last a decade. If you're running a workload full of 4KB random writes across the entire drive constantly, the 618 will be dead in less than 2 years.

    Take care,
    Anand
  • Greg512 - Thursday, November 11, 2010 - link

    Wow, I would have expected Intel to last the longest. I am going to purchase an ssd and longevity is one of my main concerns. In fact, longevity is the main reason I have not yet bought a Sandforce drive. Well, I guess that is what happens when you make assumptions. Looking forward to the article!
  • JohnBooty - Saturday, November 13, 2010 - link

    Awesome news. Looking forward to that article.

    A torture test like that is going to sell a LOT of SSDs, Anand. Because right now that's the only thing keeping businesses and a lot of "power users" from adopting them - "but won't they wear out soon?"

    That was the exact question I got when trying to get my boss to buy me one. Though I was eventually able to convince him. :)
  • Out of Box Experience - Thursday, November 11, 2010 - link

    Here's the problem

    Synthetic Benchmarks won't show you how fast the various SSD controllers handle uncompressible data

    Only a copy and paste of several hundred megabytes to and from the same drive under XP will show you what SSD's will do under actual load

    First off, due to Windows 7's caching scheme, ALL drives (Slow or Fast) seem to finish a copy and paste in the same amount of time and cannot be used for this test

    In a worst case scenario, using an ATOM computer with Windows XP and Zero SSD Tweaks, a OCZ VERTEX 2 will copy and paste data at only 3.6 Megabytes per second

    A 5400RPM laptop drive was faster than the Vertex 2 in this test because OCZ drives require massive amounts of Tweaking and highly compressible data to get the numbers they are advertizing

    A 7200RPM desktop drive was A LOT faster than the Vertex 2 in this type of test

    Anyone working with uncompressible data "already on the drive" such as video editors should avoid Sandforce SSD's and stick with the much faster desktop platter drives

    Using a slower ATOM computer for these tests will amplify the difference between slower and faster drives and give you a better idea of the "Relative" speed difference between drives

    You should use this test for ALL SSD's and compare the results to common hard drives so that end users can get a feel for the "Actual" throughput of these drives on uncompressible data

    Remember, Data on the Vertex drive's is already compressed and cannot be compressed again during a copy/paste to show you the actual throughput of the drive under XP

    Worst case scenario testing under XP is the way to go with SSD's to see what they will really do under actual workloads without endless tweaking and without getting bogus results due to Windows 7's caching scheme
  • Anand Lal Shimpi - Friday, November 12, 2010 - link

    The issue with Windows XP vs. Windows 7 doesn't have anything to do with actual load, it has to do with alignment.

    Controllers designed with modern OSes in mind (Windows 7, OS X 10.5/10.6) in mind (C300, SandForce) are optimized for 4K aligned transfers. By default, Windows XP isn't 4K aligned and thus performance suffers. See here:

    http://www.anandtech.com/show/2944/10

    If you want the best out of box XP performance for incompressible data, Intel's X25-M G2 is likely the best candidate. The G1/G2 controllers are alignment agnostic and will always deliver the same performance regardless of partition alignment. Intel's controller was designed to go after large corporate sales and, at the time it was designed, many of those companies were still on XP.

    Take care,
    Anand
  • Out of Box Experience - Saturday, November 13, 2010 - link

    Thanks Anand

    Thats good to know

    With so many XP machines out there for the foreseeable future, I would think more SSD manufacturers would target the XP market with alignment agnostic controllers instead of making the consumers jump through all these hoops to get reasonable XP performance from their SSD's

    Last question..

    Would OS agnostic garbage collection like that on the new Kingston SSD work with Sandforce controllers if the manufacturers chose to include it in firmware or is it irrelevant with Duraclass ?

    I still think SSD's should be plug and play on ALL operating Systems

    Personally, I'd rather just use the drives instead of spending all this time tweaking them
  • sheh - Thursday, November 11, 2010 - link

    This seems like a worrying trend, though time will tell how reliable SSDs are long-term. What's the situation with 2Xnm? And where does SLC fit into all that regarding reliability, performance, pricing, market usage trends?

Log in

Don't have an account? Sign up now