RAPID: PCIe-like Performance from a SATA SSD

The software story around Samsung's SSD 840 EVO is quite possibly the strongest we've ever seen from an SSD manufacturer. Samsung's SSD Magician got a major update not too long ago, giving it a downright awesome UI. Magician gives you access to SMART details about your drive and provides decent visualization of things like total host writes. I'd love to see the inclusion of total NAND writes reported somewhere, as reporting host writes alone doesn't take into account write amplification and can give a false sense of security for those users deploying drives into very write intensive environments. There's a prominent drive health indicator which is tied to NAND wear and should draw a lot of attention to itself should things get bad. Samsung's SSD Magician also includes a built in benchmark, controllable overprovisioning and secure erase functionality.

Samsung sent us a beta of the next version of its Magician software (4.2) which includes support for RAPID mode (Real-time Accelerated Processing of I/O Data). RAPID is a feature exclusive to the EVO (for now) and comes courtesy of Samsung's NVELO acquisition from last year. As NVELO focused on NAND caching software, you shouldn't be too surprised by RAPID's role in improving storage performance. Unlike traditional SSD caches however that use NAND to cache mechanical storage, RAPID is designed to further improve the performance of an SSD and not make a HDD more SSD-like. RAPID uses some of your system memory and CPU resources to cache hot data, serving it out of DRAM rather than your SSD.

The architecture is rather simple to understand. Enabling RAPID installs a filter driver on your Windows machine that keeps track of all reads/writes to a single EVO (RAPID only supports caching a single drive today). The filter driver looks at both file types/sizes and LBAs, but it fundamentally caches at the block level (it simply gets hints from the filesystem to determine what to cache). File types that are meaningless to cache are automatically excluded (think very large media files), but things like Outlook PST files are prime targets for caching. Since RAPID works at the block level you can cache frequently used parts of a file, rather than having to worry about a file being too big for the cache.

The cache resides in main memory and is allocated out of non-paged kernel memory. In fact, that's the easiest way to determine whether or not RAPID is actually working - you'll see non-paged kernel memory jump in size after about a minute of idle time on your machine:

Presently RAPID will use no more than 25% of system memory or 1GB, whichever comes first. Both reads and writes are cached, but in different ways. The read cache works as you'd expect, while RAPID more accurately does something like buffering/combining for writes. Reads are simple to cache (just look at what addresses are frequently accessed and draw those into the cache), but writes offer a different set of challenges. If you write to DRAM first and write back to the SSD you run the risk of losing a ton of data in the event of a crash or power failure. Although RAPID obeys flush commands, there's always the risk that anything pending could be lost in a system crash. Recognizing this potential, Samsung tells me that RAPID tries to instead focus on combining low queue depth writes into much larger bundles of data that can be written more like large transfers across many NAND die. To test this theory I ran our 4KB random write IOmeter test at a queue depth of 1 with RAPID enabled and disabled:

Samsung SSD 840 EVO 250GB - 4K Random Write, QD1, 8GB LBA Space
  IOPS MB/s Average Latency Max Latency CPU Utilization
RAPID Disabled 22769.31 93.26 MB/s 0.0435 ms 0.7512 ms 13.81%
RAPID Enabled 73466.28 300.92 MB/s 0.0135 ms 31.4259 ms 31.18%

Write coalescing seems to work extremely well here. With RAPID enabled the system sees even better random write performance than it would at a queue depth of 32. Average latency drops although the max observed latency was definitely higher. I've seen max latency peaks as high as 10ms on the EVO, so the increase in max latency is a bit less severe than what the data here indicates (but it's still large).

My test system uses a quad-core Sandy Bridge, so we're looking at an additional 60 - 70% CPU load on a single core when running an unconstrained IO workload. In real world scenarios I'd expect that impact to be much lower, but there's no getting around the fact that you're spending extra cycles on doing this DRAM caching. RAPID will revert into a pass-through mode if the CPU is already tied up doing other things. The technology is really designed to make use of excess CPU and DRAM in modern day PCs.

The potential performance upside is tremendous. While the EVO is ultimately limited by the performance of 6Gbps SATA, any requests serviced out of main memory are limited by the speed of your DRAM. In practice I never saw more than 4 - 5GB/s out of the cache, but that's still an order of magnitude better than what you'd get from the SSD itself. I ran a couple of tests with and without RAPID enabled to further characterize the performance gains:

Samsung SSD 840 EVO 250GB
  PCMark 7 Secondary Storage Score ATSB - Heavy 2011 Workload (Avg Data Rate) ATSB - Heavy 2011 Workload (Avg Service Time) ATSB - Light 2011 Workload (Avg Data Rate) ATSB - Light 2011 Workload (Avg Service Time)
RAPID Disabled 5414 229.6 MB/s 1101.0 µs 338.3 MB/s 331.4 µs
RAPID Enabled 5977 307.7 MB/s 247.0 µs 597.7 MB/s 145.4 µs
% Increase 10.4% 34.0%   75.0%  

The gains in these tests range from only 10% in PCMark 7 to as much as 75% in our Light 2011 workload. I'm in the process of running a RAPID enabled drive against our Destroyer benchmark to see how it fares there. In our two storage bench tests here the impact is actually mostly on the write side, average performance actually regresses slightly in both cases. I'm not entirely sure why that is other than both of these tests were designed to be a bit more write intensive than normal in order to really stress the weaknesses on SSDs at the time. To make sure that reads could indeed be cached I ran ATTO at a couple of different test sizes, starting with our standard 2GB test:

ATTO makes for a great test because we can see the impact transfer size has on RAPID's caching algorithms. Here we see pretty much no improvement until transfers get larger than 32KB, indicating an optimization for caching large block sequential reads. Note that even though ATTO's test file is 2GB in size (and RAPID's cache is limited to 1GB) we're still able to see some increase in performance. At best RAPID boosts sequential read performance by 34%, driving the 250GB EVO beyond 700MB/s. Since the test file is larger than the maximum size of the cache we're ultimately limited by the performance of the EVO itself.

Writes show a different optimization point. Here we see big uplift above 4KB transfer sizes but more or less the same performance once we move to large block sequential transfers. Again this makes sense as Samsung would want to coalesce small writes into large blocks it can burst across many NAND die, but caching large sequential transfers is just risking potential data loss in the event of a crash/unexpected power loss. Here the potential uplift is even larger - nearly 60% over the RAPID-disabled configuration.

To see what would happen if the entire workload could fit within a 1GB cache I reduced the size of ATTO's test set to 512MB and re-ran the tests:

Oh man. Here performance just shoots through the roof. Max sequential read performance tops out at 3.8GB/s. Note that once again we don't RAPID attempting to cache any smaller transfers, only large sequential transfers are of interest. Towards the end of the curve performance appears to regress when the transfer size exceeds 1MB. What's actually happening is RAPID's performance is exceeding the variable ATTO uses to store its instantaneous performance results. What we're seeing here is a 32-bit integer wrapping itself. 

Writes see similarly insane increases in performance. Here the best performance is north of 4GB/s. When the entire workload can fit in the cache, Samsung appears to relax some of its feelings about not caching large transfers unfortunately. The focus extends beyond just small file writes and we see nearly 4GB/s when we're transferring 8MB of data at a time. We're likely also seeing the same issue where RAPID's performance is so high that it's overflowing the 32-bit integer ATTO uses to report it.

While I appreciate the tremendous increase in both read and write performance, part of me wishes that Samsung would be more conservative in buffering writes. Although the cache map is stored on the C: drive and is persistent across boots, any crash or power loss with uncommitted (non-flushed) writes in the DRAM cache runs the risk of not making it to disk. Samsung is quick to point out that Windows issues flush commands regularly, so the risk should be as low as possible, but you're still risking more than had you not deployed another DRAM cache. If you've got a stable system connected to a UPS (or a notebook on a battery) this will sound like paranoia, but it's still a concern. 

If, however, you want to get PCIe-like SSD speeds without shelling out the money for a PCIe SSD, Samsung's RAPID is the closest you'll get.

TurboWrite: MLC Performance on a TLC Drive Performance Consistency & Testing TRIM
Comments Locked

137 Comments

View All Comments

  • MVR - Thursday, November 14, 2013 - link

    It will be very interesting when they start loading these up with more than 512MB of DRAM cache. Imagine a drive with 4-8+ GB on board. The response times would be insane. It is only a matter of time considering you can buy 8GB of SODIMM memory for $70. They could probably put it on board for $50 added cost to the drive - then these would truly act like PCIe SSD cards, except it would totally max out the SATA3 throughput limit.
  • MVR - Thursday, November 14, 2013 - link

    Of course SATA revision 3.2 at 16gbit/sec would sure enjoy it. Imagine a pair of those in RAID 0 :)
  • Wao - Sunday, November 24, 2013 - link

    I'm going to change my old noisy hard disk with a Samsung 840 EVO 1TB model. I am wondering if I really need to enable TRIM in OS X. I check the data sheet. It only said "Yes" about garbage collection and TRIM support. Does it meant that this model has its own garbage collection built-in, or I really need to enable TRIM in OS X. Honestly, I don't like to hack around the system files.
    Thanks !
  • iradel - Monday, November 25, 2013 - link

    In the "IMFT vs. Samsung NAND Comparison" table, how did you get a Pages per Block value of 256 for 19nm TLC (a.k.a. the 840 EVO)? 8KB * 256 pages per block would imply an erase block size of 2048KB, whereas I've read that the 840 EVO has an EBS of 1536KB (which would mean 192 pages per block).

    Where did you get the 256 value?
  • sambrightman - Sunday, September 20, 2015 - link

    I have the same question. I've read both the 840 and 840 EVO have 1536KiB EBS due to TLC, this is the only place saying 2MiB. Did you find an answer?
  • Scraps - Tuesday, November 26, 2013 - link

    What would be the optimum configuration for this situation. A MacBook Pro with 2 samsung evo 1tb. Would striped raid zero be the best ?
  • code42 - Wednesday, December 18, 2013 - link

    Can I use the Samsung 840 Pro 1TB with a NAS solution? Can some propose a nice setup? Thanks
  • Hal9009 - Wednesday, December 18, 2013 - link

    Just received my new ASUS N550JV and updated the slow HD with 840 EVO-Series 750GB SSD, 16GB of G.SKILL 16GB (2 x 8G) 204-Pin DDR3 and a fresh copy of win-7x64...could not be happier, Samsung makes great SSDs
  • 7beauties - Saturday, December 28, 2013 - link

    I bought the Samsung 840 EVO 1TB because Maximum PC gave it a 9 Kick *ss award, but they described it as being MLC. Good ole Anand tells it like it is. This is TLC. I was pretty steamed with Samsung because they describe this as their "new 3 bit MLC NAND," which I wouldn't have bought over Crucial's M500 960GB MLC SSD. Though Anand tries to calm fears of TLC's endurance, I can't understand what a "GiB" is and how I can calculate my drive's life span.
  • verjic - Thursday, February 13, 2014 - link

    I have a question. In some of the tests I found of real life use shows that Kingston V300 and Samsung a practically the same speed and even at copy 2 GB of 26000 files is slowly on samsung with about 30 %!!! Also installing a program like photoshop, takes longer on Samsung than Kingston, difference is not so big but is arou 10-15 %. Why is that? From all the test for kingston and Samsung, everyone say that Samsung is better but I don't see how? If anyone can explain to me, please

Log in

Don't have an account? Sign up now