Putting Theory to Practice: Understanding the SSD Performance Degradation Problem

Let’s look at the problem in the real world. You, me and our best friend have decided to start making SSDs. We buy up some NAND-flash and build a controller. The table below summarizes our drive’s characteristics:

  Our Hypothetical SSD
Page Size 4KB
Block Size 5 Pages (20KB)
Drive Size 1 Block (20KB
Read Speed 2 KB/s
Write Speed 1 KB/s

 

Through impressive marketing and your incredibly good looks we sell a drive. Our customer first goes to save a 4KB text file to his brand new SSD. The request comes down to our controller, which finds that all pages are empty, and allocates the first page to this text file.


Our SSD. The yellow boxes are empty pages

The user then goes and saves an 8KB JPEG. The request, once again, comes down to our controller, and fills the next two pages with the image.


The picture is 8KB and thus occupies two pages, which are thankfully empty

The OS reports that 60% of our drive is now full, which it is. Three of the five open pages are occupied with data and the remaining two pages are empty.

Now let’s say that the user goes back and deletes that original text file. This request doesn’t ever reach our controller, as far as our controller is concerned we’ve got three valid and two empty pages.

For our final write, the user wants to save a 12KB JPEG, that requires three 4KB pages to store. The OS knows that the first LBA, the one allocated to the 4KB text file, can be overwritten; so it tells our controller to overwrite that LBA as well as store the last 8KB of the image in our last available LBAs.

Now we have a problem once these requests get to our SSD controller. We’ve got three pages worth of write requests incoming, but only two pages free. Remember that the OS knows we have 12KB free, but on the drive only 8KB is actually free, 4KB is in use by an invalid page. We need to erase that page in order to complete the write request.


Uhoh, problem. We don't have enough empty pages.

Remember back to Flash 101, even though we have to erase just one page we can’t; you can’t erase pages, only blocks. We have to erase all of our data just to get rid of the invalid page, then write it all back again.

To do so we first read the entire block back into memory somewhere; if we’ve got a good controller we’ll just read it into an on-die cache (steps 1 and 2 below), if not hopefully there’s some off-die memory we can use as a scratch pad. With the block read, we can modify it, remove the invalid page and replace it with good data (steps 3 and 4). But we’ve only done that in memory somewhere, now we need to write it to flash. Since we’ve got all of our data in memory, we can erase the entire block in flash and write the new block (step 5).

Now let’s think about what’s just happened. As far as the OS is concerned we needed to write 12KB of data and it got written. Our SSD controller knows what really transpired however. In order to write that 12KB of data we had to first read 12KB then write an entire block, or 20KB.

Our SSD is quite slow, it can only write at 1KB/s and read at 2KB/s. Writing 12KB should have taken 12 seconds but since we had to read 12KB and then write 20KB the whole operation now took 26 seconds.

To the end user it would look like our write speed dropped from 1KB/s to 0.46KB/s, since it took us 26 seconds to write 12KB.

Are things starting to make sense now? This is why the Intel X25-M and other SSDs get slower the more you use them, and it’s also why the write speeds drop the most while the read speeds stay about the same. When writing to an empty page the SSD can write very quickly, but when writing to a page that already has data in it there’s additional overhead that must be dealt with thus reducing the write speeds.

The Blind SSD Free Space to the Rescue
POST A COMMENT

245 Comments

View All Comments

  • KadensDad - Tuesday, October 27, 2009 - link

    How do these drives fail? I have heard that they will just suddenly die, no more writes or reads possible. What I would like to know is what happens when it dies? Do you lose all data? Just can't write anymore? How does the OS respond? Any early warnings? What about e.g. CRC? How does possibility of data corruption compare to traditional SSD? What about RAID? Since the drives are electrical, not mechanical, this reduces the number of failure vectors and environmental concerns (e.g., ambient temperature over lifetime of the drive). Won't SSDs therefore fail closer together in time in a RAID configuration? This reduces the window of opportunity for fixing an array and also decreases the applicability of RAID, however marginal.
    Reply
  • Dewend - Friday, March 04, 2016 - link

    My partner and i also were seated for lunch, whenever i mentioned to her that I read a script each and every morning newspaper, therefore i chosen to do a little research. Thankfully, I stumbled upon this website, which helped me discover why people consider even thinking about this. http://followersteam.com Reply
  • jackeyroe - Friday, April 22, 2016 - link

    Great, that what you said I agree and I think that all your opinion are nice and smart for me. If you want any sports shoes you can feel free to my website and I will share you coupon code. And for your this opinion of SSD might I know that is it possible for my website, if it is possible I will try to find is there any more tips that could improve all the functions of my sports shoes site. If it works then will be my pleasure and I will share the great article to all my kind friend, they will be happy to share it into the network of them. Would you please take a look to my website firstly:
    http://www.kicksvovo.com/
    Reply
  • davidsmith123 - Thursday, July 07, 2016 - link

    That sure was a lot to take in! Fantastic article though, it has really opened my eyes to the possibilities that Solid State Drives provide. Probably wont be buying one in the immediate future given the so-called depression and such things, but i will certainly keep up with SSD progress.
    Thanks again for your fantastic articles - im sure im not the only one who really appreciates them :)
    http://www.hubnames.com/
    Reply
  • Rahul Ji - Sunday, August 21, 2016 - link

    Dear KadensDad, these drivers fail all the time. With the new technology and new softwares coming , everyday some drivers and some softwares go outdated. This ought to happen. But, what will not go outdated is our Microsoft xbox live codes at http://xbcodejunction.com . These codes work anywhere anytime, anybody can use them. Just so easy to get and use. So come and try your copy for free. Reply
  • adsmith82 - Monday, September 14, 2009 - link

    I need to run HDDErase on an X25-M. No matter what bootable CD or flash drive I create, HDDErase does not see either of my SATA hard drives. I already disabled AHCI in BIOS. Also, I am using version 3.3. I know that 4.0 does not work with the X25-M.

    Can someone help me troubleshoot this please? Thanks.
    Reply
  • gallde - Thursday, June 11, 2009 - link

    You point out that TRIM will only work on deletions, not on overwrites. But, couldn't a smart controller look at blocks that have a majority of invalid pages and "trim" them as well, recovering clean pages as a background process? Reply
  • forsunny - Thursday, August 13, 2009 - link

    Why not just make the SSDs capable of individual page erases instead of blocks? Problem solved. Reply
  • Ron White - Sunday, August 31, 2014 - link

    Erasing the NAND transistor in an SSD requires such a large jolt of voltage that it would affect surrounding transistors. Reply
  • lyeoh - Friday, May 29, 2009 - link

    Good and informative article.

    Regarding the shill tshen83 who claims that Anandtech cost the drive manufacturers millions of dollars in sales.

    If that is true, Anandtech has saved customers millions of dollars.

    Anandtech should care more about their readers losses than drive manufacturer losses. If Anandtech was a site for drive manufacturers and their shills we wouldn't be reading it.

    To me, if the SSD drive manufacturers lose money, it's their own fault for building crap that has higher write latencies than old fashioned drives with metal discs spinning at 7200RPM or slower. Not anandtech's.

    I can get higher sequential reads and writes by using RAID on old fashioned drives. It is much harder to get lower latency. So Anandtech did the right thing for OCZ.

    Lastly, there might be a way of making your windows machine stutter less even with a crap SSD. Note: I haven't tested the actual effect on an SSD because I don't have an SSD.

    Basically by default when Windows accesses a file on NTFS, it will WRITE to the directory the time of the access. Yep, it writes when it opens files and directories (which are just special files). That might explain the stuttering people see. For a lot of things, Windows has to open files.

    Warning! There are reasons why some people or programs would want to know the last access time of files. Me and my programs don't (and I doubt most people would).

    If you are sure that's true for you (or are willing to take the risk) set NtfsDisableLastAccessUpdate=1 as per:

    http://technet.microsoft.com/en-us/library/cc75856...">http://technet.microsoft.com/en-us/library/cc75856...
    Reply

Log in

Don't have an account? Sign up now