Live Long and Prosper: The Logical Page

Computers are all about abstraction. In the early days of computing you had to write assembly code to get your hardware to do anything. Programming languages like C and C++ created a layer of abstraction between the programmer and the hardware, simplifying the development process. The key word there is simplification. You can be more efficient writing directly for the hardware, but it’s far simpler (and much more manageable) to write high level code and let a compiler optimize it.

The same principles apply within SSDs.

The smallest writable location in NAND flash is a page; that doesn’t mean that it’s the largest size a controller can choose to write. Today I’d like to introduce the concept of a logical page, an abstraction of a physical page in NAND flash.

Confused? Let’s start with a (hopefully, I'm no artist) helpful diagram:

On one side of the fence we have how the software views storage: as a long list of logical block addresses. It’s a bit more complicated than that since a traditional hard drive is faster at certain LBAs than others but to keep things simple we’ll ignore that.

On the other side we have how NAND flash stores data, in groups of cells called pages. These days a 4KB page size is common.

In reality there’s no fence that separates the two, rather a lot of logic, several busses and eventually the SSD controller. The latter determines how the LBAs map to the NAND flash pages.

The most straightforward way for the controller to write to flash is by writing in pages. In that case the logical page size would equal the physical page size.

Unfortunately, there’s a huge downside to this approach: tracking overhead. If your logical page size is 4KB then an 80GB drive will have no less than twenty million logical pages to keep track of (20,971,520 to be exact). You need a fast controller to sort through and deal with that many pages, a lot of storage to keep tables in and larger caches/buffers.

The benefit of this approach however is very high 4KB write performance. If the majority of your writes are 4KB in size, this approach will yield the best performance.

If you don’t have the expertise, time or support structure to make a big honkin controller that can handle page level mapping, you go to a larger logical page size. One such example would involve making your logical page equal to an erase block (128 x 4KB pages). This significantly reduces the number of pages you need to track and optimize around; instead of 20.9 million entries, you now have approximately 163 thousand. All of your controller’s internal structures shrink in size and you don’t need as powerful of a microprocessor inside the controller.

The benefit of this approach is very high large file sequential write performance. If you’re streaming large chunks of data, having big logical pages will be optimal. You’ll find that most flash controllers that come from the digital camera space are optimized for this sort of access pattern where you’re writing 2MB - 12MB images all the time.

Unfortunately, the sequential write performance comes at the expense of poor small file write speed. Remember that writing to MLC NAND flash already takes 3x as long as reading, but writing small files when your controller needs large ones worsens the penalty. If you want to write an 8KB file, the controller will need to write 512KB (in this case) of data since that’s the smallest size it knows to write. Write amplification goes up considerably.

Remember the first OCZ Vertex drive based on the Indilinx Barefoot controller? Its logical page size was equal to a 512KB block. OCZ asked for a firmware that enabled page level mapping and Indilinx responded. The result was much improved 4KB write performance:

Iometer 4KB Random Writes, IOqueue=1, 8GB sector space Logical Block Size = 128 pages Logical Block Size = 1 Page
Pre-Release OCZ Vertex 0.08 MB/s 8.2 MB/s

A Quick Flash Refresher The Cleaning Lady and Write Amplification
Comments Locked

295 Comments

View All Comments

  • shabby - Monday, August 31, 2009 - link

    The 80gig g2 is $399 now!
  • gfody - Tuesday, September 1, 2009 - link

    The gen2 80gb is at $499 as of 12:00AM PST
  • maxfisher05 - Monday, August 31, 2009 - link

    As of right now (8/31) newegg has the 160GB Intel G2 listed at $899!!!!!!!!!!!!!!!!!!! To quote Anand "lolqtfbbq!"
  • siliq - Monday, August 31, 2009 - link

    Great article! Love reading this. Thanks Anand.

    We gather from this article that all the pain-in-@$$ about SSDs come from the inconsistency between the size of the read-write page and the erase block. When SSDs are reading/writing a page it's 4K, but the minimum size of erasing operation is 512K. Just wondering is there any possibility that manufacturers can come up with NAND chips that allows controllers to directly erase a 4K page without all the extra hassles. What are the obstacles that prevent manufacturers from achieving this today?
  • bji - Tuesday, September 1, 2009 - link

    It is my understanding that flash memory has already been pushed to its limit of efficiency in terms of silicon usage in order to allow for the lowest possible per-GB price. It is much cheaper to implement sophisticated controllers that hide the erase penalty as much as possible than it is to "fix" the issue in the flash memory itself.

    It is absolutely possible to make flash memory that has the characteristics you describe - 4K erase blocks - but it would require a very large number of extra gates in silicon and this would push the cost up per GB quite a bit. Just pulling numbers out of the air, let's say it would cost 2x as much per GB for flash with 4K erase blocks. People already complain about the high cost per GB of SSD drives (well I don't - because I don't steal software/music/movies so I have trouble filling even a 60 GB drive), I can't imagine that it would make market sense for any company to release an SSD based on flash memory that costs $7 per GB, especially when incredible performance can be achieved using standard flash, which is already highly optimized for price/performance/size as much as possible, as long as a sufficiently smart controller is used.

    Also - you should read up on NOR flash. This is a different technology that already exists, that has small erase blocks and is probably just what you're asking for. However, it uses 66% more silicon area than equivalent NAND flash (the flash used in SSD drives), so it is at least 66% more expensive. And no one uses it in SSDs (or other types of flash drives AFAIK) for this reason.
  • bji - Tuesday, September 1, 2009 - link

    Oh I just noticed in the Wikipedia article about NOR flash, that typical NOR flash erase block sizes are also 64, 128, or 256 KB. So the eraseblocks are just as problematic there as in NAND flash. However, NOR flash is more easily bit-addressable so would avoid some of the other penalties associated with NAND that the smart contollers have to work around.

    So to make a NAND or NOR flash with 4K eraseblocks would probably make them both 2X - 4X more expensive. No one is going to do that - it would push the price back out to where SSDs were not viable, just as they were a few years ago.
  • siliq - Tuesday, September 1, 2009 - link

    Amazing answers! Thank you very much
  • morrie - Monday, August 31, 2009 - link

    My laptop is limited to 4 GB swap. While that's enough for 99% of Linux users, I don't shut down my laptop, it's used as a desktop with dozens of apps running and hundreds of browser tabs. Therefore, after a few months of uptime, memory usage climbs above 4 GB. I have two hard drives in the laptop, and set up a software raid0 1GB swap partition, but I went with software raid1 for the other swap partition. So once the ram is used up for swap, the laptop slows noticeably, but after the raid0 swap partition fills up, the raid1 partition really slows it down. Once that fills up, it hits swap files (non raid) which slow it down more. But thanks to the kernel and the way swappiness works, once about 4 GB of Ram plus about 3 GB of physical swap is used, it really slows. I can gain a bit of speed by adding some physical swap files to increase the ratio of physical swap to ram swap (thus changing swappiness through other means), but this only works for another 1 GB of ram.

    No lectures or advice please, on how I'm using up memory or about how 4GB is more than sufficient, my uptimes are in the hundreds of days on this laptop and thanks to ADD/limited attention span, intermittent printer availability for printing out saved browser tabs and other reasons (old habits dying hard being one), my memory usage is what it is.

    So, the big question is, since the laptop has an eSATA port, can I install one of these ssd drives in an externel SATA tray, connected via eSATA to the laptop and move physical swap partitions to the ssd? I believe that swap on the ssd would be a lot faster even on the eSATA wire, than swap on the drives in the laptop (they're 7200 rpm drives btw). I'm aware that using the ssd for swap would shorten it's life, but if it lasts a year till faster laptops with more memory are available (and I get used to virtual machines and saving state so I can limit open browser windows), I'll be happy.

    Buying two of the drives and using them raided in the laptop is too costly right now, when prices drop that'll be a solution for this current laptop.

    Externel SSD over eSATA for Linux swap on a laptop? Faster than my current setup?
  • hpr - Monday, August 31, 2009 - link

    Sounds like you have some very small memory leak going on there.

    Have you tried that Firefox plugin that enables you to have your tabs but it doesn't really have a tab open in memory.


    TooManyTabs
    https://addons.mozilla.org/en-US/firefox/addon/942...">https://addons.mozilla.org/en-US/firefox/addon/942...

    Have fun filling up thousands of tabs and having low memory usage.
  • gstrickler - Monday, August 31, 2009 - link

    You should be able to use an SSD in an eSATA case, and yes, it should be faster than using your internal 7200 RPM drives. You probably want to use an Intel SSD for that (see page 19 of the article and note that the Intel drives don't drop off dramatically with usage).

    If you don't need to storage of your two internal 7200 RPM drives (or if you can get a sufficiently large SSD), you might be better off replacing one of them with an SSD and reconsider how you're allocating all your storage.

    As for printer availability, seems to me it would make more sense to use a CUPS based setup to create PDFs rather than having jobs sit in a print queue indefinitely. Then, print the PDFs at your convenience when you have a printer available. I don't know how your printing setup currently works, but it sounds like doing so would reduce your swap space usage.

Log in

Don't have an account? Sign up now