We had a chance to talk to Western Digital last week about a new hard drive technology they have coming down the pipe called Advanced Format. Ahead of the impending launch of the first Advanced Format drives, Western Digital is engaging in an education campaign about the technology, as it represents the first step in a general transition for hard drives, and is a transition that can’t be done completely behind the scenes. Specifically, Advanced Format is going to bring about some issues where older operating systems (and we’re looking almost exclusively at Windows XP here) will inadvertently harm drive performance by making certain assumptions about the drives that will no longer be true.

We’ll get in to the nitty-gritty about the impact of this in a bit, but first let’s start with Advanced Format. What is it? In a nutshell, it’s replacing the traditional 512 byte hard drive sector with a 4 kilobyte (4,096B) hard drive sector for a number of practical and technological reasons. Advanced Format is the name Western Digital is giving to drives implementing 4KB sectors.

The move to 4K-sectors has been some 10 years in the making, as research in the late 90’s identified that 512B sectors made less and less sense as drives continued to get larger. At a time when hard drives were measured in megabytes, 512B sectors were a good balance between technical factors and a desire to keep wasted space at a minimum (you can’t normally put 2 files in 1 sector), but times have changed and hard drive capacities are several orders of magnitude larger.

The crux of the problem is that there are 3 factors that are in constant need of balancing when it comes to hard drive design: areal density, the signal-to-noise ratio (SNR) in reading from drive platters, and the use of Error Correcting Code (ECC) to find and correct any errors that occur. As areal density is increases, sectors become smaller and their SNR decreases. To compensate for that, improvements are made to ECC (usually through the use of more bits) in order to maintain reliability. So for a drive maker to add more space, they ultimately need to improve their error-correction capabilities, which means the necessary ECC data requires more space. Rinse, wash, repeat.


The SNR for various areal densities

At some point during this process drive manufacturers stop gaining any usable space - that is, they have to add as much ECC data as they get out of the increase areal density in the first place - which limits their ability to develop larger drives. Drive manufacturers dislike this both because it hinders their ability to develop new drives, and because it means their overall format efficiency (the amount of space on a platter actually used to store user data) drops. Drive manufacturers want to build bigger drives, and they want to spend as little space on overhead as possible.

But all is not lost. The principle problem here is that ECC correction takes place in 512B chunks, while ECC can be more efficient when used over larger chunks of data. If ECC data is calculated against a larger sector, even though more ECC data is necessary than for a single 512B sector, less ECC data than the sum of multiple sectors is needed to maintain the same level of operational reliability. One estimate for 4K sector technology puts this at 100 bytes of ECC data needed for a 4K sector, versus 320 (40x8) for 8 512B sectors.  Furthermore the larger sectors means that larger erroneous chunks of data can be corrected (burst error correction), something that was becoming harder as greater areal densities made it easier to wipe out larger parts of a 512B sector. As a result, the need for the larger sector is born.

As for why Western Digital and other drive makers are transitioning specifically to a 4K sector, the rationale is based on the earlier problems. 4K sectors are much larger than 512B sectors, which means they benefit more from our earlier ECC optimizations, which in turn brings a greater increase in format efficiency than using smaller sectors (i.e. 1K).

4K also happens to be a magical number elsewhere when it comes to computers – this is primarily rooted in the fact that a normal page of memory on an x86 processor continues to be 4KB (ed: 4MB pages also exist). The x86 page size in turn has also lead to file system clusters (the smallest unit of storage in a file system) becoming 4KB as 4KB clusters neatly fit in to a page of memory, while the need for smaller clusters has subsided amidst a general increase in file size (i.e. fewer files are smaller than 4KB and waste space). NTFS, EXT3, and HFS+ all default to 4KB clusters on modern hard drives. So 4KB physical sectors map perfectly with 4KB file system clusters, which in turn map perfectly with 4KB memory pages. And hence 4KB is the largest practical size for a hard drive sector at this time.

  Old Sector New Sector x86 Mem Page NTFS Cluster HFS+ Cluster
Size 512B 4K 4K 4K 4K

So what’s the net benefit from all of this for consumers? At the moment, not a lot, which is why this is a low-key launch for Western Digital, and the focus of this is an education effort on what the use of 4K sectors means for older operating systems. The biggest benefit is going to be that this will enable Western Digital to more easily design drives over 2TB in size.

From a numbers perspective, Western Digital estimates that the use of 4K sectors will give them an immediate 7%-11% increase in format efficiency. ECC burst error correction stands to improve by 50%, and the overall error rate capability improves by 2 orders of magnitude. In theory these reliability benefits should immediately apply to all 4K sector drives (making the Advanced Format drives more reliable than regular drives), but Western Digital is not pushing that idea at this time.


IDEMA projected format efficiency and ECC burst sizes

With all of that said, to make this kind of advancement practical, a transition period is necessary. That transition period will be done through the use of 512B emulation technology, which will expose Advanced Format drives to the drive controller and operating system as having 512B sectors, when in reality they will have 4K sectors. Unfortunately, this method has a downside, which is the ultimate reason for Western Digital’s education campaign.


512 byte sector emulation: how the drive is laid out (physical) and what the OS sees (logical)

With the emulation of 512B sectors, there’s the risk that a partition could be misaligned compared to the 4K physical sectors - where it would be unwittingly started in the middle of such a sector. As a result, the clusters of a file system on that partition would end up straddling 4K sectors, which would cause performance problems. Specifically, in testing IDEMA (the drive manufacturers’ trade group) found that random writes in would be particularly impacted, as a Read-Modify-Write(RMW) would need to take place to update each sector, rather than a straight write(ed: This is very similar to writing to a dirty block on a SSD). Although this isn’t mechanically/electronically  harmful in any way, the performance hit compared to a straight write makes it undesirable.


A single unaligned 4K cluster straddling 2 physical sectors


A misaligned partition under WinXP. Only LBAs divisible by 8 are aligned

What Operating Systems Are Affected
Comments Locked

86 Comments

View All Comments

  • Iridium130m - Friday, December 18, 2009 - link

    We've been dealing with alignment issues in the server and SAN space for years, especially with virtualization at the vm level and storage level: wrapping data in vmdks on vmfs on wafl filesystems opens many opportunities for misaligned data and up to a 20% performance hit on saturated systems.

    thankfully microsoft FINALLY recognized the issues with server 2008 and formats at a 1mb boundary automatically, which covers the majority of storage arrays and disks.
  • cjcoats - Friday, December 18, 2009 - link

    "Caviar Green... They all have 64MB of cache -
    the first WD drives to come with that much cache..."

    What about the 2TB Caviar Black and Raid Edition drives?
    They have 64MB caches, and dual on-board cache-processors.
    Is that all there is to their performance edge?
  • Ryan Smith - Friday, December 18, 2009 - link

    That was supposed to be "the first WD Caviar Green drives...". Fixed.
  • Byte - Friday, December 18, 2009 - link

    Do the new WD drives give any space advantage? Also will there be any real performane advantage when they release the true higher level 4k clusters?
  • magreen - Friday, December 18, 2009 - link

    quote:

    One estimate for 4K sector technology puts this at 100 bytes of ECC data needed for a 4K sector, versus 320 (40x8) for 8 512B sectors.

    it sounds like you gain a whopping 220 bytes per 4Kb of storage space. That's 5%. Why does WD or anyone else care?
  • Ryan Smith - Saturday, December 19, 2009 - link

    I didn't get in to it in the article, but there's also the benefit of getting rid of other per-sector overhead: sector gaps and sector sync marks. Between that and the ECC improvement, you get the total 7%-11% improvement.
  • DanNeely - Friday, December 18, 2009 - link

    Because it means they need 5% less process shrink for their next generation platters; probably 750GB. Computing less ECC code will let them get away with a marginally cheaper controller chip as well.

    Implementing it in current models is mostly as a largescale testbed for the next generation although they could get an additional small performance gain from a 5% short stroke.
  • Spivonious - Friday, December 18, 2009 - link

    Because 5% of 1TB is 51GB? That's not exactly an insignificant amount.
  • krazyderek - Friday, December 18, 2009 - link

    It would be great for any power user that understands the 5.xx limitations of these drive's to be able to purchase them regardless. I'm building a new linux file server for work and i'd love to drop in this new technology instead of knowing about and not being able to take full advantage of it for several years simply due to the fact that microsoft dropped the ball yet again.
  • Griswold - Friday, December 18, 2009 - link

    I wonder why the "power user" would want these drives, since they come with a significant performance hit under certain conditions when writing - as mentioned in the article.

    To write a 4KB block that spans two physical sectors, the sectors have to be read, data modified and then written. That means the platter has to spin an additional time - thats a pretty big 50% performance hit when it happens.

    The benefits (reliability, capacity) are nice, but to me nothing to call home about. I backup daily and weekly. The last drive that failed me was 15 years ago. I'm content with how things are. But I'm not willing to take performance hits for these benefits. So, I hope they find a way around this issue before these drives are the only ones I can buy...

Log in

Don't have an account? Sign up now