The NAND: Going Vertical, Not 3D (Yet)

Since we are dealing with a fixed number of NAND packages (unless you want to do Mushkin's and build a dual-PCB design), there are two ways to increase the capacity. One is to increase the capacity per die, which is what has been done before. This is a logical way as the lithographies get smaller (3Xnm was 32Gb, 2Xnm 64Gb and now 1X/2Ynm 128Gb) and you can double the capacity per die while keeping the die are roughly the same as with the previous process node. However, it's not very efficient to increase the capacity per die unless there is a die shrink because otherwise the end result is a large die, which is generally bad for yields and that in turn is bad for financials. Since we are still moving to 128Gb die (Samsung and IMFT have moved already, Toshiba/SanDisk and SK Hynix are still using 64Gb), moving to 256Gb die this quickly was out of question. I do not expect 256Gb die to happen until 2015/2016 and it may well be that we will not even see manufacturers going past 128Gb at that point. We'll see a big push for 3D NAND during the next few years and I am not sure if planar NAND will get to a point where 256Gb die becomes beneficial. 

Samsung SSD 840 EVO mSATA NAND Configurations
Capacity 120GB 250GB 500GB 1TB
Raw NAND Capacity 128GiB 256GiB 512GiB 1024GiB
# of NAND Packages 2 2 4 4
# of Dies per Package 4 8 8 16
Capacity per Die 16GiB 16GiB 16GiB 16GiB
Over-Provisioning 12.7% 9.1% 9.1% 9.1%

If you cannot increase the capacity per die, the only way left is to increase the die count. So far the limit has been eight dies and with traditional packaging methods there is already some performance loss after exceeding four dies per package. That is due to the limits of the interconnects that connect the dies to the PCB and as you add more dies the signal integrity degrades and latency goes up exponentially.

Source: Micron

In order to achieve the 1TB capacity with only four NAND packages, Samsung had to squeeze sixteen NAND dies into one package. To my surprise, when I started researching a bit about Samsung's 16-die NAND, I found out that it's actually nothing new. Their always-so-up-to-date NAND part number decoder from August 2009 mentions a 16-die MLC configuration and I managed to find TechInsights' report of a 512GB SSD used in 2012 Retina MacBook Pro with x-ray shots of a 16-die NAND package. That is a SSD 830 based drive, so I circled back to check the NAND used in the 512GB SSD 830 and it indeed has sixteen 32Gb dies per package too. 

Courtesy of TechInsights

I also made a graph based on the x-ray shot since it's not exactly clear unless you know what you're looking for.

Unfortunately I couldn't find any good x-ray shots of other manufacturers' NAND to see if Samsung's packaging method is different, which would explain their ability to ship a 16-die package with no significant performance loss. However, what I was able to find suggested that others use similar packaging (i.e. an inclinated tower of dies with interconnects falling from both sides). Samsung is also very tight-lipped about their NAND and the technologies involved, so I've not been able to get any details out of them. Anand is meeting with their SSD folks at CES and there is hope that he will be able to convince them to give us even a brief overview.

I am thinking this is not strictly hardware related but software too. In the end, the problem is signal integrity and latency, both which can be overcome with high quality engineering. The two are actually related: Poor signal integrity means more errors, which in turn increases latency because it's up to the ECC engine to fix the error. The more errors there are, the longer it obviously takes. With an effective combination of DSP and ECC (and a bunch of other acronyms), it's possible to stack more dies without sacrificing performance. Samsung's control over the silicon is a huge help here -- ECC needs to be built into the hardware to be efficient and since it's up to Samsung to decide how much resources and die area they want to devote to ECC, they can make it happen.

Introduction, The Drive & The Test Performance Consistency & TRIM Validation
Comments Locked

65 Comments

View All Comments

  • Chubblez - Thursday, January 9, 2014 - link

    I have 8 5 shelf SAN's (EMC AX4-5f) all full of 1 and 2TB drives that would disagree with you.

    I also have 4 256G EVO's that get hammered pretty hard in an HP server, but I can't speak to the very long term effects. Initial data tells me your wrong on that as well.
  • Death666Angel - Sunday, January 12, 2014 - link

    I have 8 2TB drives since October 2011 running in 2 RAID 5 in my file server (consumer grade Samsung, WD, Hitachi). I have 2 1TB 2.5" drives running since 2 years in my desktop. I have a 1TB and a 1.5TB 3.5" drive for well over 4 years in external enclosures. Not one failed. So let's pit anecdotal evidence against anecdotal evidence, shall we?
  • dgingeri - Thursday, January 9, 2014 - link

    I'd like to see Samsung put out a 500Gb or 1TB SSD with a m.2 PCIe connection. That with an adapter for two m.2 PCIe drives on a PCIe x4 card would make my year. I'd finally be able to get all the drives out of my system and bring it down to just motherboard, cards, H100, and power supply.
  • romrunning - Thursday, January 9, 2014 - link

    Why does the Samsung 840 Pro score really badly in the "Destroyer" tests, but are at the top of pretty much every other benchmark shown here?
  • Kristian Vättö - Thursday, January 9, 2014 - link

    Because it only has 7% OP by default, whereas most drives have 12% nowadays. In steady-state even small differences in OP can play a major role.
  • blanarahul - Thursday, January 9, 2014 - link

    Do you think it would be beneficial to 'not' have Turbowrite at or above 500 GB and focus more on parallism?? I ask this because a 250 GB 840 Evo achieves about 260 MB/s after it runs out of TW cache. So I would expect the 500 GB or above Evo to atleast reach 500 MB/s.
  • Solid State Brain - Friday, January 10, 2014 - link

    Question: do Samsung 840 EVO SSDs have less OP than the previous 840 ones, since a percentage of the reserved space is now used for the fixed SLC cache area/TurboWrite?
    With trim enabled, that should not be a problem (as it makes the drive have "dynamic OP" with free space), but under intense workloads it might, if it's the case.
  • Kristian Vättö - Friday, January 10, 2014 - link

    Technically yes. The 840 actually has slightly better IO consistency, which the higher OP explains.

    https://dl.dropboxusercontent.com/u/128928769/IO%2...

    (The title reads Intel DC S3700 but that's because I use it's graph as a basis of other graphs to guarantee that they're all the same size)
  • romrunning - Friday, January 10, 2014 - link

    That's interesting. I guess that means if you manually over-provision the 840 Pro, then it should dominate the Destroyer benchmark as well. It would be nice to test that scenario out.
  • Kristian Vättö - Friday, January 10, 2014 - link

    Due to the nature of the Destroyer, it's unfortunately not possible to assign the test for a certain LBA range (i.e. add OP).

Log in

Don't have an account? Sign up now