Sequential Read

Intel provides no specifications for sequential access performance of the Optane SSD DC P4800X. Buying an Optane SSD for a mostly sequential workload would make very little sense given that sufficiently large flash-based SSDs or RAID arrays can offer plenty of sequential throughput. Nonetheless, it will be interesting to see how much faster the Optane SSD is with sequential transfers instead of random access.

Sequential access is usually tested with 128kB transfers, but this is more of an industry convention and is not based on any workload trend as strong as the tendency for random I/Os to be 4kB. The point of picking a size like 128kB is to have transfers be large enough that they can be striped across multiple controller channels and still involve writing a full page or more to the flash on each channel. Real-world sequential transfer sizes vary widely depending on factors like which application is moving the data or how fragmented the filesystem is.

Even without a large native page size to its 3D XPoint memory, we expect the Optane SSD DC P4800X to exhibit good performance from larger transfers. A large transfer requires the controller to process fewer operations for the same amount of user data, and fewer operations means less protocol overhead on the wire. Based on the random access tests, it appears that the Optane SSD is internally managing the 3D XPoint memory in a way that greatly benefits from transfers being at least 4kB even though the drive emulates a 512B sector size out of the box.

The drives were preconditioned with two full writes using 4kB random writes, so the data on each drive is entirely fragmented. This may limit how much prefetching of user data the drives can perform on the sequential read tests, but they can likely benefit from better locality of access to their internal mapping tables.

Queue Depth 1

The test of sequential read performance at different transfer sizes was conducted at queue depth 1. Each transfer size was used for four minutes, and the throughput was averaged over the final three minutes of each test segment.

Sequential Read
Vertical Axis scale: Linear Logarithmic

For transfer sizes up to 32kB, both Intel drives deliver similar sequential read speeds. Beyond 32kB the P3700 appears to be saturated but also highly inconsistent. The Micron 9100 is plodding along with very low but steadily growing speeds, and by the end of the test it has almost caught up with the Intel P3700. It was at least ten times slower than the Optane SSD until the transfer size reached 64kB. The Optane SSD passes 2GB/s with 128kB transfers and finishes the test at 2.3GB/s.

Queue Depth > 1

For testing sequential read speeds at different queue depths, we use the same overall test structure as for random reads: total queue depths of up to 64 are tested using a maximum of four threads. Each thread is reading sequentially but from a different region of the drive, so the read commands the drive receives are not entirely sorted by logical block address.

The Optane SSD DC P4800X starts out with a far higher QD1 sequential read speed than either flash SSD can deliver. The Optane SSD's median latency at QD1 is not significantly better than what the Intel P3700 delivers, but the P3700's 99th and 99.999th percentile latencies are at least an order of magnitude worse. Beyond QD1, the Optane SSD saturates while the Intel P3700 takes a temporary hit to throughput and a permanent hit to latency. The Micron 9100 starts out with low throughput and fairly high latency, but with increasing queue depth it manages to eventually surpass the Optane SSD's maximum throughput, albeit with ten times the latency.

Sequential Read Throughput
Vertical Axis units: IOPS MB/s

The Intel Optane SSD DC P4800X starts this test at 1.8GB/s for QD1, and delivers 2.5GB/s at all higher queue depths. The Intel P3700 performs significantly worse when a second QD1 thread is introduced, but by the time there are four threads reading from the drive the total throughput has recovered. The Intel P3700 saturates a little past QD8, which is where the Micron 9100 passes it. The Micron 9100 then goes on to surpass the Optane SSD's throughput above QD16, but it too has saturated by QD64.

Sequential Read Latency
Mean Median 99th Percentile 99.999th Percentile

The Optane SSD's latency increases modestly from QD1 to QD2, and then unavoidably increases linearly with queue depth due to the drive being saturated and unable to offer any better throughput. The Micron 9100 starts out with almost ten times the average latency, but is able to hold that mostly constant as it picks up most of its throughput. Once the 9100 passes the Optane SSD in throughput it is delivering slightly better average latency, but substantially higher 99th and 99.999th percentile latencies. The Intel P3700's 99.999th percentile latency is the worst of the three across almost all queue depths, and its 99th percentile latency is only better than the Micron 9100's during the early portions of the test.

Sequential Write

The sequential write tests are structured identically to the sequential read tests save for the direction the data is flowing. The sequential write performance of different transfer sizes is conducted with a single thread operating at queue depth 1. For testing a range of queue depths, a 128kB transfer size is used and up to four worker threads are used, each writing sequentially but to different portions of the drive. Each sub-test (transfer size or queue depth) is run for four minutes and the performance statistics ignore the first minute.

Sequential Write
Vertical Axis scale: Linear Logarithmic

As with random writes, sequential write performance doesn't begin to take off until transfer sizes reach 4kB. Below that size, all three SSDs offer dramatically lower throughput, with the Optane SSD narrowly ahead of the Intel P3700. The Optane SSD shows the steepest growth as transfer size increases, but it and the Intel P3700 begin to show diminishing returns beyond 64kB. The Optane SSD almost reaches 2GB/s by the end of the test while the Intel P3700 and the Micron 9100 reach around 1.2-1.3GB/s.

Queue Depth > 1

When testing sequential writes at varying queue depths, the Intel SSD DC P3700's performance was highly erratic. We did not have sufficient time to determine what was going wrong, so its results have been excluded from the graphs and analysis below.

The Optane SSD DC P4800X delivers better sequential write throughput at every queue depth than the Micron 9100 can deliver at any queue depth. The Optane SSD's latency increases only slightly as it reaches saturation while the Micron 9100's 99th percentile latency begins to climb steeply well before that drive reaches its maximum throughput. The Micron 9100's 99.999th percentile latency also grows substantially as throughput increases, but its growth is more evenly spread across the range of queue depths.

Sequential Write Throughput
Vertical Axis units: IOPS MB/s

The Optane SSD reaches its maximum throughput at QD2 and maintains it as more threads and higher queue depths are introduced. The Micron 9100 only provides a little over half of the throughput and requires a queue depth of around 6-8 to reach that performance.

Sequential Write Latency
Mean Median 99th Percentile 99.999th Percentile

The Micron 9100's 99th percentile latency starts out around twice that of the Optane SSD, but at QD3 it increases sharply as the drive approaches its maximum throughput until it is an order of magnitude higher than the Optane SSD. The 99.999th percentile latencies of the two drives are separated by a wide margin throughout the test.

Random Access Performance Mixed Read/Write Performance
Comments Locked

117 Comments

View All Comments

  • ddriver - Sunday, April 23, 2017 - link

    It is not expensive because it is new, it is expensive because intel and micron wasted a crapload of money on RDing it and it turned out to be mediocre - significantly weaker than good old and almost forgotten SLC. So now they hype and lie about it and sell it significantly overpriced in hopes they will see some returns of the investment.

    Also, it seems like you are quite ignorant, ignorant enough to not know what "order of magnitude" means. You just heard someone smart using it and decided to imitate, following some brilliant logic that it will make you look smart. Well, it doesn't. It does exactly the opposite. Now either stop using it, or at the very least, look it up, understand and remember what it actually means, so the next time you use it, you don't end up embarrassing yourself.
  • factual - Sunday, April 23, 2017 - link

    "significantly weaker than good old and almost forgotten SLC"

    Seriously ?! You must be getting paid to spew this bs! no one can be this ignorant!! can you read numbers ?! what part of 8.9us latency don't you understand, this is at least 10x better than the latest and greatest NVMe SSDs (be it TLC, VNAND or whatever bs marketing terms they feed idiots like you nowadays).

    what part of 95K/108K QD1 IOPS don't you understand ?! This is 3-10x compared to this best SSDs on the market.

    So I repeat again, Xpoint is orders of magnitude better performing than the latest and greatest SSDs (from Samsung or whichever company) on the market. This is a fact.

    You don't even understand basic math, stop embarrassing yourself by posting these idiotic comments!
  • ddriver - Monday, April 24, 2017 - link

    LOL, your intellect is apparently equal to that of a parrot.
  • factual - Monday, April 24, 2017 - link

    Well if this fruitless exchange is any evidence my intellect is far superior to yours. So If my intellect is equal to that of a parrot, yours must be equal to that of a maggot ... lol
  • evilpaul666 - Saturday, April 22, 2017 - link

    So where are the 32gb client ones?
  • tomatus89 - Saturday, April 22, 2017 - link

    Who is this ddriver troll? Hahaha you are hillarious. And the worse is that people keep feeding him instead of ignoring him.
  • peevee - Saturday, May 27, 2017 - link

    From your testing, looks like the drive offers real advantages on low QD, i.e. for desktop/small office server use. For these uses a normal SSD is also enough though.
    Given that modern Xeons have up to 28 cores (running 56 threads each) and server motherboards have 2 or more CPU slots, a properly loaded server will offer QD > 64 all day long, and certainly not just 4 active threads - where the Micron 9100 offers even higher performance, and if the performance is good enough there, it certainly good enough on lower QDs where it is even better PER REQUEST.
    And who cares what 99.999% latency is, as long as it is milliseconds and not seconds - network and other latencies on the accesses to these servers will be higher anyway.

    An incredibly good first attempt, but it really does not push the envelope in the market it is priced for - high-performance storage-bottlenecked servers.

Log in

Don't have an account? Sign up now