Sequential Read Performance

Our first test of sequential read performance uses short bursts of 128MB, issued as 128kB operations with no queuing. The test averages performance across eight bursts for a total of 1GB of data transferred from a drive containing 16GB of data. Between each burst the drive is given enough idle time to keep the overall duty cycle at 20%.

The Patriot Hellfire, in blue, is highlighted as an example of a last-generation Phison E7 drive. Although we didn't test it at the time, the MP500 was based on the same controller and memory.

Burst 128kB Sequential Read (Queue Depth 1)

With a sequential read burst speed just shy of 2GB/s, the Corsair Force MP510 isn't the absolute fastest TLC drive on the market, but there isn't much that can beat it.

Our test of sustained sequential reads uses queue depths from 1 to 32, with the performance and power scores computed as the average of QD1, QD2 and QD4. Each queue depth is tested for up to one minute or 32GB transferred, from a drive containing 64GB of data. This test is run twice: once with the drive prepared by sequentially writing the test data, and again after the random write test has mixed things up, causing fragmentation inside the SSD that isn't visible to the OS. These two scores represent the two extremes of how the drive would perform under real-world usage, where wear leveling and modifications to some existing data will create some internal fragmentation that degrades performance, but usually not to the extent shown here.

Sustained 128kB Sequential Read

On the longer sequential read test, the MP510's standing falls and it is one of the slower drives in the current high-end NVMe segment. However, it does handle reading fragmented data better than most of its competitors.

Sustained 128kB Sequential Read (Power Efficiency)
Power Efficiency in MB/s/W Average Power in W

The power efficiency of the MP510 on the sequential read test is similar to most of the other high-end competition, and the MP510 isn't setting any records.

The Corsair Force MP510 hits quite high sequential read speeds when operating with a high enough queue depth, but it doesn't scale well at lower queue depths with QD4 performance only slightly higher than QD1.

Sequential Write Performance

Our test of sequential write burst performance is structured identically to the sequential read burst performance test save for the direction of the data transfer. Each burst writes 128MB as 128kB operations issued at QD1, for a total of 1GB of data written to a drive containing 16GB of data.

Burst 128kB Sequential Write (Queue Depth 1)

The Corsair Force MP510 handsle bursts of sequential writes just as well as it does random writes, so its SLC cache sets another record. Only a handful of drives can manage more than 2GB/s for QD1 writes, and the MP510 exceeds 2.6GB/s on this test.

Our test of sustained sequential writes is structured identically to our sustained sequential read test, save for the direction of the data transfers. Queue depths range from 1 to 32 and each queue depth is tested for up to one minute or 32GB, followed by up to one minute of idle time for the drive to cool off and perform garbage collection. The test is confined to a 64GB span of the drive.

Sustained 128kB Sequential Write

On the longer sequential write test that adds in some higher queue depths, the MP510 falls out of first place but stays in a fairly high performance tier.

Sustained 128kB Sequential Write (Power Efficiency)
Power Efficiency in MB/s/W Average Power in W

The power efficiency of the MP510 during sequential writes is well below what the BiCS TLC NAND can manage when paired with Toshiba's controller, but still above average when considering the broader field of competition.

The Corsair Force MP510 shows unsteady performance during the sequential write test, indicating that the very fast SLC write cache does fill up and that can have a significant but temporary impact on performance. At its worst, the MP510 is still handling more than 1GB/s of writes on average, so filling up the SLC cache doesn't ruin performance.

Random Performance Mixed Read/Write Performance


View All Comments

  • Wolfclaw - Friday, October 19, 2018 - link

    Not fussed about top end speed, just cheap mass storage in raid or Microsoft Storage, that wipes the floor with HDD's and can satuate a SATA3 interface is more than enough for me. Reply
  • ATC9001 - Friday, October 19, 2018 - link

    Not bad...competition is good to drive prices down, but if I were in the market for an nvme drive I'd take the HP EX920 1TB for 199! Reply
  • euler007 - Friday, October 19, 2018 - link

    I'm really liking these prices. If RAM comes down in price a new PC is in my future. Reply
  • enzotiger - Friday, October 19, 2018 - link

    Please check your numbers. Random read IOPS of 610K is not only by far the highest IOPS among M.2, it actually beats Optane 905P. Highly suspicious. Reply
  • Billy Tallis - Monday, October 22, 2018 - link

    The 610k IOPS for random reads is the advertised specifications from Corsair, not my own measurements. I don't test consumer drives at queue depths high enough to determine whether it can actually hit 610k IOPS, because that doesn't come close to representing any real consumer workload. Reply
  • Hxx - Friday, October 19, 2018 - link

    those prices are wrong right? I see the 480 gb model for 240+ at amazon unless amazon is price gouging. Reply
  • eek2121 - Saturday, October 20, 2018 - link

    All the big retailers have algorithms to automatically shift pricing based on supply vs demand. Anandtech lists the MSRPs, but if everyone rushes out to buy the drive at once, Amazon, Newegg, etc. want to make as much money as possible while still balancing supply vs demand, so the price automatically shifts up. I'm surprised people haven't figured this out yet. That's why you wait for demand to drop before buying a product. Reply
  • ballsystemlord - Sunday, October 21, 2018 - link

    Tallis, where are the 4k sequential read and write tests? I have a use case for them! Reply
  • Billy Tallis - Monday, October 22, 2018 - link

    I doubt it. Whatever OS and filesystem you are using is likely to have a prefetch mechanism that make your small block sequential reads into mostly large block reads, and write caching that will batch up small block sequential writes. If you're trying to bypass the write cache for small block writes, then you probably need to be shopping for an enterprise SSD. Reply
  • ballsystemlord - Monday, October 22, 2018 - link

    Ok. Thanks! Reply

Log in

Don't have an account? Sign up now