Multi-Client Access - NAS Environment

We configured three of the Seagate Enterprise NAS HDD drives in a RAID-5 volume in the QNAP TS-EC1279U-SAS-RP. A CIFS share in the volume was subject to some IOMeter tests with access from up to 25 VMs simultaneously. The following four graphs show the total available bandwidth and the average response time while being subject to different types of workloads through IOMeter. IOMeter also reports various other metrics of interest such as maximum response time, read and write IOPS, separate read and write bandwidth figures etc. Some of the interesting aspects from our IOMeter benchmarking run are available here.

Seagate Enterprise NAS HDD Multi-Client CIFS Performance - 100% Sequential Reads

 

Seagate Enterprise NAS HDD Multi-Client CIFS Performance - Max Throughput - 50% Sequential Reads

 

Seagate Enterprise NAS HDD Multi-Client CIFS Performance - Random 8K - 70% Reads

 

Seagate Enterprise NAS HDD Multi-Client CIFS Performance - Real Life - 60% Random 65% Reads

We see that the sequential accesses are still limited by the network link, but, this time, on the NAS side. On the other hand, our random access tests show markedly better performance for enterprise drives. From a viewpoint of the average access times, the Seagate Enterprise Capacity v4, Seagate Enterprise NAS HDD and the HGST Ultrastar He6 all fall in the same category. Amongst these three, the Enterprise NAS HDD doesn't win out on any particular benchmark. However, that is to be expected - the other drives that are in the comparison list are targeted towards true enterprise-class / datacenter applications.

Single Client Access - NAS Benchmarks RAID-5 Benchmarking - Miscellaneous Aspects
Comments Locked

51 Comments

View All Comments

  • Communism - Wednesday, December 10, 2014 - link

    Seagate 1TB per platter drives have been the fastest (per RPM) ever since their introduction.

    Compare to WD Blacks with 1TB per platter or HGST 1TB per platter drives and in every single sequential benchmark they have been faster.

    The cache size differential between the competing drives has little to do with the sequential results.
  • Laststop311 - Thursday, December 11, 2014 - link

    The seagate did have like 20-30MB/sec faster sequential transfers but the He6 has 2-3 milliseconds faster latency on the access times. Personally I'd rather have the 2-3 milliseconds lower in access time over 20-30MB/sec higher sequential transfers. Not too mention the lower power use, less heat, less noise and hitachis unrivaled reliability. If you are building a dense NAS setup the lower heat per drive really helps out. I feel like you would notice the lower latency more than like 160MB/sec vs 130MB/sec
  • MrSpadge - Thursday, December 11, 2014 - link

    "The cache size differential between the competing drives has little to do with the sequential results."

    I know. That's exactly why I replied this to Ganesh's

    "... Seagate Enterprise Capacity v4 vs. the WD Red Pro at the 4 TB capacity point. Both of them use the same number of platters, have the same rotational speed. The only difference was the cache size."
  • romrunning - Wednesday, December 10, 2014 - link

    All of the performance test charts shown MB/sec generally in the hundreds. However, the "Real Life 60% Random 65% Reads" test shows only single digits in MB/s. Is this a chart labeling problem? If not, why isn't there any explanation about the huge difference?
  • DanNeely - Wednesday, December 10, 2014 - link

    HDDs are very fast for sequential reads/writes because as soon as it finishes reading/writing one sector, the next is underneath the read heads. They're horribly slow for random IO because most of the time is spent moving the read/write heads into place not doing data reads. This has been the case with every HDD for decades. (Possibly all the way to the beginning; but I'm not familiar with very old designs limitations.) The main advantage of SDDs is that because they don't have to move drive heads around they can be many times faster in random IO than a magnetic HDD. (They're still faster in sequential IO; read the intro to SSD articles on this site from a few years ago for details about their architecture.)
  • romrunning - Wednesday, December 10, 2014 - link

    I agree with you, but that is a serious drop-off. Shouldn't an intelligent NAS be able to have different drives look for different parts of those reads with some type of large LUT?
  • MrSpadge - Wednesday, December 10, 2014 - link

    You've just invented Raid 0 / 5 / whatever :)

    For small files the typical transfer rates of HDDs are in the low single-digit range. Even if you have 4 of them and performance scales perfectly, that's still very slow. That's why a good SSD on SATA 2 get still be 10 to 100 times faster than an HDD, depending on the actual usage case, even though their maximum transfer rates are comparable.
  • romrunning - Thursday, December 11, 2014 - link

    That's what I was thinking - the test was performed on a 3-drive RAID-5 array in the QNAP, right? So why isn't it's RAID controller more intelligent?
  • Supercell99 - Thursday, December 11, 2014 - link

    Honestly, most serious enterprises do not use SATA HDD drives for production servers. The queue depth is only 32 vs 256 for SAS drives. SATA drives are fine for backups, the just can't provide the IOPS an Enterprise server running multiple VM's or DB's. Will still need to demand SAS for better IOPS in the HDD storage arena. VSphere VSAN will choke on SATA based disk system if a hosts dies.
  • cm2187 - Thursday, December 11, 2014 - link

    Most clouds use SATA drives.

Log in

Don't have an account? Sign up now