Memory

The eight DIMM slots in the Z9PE-D8 WS allow up to 64GB of DRAM. Intel NASPT (one of our anticipated VM workloads) requires a minimum of 1GB of DRAM and doesn’t really like 4GB (as it introduces caching effects and leads to anomalous benchmarking results). Even low end clients in SMB environments come with a minimum of 2GB of DRAM nowadays, and hence, we decided to go with 2GB of DRAM for each VM. As SMB NAS speeds approach 200 MBps,  it is sometimes necessary to have data sources and sinks capable of handling such speeds for file transfer tests. One option would be to have a really fast SSD or SSDs in RAID0. However, this introduces unnecessary extra variables into the mix. A RAM disk makes for a better solution, and in our build, also acts as a common storage resource for multiple VMs.

Keeping the above factors in mind, we decided to max out the capabilities of the Z9PE-D8 WS by installing 64GB of DRAM. We used G.Skill’s RipjawsZ F3-12800CL10Q2-64GBZL (8Gx8) modules. This quad-channel DDR3 kit is rated for operation at 1.5V and 1600 MHz with a CAS latency of 10-10-10-30. For our application, no overclocking was necessary. The Z9PE-D8 WS BIOS actually set it to 1333 MHz by default. We did find the performance at that setting to be good enough for our workloads, but decided to run the DIMMs at the native setting in the final configuration. Even though G.Skill targets the X79 platform, we had no trouble using it with the Z9PE-D8 WS. The combination of high capacity and efficiency made the G.Skill RipjawsZ a good choice for our testbed.

Storage

The storage subsystem is one of the most important aspects in a build meant to host multiple VMs concurrently. In our earlier NAS testbed, we used to run our VMs off a Seagate 2TB hard drive which had the host OS and the VMs in separate partitions. However, this is not a suitable solution for running multiple VMs concurrently. Hence, we made a decision to devote one physical disk to each VM. Fortunately, the Z9PE-D8 WS had 14 SATA ports.

Our planned workload doesn’t involve the storage of massive media files or any other such data which calls for hard disk drives in the testbed. The only exception is the robocopy test where we transfer a Blu-ray folder structure (with a size of 10.7GB) to the NAS and back. It is quite easy to handle that with a RAM disk, and hence, we decided to go with an SSD-only build.

We decided to equip the testbed with a 128GB OCZ Vertex 4 SSD for the host OS (Windows Server 2008 R2) and devote one 64GB OCZ Vertex 4 SSD to each VM. During the initial installation, we found that Windows Server 2008 R2 recommended at least 80GB of disk space for the primary partition. With the necessity to store temporary files for 12 VMs at the same time, we would have run the host OS SSD very close to full capacity. To resolve this, we installed another 128GB SSD to store the VM data and other necessary softwares.

The 128GB OCZ Vertex 4 provides up to 560 MBps / 430 MBps sequential read / write performance, and 90K / 120K IOPS for 4KB random reads and writes. At idle, the power consumption is 1.3W and it shoots up to 2.5W when fully active. These numbers remain the same for the 64GB OCZ Vertex 4. However, the sequential read / write performance drops down to 460 MBps / 220 MBps and the IOPS for 4K random reads and writes come in at 70K / 85K. Our aim in going with an SSD-only build was to make sure that the system’s storage subsystem didn’t end up being a bottleneck for our VMs. The much lower power consumption (compared to several distinct hard disk drives) ends up being an added bonus.

 

Hardware Build - Motherboard, CPUs and Coolers Hardware Build - Networking
Comments Locked

74 Comments

View All Comments

  • mfed3 - Wednesday, September 5, 2012 - link

    someone didn't read the title of the article or the article itself. the purpose is to set up a testbed, not build a system with this software target in mind.
  • Zink - Wednesday, September 5, 2012 - link

    At the same time this system seems extremely over the top for the uses mentioned. It seems likely that the same tests could be run with much less hardware. I know the testbed as specced can be used for much more than testing NAS performance but the only use discussed is simulating the network utilization of a SMB environment.
    The SSDs are justified because a single HDD was "not suitable" for 12VMs but it seems there are intermediate solutions such as RAIDing two 512GB SSDs that would provide buckets of performance and a cleaner solution than 14 individual disks. I also do not understand how having a physical CPU core per VM is needed to “ensure smooth operation” if network benchmarking software is I/O bound and runs fine on a Pentium 4. Assuming you really do need 64GB of RAM for shared files and Windows VMs then it seems a 1P 2011 board would be more than up to running these benchmarks. Switch to Linux VMs for Dynamo and you could try running the benches from an even lighter system such as an i7-3770.
    On the network side would it not also be possible to virtualize the physical LAN? The clients could connect together over the internal network and the host OS on the tested perform the switch’s role and stress the NAS over a single aggregated link? For testing NAS performance specifically, what would the effect be of removing the VMs entirely and just running multiple Iometer sessions over a single aggregated link or letting Iometer use the multiple NICs from the host OS?
    NAS benchmarking would be an interesting application to try to optimize a system for. A simpler system would help you out with reducing power consumption, increasing reliability and reducing cost. You could run some experiments by changing the system configuration and benching again to see if the same NAS performance can be generated. Figuring out what other kinds of systems generate the same results would also make it possible for other editors to bench NAS units without having to purchase 14 SSDs.
    Sorry for complaining about the system configuration, I know you built it to test other hardware and not as a project in itself but I find the testbed more interesting than the NAS performance.
  • ganeshts - Thursday, September 6, 2012 - link

    Zink, Thanks for your comment. Let me try to address your concerns one-by-one, starting with the premise that the current set of tests are not the only ones we propose to run in the testbed. That premise accounts for devoting a single physical core to each VM.

    As for the single disk for each VM vs. RAIDed SSDs, that was one of the ideas we considered. However, we decided to isolate the VMs from each other as much as possible. In fact, if you re-check the build, the DRAM is the only 'hardware component' that is shared.

    We didn't go with the 'virtualizing the physical LAN' because that puts an upper limit to the number of clients which can be set up for benchmarking purpose (dependent on the host resources). In the current case, using an external switch and one physical LAN port for each VM more accurately represents real world usage. Also, in case we want to increase the number of clients, it is a simple matter of connecting more physical machines to the switch.

    Multiple IOMeter sessions: As far as we could test out / understand, IOMeter doesn't allow multiple simultaneous sessions on a given machine. One can create multiple workers, but synchronizing across them is a much more difficult job than synchronizing the dynamo processes across multiple machines. I am also not sure if the workers on one machine can operate through different network interfaces.

    As noted by another reader, 12 VMs haven't been able to max out the N4800 from Thecus. The next time around, we will probably go with the RAIDing 512 GB SSD option for storage of the VM disks. Physical NICs are probably going to remain (along with one physical CPU core or, probably, thread, for each VM).
  • bobbozzo - Thursday, September 6, 2012 - link

    Hi Ganesh,

    Could you post power consumption for the server with the CPUs loaded (with Prime95 or whatever)?

    I'm thinking of building something like this for a webserver.

    Thanks!
  • ganeshts - Friday, September 7, 2012 - link

    Power consumption with Prime95 set for maximum power consumption was 202 W with all CPU cores 100% loaded. Note that the BIOS has a TDP limit of 70W before throttling the cores down.

    However, I noticed that RAM usage in that particular scenario was only 4 GB in total out of the 64 GB available. It is possible that higher DRAM activity might result in more power usage.
  • Stahn Aileron - Wednesday, September 5, 2012 - link

    Just out of curiosity, when you run with multiple clients accessing the NAS, are they all running the (exact?) same type of workload? Or is each VM/client set to use a slightly (if not entirely) different workload?

    I'm curious since, from a home network PoV, I can see multiple access coming from say:

    -One (or more) client(s) streaming a movie (or maybe music)
    -Another (or several) doing copy (reads) from the NAS
    -Others doing writes to the NAS
    -Maybe even one client (I can't really imagine more) doing a torrent (I don't like the idea of a client using a mounted shared network device as the primary drive for torrenting, but you never know. Also, some NASes feature built-in torrent functionality as a feature.)

    I'm just wondering how much the workload from each client differs from one another, if at all, when conducting your tests/benchmarks.

    Also, for the NASes that do RAID, will you be testing how array degradation and/or rebuilding impacts client usage benchmarks?
  • ganeshts - Wednesday, September 5, 2012 - link

    Stahn,

    Thanks for your feedback. This is exactly what I am looking at from our readers.

    As for your primary question, in our benchmark case, all the VMs are running the same type of workload at a given time. The type of workload is given in the title of each graph.

    It should be possible to set up an IOMeter benchmark ICF file with the type of multiple workloads that you are mentioning. I will try to frame one and try to get it processed for the next NAS review.

    Ref. array degradation / rebuild process : Right now, we present results indicating the time taken to rebuild the array when there is no access to the NAS. I will set up a NASPT run when rebuild is in progress to get a feel of how the rebuild process affects the NAS performance.
  • Stahn Aileron - Thursday, September 6, 2012 - link

    Glad to be of some help. To be honest, benchmarking and running tests (troubleshooting) is something I used to do in the Navy as Avionics Technician. I actually do kind of miss it (especially being a tech geek.) Reminiscing aside...

    Back on-topic: what I described in my previous post was more of a home user secenario. Is there anything else you would also need/want to consider in a more work-oriented "dissimilar multi-client workload" benchmark/test? If this was a SOHO environment, I would add the following to my previous post:

    -DB access (not sure how you want to distribute the read/write workload, though I suppose leaning heavier to reads).

    I mention this now because my previous post for read/writes was more along the lines of sequential instead of random. I would guess DB access would be more random-ish in nature.

    For other work-oriented scenarios in a "dissimilar multi-client workload" benchmark, I'm not sure what else could be added. I'm mainly just a power-user. I dunno is people would really use an NAS for, say, an Exchange Server's storage or maybe a locally-hosted website. (Some NASes come with Web service funtions and features, no?)

    I'm just throwing out ideas for consideration. I don't xpect you to implement everything and anything since you don't have the time to do that. Time is your most precious resource during testing and benchmarking, after all.

    Thank you all for running a wonderful website and to Ganesh for a quick reply.

    Oh, one last thing: does disk fragmentation matter in regards to NASes? Would it affect NAS perfomance? Do any NASes defrag themselves?

    This is more of a long-term issue, so you can't really test it readily I'm guessing. (Unless you happen to have a fragmented dataset you could clone to the NAS somehow...) I haven't heard much about disk fragmentation since the advent of SSDs in the consumer space. That, and higher perfomance HDDs. This is mainly just a curiosity for me. (I do have a more personal reason for my interest, but it's a long story...)
  • insz - Wednesday, September 5, 2012 - link

    Interesting article. Would it be possible to add some pics of the final setup? It'd be interesting to see what the testbed would look like assembled and wired up.
  • ganeshts - Friday, September 7, 2012 - link

    I didn't add the pics to the article because the setup wasn't 'photogenic' after final assembly and placement in my work area :) (as the album below shows). Doesn't matter, I will just link it in this comments section

    2012 AnandTech SMB / SOHO NAS Testbed : http://imgur.com/a/h4bQR

    Individual images:

    http://i.imgur.com/hjD9qh.jpg

    http://i.imgur.com/PJ91Vh.jpg

    http://i.imgur.com/2BcEfh.jpg

    http://i.imgur.com/dvmbrh.jpg

Log in

Don't have an account? Sign up now