Conclusion

Home file servers can be enormously useful in a multi-system residence.  Adding such a specialty system to a home network is easy, and certain file server operating systems are also very simple to set up and administer.  A home file server is also an easy introduction to servers in general.  Remember, the primary purpose of a home file server is storage - so rather than building around the CPU or GPU, build a home file server around the HDDs.  Home file servers don't need powerful CPUs - instead focus on power savings and adequacy rather than raw processing muscle.  While it's difficult - perhaps impossible - to know what HDD models are most reliable, choose your case carefully and take time managing cables so you can maintain optimal environments for your drives.  Buy high-quality PSUs so your HDDs are supplied with clean power.  And of course, back up your data!

"The files are in the computer.  It's so simple!"

Hard Drives
Comments Locked

152 Comments

View All Comments

  • mino - Tuesday, September 6, 2011 - link

    For a plenty of money :)

    Basically, a SINGLE decent raid card costs ~200+ for which you have the rest of the system.

    And you need at least 2 of them for redundancy.

    Also, with a DEDICATED file server and open sourced ZFS, who needs HW RAID? ...
  • alpha754293 - Tuesday, September 6, 2011 - link

    In most cases, the speed of the drives/controller/interface is almost immaterial because you're going to be streaming it over a 1 Gbps network at most.

    And if you actually HAVE 10GoE or IB or Myrinet or any of the others, I'm pretty sure that if you can afford the $5000 switch, you'd "splurge" on the $1000 "proper" HW RAID card.

    Amusing how all these people are like "speed speed speed!!!!" forgetting that the network will likely be the bottleneck. (And wifi is even worse, 0.45 Gbps is the best you can do with wifi-n.)
  • DigitalFreak - Sunday, September 4, 2011 - link

    I've been using Dell PERC-5i cards for years. You can find them relatively cheap on E-bay, and they usually include the battery backup. I believe they're limited to 2TB drives though.
  • JohanAnandtech - Monday, September 5, 2011 - link

    "But there's the fact that software RAID (which is what you're getting on your main board) is utterly inferior to those with dedicated RAID cards"

    hmm. I am not sure those entry-level firmware thingies that have a R in front of them are so superior. They offload most processing tasks to the CPU anyway, and they tend to create problems if they break and you replace them with a new one with a newer firmware. I would be interested to know why you feel that Hardware RAID (except the high end stuff) is superior?
  • Brutalizer - Monday, September 5, 2011 - link

    When you are saying that software raid is inferior to hardware raid, I hope you are aware that hw-raid is not safe against data corruption?

    You have heard about ECC RAM? Spontaneous bit flips can occur in RAM, which is corrected by ECC memory sticks.

    Guess what, the same spontaneous bit flips occur in disks too. And hw-raid does not detect nor correct such bit flips. In other words, hw-raid has no ECC correction functionality. Data might be corrupted by hw-raid!

    Neither does NTFS, ext3, XFS, ReiserFS, etc correct bit flips. Read here for more information, there are also research papers on data corruption vs hw-raid, NTFS, JFS, etc:
    http://en.wikipedia.org/wiki/ZFS#Data_Integrity

    In my opinion, the only reason to use ZFS is because it detects and corrects such bit flips. No other solution does. Read the link for more information.
  • sor - Monday, September 5, 2011 - link

    Many RAID solutions scrub disks, comparing the data on one disk to the other disks in the array. This is not quite as robust as the filesystem being able to checksum, but as your own link points out, the chances of a hard drive flipping bits is something on the order of 1 in 1.6PB, so combined with a RAID that regularly scrubs the data I don't see home users needing to even think about this.
  • Brutalizer - Monday, September 5, 2011 - link

    You are neglecting something important here.

    Say that you repair a raid-5 array. Say that you are using 2TB disks, and you have an error rate of 1 in 10^16 (just as stated in the article). If you repair one disk, then you need to read 2 000 000 000 000 byte, every time you read a bit, an error can occur.

    The chances of at LEAST ONE ERROR, can be calculated by this wellknown formula:
    1 - (1-P)^n
    where P is the probability of an error occuring, and "n" is the number of times the error can occur.

    If you insert those numbers, then it turns out that during repair, there is something like 25% of hitting at least one read error. It might you have hit two errors, or three errors, etc. Thus, there are 25% chance of you getting read errors.

    If you repair a raid, and then run into read errors - you have lost all your data, if you are using raid-5.

    Thus, this silent corruption is a big problem. Say some bits in a video file is flipped - that is no problem. An white pixel might be red instead. Say your rar file has been affected, then you can not open it anymore. Or a database is affected. This is a huge problem for sysadmins:
    http://jforonda.blogspot.com/2006/06/silent-data-c...
  • Brutalizer - Monday, September 5, 2011 - link

    PS. There is 1 in 10^16 that the disk will not be able to recover the bit. But there are more factors involved: current spikes (no raid can do this):
    http://blogs.oracle.com/elowe/entry/zfs_saves_the_...

    bugs in firmware, loose cables, etc. Thus, the chance is much higher than 10^ 16.

    Also, raid does not scrub disks thoroughly. They only compute parity. That is not checksumming data. See here about raid problems:
    http://en.wikipedia.org/wiki/RAID#Problems_with_RA...
  • alpha754293 - Tuesday, September 6, 2011 - link

    @Brutalizer
    Bit flips

    I think that CERN was testing that and found that it was like 1 bit in 10^14 bits (read/write) or something like that. That works out (according to the CERN presentation) to be 1 BIT in 11.6 TiB.

    If a) you're concerned about silent data corruption on that scale, and b) that you're running ZFS - make sure you have tape backups. Since there ARE no offline data recovery tools available. Not even at Sun/Oracle. (I asked.)
  • sor - Monday, September 5, 2011 - link

    Inferior how? I've been doing storage evaluation for years, and I can say that software raid generally performs better, uses negligible CPU, and is easier to recover from failure (no proprietary hardware). The only reason I'd want a hardware RAID is for ease of use and the battery-backed writeback.

Log in

Don't have an account? Sign up now