Final Words

If you haven't gotten the hint by now, we'll spell it out for you: there is no place, and no need for a RAID-0 array on a desktop computer. The real world performance increases are negligible at best and the reduction in reliability, thanks to a halving of the mean time between failure, makes RAID-0 far from worth it on the desktop.

There are some exceptions, especially if you are running a particular application that itself benefits considerably from a striped array, and obviously, our comments do not apply to server-class IO of any sort. But for the vast majority of desktop users and gamers alike, save your money and stay away from RAID-0.

If you do insist on getting two drives, you are much better off putting them into a RAID-1 array to have a live backup of your data. The performance hit of RAID-1 is just as negligible as the performance gains of RAID-0, but the improvement in reliability is worthwhile...unless you're extremely unlucky and both of your drives die at the exact same time.

When Intel introduced ICH5, and now with ICH6, they effectively brought RAID to the mainstream, pushing many users finally to bite the bullet and buy two hard drives for "added performance". While we applaud Intel for bringing the technology to the mainstream, we'd caution users out there to think twice before buying two expensive Raptors or any other drive for performance reasons. Your system will most likely run just as fast with only one drive, but if you have the spare cash, a bit more reliability and peace of mind may be worth setting up a RAID-1 array.

Bottom line: RAID-0 arrays will win you just about any benchmark, but they'll deliver virtually nothing more than that for real world desktop performance. That's just the cold hard truth.

Game Loading Performance
Comments Locked

127 Comments

View All Comments

  • timw - Thursday, July 8, 2004 - link

    This isn't really anything new. As someone else mentioned, seek time and cache size with the right firmware optimizations are the most important. RAID 0 won't be able to improve that, and may actually be slower than a single drive in many instances. If you don't believe what Anandtech has to say, take a look at the latest article at storagereview.com.
  • Pumpkinierre - Thursday, July 8, 2004 - link

    Wrong again, mdrohn, Australia to be exact, but I was using an Oxford dictionary. The hyperdictionary.com (USA based I think as it advertises cheap dental insurance for US residents) gives redundancy as:

    Definition: [n] repetition of an act needlessly
    [n] the attribute of being superfluous and unneeded; "the use of industrial robots created redundancy among workers"
    [n] (electronics) a system design that duplicates components to provide alternatives in case one component fails
    [n] repetition of messages to reduce the probability of errors in transmission

    Which agrees with both our definitions. Yours is more correct electronically. Hyperdictionary has a more extensive electronic description but doesnt add much more to the above electronic definition:
    http://www.hyperdictionary.com/dictionary/redundan...

    I'm at odds with that electronic meaning of redundancy. After all the language came before the electronics.
  • mdrohn - Thursday, July 8, 2004 - link

    Heh, I guess that means you are writing from Britain, Pumpkinierre. That special meaning of 'redundancy' in the workplace context of someone losing their job is unknown here in the USA. In fact I'd never heard it in my life until watching 'The Office' on DVD last month ;) We call that layoffs or downsizing here.

    The electronics/systems meaning I posted was also taken straight from a dictionary.
  • Pumpkinierre - Thursday, July 8, 2004 - link

    Your job becomes redundant when you are no longer of use. You then get a redundancy payout based on the years worked etc.. So I think they initially used the term to describe drives that were no longer of use as they had been replaced by newer bigger drives. My dictionary has redundant as meaning superfluous which is more like your definition, #100, and I suppose you could regard a backup drive as such....until your main drive lets go. So, I dont like the usage of redundancy for duplexed or mirrored drives.
  • mdrohn - Wednesday, July 7, 2004 - link

    "Redundancy means of no further use."

    Actually, 'redundant' more precisely means 'exceeding what is required' or 'exactly duplicating the function or meaning of another', which is an important distinction.

    'Redundancy' in an electronics or systems context means 'incorporating extra components that perform the same function in order to cope with failures and errors'. Thus RAID 0 is not, strictly speaking, a 'redundant' array of disks despite its RAID name, since every drive in RAID 0 records different data. RAID 1 is classic redundancy--all the drives in a RAID 1 array are reading and writing exactly the same data.
  • Pumpkinierre - Wednesday, July 7, 2004 - link

    I dont know about that latency increase with RAID. Seek times dont seem to be much affected in reviews I've seen. If the controller reads the drives simultaneously then there shouldnt be much effect on latency.

    (will it make it to the 6th page?!)
  • masher - Wednesday, July 7, 2004 - link

    > I'm fairly certain that the performance
    > advantages of having 4 or 5 striped drives are
    > likely to be a lot better than just 2...

    No, not for ATA drives. You're still limited to the max bandwidth of 133MB/s (150 for some SATA implementations), so beyond 3 drives you don't get the full transfer rate of each drive. Plus, the latency gets worse the more drives you add...with 5 or more drives, your mean latency is essentially your max latency of any single drive.

    So a larger array is a repeat of the 2-drive situation. Its much faster in the rare case of a disk-bound app transferring huge files...and no faster (or possibly slightly slower) the rest of the time.

    You are right on one thing though. Cheaper (translation: slower) disks would tend to look a bit better here. Not a huge difference, but the slower the disk, the more likely the app is to be disk-bound.
  • kapowaz - Wednesday, July 7, 2004 - link

    Perhaps the review ought to have pointed out what RAID stands for: Redundant Array of *Inexpensive* Disks. The idea is to improve the performance or reliability of a system by using many smaller/cheaper disks compared to using a single expensive disk. Often this isn't the case (I doubt anyone would say the 15krpm disks in modern servers are 'inexpensive'), but the origin behind the technology remains applicable today.

    Maybe a better test would be to take some cheap disks and see how well they perform. Also, am I not right in thinking that SATA RAID allows for more than just two devices? I'm fairly certain that the performance advantages of having 4 or 5 striped drives are likely to be a lot better than just 2...
  • masher - Wednesday, July 7, 2004 - link

    Umm, MadAd...The "array" of drives can't fail unless a drive in it fails...and it will always fail if one of its drives does. Its just a logical grouping, not a separate entity.

    Furthermore, MTBF is the wrong statistic to use here. MTTF is the relevant one.

    Obviously the raid controller itself could fail, but this is outside the scope of the argument. And such a failure is highly unlikely to impact data in any case.
  • MadAd - Tuesday, July 6, 2004 - link

    "But when we are talking about an ARRAY of drives, the operating life of each individual drive in the array is not what is at issue. What is relevant is ARRAY failure, not DRIVE failure."

    Are you also trying to say that if the array fails without a drive failing then thats still down to the drives MTBF? wouldnt that be an array MTBF?

    If a drive fails in Raid0 then of course we expect the array will fail. If a drive does not fail but the array fails (and you can reuse the drive) then thats nothing to do with the drives MTBF is it? The drives not failed, its still got service life, the array failed. You'll need a different way to measure the chance of an array failure since (unless its connected with a drive failure) its nothing to do with the expected longevity of the components that we measure by drive manufacturers MTBF figures of service life.

Log in

Don't have an account? Sign up now