I have built and maintained linux production servers under both hardware RAID 5 on HP MSA's,
and hardware RAID 1+0 on both HP Proliant's and Dell 2950's/1950's (with a variety of software disk management [depending on the server farm standards]).
I have also configured LVM and md under SATA or iSCSI systems in both RAID 5 and RAID 10.
[I have also built Solaris servers under RAID 5/RAID 10 using SVM, and later replaced with N1 over multi-path I/O on Sun 2450's (zfs); and worked with NetApps/Redhat (xfs).]
I am a proponent of md/LVM over hardware RAID because Linux md does not handle bad block relocation; I love simply rebuilding the array, especially since drives heat up, get torged by power, and are simply not build with QA these days.
Since disk is so incredibly cheap these days (by comparison in your lower level [non enterprise] solution), the popularity of RAID 5 is offset by the full mirror and rebuild protection of RAID 1+0.
Here's the technical descriptions of each level:
RAID 3 uses a number of disks (at least 2) to store data
in a striped RAID 0 fashion. It also uses an additional redundancy
disk to store the XOR sum of the data from the data disks. Should
the redundancy disk fail, the system can continue to operate as if
nothing happened. Should any single data disk fail the system can
compute the data on this disk from the information on the redundancy
disk and all remaining disks. Any double fault will bring the whole
RAID set off-line.
RAID 3 makes sense only with at least 2 data disks (3 disks
including the redundancy disk). Theoretically there is no limit for
the number of disks in the set, but the probability of a fault
increases with the number of disks in the RAID set. Usually the
upper limit is 5 to 7 disks in a single RAID set.
Since RAID 3 stores all redundancy information on a dedicated disk
and since this information has to be updated whenever a write to any
data disk occurs, the overall write speed of a RAID 3 set is limited
by the write speed of the redundancy disk. This, too, is a limit for
the number of disks in a RAID set. The overall read speed of a RAID
3 set with all data disks up and running is that of a RAID 0 set
with that number of data disks. If the set has to reconstruct data
stored on a failed disk from redundant information, the performance
will be severely limited: All disks in the set have to be read and
XOR-ed to compute the missing information.
RAID 5 is just like RAID 3, but the redundancy
information is spread on all disks of the RAID set. This improves
write performance, because load is distributed more evenly between
all available disks. Parity data is rotated across all disks so
total net storage equals all disks minus 1.
RAID 1+0 combines striping with duplication
as striped arrays of mirrored arrays
which gives very high transfers combined with fast seeks as well as
redundancy. The disadvantage is high disk consumption as well as
the above mentioned complexity.
When you are scrounging disk space against money you are sure that two of your disks are not going down. But believe me, THEY DO, and that's it! Also RAID 1+0 out performs in a fast good way. What use is a huge slow system?
Full discussion available here that includes a complete analysis of md systems, and all the types of data loss that generally occur.
www.Obnosis.com |
http://wiki.obnosis.com |
http://hackfest.obnosis.com (503)754-4452
PLUG HACKFESTS -
http://uat.edu Second Saturday of Each Month Noon - 3PM
Date: Mon, 19 Jan 2009 07:30:08 -0700
From:
mark@phillipsmarketing.biz
To:
plug-discuss@lists.plug.phoenix.az.us
Subject: Re: Looking For RAID Hardware/Software Advice
EricThanks for the summary, and thank-you to everyone for their ideas.Based on NewEgg prices, here is some more information:
Option A
Single Disk IDE Drive - 500 GB and backups, keep OS on existing drive = $69.99
Use existing controller and just add another drive. No redundancy
Option B
RAID10 with 500 GB backup capacity and redundancy, keep OS on existing drive = $179.97
2 500 GB SATA2 Drives, new SATA2 controller
Option C
RAID10 with 750 GB backup capacity and redundancy, keep OS on existing drive = $239.97
Two 750 GB SATA2 Drives, new SATA2 controller
Option D
RAID5 with 1,000 GB backup capacity and redundancy, keep OS on existing drive = $239.97
Three 500 GB SATA2 Drives, new SATA2 controller
I am leaning towards Option C based on less power consumption with fewer drives. However, I have to rethink my budget...After some more reading, I am a little confused about the debate between RAID5 and RIAD10. I am interested in the group's opinions on which is better - RAID 5 or RAID 10 and why? What experiences have you had regarding installation, maintenance, and fixing problems? I am running Debian testing.
Thanks!MarkOn Sat, Jan 17, 2009 at 6:35 AM, Eric Shubert <
ejs@shubes.net> wrote:
Mark Phillips wrote:
> I am running out of room for my backups. I use backuppc and I have
> almost filled a 150GB drive with backups from 7 computers, and I need to
> add another 2 computers to the set. I have an old Dell Poweredge 1300
> server (Pentium III 550 Mhz, 500 MB RAM, PCI 33.3Mhz) that I could turn
> into a backup server. I am looking for suggestions/thoughts on how to
> set this up. I need to keep the cost down as much as possible; under $150.
>
> My initial thoughts:
>
> * Keep current 72 GB drive for OS (debian testing, about 68% full)
> * Add two 500 GB SATA drives and a PCI SATA controller ~$130
> * Software RAID and LVM for the two drives
> * Move current 150 GB of backups to the RAID
> * Backuppc now runs on this machine and slowly fills up the RAID
>
> My questions:
>
> 1. Should I keep the 72 GB drive for OS, or put it on the RAID?
>
> 2. I can add another CPU (P III 550 MHz) processor to the box - is it
> worth the effort to find one? I found one source for $5/CPU, I just need
> to find the heat sink and mounting hardware. Will this improve performance?
>
> 3. The box has a built-in SCSI 68-pin Ultra2/wide bus/controller, but
> SCSI drives are more expensive, at least from a cursory google search.
> Is this correct? I don't think I can use SCSI drives within my budget
> constraint.
>
> 4. Would upgrading the memory to 1GB improve performance - top shows:
> Mem: 646676k total, 639300k used 7376k free, 64548k buffers
> This would add another ~$60 to my cost.
>
> 5. Should I look at hardware RAID cards - they seem very cheap, so
> perhaps software is better?
>
> 4. Does this plan make sense, or is there a better way to proceed for
> about the same cost?
>
> Thanks!
>
> Mark
>
Good replies, all. To sum things up, I think a SATAII PCI card (2 or 4
port) and 2 drives is all the HW you need to add to the backup box you
currently have. Set up the drives with SW RAID-1 (mirrored) and you're
good to go. Migrate the data to the raid device, and keep the OS on the
existing drive.
With KeepItSimpleStupid in mind, I recommend using RAID-1 as opposed to
RAID-5. With the price of drives these days, the additional space you
get with RAID-5 isn't worth the headache you'll get when there's a
problem. With RAID-1, each drive can be mounted (and used) individually
if necessary. Not so with RAID-5.
--
-Eric 'shubes'
---------------------------------------------------
PLUG-discuss mailing list -
PLUG-discuss@lists.plug.phoenix.az.us
To subscribe, unsubscribe, or to change your mail settings:
http://lists.PLUG.phoenix.az.us/mailman/listinfo/plug-discuss
_________________________________________________________________
Windows Live™: Keep your life in sync.
http://windowslive.com/explore?ocid=TXT_TAGLM_WL_t1_allup_explore_012009---------------------------------------------------
PLUG-discuss mailing list -
PLUG-discuss@lists.plug.phoenix.az.us
To subscribe, unsubscribe, or to change your mail settings:
http://lists.PLUG.phoenix.az.us/mailman/listinfo/plug-discuss