LinuxQuestions.org

LinuxQuestions.org (/questions/)
-   Linux - Hardware (https://www.linuxquestions.org/questions/linux-hardware-18/)
-   -   raid6 questions/implementation (https://www.linuxquestions.org/questions/linux-hardware-18/raid6-questions-implementation-685648/)

jadeddog 11-23-2008 09:42 PM

raid6 questions/implementation
 
ok, so i've searched the existing raid6 threads, and cannot find all the answers to me questions, so i apologize in advance if these are simple questions but i've never set up a raid system in linux before... be kind please :)

my situation is that i need a file server that is quite large (for personal usage, not business) and i want redundancy so was thinking about raid6 as my solution... the overall useable size should be in excess of 5 TB at the minimum and preferably 7-8 TB in size... so here are my questions

1. software vs. hardware? i'm thinking about a 12 drive system... do i *need* a hardware controller to do this? assuming i can get a mobo that supports 12 drives, i was hoping to avoid the roughly $700 cost of a pci-x controller that supports 12 drives and use a software raid... is this possible? if it is, are there any real drawbacks (keeping in mind that this box will only be a file server, and doesn't need to do any other tasks)

2. is there any distro that is known for being better for raid6? since raid6 still isn't "official" (as far as i know), i'm wondering if any distro has a good reputation for raid6

3. do i really need a hot drive? since this is just for personal use, i don't much care about rebuild time, so do i really need a hot drive? is there any other advantage to having one other than reduced down time?

4. can i increase the size of a raid6 after the initial build? say i get a mobo/controller that allows 12 drives, but only want to buy 8 right now... does raid6 allow for adding the other 4 drives after the initial build? i was thinking this might be possible using LVM in some manner, but im not an expert with raid/LVM by any means, so im not sure

5. follow up question to #4. if it is possible to add drives after the initial build, do the drives have to be the exact same size and manufacturer?

6. for failed drives, do the replacement drives have to be the exact same?

thats about it, i realize that im asking a lot of you all, but any guidance help would be greatly appreciated

lazlow 11-24-2008 01:23 AM

Take a look at man mdadm (software raid).

Here is a wiki to discuss raid: http://en.wikipedia.org/wiki/Redunda...are_RAID.22.29

You will take a huge performance hit with SW raid compared to HW raid.

If you are going to Raid6 for performance (speed) reasons, forget about it. Ethernet 100 transfers at about 9.5MB/s(if you are lucky) and GigE at about 95MB/s. Most modern drives are in the 70MB/s range with quite a few in the 100MB/s range(sata/pata drives). So if your pipe cannot transfer the data as fast as the drives, it is just a waste. For a Video server application a 70MB/s drive is plenty fast. The reality is that no matter what Raid solution you use, you are going to need backups. Do not think for a minute that any raid solution will protect you. I lost a server in the basement to a baseball coming through a window and it rained before I got home. In these type cases raid will do you zero(0) good. I had a client have an aquarium break in the living room with the similar results. Just look at your cost to value ratio. For a home system, I would keep things backed up and just run 8 1Tb drives. Once you go to raid you will need twice as many drives to get the same storage. Plus the cost ($ for HW or cpu performance for SW) of the Raid itself.

jadeddog 11-24-2008 02:54 AM

Quote:

Originally Posted by lazlow (Post 3352470)
Take a look at man mdadm (software raid).

Here is a wiki to discuss raid: http://en.wikipedia.org/wiki/Redunda...are_RAID.22.29

You will take a huge performance hit with SW raid compared to HW raid.

If you are going to Raid6 for performance (speed) reasons, forget about it. Ethernet 100 transfers at about 9.5MB/s(if you are lucky) and GigE at about 95MB/s. Most modern drives are in the 70MB/s range with quite a few in the 100MB/s range(sata/pata drives). So if your pipe cannot transfer the data as fast as the drives, it is just a waste. For a Video server application a 70MB/s drive is plenty fast. The reality is that no matter what Raid solution you use, you are going to need backups. Do not think for a minute that any raid solution will protect you. I lost a server in the basement to a baseball coming through a window and it rained before I got home. In these type cases raid will do you zero(0) good. I had a client have an aquarium break in the living room with the similar results. Just look at your cost to value ratio. For a home system, I would keep things backed up and just run 8 1Tb drives. Once you go to raid you will need twice as many drives to get the same storage. Plus the cost ($ for HW or cpu performance for SW) of the Raid itself.

no i wasn't going to raid for performance... moreso for data backup to be honest... i know, i know, raid isn't great for backup, but i can't afford to double up my drives for a true backup (which is really just raid 1 and doesn't provide good backups anyhow) and i can't afford to spend $2000 on a tape backup system... unless you know of a backup system for around $500-750 that would allow me to backup around 7-8 TB of data??

lazlow 11-24-2008 11:57 AM

That is just the point. Any backup raid solution will require twice (ballpark) as much hard drive space in order to get backup. In your case in order to get 8TB of storage you will have to have 16TB(raid6) of disk space(16 1Tb drives). Take a look at the link I provided. The most efficient (space wise) raids are 3,4, and 5. They are only 62% efficient. 8tb/62=13Tb(close but not quite 2X). Raid 6 is less efficient (roughly 50%), so you will need 2X the drives. Assuming that the vast majority of the data will not change over time(video collection, etc) the easiest way out is just to use backup drives in a singular fashion(either usb or drive slides). Add new stuff to main drives and backups at the same time. When the backup drive is full, move to the next.

jadeddog 11-24-2008 08:16 PM

yeah i guess i could just go with a raid1 with LVM, so i can increase the size as needed .... im just trying to get around having to spend double on the drives... grrr

lazlow 11-24-2008 08:31 PM

Not a big fan of LVMs either. They are fine until they break. When they break, hold on to your shorts. IF you need HUGE files (file size larger than drive size) then you have no choice. What you can do(assuming normal size files)is just put the mount points of the added drives within the current drives. This way each new drive (or drive set) appears as a sub directory of the main drive(or drive set). If anything catastrophic does occur(using this system) you can always mount the added drive sets independently from the master drive. Whereas if a LVM breaks, everything is broke. This is the way I prefer to setup drives. First set(2 drive array) as master and then each following set (again 2 drives) mounted as a sub directory of the master. If the master breaks you still have access (manual mount elsewhere) to all your other drives. If any of the secondary sets break, you just unmount them and repair, without interfering with the rest of the data. I also keep the OS on separate drives from the data.

uberNUT69 11-25-2008 07:10 PM

Take a look at LVM on Linux raid10. (Still 2x the number of drives)


All times are GMT -5. The time now is 03:10 PM.