LinuxQuestions.org
Share your knowledge at the LQ Wiki.
Go Back   LinuxQuestions.org > Forums > Linux Forums > Linux - Server
User Name
Password
Linux - Server This forum is for the discussion of Linux Software used in a server related context.

Notices

Reply
 
Search this Thread
Old 09-22-2009, 11:11 PM   #1
horde
LQ Newbie
 
Registered: Jan 2005
Posts: 19

Rep: Reputation: 0
RAID 5 --ADD Not completing at 100%


Hi All,

I had a disk fail on a raid array and rebuilt it thus:

I unmounted the filesystem and then issued:

mdadm /dev/md1 --fail /dev/sdf1
mdadm /dev/md1 --remove /dev/sdf1
mdadm /dev/md1 --add /dev/sdf1

And then monitored it using mdadm --detail and /proc/mdstat

> mdadm --detail /dev/md1
/dev/md1:
Version : 0.90
Creation Time : Sat Nov 22 13:42:02 2008
Raid Level : raid5
Array Size : 3907039744 (3726.04 GiB 4000.81 GB)
Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
Raid Devices : 5
Total Devices : 5
Preferred Minor : 1
Persistence : Superblock is persistent

Intent Bitmap : Internal

Update Time : Wed Sep 23 13:45:02 2009
State : active, degraded, recovering
Active Devices : 4
Working Devices : 5
Failed Devices : 0
Spare Devices : 1

Layout : left-symmetric
Chunk Size : 128K

Rebuild Status : 91% complete

UUID : 0f2561a9:81e8cd4a:10e4ddb4:49a3b4a7
Events : 0.693864

Number Major Minor RaidDevice State
0 8 17 0 active sync /dev/sdb1
1 8 49 1 active sync /dev/sdd1
2 8 65 2 active sync /dev/sde1
5 8 81 3 spare rebuilding /dev/sdf1
4 8 33 4 active sync /dev/sdc1

> cat /proc/mdstat
Personalities : [raid6] [raid5] [raid4]
md1 : active raid5 sdf1[5] sdb1[0] sdc1[4] sde1[2] sdd1[1]
3907039744 blocks level 5, 128k chunk, algorithm 2 [5/4] [UUU_U]
[===================>.] recovery = 99.9% (488036992/488379968) finish=0.0min speed=56148K/sec
bitmap: 10/233 pages [40KB], 2048KB chunk

unused devices: <none>
> cat /proc/mdstat
Personalities : [raid6] [raid5] [raid4]
md1 : active raid5 sdf1[5] sdb1[0] sdc1[4] sde1[2] sdd1[1]
3907039744 blocks level 5, 128k chunk, algorithm 2 [5/4] [UUU_U]
[====================>] recovery =100.0% (488392576/488379968) finish=2802575022411.7min speed=54850K/sec
bitmap: 10/233 pages [40KB], 2048KB chunk

unused devices: <none>
> cat /proc/mdstat
Personalities : [raid6] [raid5] [raid4]
md1 : active raid5 sdf1[5] sdb1[0] sdc1[4] sde1[2] sdd1[1]
3907039744 blocks level 5, 128k chunk, algorithm 2 [5/4] [UUU_U]
[====================>] recovery =100.1% (489231232/488379968) finish=2812115125554.9min speed=54664K/sec
bitmap: 10/233 pages [40KB], 2048KB chunk

> mdadm --detail /dev/md1
/dev/md1:
Version : 0.90
Creation Time : Sat Nov 22 13:42:02 2008
Raid Level : raid5
Array Size : 3907039744 (3726.04 GiB 4000.81 GB)
Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
Raid Devices : 5
Total Devices : 5
Preferred Minor : 1
Persistence : Superblock is persistent

Intent Bitmap : Internal

Update Time : Wed Sep 23 14:00:02 2009
State : active, degraded
Active Devices : 4
Working Devices : 5
Failed Devices : 0
Spare Devices : 1

Layout : left-symmetric
Chunk Size : 128K

UUID : 0f2561a9:81e8cd4a:10e4ddb4:49a3b4a7
Events : 0.693868

Number Major Minor RaidDevice State
0 8 17 0 active sync /dev/sdb1
1 8 49 1 active sync /dev/sdd1
2 8 65 2 active sync /dev/sde1
5 8 81 3 spare rebuilding /dev/sdf1
4 8 33 4 active sync /dev/sdc1


It is no longer recovering. How do I get the spare activated?
 
Old 09-24-2009, 05:43 AM   #2
horde
LQ Newbie
 
Registered: Jan 2005
Posts: 19

Original Poster
Rep: Reputation: 0
It has finally recovered. All disks are now active - it took about 3hrs after reachine 100% (whcih it reached after 2 hrs) to eventually marlk all as active.

Originally it was active, degraded, recovering

after about 2 hrs it moved to active, degraded

and then after another 3 hours it moved to active

The raid-5 array was 5 X 1Tb disks so working with 4Gb of data

Hope this information helps someone in the future
 
Old 09-24-2009, 08:16 PM   #3
chrism01
Guru
 
Registered: Aug 2004
Location: Sydney
Distribution: Centos 6.5, Centos 5.10
Posts: 16,226

Rep: Reputation: 2022Reputation: 2022Reputation: 2022Reputation: 2022Reputation: 2022Reputation: 2022Reputation: 2022Reputation: 2022Reputation: 2022Reputation: 2022Reputation: 2022
That's the prob with large disks now avail on std PCs; the disk & backplane HW isn't up to the job really.
Fast HW costs more....
 
Old 10-01-2009, 03:06 AM   #4
horde
LQ Newbie
 
Registered: Jan 2005
Posts: 19

Original Poster
Rep: Reputation: 0
and to be honest is the trade off we make - speed vs cash - in this cash I was willing to wait the time - it just would have been nice if /proc/mdstat showed me the correct expected time and not finish=2812115125554.9min (which by my estimates is about 50K years - a bit too long for me to wait)
 
  


Reply

Tags
raid


Thread Tools Search this Thread
Search this Thread:

Advanced Search

Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

BB code is On
Smilies are On
[IMG] code is Off
HTML code is Off


Similar Threads
Thread Thread Starter Forum Replies Last Post
Help with python application to add number until 100 or more, new user. Michael_aust Programming 6 09-24-2006 01:35 PM
Rocket Raid 100 jonthelam Linux - Hardware 10 09-01-2004 05:11 PM
Setting up Highpoint Rocket Raid 100 jonthelam Slackware 6 08-31-2004 02:07 PM
Mandrake9.1 and Fasttrak 100 Raid controller ? frasier642 Mandriva 1 09-01-2003 08:27 AM
ata-100 raid and redhat 7.1 dcwheel Linux - Software 0 09-16-2001 11:45 PM


All times are GMT -5. The time now is 07:32 AM.

Main Menu
My LQ
Write for LQ
LinuxQuestions.org is looking for people interested in writing Editorials, Articles, Reviews, and more. If you'd like to contribute content, let us know.
Main Menu
Syndicate
RSS1  Latest Threads
RSS1  LQ News
Twitter: @linuxquestions
identi.ca: @linuxquestions
Facebook: linuxquestions Google+: linuxquestions
Open Source Consulting | Domain Registration