Linux - HardwareThis forum is for Hardware issues.
Having trouble installing a piece of hardware? Want to know if that peripheral is compatible with Linux?
Notices
Welcome to LinuxQuestions.org, a friendly and active Linux Community.
You are currently viewing LQ as a guest. By joining our community you will have the ability to post topics, receive our newsletter, use the advanced search, subscribe to threads and access many other special features. Registration is quick, simple and absolutely free. Join our community today!
Note that registered members see fewer ads, and ContentLink is completely disabled once you log in.
If you have any problems with the registration process or your account login, please contact us. If you need to reset your password, click here.
Having a problem logging in? Please visit this page to clear all LQ-related cookies.
Get a virtual cloud desktop with the Linux distro that you want in less than five minutes with Shells! With over 10 pre-installed distros to choose from, the worry-free installation life is here! Whether you are a digital nomad or just looking for flexibility, Shells can put your Linux machine on the device that you want to use.
Exclusive for LQ members, get up to 45% off per month. Click here for more info.
Hi all. I have been having some hard drive problems and I don't know what to do. I am really hoping there is some way to recover the data off these drives.
I had 2 500GB Western Digital Cavier SE16 drives set up as RAID 1 under Ubuntu Linux 8.04. I started hearing a clicking noise when I booted, and the BIOS gave me the message "Primary Slave Hard Disk Error, Press F1 to continue". At that point, the system would boot very slowly, but it would still mount the array and I could still read and write to it; however, being nervous about the disks, I shut it down until I was ready for backups.
I set up an old PC to rsync the files over to before the drives died altogether. When I tried to boot my main system again, it took an extremely long time, I got the "hard disk error" message again, and now Ubuntu doesn't mount the array. Both disks seem to be dead.
I tried throwing each of the drives (one at a time) in another Ubuntu machine. dmesg reports the following pertinent output:
Code:
[ 100.414735] ata4: PATA max UDMA/100 cmd 0xef00 ctl 0xee00 bmdma 0xed08 irq 17
[ 105.769318] ata4: port is slow to respond, please be patient (Status 0xc1)
[ 110.582476] ata4: SRST failed (errno=-16)
[ 115.775420] ata4: port is slow to respond, please be patient (Status 0xc1)
[ 118.246122] ata4.00: failed to IDENTIFY (INIT_DEV_PARAMS failed, err_mask=0x80)
[ 118.246175] ata4: failed to recover some devices, retrying in 5 secs
[ 128.439931] ata4: port is slow to respond, please be patient (Status 0xc0)
[ 133.253088] ata4: SRST failed (errno=-16)
[ 138.446022] ata4: port is slow to respond, please be patient (Status 0xc0)
[ 141.252535] ata4.00: failed to IDENTIFY (INIT_DEV_PARAMS failed, err_mask=0x80)
[ 141.252589] ata4: failed to recover some devices, retrying in 5 secs
[ 151.446356] ata4: port is slow to respond, please be patient (Status 0xc0)
[ 156.259502] ata4: SRST failed (errno=-16)
[ 161.452438] ata4: port is slow to respond, please be patient (Status 0xc0)
[ 164.258950] ata4.00: failed to IDENTIFY (INIT_DEV_PARAMS failed, err_mask=0x80)
[ 164.259003] ata4: failed to recover some devices, retrying in 5 secs
[ 174.452759] ata4: port is slow to respond, please be patient (Status 0xc0)
[ 179.265917] ata4: SRST failed (errno=-16)
[ 184.458850] ata4: port is slow to respond, please be patient (Status 0xc0)
Are my drives dead, or are there any tools that might help me out? I don't care about the physical drives, I just want to recover the data. I would prefer an open source or free solution, but I would be willing to shell out a few dollars for proprietary software if that's what it takes.
If it is Raid 1 as reported why not remove Disk A to see if Disk B is good. Repeat the same step by running Disk A to find out which disk is bad.
You can then just run the good disk alone and forget about the Raid. Isn't this what Raid 1 supposed to protect you against hardware fault by running a mirror image permanently?
Sorry if I was unclear, I have tried mounting the drives seperately, and I still get the same output in dmesg. They seem to have both died within a very short window of time; I've certainly learned my lesson about using two disks purchased at the same time from the same manufacturer for RAID 1
To clarify: the dmesg output above is from when I stuck one of the drives in another machine. The other drive produced identical output when I swapped them.
The clicking noise and the errors indicate the drives are kaput. You can go to the drive MFG website and download any diagnostic tool that can run under a bootable disk to verify.
If the data is important, get a copy of Spinrite and give it a whirl.
There is a very good chance that it will restore your drive(s) at least long enough for you to recover the data. Spinrite isn't cheap, though. But it does work miracles presuming you are not having a bearing failure.
I do consider it to be more than a bit strange that two drives failed at the same time...I would be wondering about some environmental condition. Coincidences do happen, though.
Thank you for the suggestion. Spinrite looks promising, and they have a money-back guarantee. $90 is not so bad if it works as well as they claim.
I agree, both drives failing at the same time is very strange. I guess the moral of the story is, don't use identical drives from the same lot for RAID, and always do regular backups.
In my personal experience (and I have on many occasions loaned my copy of spinrite to other people...my experience includes theirs), if the drive will spin up and the bearings are OK, and the controller electronics have not failed, and the heads work, then spinrite WILL recover the drive.
LinuxQuestions.org is looking for people interested in writing
Editorials, Articles, Reviews, and more. If you'd like to contribute
content, let us know.