Problem with mounting nfs shares after sudden poweroutage
Linux - NetworkingThis forum is for any issue related to networks or networking.
Routing, network cards, OSI, etc. Anything is fair game.
Notices
Welcome to LinuxQuestions.org, a friendly and active Linux Community.
You are currently viewing LQ as a guest. By joining our community you will have the ability to post topics, receive our newsletter, use the advanced search, subscribe to threads and access many other special features. Registration is quick, simple and absolutely free. Join our community today!
Note that registered members see fewer ads, and ContentLink is completely disabled once you log in.
If you have any problems with the registration process or your account login, please contact us. If you need to reset your password, click here.
Having a problem logging in? Please visit this page to clear all LQ-related cookies.
Get a virtual cloud desktop with the Linux distro that you want in less than five minutes with Shells! With over 10 pre-installed distros to choose from, the worry-free installation life is here! Whether you are a digital nomad or just looking for flexibility, Shells can put your Linux machine on the device that you want to use.
Exclusive for LQ members, get up to 45% off per month. Click here for more info.
Problem with mounting nfs shares after sudden poweroutage
Hi!
I am already experiencing this for the second time and i feel that this is a strange issue: I have AMD Ryzen desktop with Debian Bullseye (before Buster) and a JBOD disk tower with four fully occupied slots. On my desktop i mount the 4 disks as NFS shares, which used to work nicely before the most recent sudden power outage. After that the drives had to be checked and the inodes repaired. One disk in use was lost and restored through backup by copying the backup on the repaired disk. I also changed the file permissions and ownership after the copy procedure. After that the mount procedure during system startup happening in /etc/fstab did not mount anymore my nfs shares correctly. Now my mount table looks like this:
My /etc/fstab has this four entries related to the nfs shares
One can see that /mnt/WD01 gets mounted under /mnt/WD03 and there is no mount entry for /mnt/WD01. What is wrong here? Reboot does not change anything. This strange constellation is carried over from reboot to reboot and i don't know where i can find the file that is jumbled up.
The mount procedure is waiting for 1,5 minutes for mounting WD01 and WD03 suddenly. The drives itself are okay on my nfs server 10.10.10.2 and exportfs also lists ok:
How did you fix the issue with the old Debian installation?
It seems that maybe the nfs server is exporting WD03 in such a way that the client sees it wrong and mounts it at WD01
Also, I wonder about the issue of mounting the same device on different systems at the same time, especially in different subnets. The risk of multiple users altering the same file at the same time does exist and may cause corruption.
Lastly, exporting WD01 to 10.10.10.1 and 10.10.10.0/24 is redundant and possibly conflicting. The first only allows one machine to access it, and the other allows the entire subnet, including that one machine to access it. There may be a conflict introduced that does not appear obvious.
You can possibly avoid the extended delay by adding options in fstab for each of those entries such as
nofail & _netdev
How did you fix the issue with the old Debian installation?
I did not. I could access WD01 and WD02 from my desktop though the mount was jumbled. WD03 and WD04 i use just occasionally.
Quote:
It seems that maybe the nfs server is exporting WD03 in such a way that the client sees it wrong and mounts it at WD01
It was working before nicely with all the different export cases. In my old Debian Buster i was also suspecting a problem with the mount export. But the jumbling of the mounts happened only after the sudden power outage and disk recovery of WD01 afterwards. As WD01 was operated on, the complete file system was destroyed and i had to restore it from my backup of WD02 which god thanks was not destroyed but could be repaired from the ext4-journal
Quote:
Also, I wonder about the issue of mounting the same device on different systems at the same time, especially in different subnets. The risk of multiple users altering the same file at the same time does exist and may cause corruption.
I think that risk is should be managed by a ressource lock which is a quite common method in operating systems to guarantuee exclusive access to a ressource. But you are right, in the sudden power outage the consistency of the file system got destroyed badly beyond recovery at least on WD01. So the IO data transfer to my JBOD array is in a way risky.
Quote:
Lastly, exporting WD01 to 10.10.10.1 and 10.10.10.0/24 is redundant and possibly conflicting. The first only allows one machine to access it, and the other allows the entire subnet, including that one machine to access it. There may be a conflict introduced that does not appear obvious.
Maybe. Though in my case the 10.10.10.1 export is just a widening of the 10.10.10.0/24 export. Let us put it straight: I got no error message...
Quote:
You can possibly avoid the extended delay by adding options in fstab for each of those entries such as
nofail & _netdev
I did with nofail. The _netdev option was just necessary on Ubuntu and under certain conditions. Usually, "ro,auto,nofail 0 0" works reliably.
I have filed a kernel bug report and a Debian bug report (Bug#992866) since this thing is bothering me already for too long. I think something is wrong with the kernel mount data structures. They are storing information from reboot to reboot, but they don't check properly for consistence.
Last edited by Thomas Korimort; 08-29-2021 at 11:19 PM.
LinuxQuestions.org is looking for people interested in writing
Editorials, Articles, Reviews, and more. If you'd like to contribute
content, let us know.