Linux - GeneralThis Linux forum is for general Linux questions and discussion.
If it is Linux Related and doesn't seem to fit in any other forum then this is the place.
Welcome to LinuxQuestions.org, a friendly and active Linux Community.
You are currently viewing LQ as a guest. By joining our community you will have the ability to post topics, receive our newsletter, use the advanced search, subscribe to threads and access many other special features. Registration is quick, simple and absolutely free. Join our community today!
Note that registered members see fewer ads, and ContentLink is completely disabled once you log in.
If you have any problems with the registration process or your account login, please contact us. If you need to reset your password, click here.
Having a problem logging in? Please visit this page to clear all LQ-related cookies.
Introduction to Linux - A Hands on Guide
This guide was created as an overview of the Linux Operating System, geared toward new users as an exploration tour and getting started guide, with exercises at the end of each chapter.
For more advanced trainees it can be a desktop reference, and a collection of the base knowledge needed to proceed with system and network administration. This book contains many real life examples derived from the author's experience as a Linux system and network administrator, trainer and consultant. They hope these examples will help you to get a better understanding of the Linux system and that you feel encouraged to try out things on your own.
Click Here to receive this Complete Guide absolutely free.
A professional server that has gone without a reboot for 93 days: You need to plan a reboot real soon now. Or are you 100% sure there haven't been any kernel related (security) patches in the last 93 days?
Besides the kernel; The longer a machine goes without a reboot the greater the chance that it will not boot or come up as expected. Scripts/config files that might have changed (long ago, and possibly forgotten about), HD's that "spontaneously" die after spinning for a long time. And there are probably other reasons. Not something I would like to run into after, for whatever reason, the machine does need to get rebooted.
A high uptime might look cool, but isn't a holy grail whatsoever and in my opinion not of any interest when it comes to professional environments.
If this is a home environment: Go nuts, it is your box (but the above risks still apply).
This must be a reference to regular linux/dell/hp hardware where you reboot to patch the kernel.
I remember Solaris boxes with hot-swappable,.. well, everything. In place kernel patches... etc. So, its completely possible to have a secure/safe machine on for longer than 90 days.
"The longer a machine goes without a reboot etc,.." if its in production you should have 3t+1 servers as backup, so this shouldnt be a problem. I hope everyone out there is replicating servers/load balancing for failure.
I high uptime isn't cool. Nor is it a status symbol. It is just interesting to see what people have, why they have it, and what they plan on doing.
And in Solaris environments, it is of great interest to professionals how long the equipment has been running. I've had to turn in reports with server patchlevel/uptime listed, so it must be important to someone.
@szboardstretcher: Its more then having a HA/Fault tolerant environment.
The following is from own experience:
I've worked with large TandemNonStop environments. Great stuff to work with: Everything hot-swappable, extra cpu/memory boards could be inserted on running systems, all hardware was double/quadruple and all was patchable on a running system. A system build to "never go down" (mind the quotes).
Uptimes went through the roof 'cause there was no need for a reboot.
At a certain point it was decided (for certification if I remember correctly) to do a power-failure test. Simple in principle: Pull the plug and see if the generators kick in before reserve power is drained. The generators did not kick in and a bunch of tandems went down. We are running on one leg at this moment.
The fact that "a few disks" might not come up correctly was foreseen. 25 brand new disks were on site, 10% of the total amount. The fact that almost 90 disks died was not foreseen. The extra disks needed could not be gotten from Europe, no-one had that many. The company had to rent a private jet to get them flown in from the US (time is of the essence, still running on one leg).
The services provided never went down, so the customers didn't notice a thing. This whole incident did, however, cost millions.
From that moment on it was decided to do a controlled reboot every six months.
I had a couple of machines hovering around the 2-3 year mark not long ago (old machines, no mission-critical operations, OS long out of date, just acting as local data servers isolated from the internet until we decide to decommision them). About 3 months ago we had an extended power outage (~12 hours) that shut down the office, now everything has a 101 day uptime.