Linux - SoftwareThis forum is for Software issues.
Having a problem installing a new program? Want to know which application is best for the job? Post your question in this forum.
Notices
Welcome to LinuxQuestions.org, a friendly and active Linux Community.
You are currently viewing LQ as a guest. By joining our community you will have the ability to post topics, receive our newsletter, use the advanced search, subscribe to threads and access many other special features. Registration is quick, simple and absolutely free. Join our community today!
Note that registered members see fewer ads, and ContentLink is completely disabled once you log in.
If you have any problems with the registration process or your account login, please contact us. If you need to reset your password, click here.
Having a problem logging in? Please visit this page to clear all LQ-related cookies.
Get a virtual cloud desktop with the Linux distro that you want in less than five minutes with Shells! With over 10 pre-installed distros to choose from, the worry-free installation life is here! Whether you are a digital nomad or just looking for flexibility, Shells can put your Linux machine on the device that you want to use.
Exclusive for LQ members, get up to 45% off per month. Click here for more info.
Can anyone give me some tips on a program for performance monitoring linux machines? Need the usual: CPU, memory, disk I/O, disk buffer, swap, network.
But I also need history reports, or graphics.
I´ll start running some tests with Ganglia. Hope it helps me out.
MRTG can monitor just about anything via SNMP. I also like Ganglia for this task. There are a couple other solutions (some free and some not) if you Google for Linux performance monitoring or the like.
"Routers are only the beginning. MRTG is being used to graph all sorts of network devices as well as everything else from weather data to vending machines."
I have MRTG running on my server, where it logs things like memory usage, disk space and uptime (in addition to the standard logs, i.e. traffic and such). The Perl scripts to do the logging are fairly easy to find on the Internet, or write yourself.
Hum...
I decided on MRTG since I need to generate those reports asap.
The problem is: MRTG only checks the data every 5 minutes, and if I try to run it every minute, it seems to be no changes at all to the graphs (in this case, I´m talking about CPU checks).
Anything I can do to make it work every minute?
Note that unless you are using rrdtool you can not set Interval to less than 5 minutes.
If you are using rrdtool you can set interval down to 1 minute.
Note though, setting the Interval for an rrdtool/mrtg setup will influence the initial creation of the database. If you change the interval later, all existing databases will remain at the resolution they were initially created with.
How often do you call mrtg? The default is 5 minutes. If you call it less often, you should specify it here. This does two things:
* The generated HTML page contains the right information about the calling interval ...
* A META header in the generated HTML page will instruct caches about the time-to-live of this page .....
In this example, we tell mrtg that we will be calling it every 10 minutes. If you are calling mrtg every 5 minutes, you can leave this line commented out.
Example:
Interval: 10
Note that unless you are using rrdtool you can not set Interval to less than 5 minutes. If you are using rrdtool you can set interval in the format
Interval: MM[:SS]
Down to 1 second. Note though, setting the Interval for an rrdtool/mrtg setup will influence the initial creation of the database. If you change the interval later, all existing databases will remain at the resolution they were initially created with. Also note that you must make sure that your mrtg-rrd Web-frontend can deal with this kind of Interval setting.
I realize this post is a little late, but I have issues with centralized monitoring for several reasons:
- it doesn't scale, at least not to 100s or 1000s of systems
- if there's a network problem you lose the data you most need
- monitoring interval over a minute aren't worth the bother for diagnostic purposes
I had open sourced collectl several years ago after using it internally and with HPC customers systemes for several years before then so it has been road tested and pretty stable. It's always been able to export its data over a socket to a remote application but recently I added support for it to send data directly to a ganglia gmond for Pacific Northwest National Labs to use for monitoring their 2300 node cluster! This gives them the ability to collect data at the rate of one sample every 10 seconds on each machine but only send a subset and at a lower rate to ganglia. However they don't write it to rrd because they'd overwhelm it and only use ganglia as the transport to get the data to their own presentation engine.
The reason I mention this is that as more people try to use collectl with ganglia it will help solidify the interface between the two.
btw - HP recently started a blog on High Performance Clusters to which I've added some further discussion on HCP monitoring in case anyone is interested. see: http://www.communities.hp.com/online...e/default.aspx
LinuxQuestions.org is looking for people interested in writing
Editorials, Articles, Reviews, and more. If you'd like to contribute
content, let us know.