[Linux-cluster] OOM failures with GFS, NFS and Samba on a cluster with RHEL3-AS

Jonathan Woytek woytek+ at cmu.edu
Sun Jan 23 18:45:28 UTC 2005


Additional information:

I enabled full output on lock_gulmd, since my dead top sessions would 
often show that process near the top of the list around the time of 
crashes.  The machine was rebooted around 10:50AM, and was down again at 
12:44.  In the span of less than a minute, the machine plowed through 
over 3GB of memory and crashed.  The extra debugging information from 
lock_gulmd said nothing, except that there was a successful heartbeat. 
The OOM messages began at 12:44:01, and the machine was dead somewhere 
around 12:44:40.  Nobody should be using the machine during this time. 
A cron job that was scheduled to fire off at 12:44 (it runs every two 
minutes to check memory usage, specifically to try to track this 
problem) did not run (or at least was not logged if it did).  I took 
that job out of cron just to make sure that it isn't part of the 
problem.  The low-memory-check that ran at 12:42 reported nothing, and 
my threshold for that is set at 512MB.

The span between crashes this weekend has been between three and eight 
hours.  Yesterday, the machine rebooted (looking at lastlog, not last 
message before restart in /var/log/messages, but I'll be looking at that 
in a bit) at 15:20 (after being up since 23:50 on Friday), 18:27, 21:43, 
  onto sunday at 01:14, 04:33, and finally 12:48.  Something seems quite 
wrong with this.

jonathan


Jonathan Woytek wrote:

> I have been experiencing OOM failures (followed by reboots) on a cluster 
> running Dell PowerEdge 1860's (dual-proc, 4GB RAM) with RHEL3-AS with 
> all current updates.
> 
> The system is configured as a two-member cluster, running GFS 6.0.2-25 
> (RH SRPM) and cluster services 1.2.16-1 (also RH SRPM).  My original 
> testing went fine with the cluster, including service fail-over and all 
> that stuff (only one lock_gulmd, so if the master goes down, the world 
> explodes--but I expected that).
> 
> Use seemed to be okay, but there weren't a whole lot of users. Recently, 
> a project wanted to serve some data from their space in GFS via their 
> own machine.  We mounted their space via NFS from the cluster, and they 
> serve their data via samba from their machine.  Shortly thereafter, two 
> things happened:  more people started to access the data, and the 
> cluster machines started to crash.  The symptoms are that free memory 
> drops extremely quickly (sometimes more than 3GB disappears in less than 
> two minutes).  Load average usually goes up quickly (when I can see 
> it).  NFS processes are normally at the top of top, along with kswapd.  
> At some point, around this time, the kernel starts to spit out OOM 
> messages and it starts to kill bunches of processes.  The machine 
> eventually reboots itself and comes back up cleanly.
> 
> Space of outages seems to be dependent on how many people are using the 
> system, but I've also seen the machine go down when the backup system 
> runs a few backups on the machine.  One of the things I've noticed, 
> though, is that the backup system doesn't actually cause the machine to 
> crash if the system has been recently rebooted, and memory usage returns 
> to normal after the backup is finished.  Memory usage usually does NOT 
> return to completely normal after the gigabytes of memory become used 
> (when that happens, the machine will sit there and keep running for a 
> while with only 20MB or less free, until something presumably tries to 
> use that memory and the machine flips out).  That is the only time I've 
> seen the backup system cause the system to crash--after it has endured 
> significant usage during the day and there are 20MB or less free.
> 
> I'll usually get a call from the culprits telling me that they were 
> copying either a) lots of files or b) large files to the cluster.
> 
> Any ideas here?  Anything I can look at to tune?
> 
> jonathan
> 
> -- 
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> http://www.redhat.com/mailman/listinfo/linux-cluster




More information about the Linux-cluster mailing list