[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]

Re: [Linux-cluster] GFS lock cache or bug?



Hi Wendy:

Thank you very much for the kind answer.

Unfortunately, I am using Red Hat Enterprise Linux WS
release 4 (Nahant Update 5) 2.6.9-42.ELsmp.

When I ran gfs_tool gettune /mnt/ABC, I got:

ilimit1 = 100
ilimit1_tries = 3
ilimit1_min = 1
ilimit2 = 500
ilimit2_tries = 10
ilimit2_min = 3
demote_secs = 300
incore_log_blocks = 1024
jindex_refresh_secs = 60
depend_secs = 60
scand_secs = 5
recoverd_secs = 60
logd_secs = 1
quotad_secs = 5
inoded_secs = 15
quota_simul_sync = 64
quota_warn_period = 10
atime_quantum = 3600
quota_quantum = 60
quota_scale = 1.0000   (1, 1)
quota_enforce = 1
quota_account = 1
new_files_jdata = 0
new_files_directio = 0
max_atomic_write = 4194304
max_readahead = 262144
lockdump_size = 131072
stall_secs = 600
complain_secs = 10
reclaim_limit = 5000
entries_per_readdir = 32
prefetch_secs = 10
statfs_slots = 64
max_mhc = 10000
greedy_default = 100
greedy_quantum = 25
greedy_max = 250
rgrp_try_threshold = 100

There is no glock_purge option. I will try to tune
demote_secs, but I don't think it will fix 'ls -la'
issue.


By the way, could you please kindly direct me to a
place where I can find detailed explanations of these
tunable options?


Best,

Jas



--- Wendy Cheng <s wendy cheng gmail com> wrote:

> Ja S wrote:
> > Hi, All:
> >   
> 
> I have an old write-up about GFS lock cache issues.
> Shareroot people had 
> pulled it into their web site:
>
http://open-sharedroot.org/Members/marc/blog/blog-on-gfs/glock-trimming-patch/?searchterm=gfs
> 
> It should explain some of your confusions. The
> tunables described in 
> that write-up are formally included into RHEL 5.1
> and RHEL 4.6 right now 
> (so no need to ask for private patches).
> 
> There is a long story about GFS(1)'s "ls -la"
> problem that one time I 
> did plan to do something about it. Unfortunately I'm
> having a new job 
> now so the better bet is probably going for GFS2.
> 
> Will pass some thoughts about GFS1's "ls -la" when I
> have some spare 
> time next week.
> 
> -- Wendy
> 
> > I used to 'ls -la' a subdirecotry, which contains
> more
> > than 30,000 small files, on a SAN storage long
> time
> > ago just once from Node 5, which sits in the
> cluster
> > but does nothing. In other words, Node 5 is an
> idel
> > node.
> >
> > Now when I looked at /proc/cluster/dlm_locks on
> the
> > node, I realised that there are many PR locks and
> the
> > number of PR clocks is pretty much the same as the
> > number of files in the subdirectory I used to
> list. 
> >
> > Then I randomly picked up some lock resources and
> > converted the second part (hex number) of the name
> of
> > the lock resources to decimal numbers, which are
> > simply the inode numbers. Then I searched the
> > subdirectory and confirmed that these inode
> numbers
> > match the files in the subdirectory.
> >
> >
> > Now, my questions are:
> >
> > 1) how can I find out which unix command requires
> what
> > kind of locks? Does the ls command really need PR
> > lock? 
> >
> > 2) how long GFS caches the locks?
> >
> > 3) whether we can configure the caching period?
> >
> > 4) if GFS should not cache the lock for so many
> days,
> > then does it mean this is a bug?
> >
> > 5) Is that a way to find out which process
> requires a
> > particular lock? Below is a typical record in
> > dlm_locks on Node 5. Is any piece of information
> > useful for identifing the process? 
> >
> > Resource d95d2ccc (parent 00000000). Name (len=24)
> "  
> >     5          cb5d35"
> > Local Copy, Master is node 1
> > Granted Queue
> > 137203da PR Master:     73980279
> > Conversion Queue
> > Waiting Queue
> >
> >
> > 6) If I am sure that no processes or applications
> are
> > accessing the subdirectory, then how I can force
> GFS
> > release these PR locks so that DLM can release the
> > corresponding lock resources as well.
> >
> >
> > Thank you very much for reading the questions and
> look
> > forward to hearing from you.
> >
> > Jas
> >
> >
> >      
>
____________________________________________________________________________________
> > Be a better friend, newshound, and 
> > know-it-all with Yahoo! Mobile.  Try it now. 
>
http://mobile.yahoo.com/;_ylt=Ahu06i62sR8HDtDypao8Wcj9tAcJ
> >
> > --
> > Linux-cluster mailing list
> > Linux-cluster redhat com
> >
>
https://www.redhat.com/mailman/listinfo/linux-cluster
> >   
> 
> --
> Linux-cluster mailing list
> Linux-cluster redhat com
>
https://www.redhat.com/mailman/listinfo/linux-cluster
> 



      ____________________________________________________________________________________
Be a better friend, newshound, and 
know-it-all with Yahoo! Mobile.  Try it now.  http://mobile.yahoo.com/;_ylt=Ahu06i62sR8HDtDypao8Wcj9tAcJ


[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]