[Linux-cluster] GFS and httpd process with stat "D"

Mark Hlawatschek hlawatschek at atix.de
Wed Jan 18 13:53:01 UTC 2006


Hi Bartolomé ,

this can happen, if you are using php session ids. The httpd processes can run 
into a deadlock situation when two httpd processes try to flock the same 
session id file. In that case they stay in the "D" state which means waiting 
for I/O.   
To prevent this, we wrote a patch for php with the following features:
- make transaction based file locking on session id files.
- prevent php from flocking session ids (this is the NFS like behaviour, as 
NFS does not support flocks.)

You can download the patch at http://open-sharedroot.org 

Have Fun,
Mark

 

On Wednesday 18 January 2006 11:42, Bartolomé Rodríguez wrote:
>     Hi list,
>
>     Release: Red Hat Enterprise Linux ES release 4 (Nahant Update 2)
>     GFS-6.1.2-0
>     GFS-kernel-smp-2.6.9-42.2
>     Kernel running: 2.6.9-22.0.1.ELsmp
>     Apache: httpd-2.0.52-19.ent
>
>     I have 4 servers with GFS. These servers have shared disk with GFS
> (The disks are in a san with fiberchannel). This scene is in production.
> 2 times in 2 weeks the httpd process become stat "D". We thinks GFS is
> blocking something (file or inode... i dont know). And when httpd try
> read or writhe there, then become stat "D":
>
> apache   14383  1.5  0.3 24004 13300 ?       D    17:36   3:57
> /usr/sbin/httpd
> apache   14503  1.6  0.3 23996 13296 ?       D    17:49   3:53
> /usr/sbin/httpd
> apache   14516  1.6  0.3 23988 13288 ?       D    17:50   3:58
> /usr/sbin/httpd
> apache   14522  1.6  0.3 24044 13340 ?       D    17:50   3:54
> /usr/sbin/httpd
> apache   14635  1.5  0.3 24092 13392 ?       D    18:00   3:23
> /usr/sbin/httpd
> apache   14661  1.6  0.3 24048 13344 ?       D    18:03   3:33
> /usr/sbin/httpd
> apache   14662  1.7  0.3 24096 13392 ?       D    18:03   3:48
> /usr/sbin/httpd
> apache   14671  1.4  0.3 24044 13340 ?       D    18:03   3:12
> /usr/sbin/httpd
> apache   14687  1.6  0.3 24020 13320 ?       D    18:03   3:42
> /usr/sbin/httpd
> apache   17329  0.8  0.3 24356 13640 ?       D    19:42   1:04
> /usr/sbin/httpd
> apache   17331  0.9  0.3 24000 13284 ?       D    19:42   1:05
> /usr/sbin/httpd
> apache   17332  1.2  0.3 24156 13456 ?       D    19:43   1:29
> /usr/sbin/httpd
> apache   17335  0.9  0.3 24128 13412 ?       D    19:43   1:06
> /usr/sbin/httpd
> apache   17345  0.9  0.3 23908 13120 ?       D    19:43   1:07
> /usr/sbin/httpd
> apache   17347  0.9  0.3 23608 12896 ?       D    19:43   1:07
> /usr/sbin/httpd
> apache   17385  0.8  0.3 24120 13416 ?       D    19:45   1:01
> /usr/sbin/httpd
> apache   17386  1.2  0.3 23932 13228 ?       D    19:46   1:26
> /usr/sbin/httpd
> apache   17387  1.1  0.3 23904 13200 ?       D    19:46   1:17
> /usr/sbin/httpd
> apache   17398  0.9  0.3 24064 13360 ?       D    19:49   1:04
> /usr/sbin/httpd
> apache   17599  1.3  0.3 24392 13676 ?       D    20:41   0:52
> /usr/sbin/httpd
> apache   17606  0.3  0.3 23472 12632 ?       D    20:42   0:11
> /usr/sbin/httpd
> apache   17607  0.6  0.3 24024 13312 ?       D    20:42   0:23
> /usr/sbin/httpd
> apache   17608  1.1  0.3 23880 13164 ?       D    20:42   0:44
> /usr/sbin/httpd
> apache   17609  0.3  0.3 23556 12736 ?       D    20:42   0:11
> /usr/sbin/httpd
> apache   17620  0.5  0.3 23928 13204 ?       D    20:43   0:21
> /usr/sbin/httpd
> apache   17632  0.0  0.2 23168 12268 ?       D    20:46   0:03
> /usr/sbin/httpd
> apache   17633  0.6  0.3 23448 12704 ?       D    20:47   0:20
> /usr/sbin/httpd
> apache   17635  0.0  0.3 23464 12592 ?       D    20:48   0:01
> /usr/sbin/httpd
> apache   17636  0.0  0.3 23448 12584 ?       D    20:49   0:02
> /usr/sbin/httpd
> apache   17637  0.2  0.3 23472 12640 ?       D    20:49   0:07
> /usr/sbin/httpd
> apache   17638  0.2  0.3 23508 12772 ?       D    20:49   0:07
> /usr/sbin/httpd
> apache   17639  0.2  0.3 23488 12672 ?       D    20:49   0:07
> /usr/sbin/httpd
> apache   17643  0.1  0.3 23860 13104 ?       D    20:50   0:03
> /usr/sbin/httpd
> apache   17644  0.0  0.2 23156 12376 ?       D    20:51   0:02
> /usr/sbin/httpd
> apache   17645  1.2  0.3 23472 12680 ?       D    20:51   0:39
> /usr/sbin/httpd
> apache   17647  0.6  0.2 23184 12428 ?       D    20:52   0:19
> /usr/sbin/httpd
> apache   17648  0.5  0.3 23496 12760 ?       D    20:52   0:18
> /usr/sbin/httpd
> apache   17665  0.4  0.3 23524 12700 ?       D    20:54   0:11
> /usr/sbin/httpd
> apache   17667  0.7  0.3 23972 13240 ?       D    20:54   0:20
> /usr/sbin/httpd
> apache   17668  0.1  0.3 23500 12652 ?       D    20:54   0:03
> /usr/sbin/httpd
> apache   17669  0.2  0.3 23488 12652 ?       D    20:54   0:06
> /usr/sbin/httpd
> apache   17671  0.8  0.3 23696 12888 ?       D    20:55   0:25
> /usr/sbin/httpd
> apache   17732  0.1  0.2 23188 12264 ?       D    21:01   0:03
> /usr/sbin/httpd
> apache   17755  0.4  0.3 23588 12768 ?       D    21:02   0:11
> /usr/sbin/httpd
>
>     This httpd process appear in the 4 servers. The first time, we
> reboot all the cluster´s members, but the servers cant unmount GFS
> filesystem, therefore the only solution is a fisical poweroff (botton).
> The second time we saw the first httpd process with stat "D" and then
> reboot only this server.  Command reboot dont work again and fisical
> poweroff is required. But then, the others 3 servers fix without reboot.
> In this 3 servers, the httpd process in stat "D" became stat "S", all ok.
>     I had read this mail:
> https://www.redhat.com/archives/linux-cluster/2005-December/msg00054.html
>     I think is very similar, but in my log dont appear explicitily
> "stuck in gfs" i dont know if i have the max logging for gfs. I only see
> in my logs the try of unmount or mount:
>
> Jan 16 21:53:54 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_htdocs.2:
> withdrawing from cluster at user's request
> Jan 16 21:53:54 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_htdocs.2: about
> to withdraw from the cluster
> Jan 16 21:53:54 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_htdocs.2:
> waiting for outstanding I/O
> Jan 16 21:53:54 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_htdocs.2:
> telling LM to withdraw
> Jan 16 21:53:55 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_htdocs.2: withdrawn
> Jan 16 21:58:45 sf-03 gfs: Unmounting GFS filesystems:  failed
> Jan 16 22:02:01 sf-03 gfs: Unmounting GFS filesystems:  failed
> Jan 16 22:02:13 sf-03 gfs: Unmounting GFS filesystems:  failed
> Jan 16 22:02:25 sf-03 gfs: Unmounting GFS filesystems:  failed
> Jan 16 22:02:37 sf-03 gfs: Unmounting GFS filesystems:  failed
> Jan 16 22:02:49 sf-03 gfs: Unmounting GFS filesystems:  failed
> Jan 16 22:03:01 sf-03 gfs: Unmounting GFS filesystems:  failed
> Jan 16 22:03:14 sf-03 vgchange:   Can't deactivate volume group
> "vg_gfs1" with 2 open logical volume(s)
> Jan 16 22:03:14 sf-03 clvmd: Deactivating VG vg_gfs1: failed
> Jan 16 22:03:14 sf-03 vgchange:   Can't deactivate volume group
> "vg_gfs2" with 1 open logical volume(s)
> Jan 16 22:03:14 sf-03 clvmd: Deactivating VG vg_gfs2: failed
> Jan 16 22:10:36 sf-03 ccsd[3229]: cluster.conf (cluster name =
> gfs_cluster, version = 2) found.
> Jan 16 22:11:59 sf-03 vgchange:   3 logical volume(s) in volume group
> "vg_gfs1" now active
> Jan 16 22:11:59 sf-03 vgchange:   3 logical volume(s) in volume group
> "vg_gfs2" now active
> Jan 16 22:12:20 sf-03 kernel: GFS 2.6.9-42.2 (built Oct 21 2005
> 11:57:26) installed
> Jan 16 22:12:20 sf-03 kernel: GFS: Trying to join cluster "lock_dlm",
> "gfs_cluster:gfs_raiz"
> Jan 16 22:12:22 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_raiz.2: Joined
> cluster. Now mounting FS...
> Jan 16 22:12:22 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_raiz.2: jid=2:
> Trying to acquire journal lock...
> Jan 16 22:12:22 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_raiz.2: jid=2:
> Looking at journal...
> Jan 16 22:12:22 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_raiz.2: jid=2: Done
> Jan 16 22:12:22 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_raiz.2: Scanning
> for log elements...
> Jan 16 22:12:22 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_raiz.2: Found 0
> unlinked inodes
> Jan 16 22:12:22 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_raiz.2: Found
> quota changes for 0 IDs
> Jan 16 22:12:22 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_raiz.2: Done
> Jan 16 22:12:22 sf-03 kernel: GFS: Trying to join cluster "lock_dlm",
> "gfs_cluster:gfs_cache"
> Jan 16 22:12:25 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_cache.2: Joined
> cluster. Now mounting FS...
> Jan 16 22:12:25 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_cache.2: jid=2:
> Trying to acquire journal lock...
> Jan 16 22:12:25 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_cache.2: jid=2:
> Looking at journal...
> Jan 16 22:12:25 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_cache.2: jid=2:
> Done Jan 16 22:12:25 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_cache.2:
> Scanning for log elements...
> Jan 16 22:12:26 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_cache.2: Found 3
> unlinked inodes
> Jan 16 22:12:26 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_cache.2: Found
> quota changes for 0 IDs
> Jan 16 22:12:26 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_cache.2: Done
> Jan 16 22:12:26 sf-03 kernel: GFS: Trying to join cluster "lock_dlm",
> "gfs_cluster:gfs_htdocs"
> Jan 16 22:12:28 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_htdocs.2: Joined
> cluster. Now mounting FS...
> Jan 16 22:12:28 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_htdocs.2: jid=2:
> Trying to acquire journal lock...
> Jan 16 22:12:28 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_htdocs.2: jid=2:
> Looking at journal...
> Jan 16 22:12:28 sf-03 kernel: GFS: fsid=gfs_cluster:gfs_htdocs.2: jid=2:
> Done
> Jan 16 22:12:28 sf-03 kernel: GFS: Trying to join cluster "lock_dlm",
> "gfs_cluster:lv_webcalle20"
> Jan 16 22:12:30 sf-03 kernel: GFS: fsid=gfs_cluster:lv_webcalle20.2:
> Joined cluster. Now mounting FS...
> Jan 16 22:12:30 sf-03 kernel: GFS: fsid=gfs_cluster:lv_webcalle20.2:
> jid=2: Trying to acquire journal lock...
> Jan 16 22:12:30 sf-03 kernel: GFS: fsid=gfs_cluster:lv_webcalle20.2:
> jid=2: Looking at journal...
> Jan 16 22:12:30 sf-03 kernel: GFS: fsid=gfs_cluster:lv_webcalle20.2:
> jid=2: Done
> Jan 16 22:12:30 sf-03 gfs: Mounting GFS filesystems:  succeeded
>
>     At the end we saw a gfs_tool option interesting "lockdump" and
> before reboot ran it. At the end there are some output of my mountpoint
>     If you need more information about configuration or something please
> say me.
>
>     gfs_tool lockdump /mountpoint:
>
> Glock (5, 2114727)
>   gl_flags =
>   gl_count = 2
>   gl_state = 3
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = no
>   ail_bufs = no
>   Holder
>     owner = -1
>     gh_state = 3
>     gh_flags = 5 7
>     error = 0
>     gh_iflags = 1 6 7
> Glock (2, 119206)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 119206/119206
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (2, 228692)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 228692/228692
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (5, 735842)
>   gl_flags =
>   gl_count = 2
>   gl_state = 3
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = no
>   ail_bufs = no
>   Holder
>     owner = -1
>     gh_state = 3
>     gh_flags = 5 7
>     error = 0
>     gh_iflags = 1 6
> Glock (5, 1418402)
>   gl_flags =
>   gl_count = 2
>   gl_state = 3
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = no
>   ail_bufs = no
>   Holder
>     owner = -1
>     gh_state = 3
>     gh_flags = 5 7
>     error = 0
>     gh_iflags = 1 6 7
> Glock (2, 1885491)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 1885491/1885491
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (2, 399729)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 399729/399729
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (5, 1386646)
>   gl_flags =
>   gl_count = 2
>   gl_state = 3
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = no
>   ail_bufs = no
>   Holder
>     owner = -1
>     gh_state = 3
>     gh_flags = 5 7
>     error = 0
>     gh_iflags = 1 6 7
> Glock (5, 241672)
>   gl_flags =
>   gl_count = 2
>   gl_state = 3
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = no
>   ail_bufs = no
>   Holder
>     owner = -1
>     gh_state = 3
>     gh_flags = 5 7
>     error = 0
>     gh_iflags = 1 6 7
> Glock (5, 207713)
>   gl_flags =
>   gl_count = 2
>   gl_state = 3
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = no
>   ail_bufs = no
>   Holder
>     owner = -1
>     gh_state = 3
>     gh_flags = 5 7
>     error = 0
>     gh_iflags = 1 6 7
> Glock (2, 946688)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 946688/946688
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (5, 30184)
>   gl_flags =
>   gl_count = 2
>   gl_state = 3
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = no
>   ail_bufs = no
>   Holder
>     owner = -1
>     gh_state = 3
>     gh_flags = 5 7
>     error = 0
>     gh_iflags = 1 6 7
> Glock (2, 8340)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 8340/8340
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (2, 2308745)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 2308745/2308745
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (2, 949390)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 949390/949390
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (5, 548987)
>   gl_flags =
>   gl_count = 2
>   gl_state = 3
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = no
>   ail_bufs = no
>   Holder
>     owner = -1
>     gh_state = 3
>     gh_flags = 5 7
>     error = 0
>     gh_iflags = 1 6 7
> Glock (2, 1437881)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 1437881/1437881
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (5, 139108)
>   gl_flags =
>   gl_count = 2
>   gl_state = 3
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = no
>   ail_bufs = no
>   Holder
>     owner = -1
>     gh_state = 3
>     gh_flags = 5 7
>     error = 0
>     gh_iflags = 1 6 7
> Glock (2, 261765)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 261765/261765
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (2, 2530374)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 2530374/2530374
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (5, 33091)
>   gl_flags =
>   gl_count = 2
>   gl_state = 3
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = no
>   ail_bufs = no
>   Holder
>     owner = -1
>     gh_state = 3
>     gh_flags = 5 7
>     error = 0
>     gh_iflags = 1 6 7
> Glock (2, 863848)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 863848/863848
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (5, 208549)
>   gl_flags =
>   gl_count = 2
>   gl_state = 3
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = no
>   ail_bufs = no
>   Holder
>     owner = -1
>     gh_state = 3
>     gh_flags = 5 7
>     error = 0
>     gh_iflags = 1 6 7
> Glock (2, 51708)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 51708/51708
>     type = 2
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (5, 1887878)
>   gl_flags =
>   gl_count = 2
>   gl_state = 3
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = no
>   ail_bufs = no
>   Holder
>     owner = -1
>     gh_state = 3
>     gh_flags = 5 7
>     error = 0
>     gh_iflags = 1 6 7
> Glock (5, 864369)
>   gl_flags =
>   gl_count = 2
>   gl_state = 3
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = no
>   ail_bufs = no
>   Holder
>     owner = -1
>     gh_state = 3
>     gh_flags = 5 7
>     error = 0
>     gh_iflags = 1 6 7
> Glock (2, 1436746)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 1436746/1436746
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (2, 608211)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 608211/608211
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (2, 609430)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 609430/609430
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
> Glock (2, 893532)
>   gl_flags =
>   gl_count = 2
>   gl_state = 0
>   req_gh = no
>   req_bh = no
>   lvb_count = 0
>   object = yes
>   new_le = no
>   incore_le = no
>   reclaim = no
>   aspace = 0
>   ail_bufs = no
>   Inode:
>     num = 893532/893532
>     type = 1
>     i_count = 1
>     i_flags =
>     vnode = yes
>
> Thanks in advance.

-- 
Gruss / Regards,

Dipl.-Ing. Mark Hlawatschek
Phone: +49-89 121 409-55
http://www.atix.de/

**
ATIX - Ges. fuer Informationstechnologie und Consulting mbH
Einsteinstr. 10 - 85716 Unterschleissheim - Germany





More information about the Linux-cluster mailing list