[Linux-cluster] Kernel oops when GFS2 used as localfs

Kevin Anderson kanderso at redhat.com
Sat Mar 17 09:36:33 UTC 2007


Just opened up bugzilla 229831, you should be able to see it now.

Kevin

On Sat, 2007-03-17 at 00:11 -0600, David wrote:
> Once you get it, feel free to share with all of us.
> 
> David 
> 
> -----Original Message-----
> From: linux-cluster-bounces at redhat.com
> [mailto:linux-cluster-bounces at redhat.com] On Behalf Of Sridhar Ramaswamy
> (srramasw)
> Sent: Wednesday, March 14, 2007 12:20 PM
> To: linux clustering
> Subject: RE: [Linux-cluster] Kernel oops when GFS2 used as localfs
> 
> 
> Hi folks,
> 
> This bugzilla 229831 seems "restricted" :(  Can someone please forward
> its details? I'm interested in understanding the actual problem and its
> proposed fix.
> 
> Meanwhile I'll try an upstream kernel. Should 2.6.21.rc1 be fine?
> 
> Thanks,
> Sridhar
> 
> > -----Original Message-----
> > From: linux-cluster-bounces at redhat.com
> > [mailto:linux-cluster-bounces at redhat.com] On Behalf Of Wendy Cheng
> > Sent: Wednesday, March 14, 2007 6:33 AM
> > To: linux clustering
> > Subject: Re: [Linux-cluster] Kernel oops when GFS2 used as localfs
> > 
> > Steven Whitehouse wrote:
> > > Hi,
> > >
> > > This looks like Red Hat bugzilla 229831 and if so then
> > there is already
> > > a fix in Linus' upstream kernel and in the latest kernel build for 
> > > Fedora (both 5 and 6),
> > >
> > >   
> > yeah... agree. I missed the unlink part of the trace in previous post
> > and thought it might have something to do with my ino number 
> > hacking in 
> > the lookup code.
> > 
> > -- Wendy
> > > Steve.
> > >
> > > On Wed, 2007-03-14 at 00:15 -0700, Sridhar Ramaswamy
> > (srramasw) wrote:
> > >   
> > >> Sure Wendy. Here it is,
> > >>
> > >> "fs/gfs2/inode.c" 1256
> > >> struct inode *gfs2_ilookup(struct super_block *sb, struct
> > gfs2_inum_host
> > >> *inum)
> > >> {
> > >>         return ilookup5(sb, (unsigned long)inum->no_formal_ino,
> > >>                         iget_test, inum);
> > >> }
> > >>
> > >> static struct inode *gfs2_iget(struct super_block *sb, struct 
> > >> gfs2_inum_host *inum) {
> > >>         return iget5_locked(sb, (unsigned long)inum->no_formal_ino,
> > >>                      iget_test, iget_set, inum);
> > >> }
> > >>
> > >> BTW - this code is from kernel version 2.6.20.1.
> > >>
> > >> Thanks,
> > >> Sridhar
> > >>
> > >>     
> > >>> -----Original Message-----
> > >>> From: linux-cluster-bounces at redhat.com
> > >>> [mailto:linux-cluster-bounces at redhat.com] On Behalf Of Wendy Cheng
> > >>> Sent: Tuesday, March 13, 2007 10:29 PM
> > >>> To: linux clustering
> > >>> Subject: Re: [Linux-cluster] Kernel oops when GFS2 used as localfs
> > >>>
> > >>> Sridhar Ramaswamy (srramasw) wrote:
> > >>>
> > >>>       
> > >>>> Mar 13 15:25:56 cfs1 kernel: ------------[ cut here ]------------
> 
> > >>>> Mar 13 15:25:56 cfs1 kernel: kernel BUG at fs/gfs2/meta_io.c:474!
> > >>>>         
> > >>> I don't have time to pull kernel.org source code right
> > now. Could you
> > >>> cut-and-paste the following two routines (from fs/gfs2/inode.c):
> > >>> gfs2_ilookup() and gfs2_iget() so we can look into this
> > >>> quickly ? They 
> > >>> are all one-liner so shouldn't be too much troubles. GFS2 
> > >>> currently has 
> > >>> ino issue with its lookup code.
> > >>>
> > >>> -- Wendy
> > >>>
> > >>>       
> > >>>> Mar 13 15:25:56 cfs1 kernel: invalid opcode: 0000 [#1] Mar 13 
> > >>>> 15:25:56 cfs1 kernel: SMP Mar 13 15:25:56 cfs1 kernel: Modules 
> > >>>> linked in: lock_nolock gfs2 reiserfs nfsd exportfs nfs lockd 
> > >>>> nfs_acl ipv6 parport_pc
> > lp parport
> > >>>> autofs4 sunrpc dm_mirror dm_mod button battery ac
> > uhci_hcd ehci_hcd
> > >>>> intel_rng rng_core i2c_i801 i2c_core e1000 e100 mii
> > floppy ext3 jbd
> > >>>> Mar 13 15:25:56 cfs1 kernel: CPU:    1
> > >>>> Mar 13 15:25:56 cfs1 kernel: EIP:    0060:[<e0c88da5>]    
> > >>>>         
> > >>> Not tainted VLI
> > >>>       
> > >>>> Mar 13 15:25:56 cfs1 kernel: EFLAGS: 00010246   (2.6.20.1 #1)
> > >>>> Mar 13 15:25:56 cfs1 kernel: EIP is at
> > >>>> gfs2_meta_indirect_buffer+0x4c/0x278 [gfs2]
> > >>>> Mar 13 15:25:56 cfs1 kernel: eax: 00000000   ebx: 
> > 00012bf5   ecx: 
> > >>>> ce5a6dd4   edx: dc5eae00
> > >>>> Mar 13 15:25:56 cfs1 kernel: esi: 00000000   edi: 
> > 00000000   ebp: 
> > >>>> dc5ea9a8   esp: ce5a6d58
> > >>>> Mar 13 15:25:56 cfs1 kernel: ds: 007b   es: 007b   ss: 0068
> > >>>> Mar 13 15:25:56 cfs1 kernel: Process bonnie++ (pid: 5509,
> > >>>>         
> > >>> ti=ce5a6000
> > >>>       
> > >>>> task=dc320570 task.ti=ce5a6000)
> > >>>> Mar 13 15:25:56 cfs1 kernel: Stack: c156d274 ce5a6dd4 c016eb91
> > >>>> ce5a6dd4 00000000 dc5eae00 00000000 d6f34000
> > >>>> Mar 13 15:25:56 cfs1 kernel:        00000000 00000000 dc5ea9a8 
> > >>>> d57794a8 ce5a6e08 e0c83ad3 00012bf5 00000000
> > >>>> Mar 13 15:25:56 cfs1 kernel:        00000000 ce5a6da0 00000000 
> > >>>> 00000000 dc5ea9a8 d57794a8 e0c84cc9 ce5a6dd4
> > >>>> Mar 13 15:25:56 cfs1 kernel: Call Trace:
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<c016eb91>] iget5_locked+0x3d/0x67
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<e0c83ad3>] 
> > >>>> gfs2_inode_refresh+0x34/0xfe [gfs2]
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<e0c84cc9>] 
> > >>>>         
> > >>> gfs2_createi+0x12c/0x191 [gfs2]
> > >>>       
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<e0c8da3c>]
> > >>>>         
> > >>> gfs2_create+0x5c/0x103 [gfs2]
> > >>>       
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<e0c84be7>]
> > >>>>         
> > >>> gfs2_createi+0x4a/0x191 [gfs2]
> > >>>       
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<e0c820c4>]
> > >>>>         
> > >>> gfs2_glock_nq_num+0x3f/0x64
> > >>>       
> > >>>> [gfs2]
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<c016552e>] vfs_create+0xc3/0x126 
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<c01657f2>]
> > >>>>         
> > >>> open_namei_create+0x47/0x88
> > >>>       
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<c016597d>] open_namei+0x14a/0x539
> 
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<c015d27b>] do_filp_open+0x25/0x39
> 
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<c01d200a>]
> > >>>>         
> > >>> strncpy_from_user+0x3c/0x5b
> > >>>       
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<c015d431>]
> > get_unused_fd+0xa8/0xb1
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<c015d509>] do_sys_open+0x42/0xbe 
> > >>>> Mar 13 15:25:56 cfs1 kernel:  [<c015d59f>] sys_open+0x1a/0x1c Mar
> 
> > >>>> 13 15:25:56 cfs1 kernel:  [<c015d5dd>] sys_creat+0x1f/0x23 Mar 13
> 
> > >>>> 15:25:56 cfs1 kernel:  [<c0103410>]
> > >>>>         
> > >>> sysenter_past_esp+0x5d/0x81
> > >>>       
> > >>>> Mar 13 15:25:56 cfs1 kernel:  ======================= Mar 13 
> > >>>> 15:25:56 cfs1 kernel: Code: 80 a8 01 00 00 89 44 24
> > >>>>         
> > >>> 1c 8b 85 f0
> > >>>       
> > >>>> 01 00 00 c7 44 24 20 00 00 00 00 89 54 24 14 85 c0 89 44 24
> > >>>>         
> > >>> 18 c7 44
> > >>>       
> > >>>> 24 10 00 00 00 00 75 04 <0f> 0b eb fe 83 7c 24 1c 00 75 04
> > >>>>         
> > >>> 0f 0b eb fe
> > >>>       
> > >>>> 8d 85 24 04 00 00
> > >>>> Mar 13 15:25:57 cfs1 kernel: EIP: [<e0c88da5>]
> > >>>> gfs2_meta_indirect_buffer+0x4c/0x278 [gfs2] SS:ESP 0068:ce5a6d58
> > >>>>  
> > >>>>  
> > >>>> (2)
> > >>>>  
> > >>>> Mar 13 17:00:30 cfs1 kernel: Call Trace:
> > >>>> Mar 13 17:00:30 cfs1 kernel:  [<e0badde1>]
> > >>>>         
> > >>> gfs2_unlink+0x53/0xe0 [gfs2]
> > >>>       
> > >>>> Mar 13 17:00:30 cfs1 kernel:  [<e0baddc8>]
> > >>>>         
> > >>> gfs2_unlink+0x3a/0xe0 [gfs2]
> > >>>       
> > >>>> Mar 13 17:00:30 cfs1 kernel:  [<e0badde1>]
> > >>>>         
> > >>> gfs2_unlink+0x53/0xe0 [gfs2]
> > >>>       
> > >>>> Mar 13 17:00:30 cfs1 kernel:  [<c0166572>] vfs_unlink+0xa1/0xc5 
> > >>>> Mar 13 17:00:30 cfs1 kernel:  [<c016662b>] do_unlinkat+0x95/0xf5 
> > >>>> Mar 13 17:00:30 cfs1 kernel:  [<c01187de>]
> > scheduler_tick+0x8f/0x95
> > >>>> Mar 13 17:00:30 cfs1 kernel:  [<c0103410>]
> > >>>>         
> > >>> sysenter_past_esp+0x5d/0x81
> > >>>       
> > >>>> Mar 13 17:00:30 cfs1 kernel:  =======================
> > >>>>
> > >>>> -------------------------------------------------------------
> > >>>>         
> > >>> -----------
> > >>>       
> > >>>> --
> > >>>> Linux-cluster mailing list
> > >>>> Linux-cluster at redhat.com 
> > >>>> https://www.redhat.com/mailman/listinfo/linux-cluster
> > >>>>
> > >>>>         
> > >>> --
> > >>> Linux-cluster mailing list
> > >>> Linux-cluster at redhat.com 
> > >>> https://www.redhat.com/mailman/listinfo/linux-cluster
> > >>>
> > >>>       
> > >> --
> > >> Linux-cluster mailing list
> > >> Linux-cluster at redhat.com 
> > >> https://www.redhat.com/mailman/listinfo/linux-cluster
> > >>     
> > >
> > >   
> > 
> > --
> > Linux-cluster mailing list
> > Linux-cluster at redhat.com 
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> > 
> 
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
> 
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20070317/ed969d3b/attachment.htm>


More information about the Linux-cluster mailing list