[Linux-cluster] Kernel oops when GFS2 used as localfs

Sridhar Ramaswamy (srramasw) srramasw at cisco.com
Wed Mar 21 01:53:35 UTC 2007


The problen is solved! Kernel version I tried earlier - 2.6.21-rc1 -
doesn't have the fix for bugzilla# 229831.
 
I moved to Fedora Core 6 (to resolve all the udev problem) and picked
the latest kernel (2.6.21-rc4).  With this bonnie tests run fine. 
 
thanks,
Sridhar


________________________________

	From: linux-cluster-bounces at redhat.com
[mailto:linux-cluster-bounces at redhat.com] On Behalf Of Sridhar Ramaswamy
(srramasw)
	Sent: Sunday, March 18, 2007 8:26 PM
	To: linux clustering
	Subject: RE: [Linux-cluster] Kernel oops when GFS2 used as
localfs
	
	
	Thanks Kevin! That helps.
	 
	- Sridhar


________________________________

		From: linux-cluster-bounces at redhat.com
[mailto:linux-cluster-bounces at redhat.com] On Behalf Of Kevin Anderson
		Sent: Saturday, March 17, 2007 2:37 AM
		To: linux clustering
		Subject: RE: [Linux-cluster] Kernel oops when GFS2 used
as localfs
		
		
		Just opened up bugzilla 229831, you should be able to
see it now.
		
		Kevin
		
		On Sat, 2007-03-17 at 00:11 -0600, David wrote: 

			Once you get it, feel free to share with all of
us.
			
			David 
			
			-----Original Message-----
			From: linux-cluster-bounces at redhat.com
			[mailto:linux-cluster-bounces at redhat.com] On
Behalf Of Sridhar Ramaswamy
			(srramasw)
			Sent: Wednesday, March 14, 2007 12:20 PM
			To: linux clustering
			Subject: RE: [Linux-cluster] Kernel oops when
GFS2 used as localfs
			
			
			Hi folks,
			
			This bugzilla 229831 seems "restricted" :(  Can
someone please forward
			its details? I'm interested in understanding the
actual problem and its
			proposed fix.
			
			Meanwhile I'll try an upstream kernel. Should
2.6.21.rc1 be fine?
			
			Thanks,
			Sridhar
			
			> -----Original Message-----
			> From: linux-cluster-bounces at redhat.com
			> [mailto:linux-cluster-bounces at redhat.com] On
Behalf Of Wendy Cheng
			> Sent: Wednesday, March 14, 2007 6:33 AM
			> To: linux clustering
			> Subject: Re: [Linux-cluster] Kernel oops when
GFS2 used as localfs
			> 
			> Steven Whitehouse wrote:
			> > Hi,
			> >
			> > This looks like Red Hat bugzilla 229831 and
if so then
			> there is already
			> > a fix in Linus' upstream kernel and in the
latest kernel build for 
			> > Fedora (both 5 and 6),
			> >
			> >   
			> yeah... agree. I missed the unlink part of the
trace in previous post
			> and thought it might have something to do with
my ino number 
			> hacking in 
			> the lookup code.
			> 
			> -- Wendy
			> > Steve.
			> >
			> > On Wed, 2007-03-14 at 00:15 -0700, Sridhar
Ramaswamy
			> (srramasw) wrote:
			> >   
			> >> Sure Wendy. Here it is,
			> >>
			> >> "fs/gfs2/inode.c" 1256
			> >> struct inode *gfs2_ilookup(struct
super_block *sb, struct
			> gfs2_inum_host
			> >> *inum)
			> >> {
			> >>         return ilookup5(sb, (unsigned
long)inum->no_formal_ino,
			> >>                         iget_test, inum);
			> >> }
			> >>
			> >> static struct inode *gfs2_iget(struct
super_block *sb, struct 
			> >> gfs2_inum_host *inum) {
			> >>         return iget5_locked(sb, (unsigned
long)inum->no_formal_ino,
			> >>                      iget_test, iget_set,
inum);
			> >> }
			> >>
			> >> BTW - this code is from kernel version
2.6.20.1.
			> >>
			> >> Thanks,
			> >> Sridhar
			> >>
			> >>     
			> >>> -----Original Message-----
			> >>> From: linux-cluster-bounces at redhat.com
			> >>> [mailto:linux-cluster-bounces at redhat.com]
On Behalf Of Wendy Cheng
			> >>> Sent: Tuesday, March 13, 2007 10:29 PM
			> >>> To: linux clustering
			> >>> Subject: Re: [Linux-cluster] Kernel oops
when GFS2 used as localfs
			> >>>
			> >>> Sridhar Ramaswamy (srramasw) wrote:
			> >>>
			> >>>       
			> >>>> Mar 13 15:25:56 cfs1 kernel:
------------[ cut here ]------------
			
			> >>>> Mar 13 15:25:56 cfs1 kernel: kernel BUG
at fs/gfs2/meta_io.c:474!
			> >>>>         
			> >>> I don't have time to pull kernel.org
source code right
			> now. Could you
			> >>> cut-and-paste the following two routines
(from fs/gfs2/inode.c):
			> >>> gfs2_ilookup() and gfs2_iget() so we can
look into this
			> >>> quickly ? They 
			> >>> are all one-liner so shouldn't be too much
troubles. GFS2 
			> >>> currently has 
			> >>> ino issue with its lookup code.
			> >>>
			> >>> -- Wendy
			> >>>
			> >>>       
			> >>>> Mar 13 15:25:56 cfs1 kernel: invalid
opcode: 0000 [#1] Mar 13 
			> >>>> 15:25:56 cfs1 kernel: SMP Mar 13 15:25:56
cfs1 kernel: Modules 
			> >>>> linked in: lock_nolock gfs2 reiserfs nfsd
exportfs nfs lockd 
			> >>>> nfs_acl ipv6 parport_pc
			> lp parport
			> >>>> autofs4 sunrpc dm_mirror dm_mod button
battery ac
			> uhci_hcd ehci_hcd
			> >>>> intel_rng rng_core i2c_i801 i2c_core
e1000 e100 mii
			> floppy ext3 jbd
			> >>>> Mar 13 15:25:56 cfs1 kernel: CPU:    1
			> >>>> Mar 13 15:25:56 cfs1 kernel: EIP:
0060:[<e0c88da5>]    
			> >>>>         
			> >>> Not tainted VLI
			> >>>       
			> >>>> Mar 13 15:25:56 cfs1 kernel: EFLAGS:
00010246   (2.6.20.1 #1)
			> >>>> Mar 13 15:25:56 cfs1 kernel: EIP is at
			> >>>> gfs2_meta_indirect_buffer+0x4c/0x278
[gfs2]
			> >>>> Mar 13 15:25:56 cfs1 kernel: eax:
00000000   ebx: 
			> 00012bf5   ecx: 
			> >>>> ce5a6dd4   edx: dc5eae00
			> >>>> Mar 13 15:25:56 cfs1 kernel: esi:
00000000   edi: 
			> 00000000   ebp: 
			> >>>> dc5ea9a8   esp: ce5a6d58
			> >>>> Mar 13 15:25:56 cfs1 kernel: ds: 007b
es: 007b   ss: 0068
			> >>>> Mar 13 15:25:56 cfs1 kernel: Process
bonnie++ (pid: 5509,
			> >>>>         
			> >>> ti=ce5a6000
			> >>>       
			> >>>> task=dc320570 task.ti=ce5a6000)
			> >>>> Mar 13 15:25:56 cfs1 kernel: Stack:
c156d274 ce5a6dd4 c016eb91
			> >>>> ce5a6dd4 00000000 dc5eae00 00000000
d6f34000
			> >>>> Mar 13 15:25:56 cfs1 kernel:
00000000 00000000 dc5ea9a8 
			> >>>> d57794a8 ce5a6e08 e0c83ad3 00012bf5
00000000
			> >>>> Mar 13 15:25:56 cfs1 kernel:
00000000 ce5a6da0 00000000 
			> >>>> 00000000 dc5ea9a8 d57794a8 e0c84cc9
ce5a6dd4
			> >>>> Mar 13 15:25:56 cfs1 kernel: Call Trace:
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c016eb91>] iget5_locked+0x3d/0x67
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<e0c83ad3>] 
			> >>>> gfs2_inode_refresh+0x34/0xfe [gfs2]
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<e0c84cc9>] 
			> >>>>         
			> >>> gfs2_createi+0x12c/0x191 [gfs2]
			> >>>       
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<e0c8da3c>]
			> >>>>         
			> >>> gfs2_create+0x5c/0x103 [gfs2]
			> >>>       
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<e0c84be7>]
			> >>>>         
			> >>> gfs2_createi+0x4a/0x191 [gfs2]
			> >>>       
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<e0c820c4>]
			> >>>>         
			> >>> gfs2_glock_nq_num+0x3f/0x64
			> >>>       
			> >>>> [gfs2]
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c016552e>] vfs_create+0xc3/0x126 
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c01657f2>]
			> >>>>         
			> >>> open_namei_create+0x47/0x88
			> >>>       
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c016597d>] open_namei+0x14a/0x539
			
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c015d27b>] do_filp_open+0x25/0x39
			
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c01d200a>]
			> >>>>         
			> >>> strncpy_from_user+0x3c/0x5b
			> >>>       
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c015d431>]
			> get_unused_fd+0xa8/0xb1
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c015d509>] do_sys_open+0x42/0xbe 
			> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c015d59f>] sys_open+0x1a/0x1c Mar
			
			> >>>> 13 15:25:56 cfs1 kernel:  [<c015d5dd>]
sys_creat+0x1f/0x23 Mar 13
			
			> >>>> 15:25:56 cfs1 kernel:  [<c0103410>]
			> >>>>         
			> >>> sysenter_past_esp+0x5d/0x81
			> >>>       
			> >>>> Mar 13 15:25:56 cfs1 kernel:
======================= Mar 13 
			> >>>> 15:25:56 cfs1 kernel: Code: 80 a8 01 00
00 89 44 24
			> >>>>         
			> >>> 1c 8b 85 f0
			> >>>       
			> >>>> 01 00 00 c7 44 24 20 00 00 00 00 89 54 24
14 85 c0 89 44 24
			> >>>>         
			> >>> 18 c7 44
			> >>>       
			> >>>> 24 10 00 00 00 00 75 04 <0f> 0b eb fe 83
7c 24 1c 00 75 04
			> >>>>         
			> >>> 0f 0b eb fe
			> >>>       
			> >>>> 8d 85 24 04 00 00
			> >>>> Mar 13 15:25:57 cfs1 kernel: EIP:
[<e0c88da5>]
			> >>>> gfs2_meta_indirect_buffer+0x4c/0x278
[gfs2] SS:ESP 0068:ce5a6d58
			> >>>>  
			> >>>>  
			> >>>> (2)
			> >>>>  
			> >>>> Mar 13 17:00:30 cfs1 kernel: Call Trace:
			> >>>> Mar 13 17:00:30 cfs1 kernel:
[<e0badde1>]
			> >>>>         
			> >>> gfs2_unlink+0x53/0xe0 [gfs2]
			> >>>       
			> >>>> Mar 13 17:00:30 cfs1 kernel:
[<e0baddc8>]
			> >>>>         
			> >>> gfs2_unlink+0x3a/0xe0 [gfs2]
			> >>>       
			> >>>> Mar 13 17:00:30 cfs1 kernel:
[<e0badde1>]
			> >>>>         
			> >>> gfs2_unlink+0x53/0xe0 [gfs2]
			> >>>       
			> >>>> Mar 13 17:00:30 cfs1 kernel:
[<c0166572>] vfs_unlink+0xa1/0xc5 
			> >>>> Mar 13 17:00:30 cfs1 kernel:
[<c016662b>] do_unlinkat+0x95/0xf5 
			> >>>> Mar 13 17:00:30 cfs1 kernel:
[<c01187de>]
			> scheduler_tick+0x8f/0x95
			> >>>> Mar 13 17:00:30 cfs1 kernel:
[<c0103410>]
			> >>>>         
			> >>> sysenter_past_esp+0x5d/0x81
			> >>>       
			> >>>> Mar 13 17:00:30 cfs1 kernel:
=======================
			> >>>>
			> >>>>
-------------------------------------------------------------
			> >>>>         
			> >>> -----------
			> >>>       
			> >>>> --
			> >>>> Linux-cluster mailing list
			> >>>> Linux-cluster at redhat.com 
			> >>>>
https://www.redhat.com/mailman/listinfo/linux-cluster
			> >>>>
			> >>>>         
			> >>> --
			> >>> Linux-cluster mailing list
			> >>> Linux-cluster at redhat.com 
			> >>>
https://www.redhat.com/mailman/listinfo/linux-cluster
			> >>>
			> >>>       
			> >> --
			> >> Linux-cluster mailing list
			> >> Linux-cluster at redhat.com 
			> >>
https://www.redhat.com/mailman/listinfo/linux-cluster
			> >>     
			> >
			> >   
			> 
			> --
			> Linux-cluster mailing list
			> Linux-cluster at redhat.com 
			>
https://www.redhat.com/mailman/listinfo/linux-cluster
			> 
			
			--
			Linux-cluster mailing list
			Linux-cluster at redhat.com
	
https://www.redhat.com/mailman/listinfo/linux-cluster
			
			--
			Linux-cluster mailing list
			Linux-cluster at redhat.com
	
https://www.redhat.com/mailman/listinfo/linux-cluster

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20070320/97cfca69/attachment.htm>


More information about the Linux-cluster mailing list