[Linux-cluster] Kernel oops when GFS2 used as localfs
Sridhar Ramaswamy (srramasw)
srramasw at cisco.com
Wed Mar 21 01:53:35 UTC 2007
The problen is solved! Kernel version I tried earlier - 2.6.21-rc1 -
doesn't have the fix for bugzilla# 229831.
I moved to Fedora Core 6 (to resolve all the udev problem) and picked
the latest kernel (2.6.21-rc4). With this bonnie tests run fine.
thanks,
Sridhar
________________________________
From: linux-cluster-bounces at redhat.com
[mailto:linux-cluster-bounces at redhat.com] On Behalf Of Sridhar Ramaswamy
(srramasw)
Sent: Sunday, March 18, 2007 8:26 PM
To: linux clustering
Subject: RE: [Linux-cluster] Kernel oops when GFS2 used as
localfs
Thanks Kevin! That helps.
- Sridhar
________________________________
From: linux-cluster-bounces at redhat.com
[mailto:linux-cluster-bounces at redhat.com] On Behalf Of Kevin Anderson
Sent: Saturday, March 17, 2007 2:37 AM
To: linux clustering
Subject: RE: [Linux-cluster] Kernel oops when GFS2 used
as localfs
Just opened up bugzilla 229831, you should be able to
see it now.
Kevin
On Sat, 2007-03-17 at 00:11 -0600, David wrote:
Once you get it, feel free to share with all of
us.
David
-----Original Message-----
From: linux-cluster-bounces at redhat.com
[mailto:linux-cluster-bounces at redhat.com] On
Behalf Of Sridhar Ramaswamy
(srramasw)
Sent: Wednesday, March 14, 2007 12:20 PM
To: linux clustering
Subject: RE: [Linux-cluster] Kernel oops when
GFS2 used as localfs
Hi folks,
This bugzilla 229831 seems "restricted" :( Can
someone please forward
its details? I'm interested in understanding the
actual problem and its
proposed fix.
Meanwhile I'll try an upstream kernel. Should
2.6.21.rc1 be fine?
Thanks,
Sridhar
> -----Original Message-----
> From: linux-cluster-bounces at redhat.com
> [mailto:linux-cluster-bounces at redhat.com] On
Behalf Of Wendy Cheng
> Sent: Wednesday, March 14, 2007 6:33 AM
> To: linux clustering
> Subject: Re: [Linux-cluster] Kernel oops when
GFS2 used as localfs
>
> Steven Whitehouse wrote:
> > Hi,
> >
> > This looks like Red Hat bugzilla 229831 and
if so then
> there is already
> > a fix in Linus' upstream kernel and in the
latest kernel build for
> > Fedora (both 5 and 6),
> >
> >
> yeah... agree. I missed the unlink part of the
trace in previous post
> and thought it might have something to do with
my ino number
> hacking in
> the lookup code.
>
> -- Wendy
> > Steve.
> >
> > On Wed, 2007-03-14 at 00:15 -0700, Sridhar
Ramaswamy
> (srramasw) wrote:
> >
> >> Sure Wendy. Here it is,
> >>
> >> "fs/gfs2/inode.c" 1256
> >> struct inode *gfs2_ilookup(struct
super_block *sb, struct
> gfs2_inum_host
> >> *inum)
> >> {
> >> return ilookup5(sb, (unsigned
long)inum->no_formal_ino,
> >> iget_test, inum);
> >> }
> >>
> >> static struct inode *gfs2_iget(struct
super_block *sb, struct
> >> gfs2_inum_host *inum) {
> >> return iget5_locked(sb, (unsigned
long)inum->no_formal_ino,
> >> iget_test, iget_set,
inum);
> >> }
> >>
> >> BTW - this code is from kernel version
2.6.20.1.
> >>
> >> Thanks,
> >> Sridhar
> >>
> >>
> >>> -----Original Message-----
> >>> From: linux-cluster-bounces at redhat.com
> >>> [mailto:linux-cluster-bounces at redhat.com]
On Behalf Of Wendy Cheng
> >>> Sent: Tuesday, March 13, 2007 10:29 PM
> >>> To: linux clustering
> >>> Subject: Re: [Linux-cluster] Kernel oops
when GFS2 used as localfs
> >>>
> >>> Sridhar Ramaswamy (srramasw) wrote:
> >>>
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel:
------------[ cut here ]------------
> >>>> Mar 13 15:25:56 cfs1 kernel: kernel BUG
at fs/gfs2/meta_io.c:474!
> >>>>
> >>> I don't have time to pull kernel.org
source code right
> now. Could you
> >>> cut-and-paste the following two routines
(from fs/gfs2/inode.c):
> >>> gfs2_ilookup() and gfs2_iget() so we can
look into this
> >>> quickly ? They
> >>> are all one-liner so shouldn't be too much
troubles. GFS2
> >>> currently has
> >>> ino issue with its lookup code.
> >>>
> >>> -- Wendy
> >>>
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel: invalid
opcode: 0000 [#1] Mar 13
> >>>> 15:25:56 cfs1 kernel: SMP Mar 13 15:25:56
cfs1 kernel: Modules
> >>>> linked in: lock_nolock gfs2 reiserfs nfsd
exportfs nfs lockd
> >>>> nfs_acl ipv6 parport_pc
> lp parport
> >>>> autofs4 sunrpc dm_mirror dm_mod button
battery ac
> uhci_hcd ehci_hcd
> >>>> intel_rng rng_core i2c_i801 i2c_core
e1000 e100 mii
> floppy ext3 jbd
> >>>> Mar 13 15:25:56 cfs1 kernel: CPU: 1
> >>>> Mar 13 15:25:56 cfs1 kernel: EIP:
0060:[<e0c88da5>]
> >>>>
> >>> Not tainted VLI
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel: EFLAGS:
00010246 (2.6.20.1 #1)
> >>>> Mar 13 15:25:56 cfs1 kernel: EIP is at
> >>>> gfs2_meta_indirect_buffer+0x4c/0x278
[gfs2]
> >>>> Mar 13 15:25:56 cfs1 kernel: eax:
00000000 ebx:
> 00012bf5 ecx:
> >>>> ce5a6dd4 edx: dc5eae00
> >>>> Mar 13 15:25:56 cfs1 kernel: esi:
00000000 edi:
> 00000000 ebp:
> >>>> dc5ea9a8 esp: ce5a6d58
> >>>> Mar 13 15:25:56 cfs1 kernel: ds: 007b
es: 007b ss: 0068
> >>>> Mar 13 15:25:56 cfs1 kernel: Process
bonnie++ (pid: 5509,
> >>>>
> >>> ti=ce5a6000
> >>>
> >>>> task=dc320570 task.ti=ce5a6000)
> >>>> Mar 13 15:25:56 cfs1 kernel: Stack:
c156d274 ce5a6dd4 c016eb91
> >>>> ce5a6dd4 00000000 dc5eae00 00000000
d6f34000
> >>>> Mar 13 15:25:56 cfs1 kernel:
00000000 00000000 dc5ea9a8
> >>>> d57794a8 ce5a6e08 e0c83ad3 00012bf5
00000000
> >>>> Mar 13 15:25:56 cfs1 kernel:
00000000 ce5a6da0 00000000
> >>>> 00000000 dc5ea9a8 d57794a8 e0c84cc9
ce5a6dd4
> >>>> Mar 13 15:25:56 cfs1 kernel: Call Trace:
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c016eb91>] iget5_locked+0x3d/0x67
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<e0c83ad3>]
> >>>> gfs2_inode_refresh+0x34/0xfe [gfs2]
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<e0c84cc9>]
> >>>>
> >>> gfs2_createi+0x12c/0x191 [gfs2]
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<e0c8da3c>]
> >>>>
> >>> gfs2_create+0x5c/0x103 [gfs2]
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<e0c84be7>]
> >>>>
> >>> gfs2_createi+0x4a/0x191 [gfs2]
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<e0c820c4>]
> >>>>
> >>> gfs2_glock_nq_num+0x3f/0x64
> >>>
> >>>> [gfs2]
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c016552e>] vfs_create+0xc3/0x126
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c01657f2>]
> >>>>
> >>> open_namei_create+0x47/0x88
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c016597d>] open_namei+0x14a/0x539
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c015d27b>] do_filp_open+0x25/0x39
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c01d200a>]
> >>>>
> >>> strncpy_from_user+0x3c/0x5b
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c015d431>]
> get_unused_fd+0xa8/0xb1
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c015d509>] do_sys_open+0x42/0xbe
> >>>> Mar 13 15:25:56 cfs1 kernel:
[<c015d59f>] sys_open+0x1a/0x1c Mar
> >>>> 13 15:25:56 cfs1 kernel: [<c015d5dd>]
sys_creat+0x1f/0x23 Mar 13
> >>>> 15:25:56 cfs1 kernel: [<c0103410>]
> >>>>
> >>> sysenter_past_esp+0x5d/0x81
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel:
======================= Mar 13
> >>>> 15:25:56 cfs1 kernel: Code: 80 a8 01 00
00 89 44 24
> >>>>
> >>> 1c 8b 85 f0
> >>>
> >>>> 01 00 00 c7 44 24 20 00 00 00 00 89 54 24
14 85 c0 89 44 24
> >>>>
> >>> 18 c7 44
> >>>
> >>>> 24 10 00 00 00 00 75 04 <0f> 0b eb fe 83
7c 24 1c 00 75 04
> >>>>
> >>> 0f 0b eb fe
> >>>
> >>>> 8d 85 24 04 00 00
> >>>> Mar 13 15:25:57 cfs1 kernel: EIP:
[<e0c88da5>]
> >>>> gfs2_meta_indirect_buffer+0x4c/0x278
[gfs2] SS:ESP 0068:ce5a6d58
> >>>>
> >>>>
> >>>> (2)
> >>>>
> >>>> Mar 13 17:00:30 cfs1 kernel: Call Trace:
> >>>> Mar 13 17:00:30 cfs1 kernel:
[<e0badde1>]
> >>>>
> >>> gfs2_unlink+0x53/0xe0 [gfs2]
> >>>
> >>>> Mar 13 17:00:30 cfs1 kernel:
[<e0baddc8>]
> >>>>
> >>> gfs2_unlink+0x3a/0xe0 [gfs2]
> >>>
> >>>> Mar 13 17:00:30 cfs1 kernel:
[<e0badde1>]
> >>>>
> >>> gfs2_unlink+0x53/0xe0 [gfs2]
> >>>
> >>>> Mar 13 17:00:30 cfs1 kernel:
[<c0166572>] vfs_unlink+0xa1/0xc5
> >>>> Mar 13 17:00:30 cfs1 kernel:
[<c016662b>] do_unlinkat+0x95/0xf5
> >>>> Mar 13 17:00:30 cfs1 kernel:
[<c01187de>]
> scheduler_tick+0x8f/0x95
> >>>> Mar 13 17:00:30 cfs1 kernel:
[<c0103410>]
> >>>>
> >>> sysenter_past_esp+0x5d/0x81
> >>>
> >>>> Mar 13 17:00:30 cfs1 kernel:
=======================
> >>>>
> >>>>
-------------------------------------------------------------
> >>>>
> >>> -----------
> >>>
> >>>> --
> >>>> Linux-cluster mailing list
> >>>> Linux-cluster at redhat.com
> >>>>
https://www.redhat.com/mailman/listinfo/linux-cluster
> >>>>
> >>>>
> >>> --
> >>> Linux-cluster mailing list
> >>> Linux-cluster at redhat.com
> >>>
https://www.redhat.com/mailman/listinfo/linux-cluster
> >>>
> >>>
> >> --
> >> Linux-cluster mailing list
> >> Linux-cluster at redhat.com
> >>
https://www.redhat.com/mailman/listinfo/linux-cluster
> >>
> >
> >
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
>
https://www.redhat.com/mailman/listinfo/linux-cluster
>
--
Linux-cluster mailing list
Linux-cluster at redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster
--
Linux-cluster mailing list
Linux-cluster at redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20070320/97cfca69/attachment.htm>
More information about the Linux-cluster
mailing list