[Linux-cluster] Kernel oops when GFS2 used as localfs
Sridhar Ramaswamy (srramasw)
srramasw at cisco.com
Mon Mar 19 03:26:06 UTC 2007
Thanks Kevin! That helps.
- Sridhar
________________________________
From: linux-cluster-bounces at redhat.com
[mailto:linux-cluster-bounces at redhat.com] On Behalf Of Kevin Anderson
Sent: Saturday, March 17, 2007 2:37 AM
To: linux clustering
Subject: RE: [Linux-cluster] Kernel oops when GFS2 used as
localfs
Just opened up bugzilla 229831, you should be able to see it
now.
Kevin
On Sat, 2007-03-17 at 00:11 -0600, David wrote:
Once you get it, feel free to share with all of us.
David
-----Original Message-----
From: linux-cluster-bounces at redhat.com
[mailto:linux-cluster-bounces at redhat.com] On Behalf Of
Sridhar Ramaswamy
(srramasw)
Sent: Wednesday, March 14, 2007 12:20 PM
To: linux clustering
Subject: RE: [Linux-cluster] Kernel oops when GFS2 used
as localfs
Hi folks,
This bugzilla 229831 seems "restricted" :( Can someone
please forward
its details? I'm interested in understanding the actual
problem and its
proposed fix.
Meanwhile I'll try an upstream kernel. Should 2.6.21.rc1
be fine?
Thanks,
Sridhar
> -----Original Message-----
> From: linux-cluster-bounces at redhat.com
> [mailto:linux-cluster-bounces at redhat.com] On Behalf Of
Wendy Cheng
> Sent: Wednesday, March 14, 2007 6:33 AM
> To: linux clustering
> Subject: Re: [Linux-cluster] Kernel oops when GFS2
used as localfs
>
> Steven Whitehouse wrote:
> > Hi,
> >
> > This looks like Red Hat bugzilla 229831 and if so
then
> there is already
> > a fix in Linus' upstream kernel and in the latest
kernel build for
> > Fedora (both 5 and 6),
> >
> >
> yeah... agree. I missed the unlink part of the trace
in previous post
> and thought it might have something to do with my ino
number
> hacking in
> the lookup code.
>
> -- Wendy
> > Steve.
> >
> > On Wed, 2007-03-14 at 00:15 -0700, Sridhar Ramaswamy
> (srramasw) wrote:
> >
> >> Sure Wendy. Here it is,
> >>
> >> "fs/gfs2/inode.c" 1256
> >> struct inode *gfs2_ilookup(struct super_block *sb,
struct
> gfs2_inum_host
> >> *inum)
> >> {
> >> return ilookup5(sb, (unsigned
long)inum->no_formal_ino,
> >> iget_test, inum);
> >> }
> >>
> >> static struct inode *gfs2_iget(struct super_block
*sb, struct
> >> gfs2_inum_host *inum) {
> >> return iget5_locked(sb, (unsigned
long)inum->no_formal_ino,
> >> iget_test, iget_set, inum);
> >> }
> >>
> >> BTW - this code is from kernel version 2.6.20.1.
> >>
> >> Thanks,
> >> Sridhar
> >>
> >>
> >>> -----Original Message-----
> >>> From: linux-cluster-bounces at redhat.com
> >>> [mailto:linux-cluster-bounces at redhat.com] On
Behalf Of Wendy Cheng
> >>> Sent: Tuesday, March 13, 2007 10:29 PM
> >>> To: linux clustering
> >>> Subject: Re: [Linux-cluster] Kernel oops when GFS2
used as localfs
> >>>
> >>> Sridhar Ramaswamy (srramasw) wrote:
> >>>
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel: ------------[ cut
here ]------------
> >>>> Mar 13 15:25:56 cfs1 kernel: kernel BUG at
fs/gfs2/meta_io.c:474!
> >>>>
> >>> I don't have time to pull kernel.org source code
right
> now. Could you
> >>> cut-and-paste the following two routines (from
fs/gfs2/inode.c):
> >>> gfs2_ilookup() and gfs2_iget() so we can look into
this
> >>> quickly ? They
> >>> are all one-liner so shouldn't be too much
troubles. GFS2
> >>> currently has
> >>> ino issue with its lookup code.
> >>>
> >>> -- Wendy
> >>>
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel: invalid opcode: 0000
[#1] Mar 13
> >>>> 15:25:56 cfs1 kernel: SMP Mar 13 15:25:56 cfs1
kernel: Modules
> >>>> linked in: lock_nolock gfs2 reiserfs nfsd
exportfs nfs lockd
> >>>> nfs_acl ipv6 parport_pc
> lp parport
> >>>> autofs4 sunrpc dm_mirror dm_mod button battery ac
> uhci_hcd ehci_hcd
> >>>> intel_rng rng_core i2c_i801 i2c_core e1000 e100
mii
> floppy ext3 jbd
> >>>> Mar 13 15:25:56 cfs1 kernel: CPU: 1
> >>>> Mar 13 15:25:56 cfs1 kernel: EIP:
0060:[<e0c88da5>]
> >>>>
> >>> Not tainted VLI
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel: EFLAGS: 00010246
(2.6.20.1 #1)
> >>>> Mar 13 15:25:56 cfs1 kernel: EIP is at
> >>>> gfs2_meta_indirect_buffer+0x4c/0x278 [gfs2]
> >>>> Mar 13 15:25:56 cfs1 kernel: eax: 00000000 ebx:
> 00012bf5 ecx:
> >>>> ce5a6dd4 edx: dc5eae00
> >>>> Mar 13 15:25:56 cfs1 kernel: esi: 00000000 edi:
> 00000000 ebp:
> >>>> dc5ea9a8 esp: ce5a6d58
> >>>> Mar 13 15:25:56 cfs1 kernel: ds: 007b es: 007b
ss: 0068
> >>>> Mar 13 15:25:56 cfs1 kernel: Process bonnie++
(pid: 5509,
> >>>>
> >>> ti=ce5a6000
> >>>
> >>>> task=dc320570 task.ti=ce5a6000)
> >>>> Mar 13 15:25:56 cfs1 kernel: Stack: c156d274
ce5a6dd4 c016eb91
> >>>> ce5a6dd4 00000000 dc5eae00 00000000 d6f34000
> >>>> Mar 13 15:25:56 cfs1 kernel: 00000000
00000000 dc5ea9a8
> >>>> d57794a8 ce5a6e08 e0c83ad3 00012bf5 00000000
> >>>> Mar 13 15:25:56 cfs1 kernel: 00000000
ce5a6da0 00000000
> >>>> 00000000 dc5ea9a8 d57794a8 e0c84cc9 ce5a6dd4
> >>>> Mar 13 15:25:56 cfs1 kernel: Call Trace:
> >>>> Mar 13 15:25:56 cfs1 kernel: [<c016eb91>]
iget5_locked+0x3d/0x67
> >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c83ad3>]
> >>>> gfs2_inode_refresh+0x34/0xfe [gfs2]
> >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c84cc9>]
> >>>>
> >>> gfs2_createi+0x12c/0x191 [gfs2]
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c8da3c>]
> >>>>
> >>> gfs2_create+0x5c/0x103 [gfs2]
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c84be7>]
> >>>>
> >>> gfs2_createi+0x4a/0x191 [gfs2]
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c820c4>]
> >>>>
> >>> gfs2_glock_nq_num+0x3f/0x64
> >>>
> >>>> [gfs2]
> >>>> Mar 13 15:25:56 cfs1 kernel: [<c016552e>]
vfs_create+0xc3/0x126
> >>>> Mar 13 15:25:56 cfs1 kernel: [<c01657f2>]
> >>>>
> >>> open_namei_create+0x47/0x88
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel: [<c016597d>]
open_namei+0x14a/0x539
> >>>> Mar 13 15:25:56 cfs1 kernel: [<c015d27b>]
do_filp_open+0x25/0x39
> >>>> Mar 13 15:25:56 cfs1 kernel: [<c01d200a>]
> >>>>
> >>> strncpy_from_user+0x3c/0x5b
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel: [<c015d431>]
> get_unused_fd+0xa8/0xb1
> >>>> Mar 13 15:25:56 cfs1 kernel: [<c015d509>]
do_sys_open+0x42/0xbe
> >>>> Mar 13 15:25:56 cfs1 kernel: [<c015d59f>]
sys_open+0x1a/0x1c Mar
> >>>> 13 15:25:56 cfs1 kernel: [<c015d5dd>]
sys_creat+0x1f/0x23 Mar 13
> >>>> 15:25:56 cfs1 kernel: [<c0103410>]
> >>>>
> >>> sysenter_past_esp+0x5d/0x81
> >>>
> >>>> Mar 13 15:25:56 cfs1 kernel:
======================= Mar 13
> >>>> 15:25:56 cfs1 kernel: Code: 80 a8 01 00 00 89 44
24
> >>>>
> >>> 1c 8b 85 f0
> >>>
> >>>> 01 00 00 c7 44 24 20 00 00 00 00 89 54 24 14 85
c0 89 44 24
> >>>>
> >>> 18 c7 44
> >>>
> >>>> 24 10 00 00 00 00 75 04 <0f> 0b eb fe 83 7c 24 1c
00 75 04
> >>>>
> >>> 0f 0b eb fe
> >>>
> >>>> 8d 85 24 04 00 00
> >>>> Mar 13 15:25:57 cfs1 kernel: EIP: [<e0c88da5>]
> >>>> gfs2_meta_indirect_buffer+0x4c/0x278 [gfs2]
SS:ESP 0068:ce5a6d58
> >>>>
> >>>>
> >>>> (2)
> >>>>
> >>>> Mar 13 17:00:30 cfs1 kernel: Call Trace:
> >>>> Mar 13 17:00:30 cfs1 kernel: [<e0badde1>]
> >>>>
> >>> gfs2_unlink+0x53/0xe0 [gfs2]
> >>>
> >>>> Mar 13 17:00:30 cfs1 kernel: [<e0baddc8>]
> >>>>
> >>> gfs2_unlink+0x3a/0xe0 [gfs2]
> >>>
> >>>> Mar 13 17:00:30 cfs1 kernel: [<e0badde1>]
> >>>>
> >>> gfs2_unlink+0x53/0xe0 [gfs2]
> >>>
> >>>> Mar 13 17:00:30 cfs1 kernel: [<c0166572>]
vfs_unlink+0xa1/0xc5
> >>>> Mar 13 17:00:30 cfs1 kernel: [<c016662b>]
do_unlinkat+0x95/0xf5
> >>>> Mar 13 17:00:30 cfs1 kernel: [<c01187de>]
> scheduler_tick+0x8f/0x95
> >>>> Mar 13 17:00:30 cfs1 kernel: [<c0103410>]
> >>>>
> >>> sysenter_past_esp+0x5d/0x81
> >>>
> >>>> Mar 13 17:00:30 cfs1 kernel:
=======================
> >>>>
> >>>>
-------------------------------------------------------------
> >>>>
> >>> -----------
> >>>
> >>>> --
> >>>> Linux-cluster mailing list
> >>>> Linux-cluster at redhat.com
> >>>>
https://www.redhat.com/mailman/listinfo/linux-cluster
> >>>>
> >>>>
> >>> --
> >>> Linux-cluster mailing list
> >>> Linux-cluster at redhat.com
> >>>
https://www.redhat.com/mailman/listinfo/linux-cluster
> >>>
> >>>
> >> --
> >> Linux-cluster mailing list
> >> Linux-cluster at redhat.com
> >>
https://www.redhat.com/mailman/listinfo/linux-cluster
> >>
> >
> >
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>
--
Linux-cluster mailing list
Linux-cluster at redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster
--
Linux-cluster mailing list
Linux-cluster at redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20070318/88adcce7/attachment.htm>
More information about the Linux-cluster
mailing list