[Linux-cluster] GFS2 error recovering journal 0
Steven Whitehouse
swhiteho at redhat.com
Thu Dec 11 09:41:16 UTC 2008
Hi,
The first thing to try is running fsck on it, the more recent the
version of fsck, the better. The filesystem is refusing to mount because
it thinks that there is something wrong with the journal, so it looks
like it needs manual correction,
Steve.
On Wed, 2008-12-10 at 19:11 -0600, Nathan Stratton wrote:
> I have a production system that is down right now, any help would be
> greatly appreciated.
>
> I get a panic when I try to mount:
>
> Dec 10 18:53:41 xen0 kernel: GFS2: fsid=xen_sjc:share.0: Joined cluster.
> Now mounting FS...
> Dec 10 18:53:41 xen0 kernel: GFS2: fsid=xen_sjc:share.0: jid=0, already
> locked for use
> Dec 10 18:53:41 xen0 kernel: GFS2: fsid=xen_sjc:share.0: jid=0: Looking at
> journal...
> Dec 10 18:53:41 xen0 kernel: GFS2: fsid=xen_sjc:share.0: fatal: filesystem
> consistency error
> Dec 10 18:53:41 xen0 kernel: GFS2: fsid=xen_sjc:share.0: inode = 4 53
> Dec 10 18:53:41 xen0 kernel: GFS2: fsid=xen_sjc:share.0: function =
> jhead_scan, file = fs/gfs2/recovery.c, line = 239
> Dec 10 18:53:41 xen0 kernel: GFS2: fsid=xen_sjc:share.0: about to withdraw
> this file system
> Dec 10 18:53:41 xen0 kernel: GFS2: fsid=xen_sjc:share.0: telling LM to
> withdraw
> Dec 10 18:53:41 xen0 kernel: GFS2: fsid=xen_sjc:share.0: withdrawn
> Dec 10 18:53:41 xen0 kernel:
> Dec 10 18:53:41 xen0 kernel: Call Trace:
> Dec 10 18:53:41 xen0 kernel: [<ffffffff8863d0ee>]
> :gfs2:gfs2_lm_withdraw+0xc1/0xd0
> Dec 10 18:53:41 xen0 kernel: [<ffffffff886497bc>]
> :gfs2:gfs2_replay_read_block+0x78/0x89
> Dec 10 18:53:41 xen0 kernel: [<ffffffff8864986a>]
> :gfs2:get_log_header+0x9d/0xe7
> Dec 10 18:53:41 xen0 kernel: [<ffffffff8864ee4f>]
> :gfs2:gfs2_consist_inode_i+0x43/0x48
> Dec 10 18:53:41 xen0 kernel: [<ffffffff88649a12>]
> :gfs2:gfs2_find_jhead+0xf5/0x119
> Dec 10 18:53:41 xen0 kernel: [<ffffffff88649b77>]
> :gfs2:gfs2_recover_journal+0x141/0x837
> Dec 10 18:53:41 xen0 kernel: [<ffffffff88640558>]
> :gfs2:gfs2_meta_read+0x17/0x65
> Dec 10 18:53:41 xen0 kernel: [<ffffffff802648f1>]
> _spin_lock_irqsave+0x9/0x14
> Dec 10 18:53:41 xen0 kernel: [<ffffffff80222b5c>] __up_read+0x19/0x7f
> Dec 10 18:53:41 xen0 kernel: [<ffffffff88631111>]
> :gfs2:gfs2_block_map+0x32b/0x33e
> Dec 10 18:53:41 xen0 kernel: [<ffffffff8864425c>]
> :gfs2:map_journal_extents+0x6f/0x13b
> Dec 10 18:53:41 xen0 kernel: [<ffffffff886312cd>]
> :gfs2:gfs2_write_alloc_required+0xfd/0x122
> Dec 10 18:53:41 xen0 kernel: [<ffffffff886445d5>]
> :gfs2:init_journal+0x2ad/0x40c
> Dec 10 18:53:41 xen0 kernel: [<ffffffff8864cc3a>]
> :gfs2:gfs2_jindex_hold+0x54/0x19c
> Dec 10 18:53:41 xen0 kernel: [<ffffffff88644793>]
> :gfs2:init_inodes+0x5f/0x1d3
> Dec 10 18:53:41 xen0 kernel: [<ffffffff88644d27>]
> :gfs2:fill_super+0x420/0x571
> Dec 10 18:53:41 xen0 kernel: [<ffffffff8863a56b>]
> :gfs2:gfs2_glock_nq_num+0x3b/0x68
> Dec 10 18:53:41 xen0 kernel: [<ffffffff802cf7bb>] set_bdev_super+0x0/0xf
> Dec 10 18:53:41 xen0 kernel: [<ffffffff802cf7ca>] test_bdev_super+0x0/0xd
> Dec 10 18:53:41 xen0 kernel: [<ffffffff88644907>]
> :gfs2:fill_super+0x0/0x571
> Dec 10 18:53:41 xen0 kernel: [<ffffffff802d077e>] get_sb_bdev+0x10a/0x164
> Dec 10 18:53:41 xen0 kernel: [<ffffffff802cac1d>] __kmalloc+0x8f/0x9f
> Dec 10 18:53:41 xen0 kernel: [<ffffffff886439e5>]
> :gfs2:gfs2_get_sb+0x13/0x2f
> Dec 10 18:53:41 xen0 kernel: [<ffffffff802d011b>]
> vfs_kern_mount+0x93/0x11a
> Dec 10 18:53:41 xen0 kernel: [<ffffffff802d01e4>] do_kern_mount+0x36/0x4d
> Dec 10 18:53:41 xen0 kernel: [<ffffffff802d9866>] do_mount+0x6a7/0x717
> Dec 10 18:53:41 xen0 kernel: [<ffffffff8020622a>]
> hypercall_page+0x22a/0x1000
> Dec 10 18:53:41 xen0 kernel: [<ffffffff8020ba3a>]
> free_hot_cold_page+0x107/0x14d
> Dec 10 18:53:41 xen0 kernel: [<ffffffff8020ae99>]
> get_page_from_freelist+0x32e/0x3bc
> Dec 10 18:53:41 xen0 kernel: [<ffffffff80264997>] _read_lock_irq+0x9/0x19
> Dec 10 18:53:41 xen0 kernel: [<ffffffff802071e1>] find_get_page+0x4d/0x54
> Dec 10 18:53:41 xen0 kernel: [<ffffffff80213b9c>]
> filemap_nopage+0x188/0x322
> Dec 10 18:53:41 xen0 kernel: [<ffffffff802090c2>]
> __handle_mm_fault+0x755/0x11bd
> Dec 10 18:53:41 xen0 kernel: [<ffffffff802648f1>]
> _spin_lock_irqsave+0x9/0x14
> Dec 10 18:53:41 xen0 kernel: [<ffffffff80222b5c>] __up_read+0x19/0x7f
> Dec 10 18:53:41 xen0 kernel: [<ffffffff8020ad5f>]
> get_page_from_freelist+0x1f4/0x3bc
> Dec 10 18:53:41 xen0 kernel: [<ffffffff8020f412>]
> __alloc_pages+0x65/0x2ce
> Dec 10 18:53:41 xen0 kernel: [<ffffffff8022b52a>] iput+0x4b/0x84
> Dec 10 18:53:41 xen0 kernel: [<ffffffff8024de16>] sys_mount+0x8a/0xcd
> Dec 10 18:53:41 xen0 kernel: [<ffffffff80260106>] system_call+0x86/0x8b
> Dec 10 18:53:41 xen0 kernel: [<ffffffff80260080>] system_call+0x0/0x8b
> Dec 10 18:53:41 xen0 kernel:
> Dec 10 18:53:41 xen0 kernel: GFS2: fsid=xen_sjc:share.0: jid=0: Failed
> Dec 10 18:53:41 xen0 kernel: GFS2: fsid=xen_sjc:share.0: error recovering
> journal 0: -5
>
>
>
> ><>
> Nathan Stratton CTO, BlinkMind, Inc.
> nathan at robotics.net nathan at blinkmind.com
> http://www.robotics.net http://www.blinkmind.com
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
More information about the Linux-cluster
mailing list