[Linux-cluster] lm_dlm_cancel
James Chamberlain
jamesc at exa.com
Mon Sep 1 23:55:48 UTC 2008
Hi all,
Since I sent the below, the aforementioned cluster crashed. Now I
can't mount the scratch112 filesystem. Attempts to do so crash the
node trying to mount it. If I run gfs_fsck against it, I see the
following:
# gfs_fsck -nv /dev/s12/scratch112
Initializing fsck
Initializing lists...
Initializing special inodes...
Validating Resource Group index.
Level 1 check.
5834 resource groups found.
(passed)
Setting block ranges...
Can't seek to last block in file system: 4969529913
Unable to determine the boundaries of the file system.
Freeing buffers.
Not being able to determine the boundaries of the file system seems
like a very bad thing. However, LVM didn't complain in the slightest
when I expanded the logical volume. How can I recover from this?
Thanks,
James
On Aug 29, 2008, at 9:19 PM, James Chamberlain wrote:
> Hi all,
>
> I'm trying to grow a GFS filesystem. I've grown this filesystem
> before and everything went fine. However, when I issued gfs_grow
> this time, I saw the following messages in my logs:
>
> Aug 29 21:04:13 s12n02 kernel: lock_dlm: lm_dlm_cancel 2,17 flags 80
> Aug 29 21:04:13 s12n02 kernel: lock_dlm: lm_dlm_cancel skip 2,17
> flags 100
> Aug 29 21:04:14 s12n02 kernel: lock_dlm: lm_dlm_cancel 2,17 flags 80
> Aug 29 21:04:14 s12n02 kernel: dlm: scratch112: (14239) dlm_unlock:
> 10241 busy 2
> Aug 29 21:04:14 s12n02 kernel: lock_dlm: lm_dlm_cancel rv -16 2,17
> flags 40080
>
> The last three lines of these log entries repeat themselves once a
> second until I hit ^C. The filesystem appears to still be up and
> accessible. Any thoughts on what's going on here and what I can do
> about it?
>
> Thanks,
>
> James
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
More information about the Linux-cluster
mailing list