[Linux-cluster] Linux-cluster Digest, Vol 97, Issue 5
emmanuel segura
emi2fast at gmail.com
Wed May 16 18:20:08 UTC 2012
Yes Randy
I rember in my jobs i found this problem in a cluster
This it's wrong
==============================================
<service autostart="1" domain="nfs1-domain" exclusive="0" name="nfs1"
nfslock="1" recovery="relocate">
<ip ref="192.168.1.1">
<fs __independent_subtree="1"
ref="volume01">
<nfsexport name="nfs-volume01">
<nfsclient name=" "
ref="local-subnet"/>
</nfsexport>
</fs>
</ip>
================================================
it's must be
================================================
<service autostart="1" domain="nfs1-domain" exclusive="0" name="nfs1"
nfslock="1" recovery="relocate">
<ip ref="192.168.1.1"/>
<fs __independent_subtree="1"
ref="volume01">
<nfsexport name="nfs-volume01">
<nfsclient name=" "
ref="local-subnet"/>
</nfsexport>
</fs>
================================================
A give a little explaination, Redhat have a internar order and knows i
which sequense start the resource
For more information read the script /usr/share/cluster/service.sh under
the metadata session
2012/5/16 Randy Zagar <zagar at arlut.utexas.edu>
> Also, it looks like the resource manager tries to disable the IP address
> when it's a child of the nfsclient resource. Is that going to be a problem
> when I have 16 NFS exports hosted on a single IP?
>
>
> -RZ
>
> On 05/16/2012 11:00 AM, fdinitto at redhat.com wrote:
>
> On 05/15/2012 07:33 PM, Randy Zagar wrote:
>
>
> > <resources>> <ip address="192.168.1.1" monitor_link="1"/>> <ip address="192.168.1.2" monitor_link="1"/>> <ip address="192.168.1.3" monitor_link="1"/>> <fs device="/dev/cvg00/volume01" force_fsck="0" force_unmount="1" fsid="49388" fstype="ext3" mountpoint="/lvm/volume01" name="volume01" self_fence="0"/>> <fs device="/dev/cvg00/volume02" force_fsck="0" force_unmount="1" fsid="58665" fstype="ext3" mountpoint="/lvm/volume01" name="volume01" self_fence="0"/>> <fs device="/dev/cvg00/volume03" force_fsck="0" force_unmount="1" fsid="61028" fstype="ext3" mountpoint="/lvm/volume01" name="volume01" self_fence="0"/>> <nfsclient allow_recover="1" name="local-subnet" options="rw,insecure" target="192.168.1.0/24"/>> </resources>
>
> For the <fs resources you want nfslock="1" option too.
>
>
> > <service autostart="1" domain="nfs1-domain" exclusive="0" name="nfs1" nfslock="1" recovery="relocate">> <ip ref="192.168.1.1">> <fs __independent_subtree="1" ref="volume01">> <nfsexport name="nfs-volume01">> <nfsclient name=" " ref="local-subnet"/>> </nfsexport>> </fs>> </ip>
>
> For all services you need to change the order.
>
> <fs..
> <nfsexport..
> <nfsclient..
> <ip..
> </nfsclient..
> </nfsexport..
> </fs
>
> This solves different issues at startup, relocation and recovery
>
> Also note that there is known limitation in nfsd (both rhel5/6) that
> could cause some problems in some conditions in your current
> configuration. A permanent fix is being worked on atm.
>
> Without extreme details, you might have 2 of those services running on
> the same node and attempting to relocate one of them can fail because
> the fs cannot be unmounted. This is due to nfsd holding a lock (at
> kernel level) to the FS. Changing config to the suggested one, mask the
> problem pretty well, but more testing for a real fix is in progress.
>
> Fabio
>
>
> --
> Randy Zagar Sr. Unix Systems Administrator
> E-mail: zagar at arlut.utexas.edu Applied Research Laboratories
> Phone: 512 835-3131 Univ. of Texas at Austin
>
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>
--
esta es mi vida e me la vivo hasta que dios quiera
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20120516/b03dd46f/attachment.htm>
More information about the Linux-cluster
mailing list