[Linux-cluster] How to run same service in parallel in RedHat Cluster 5.0

Ruben Sajnovetzky rsajnove at cisco.com
Wed Sep 28 17:02:06 UTC 2011


We crossed e-mails :)
I sent the new-fresh configuration.
I thought about fencing, the problem is that we have a very "odd"
configuration because I don't really need fence anything ... Maybe I
can establish kind of rule like "If Central is not working, Collector cant
work" or similar, will think on it.

The IP will not be migrated, we have a requirement to have Virtual IP for
the Central (only), that is why it is configured.

Regarding FS (other member asked why GFS), I don't see the advantage
because, again, we are not "sharing" the storage.

Thanks



On 28-Sep-2011 12:50 PM, "Digimer" <linux at alteeve.com> wrote:

> On 09/28/2011 06:20 AM, Ruben Sajnovetzky wrote:
>> <?xml version="1.0"?>
>> <cluster alias="PPM_CL1" config_version="81" name="PPM_CL1">
>>         <fence_daemon clean_start="0" post_fail_delay="0"
>> post_join_delay="30"/>
>>         <clusternodes>
>>                 <clusternode name="server-87111" nodeid="1" votes="2">
>>                         <fence/>
>>                 </clusternode>
>>                 <clusternode name="server-87112" nodeid="2" votes="1">
>>                         <fence/>
>>                 </clusternode>
>>         </clusternodes>
>>         <cman expected_votes="1">
>>                 <multicast addr="224.4.5.6"/>
>>         </cman>
>>         <fencedevices/>
>>         <rm>
>>                 <failoverdomains>
>>                         <failoverdomain name="PPM_GW_FDN" nofailback="1"
>> ordered="0" restricted="1">
>>                                 <failoverdomainnode name="server-87111"
>> priority="1"/>
>>                         </failoverdomain>
>>                         <failoverdomain name="PPM_UNIT_FDN" nofailback="1"
>> ordered="0" restricted="1">
>>                                 <failoverdomainnode name="server-87112"
>> priority="1"/>
>>                         </failoverdomain>
>>                 </failoverdomains>
>>                 <resources>
>>                         <ip address="10.116.159.105" monitor_link="1"/>
>>                         <script file="/usr/local/bin/cluster/ppm_gw_ha"
>> name="PPM_GW"/>
>>                         <fs device="/dev/mapper/VolGroup00-optvol"
>> force_fsck="1" force_unmount="0" fsid="36845" fstype="ext3"
>> mountpoint="/opt" name="PPM_OPT_FS" self_fence="0"/>
>>                         <apache config_file="conf/httpd.conf" name="web"
>> server_root="/etc/httpd" shutdown_wait="0"/>
>>                         <clusterfs device="/dev/hdd" force_unmount="0"
>> fsid="20023" fstype="gfs" mountpoint="/mnt" name="pru1" self_fence="0"/>
>>                 </resources>
>>                 <service autostart="0" exclusive="0" name="PPM_PRUEBA"/>
>>                 <service autostart="0" domain="PPM_UNIT_FDN" exclusive="0"
>> name="PPM Units">
>>                         <fs ref="PPM_OPT_FS"/>
>>                 </service>
>>                 <service autostart="0" domain="PPM_GW_FDN" exclusive="0"
>> name="PPM Gateway">
>>                         <fs ref="PPM_OPT_FS">
>>                                 <ip ref="10.116.159.105">
>>                                         <script ref="PPM_GW"/>
>>                                 </ip>
>>                         </fs>
>>                 </service>
>>         </rm>
>> </cluster>
> 
> Ah, here it is.
> 
> If I can recommend, remove the parts you're not yet using (the empty
> service, the apache resource, etc). Remove any options that aren't
> critical... make it as simple as possible. Once you get it working, you
> can start adding things back.
> 
> As an aside, with the config you have, the IP address will never
> migrate. It will come up on and only ever run on PPM_GW_FDN. Also, it
> will not come up if the fs resource fails as it is a child. This may be
> what you want though.
> 
> Also, you *must* configure fencing. Even without a shared file system, a
> fence call will hang the cluster. It can cause many odd symptoms. Get
> fencing configured and working, strip down the config and try again. If
> it still fails, repost the config as it is at that point.




More information about the Linux-cluster mailing list