[Linux-cluster] CLVM in a 3-node cluster

Digimer lists at alteeve.ca
Wed Jul 11 14:04:09 UTC 2012


Please start a new thread, with a new subject, and include your
cluster.conf file please.

Digimer

On 07/11/2012 05:57 AM, AKIN ÿffffffffffd6ZTOPUZ wrote:
> Hi
>  
> I have 2-nodes cluster without quorum disks.ı noticed a problem at below:
>  
>  
> when I want to move resources to other node it is failed   to relocate
> services to other node and again services   run the orginal node.
>  
> but when I want to restart node it is ok
>  
> could you have any ideas?
> 
> *From:* Fabio M. Di Nitto <fdinitto at redhat.com>
> *To:* linux-cluster at redhat.com
> *Sent:* Tuesday, July 3, 2012 7:04 AM
> *Subject:* Re: [Linux-cluster] CLVM in a 3-node cluster
> 
> On 07/02/2012 11:39 PM, urgrue wrote:
>> On 2/7/12 19:14, Digimer wrote:
>>> On 07/02/2012 01:08 PM, urgrue wrote:
>>>> I'm trying to set up a 3-node cluster with clvm. Problem is, one node
>>>> can't access the storage, and I'm getting:
>>>> Error locking on node node3: Volume group for uuid not found: <snip>
>>>> whenever I try to activate the LVs on one of the working nodes.
>>>>
>>>> This can't be "by design", can it?
>>>
>>> Does pvscan show the right device? Are all nodes in the cluster? What
>>> does 'cman_tool status' and 'dlm_tool ls' show?
>>>
>>
>> Sorry, I realize now I was misleading, let me clarify:
>> The third node cannot access the storage, this is by design. I have
>> three datacenters but only two have access to the active storage. The
>> third datacenter only has an async copy, and will only activate
>> (manually) in case of a massive disaster (failure of both the other
>> datacenters).
>> So I deliberately have a failover domain with only node1 and node2.
>> node3's function is to provide quorum, but also be able to be activated
>> (manually is fine) in case of a massive disaster.
>> In other words node3 is part of the cluster, but it can't see the
>> storage during normal operation.
>> Looking at it another way, it's kind of as if we had a 3-node cluster
>> where one node had an HBA failure but is otherwise working. Surely node1
>> and node2 should be able to continue running the services?
>> So my question is, do I have an error somehwere, or is clvm really
>> actually not able to function without all nodes being active and able to
>> access storage?
> 
> CLVM requires a consistent view of the storage from all nodes in the
> cluster. This is by design.
> 
> A storage failure during operations (aka you start with all nodes able
> to access the storage and then downgrade) is handle correctly.
> 
> Fabio
> 
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com <mailto:Linux-cluster at redhat.com>
> https://www.redhat.com/mailman/listinfo/linux-cluster
> 
> 
> 
> 
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
> 


-- 
Digimer
Papers and Projects: https://alteeve.com





More information about the Linux-cluster mailing list