[Linux-cluster] Storage Cluster Newbie Questions - any help with answers greatly appreciated!

Michael @ Professional Edge LLC m3 at professionaledgellc.com
Mon Mar 15 22:04:57 UTC 2010


Ok... I don't feel quite as silly... as I had roasted one of my test 
machines (unable to mount /) - I had to do a full rebuild from scratch.

Well... It seems I had 2 of 3.

For anyone who ever needs to disable their Qlogic fiber card from 
auto-starting in the future - do all of the following 3 changes.

1. Modify /etc/modprobe.conf - and comment out the line "#alias 
scsi_hostadapter2 qla2xxx"
2. Rebuild the init - I use this - "mkinitrd -f /boot/initrd-$(uname 
-r).img $(uname -r)"
3. Add "blacklist qla2xxx" - to /etc/modprobe.d/blacklist

-Michael

Michael @ Professional Edge LLC wrote, On 3/15/2010 11:05 AM:
> DOH... ok I feel stupid... :-)
>
> Well 'blacklist scsi_transport_fc' didn't work... but - 'blacklist 
> qla2xxx' works fine.
>
> Amazing how all the docs I found said to use - remove qla2xxx modprobe 
> -r - kind of stuff... and all that I really had to do was add the 
> silly thing to the blacklist... :-)
>
> OK... at least I can proceed with my testing again.  Thank You!
>
> -Michael
>
> Kaloyan Kovachev wrote, On 3/15/2010 1:41 AM:
>> Hello,
>>
>> On Sun, 14 Mar 2010 22:28:20 -0700, Michael @ Professional Edge LLC 
>> wrote
>>> Kaloyan,
>>>
>>> I agree - disabling the qla2xxx driver (Qlogic HBA) from starting at
>>> boot would be the simple method of handling the issue.  Then I just put
>>> all the commands to load the driver, multipath, mdadm, etc... inside
>>> cluster scripts.
>>>
>>> Amusingly it seems I am missing something very basic - as I can't seem
>>> to figure out how to not load the qla2xxx driver.
>>>
>>> Do you happen to know the syntax to make the qla2xxx driver not load at
>>> boot automatically?
>>>
>>> I've been messing with /etc/modprobe.conf - and mkinird - but no
>>> combination has resulted in the - qla2xxx being properly disabled 
>>> during
>>> boot - I did accomplish making one of my nodes unable to mount it's 
>>> root
>>> partition - but I don't consider that success. :-)
>>>
>> Is 'blackist scsi_transport_fc' not enough? What other modules are 
>> loaded? If
>> you blacklist the one, that most others depend on they should not load.
>>
>>> As for your 2nd idea; I have seen folks doing something similar in that
>>> mode; when the disks are local to the node.  But in my case - all nodes
>>> - can already see all LUNs - so I dont really have any need to do an
>>> iSCSI export - appreciate the thought though.
>>>
>> The idea was actually not to export them, but to run mdamd 
>> simultaneously on
>> both nodes. But the problem is when just one of the nodes looses its 
>> link, to
>> just one of the arrays
>>
>>> -Michael
>>>
>>> Kaloyan Kovachev wrote, On 3/4/2010 10:28 AM:
>>>> On Thu, 04 Mar 2010 09:26:35 -0800, Michael @ Professional Edge LLC 
>>>> wrote
>>>>
>>>>> Hello Kaloyan,
>>>>>
>>>>> Thank you for the thoughts.
>>>>>
>>>>> You are correct when I said - "Active / Passive" - I simply meant 
>>>>> that I
>>>>> had no need for "Active / Active" - and floating IP on the NFS share
>>>>> would be exactly what I had in mind.
>>>>>
>>>>> The software raid - of any type, raid1,5,6 etc... is the issue.  From
>>>>> what I have read - mdadm - is not cluster aware... and... since all
>>>>> disks are seen by all RHEL nodes. - As Leo mentioned; some method to
>>>>> disable the kernel from finding detecting and attempting to 
>>>>> assemble all
>>>>> the available software raids - is a major problem.  This is why I was
>>>>> asking if perhaps - CLVM w/mirroring would be a better method.  
>>>>> Although
>>>>> since it was just introduced in RHEL 5.3 - I am a bit leery.
>>>>>
>>>>>
>>>> I am not common with FC, so maybe completely wrong here, but if you 
>>>> do not
>>>> start multipath and load your HBA drivers on boot, how the FC disks 
>>>> based
>>>> software raid will start at all?
>>>>
>>>> even if started you may still issue 'mdadm --stop /dev/mdX' in S00 as
>>>> suggested from Leo and assemble it again as a cluster service later
>>>>
>>>>
>>>>> Sorry for being confusing - yes - the linux machines will have a
>>>>> completely different filesystem share; than the windows machines.  My
>>>>> original thought was I would do "node#1 primary nfs share (floating
>>>>> ip#1) to linux machines w/node#2 backup" - and then "node#2 
>>>>> primary nfs
>>>>> or samba share (floating ip#2) to windows machines w/node#1 backup".
>>>>>
>>>>> Any more thoughts you have would be appreciated... as my original 
>>>>> plan
>>>>> with MDADM w/HA-LVM - so far doesn't seem very possible.
>>>>>
>>>>>
>>>> Then there are two services each with its own raid array and ip, 
>>>> but basically
>>>> the same
>>>>
>>>> another idea ... not using it in production, but i had good results 
>>>> (testing)
>>>> with (small) software raid5 array from 3 nodes ... Local device on 
>>>> each node
>>>> exported via iSCSI and software RAID5 over the imported ones which 
>>>> is then
>>>> used from LVM. Weird, but worked and the only problem was that on 
>>>> every reboot
>>>> of any node the raid is rebuilt, which i won't happen in your case 
>>>> as you will
>>>> see all the disks in sync (after the initial sync done on only one 
>>>> of them)
>>>> ... you may give it a try
>>>>
>>>>
>>>>> -Michael
>>>>>
>>>>> Kaloyan Kovachev wrote, On 3/4/2010 8:52 AM:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> On Wed, 03 Mar 2010 11:16:07 -0800, Michael @ Professional Edge 
>>>>>> LLC wrote
>>>>>>
>>>>>>
>>>>>>> Hail Linux Cluster gurus,
>>>>>>>
>>>>>>> I have researched myself into a corner and am looking for 
>>>>>>> advice.  I've
>>>>>>> never been a "clustered storage guy", so I apologize for the 
>>>>>>> potentially
>>>>>>> naive set of questions.  ( I am savvy on most other aspects of 
>>>>>>> networks,
>>>>>>> hardware, OS's etc... but not storage systems).
>>>>>>>
>>>>>>> I've been handed ( 2 ) x86-64 boxes w/2 local disks each; and ( 2 )
>>>>>>> FC-AL disk shelves w/14 disks each; and told to make a mini 
>>>>>>> NAS/SAN (NFS
>>>>>>> required, GFS optional).  If I can get this working reliably 
>>>>>>> then there
>>>>>>> appear to be about another ( 10 ) FC-AL shelves and a couple of 
>>>>>>> Fiber
>>>>>>> Switches laying around that will be handed to me.
>>>>>>>
>>>>>>> NFS filesystems will be mounted by several (less than 6) linux 
>>>>>>> machines,
>>>>>>> and a few (less than 4) windows machines [[ microsoft nfs client 
>>>>>>> ]] -
>>>>>>> all more or less doing web server type activities (so lots of 
>>>>>>> reads from
>>>>>>> a shared filesystem - log files not on NFS so no issue with high IO
>>>>>>> writes).  I'm locked into NFS v3 for various reasons.  
>>>>>>> Optionally the
>>>>>>> linux machines can be clustered and GFS'd instead - but I would 
>>>>>>> still
>>>>>>> need to come up with a solution for the windows machines - so a NAS
>>>>>>> solution is still required even if I do GFS to the linux boxes.
>>>>>>>
>>>>>>> Active / Passive on the NFS is fine.
>>>>>>>
>>>>>>>
>>>>>> Why not start NFS/Samba on both machines with only the IP 
>>>>>> floating between
>>>>>> them then?
>>>>>>
>>>>>>
>>>>>>
>>>>>>> * Each of the ( 2 ) x86-64 machines have a Qlogic dual HBA 1 fiber
>>>>>>> direct connected to each shelf  (no fiber switches yet - but 
>>>>>>> will have
>>>>>>> them later if I can make this all work); I've loaded RHEL 5.4 
>>>>>>> x86-64.
>>>>>>>
>>>>>>> * Each of the ( 2 ) RHEL 5.4 boxes - used the 2 local disks 
>>>>>>> w/onboard
>>>>>>> fake raid1 = /dev/sda - basic install so /boot and LVM for the 
>>>>>>> rest -
>>>>>>> nothing special here (didn't do mdadm basically for simplicity 
>>>>>>> of /dev/sda)
>>>>>>>
>>>>>>> * Each of the ( 2 ) RHEL 5.4 boxes can see all the disks on both 
>>>>>>> shelves
>>>>>>> - and since I don't have Fiber Switches yet - at the moment 
>>>>>>> there is
>>>>>>> only 1 path to each disk; however as I assume I will figure out 
>>>>>>> a method
>>>>>>> to make this work - I have enabled multipath - and therefore I have
>>>>>>> consistent names to 28 disks.
>>>>>>>
>>>>>>> Here's my dilemma.  How do I best add Redundancy to the Disks, 
>>>>>>> removing
>>>>>>> as many single points of failure, and preserving as much 
>>>>>>> diskspace as
>>>>>>> possible?
>>>>>>>
>>>>>>> My initial thought was - to take "shelf1:disk1 and shelf2:disk1" 
>>>>>>> and put
>>>>>>> them into a software raid1 - mdadm; then put the resulting 
>>>>>>> /dev/md0 into
>>>>>>> a LVM.  When I need more diskspace, I just then create 
>>>>>>> "shelf1:disk2 and
>>>>>>> shelf2:disk2" as another software raid1 then just add the new 
>>>>>>> "/dev/md1"
>>>>>>> into the LVM and expand the FS. This handles a couple things in 
>>>>>>> my mind:
>>>>>>>
>>>>>>> 1. Each shelf is really a FC-AL so it's possible that a single disk
>>>>>>> going nuts could flood the FC-AL and all the disks in that shelf 
>>>>>>> go poof
>>>>>>> until the controller can figure itself out and/or the bad disk 
>>>>>>> is removed.
>>>>>>>
>>>>>>> 2. Efficient I am retaining 50% storage capacity after 
>>>>>>> redundancy - if I
>>>>>>> can do the "shelf1:disk1 + shelf2:disk2" mirrors; plus all 
>>>>>>> bandwidth
>>>>>>> used is spread across the 2 HBA fibers and nothing goes over the 
>>>>>>> TCP
>>>>>>> network.  Conversely DRBD doesn't excite me much - as I then 
>>>>>>> have to do
>>>>>>> both raid in the shelf (probably still with MDADM) and then I 
>>>>>>> add TCP
>>>>>>> (ethernet) based RAID1 between the nodes - and when all is said 
>>>>>>> and done
>>>>>>> - I only the have 25% of storage capacity still available after 
>>>>>>> redundancy.
>>>>>>>
>>>>>>> 3. I easy to add more diskspace - as each new mirror (software 
>>>>>>> raid1)
>>>>>>> can just be added to an existing LVM.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>> You may create RAID1 (between the two shelfs) over RAID6 (on the 
>>>>>> disks from
>>>>>> the same shelf), so you will loose only 2 more disks per shelf or 
>>>>>> about 40%
>>>>>> storage space left, but more stable and faster. Or several RAID6 
>>>>>> arrays with
>>>>>> 2+2 disks from each shelf - again 50% storage space, but better 
>>>>>> performance
>>>>>> with the same chance for data loss like with several RAID1 ... 
>>>>>> the resulting
>>>>>> mdX you may add to LVM and use the logical volumes
>>>>>>
>>>>>>
>>>>>>
>>>>>>>        From what I can find messing with Luci (Conga) though... 
>>>>>>> is - I don't
>>>>>>> see any resource scripts listed for - "mdadm" (on RHEL 5.4) - so 
>>>>>>> would
>>>>>>> my idea even work  (I have found some posts asking for a mdadm 
>>>>>>> resource
>>>>>>> script but I've seen no response)?  I also see with RHEL 5.3 LVM 
>>>>>>> has
>>>>>>> mirrors that can be clustered now - is this the right answer?  
>>>>>>> I've done
>>>>>>> a ton of reading but everything I've dug up so far; assumes that 
>>>>>>> the
>>>>>>> fiber devices are being presented by a SAN that is doing the 
>>>>>>> redundancy
>>>>>>> before the RHEL box sees the disk... or... there are a ton of 
>>>>>>> examples
>>>>>>> of where fiber is not in the picture and there are a bunch of 
>>>>>>> locally
>>>>>>> attached hosts presenting storage onto the TCP (ethernet) - but 
>>>>>>> I've not
>>>>>>> found nearly anything on my situation...
>>>>>>>
>>>>>>> So... here I am... :-)  I really just have 2 nodes - who can 
>>>>>>> both see -
>>>>>>> a bunch of disks (JBOD) and I want to present them to multiple 
>>>>>>> hosts via
>>>>>>> NFS (required) or GFS (to linux boxes only).
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>> if the Windows and Linux data are different volumes it is better to
>> leave the
>>>>>> GFS partition(s) available only via iSCSI to the linux nodes
>> participating in
>>>>>> the cluster and not to mount it/them locally for the NFS/Samba 
>>>>>> shares,
>> but if
>>>>>> the data should be the same you may go even Active/Active with 
>>>>>> GFS over
>> iSCSI
>>>>>> [over CLVM and/or] [over DRBD] over RAID and use NFS/Samba over 
>>>>>> GFS as a
>>>>>> service in the cluster. It all depends on how the data will be 
>>>>>> used from the
>>>>>> storage
>>>>>>
>>>>>>
>>>>>>
>>>>>>> All ideas - are greatly appreciated!
>>>>>>>
>>>>>>> -Michael
>>>>>>>
>>>>>>> -- 
>>>>>>> Linux-cluster mailing list
>>>>>>> Linux-cluster at redhat.com
>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>
>>>>>>>
>>>>>>
>
> -- 
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster





More information about the Linux-cluster mailing list