[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]



 Sorry i dont have time time thats why i mailed it in urgent if you
get strucked up anywhere mail me, sorry if you  find  any words are

 For GFS to work u need to install all cluster related rpms and
configure the simple running cluster with following things configured.

Cluster rpms and Deps

 rpm -ivh ccs-1.0.10-0.i686.rpm cluster-cim-0.9.1-8.i386.rpm
cluster-snmp-0.9.1-8.i386.rpm cman-1.0.17-0.i686.rpm
cman-kernel-2.6.9-50.2.i686.rpm cman-kernel-smp-2.6.9-50.2.i686.rpm
cman-kernheaders-2.6.9-50.2.i686.rpm dlm-1.0.3-1.i686.rpm
dlm-kernel-2.6.9-46.16.i686.rpm dlm-kernel-smp-2.6.9-46.16.i686.rpm
dlm-kernheaders-2.6.9-46.16.i686.rpm fence-1.32.45-1.i686.rpm
iddev-2.0.0-4.i686.rpm ipvsadm-1.24-6.i386.rpm luci-0.9.1-8.i386.rpm
magma-1.0.7-1.i686.rpm magma-devel-1.0.7-1.i686.rpm
magma-plugins-1.0.12-0.i386.rpm modcluster-0.9.1-8.i386.rpm
perl-Net-Telnet-3.03-3.noarch.rpm rgmanager-1.9.68-1.i386.rpm
system-config-cluster-1.0.45-1.0.noarch.rpm gulm-1.0.10-0.i686.rpm


rpm -ivh cmirror-1.0.1-1.i386.rpm cmirror-kernel-2.6.9-32.0.i686.rpm
cmirror-kernel-smp-2.6.9-32.0.i686.rpm GFS-6.1.14-0.i386.rpm
GFS-kernel-2.6.9-72.2.i686.rpm GFS-kernel-smp-2.6.9-72.2.i686.rpm
warning: cmirror-1.0.1-1.i386.rpm:

modprobe -v gfs


1) cluster name --- apps_cluster

2)  clusternode name  --- node1
    clusternode name  ---node2

3) fencedevices

  fencedevice agent="fence_manual" name="test"

  4)  failoverdomains

 failoverdomain name="apps" ordered="0" restricted="0"

failoverdomainnode name="node1" priority="1"
failoverdomainnode name="node2" priority="1"

 Then start the cluster , once it is started up and running without error.

Then for GFS LVM is required. execute the below commands from  node1
or node 2 provided  storage lun's should be presented to both the

1) pvcreate  /dev/sdd1
2) pvdisplay
3) vgcreate testapps /dev/sdd1
4) vgdisplay
5) lvcreate  -L 135G -n data testapps

 To format file system with GFS you need below details.

6) cman_tool  status

Cluster name: apps_cluster

apps_cluster is the name of the cluster and u can get it using the above command

:data is the logical volume name used during above lvcreate command

gfs_mkfs -p lock_dlm -t apps_cluster:data -j 7 /dev/testapps/data

-p LockProtoName

-t LockTableName
    The lock table field appropriate to the lock module you're using.
It is clustername:fsname. Clustername must match that in cluster.conf;


Specifies the number of journals to be created by the gfs_mkfs
command. One journal is required for each node that mounts the file
system. (More journals than are needed can be specified at creation
time to allow for future expansion.)

After that u can mount it to ur desired mount point in our case we
created /data-new using mkdir

mount -t gfs /dev/testapps/data  /data-new/

2008/10/8 Ryan Golhar <golharam umdnj edu>:
> Has anyone successfully setup GFS?  I have SAN connected to several
> computers by fibre, and it appears that GFS is the way to go as opposed to
> use an NFS server.
> Do I really need to set up all the other aspects of a Redhat cluster to get
> GFS to work?  There doesn't seem to be a good HOW-TO of this anywhere, and
> the RedHat docs are not as helpful as I would have liked.
> --
> redhat-list mailing list
> unsubscribe mailto:redhat-list-request redhat com?subject=unsubscribe
> https://www.redhat.com/mailman/listinfo/redhat-list

[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]