[Linux-cluster] How to test GFS-6.0.2.20-1 on RHELAS3.0 ?

Aydin SASMAZ sasmaz at itu.edu.tr
Mon Jul 4 15:38:05 UTC 2005


Hi
 
I would like to be sure which method to test GFS-6.0.2.20-1installed 3 node
file cluster doesn't corrupt data while multiple nodes writing on the same
file. At the same time how LOCK_GULM manage filesystem ? Is there any known
test? or benchmark other than bonnie++ . I would like to deploy 3
filesystem, each one on one gfs server and export these file systems with
smb to other smb clients. I plan to use 3 smb instance for all 3 filesystem
on every server node and load balancer in front of them.
 
Platform            : 3x HP-DL380 G4
OP                   : RedHat Enterprise Linux Advanced Server 3.0 Update 4
Cluster Suite     : 3.0, clumanager-1.2.22-2
                            redhat-config-cluster-1.0.3-1.noarch.rpm
GFS                 :    GFS-devel-6.0.2.20-1
                            GFS-modules-smp-6.0.2.20-1
                            GFS-6.0.2.20-1
Fencing Dev      : fence_ilo
 
 
Any advice would be appreciated. 
 
Some notes about system :
------------------------------------------
 
[root at gfs-test2 root]# gfs_tool df
/users/lnxsrv1:
  SB lock proto = "lock_gulm"
  SB lock table = "gfs-test:lnxsrv1"
  SB ondisk format = 1308
  SB multihost format = 1401
  Block size = 4096
  Journals = 8
  Resource Groups = 1596
  Mounted lock proto = "lock_gulm"
  Mounted lock table = "gfs-test:lnxsrv1"
  Mounted host data = ""
  Journal number = 1
  Lock module flags = async 
  Local flocks = FALSE
  Local caching = FALSE
 
  Type           Total          Used           Free           use%

  ------------------------------------------------------------------------
  inodes         8              8              0              100%
  metadata       92131          60723          31408          66%
  data           104492113      30341120       74150993       29%
 
/users/lnxsrv2:
  SB lock proto = "lock_gulm"
  SB lock table = "gfs-test:lnxsrv2"
  SB ondisk format = 1308
  SB multihost format = 1401
  Block size = 4096
  Journals = 8
  Resource Groups = 1596
  Mounted lock proto = "lock_gulm"
  Mounted lock table = "gfs-test:lnxsrv2"
  Mounted host data = ""
  Journal number = 1
  Lock module flags = async 
  Local flocks = FALSE
  Local caching = FALSE
 
  Type           Total          Used           Free           use%

  ------------------------------------------------------------------------
  inodes         5              5              0              100%
  metadata       38             38             0              100%
  data           104584209      0              104584209      0%
 
/users/lnxsrv3:
  SB lock proto = "lock_gulm"
  SB lock table = "gfs-test:lnxsrv3"
  SB ondisk format = 1308
  SB multihost format = 1401
  Block size = 4096
  Journals = 8
  Resource Groups = 396
  Mounted lock proto = "lock_gulm"
  Mounted lock table = "gfs-test:lnxsrv3"
  Mounted host data = ""
  Journal number = 1
  Lock module flags = async 
  Local flocks = FALSE
  Local caching = FALSE
 
  Type           Total          Used           Free           use%

  ------------------------------------------------------------------------
  inodes         5              5              0              100%
  metadata       10             10             0              100%
  data           25949437       0              25949437       0%

[root at gfs-test1 root]# gulm_tool nodeinfo gfs-test{3,1}.-----
 Name: gfs-test1.-----
  ip    = 160.75.100.22
  state = Logged in
  mode = Slave
  missed beats = 0
  last beat = 1120235791271378
  delay avg = 6672475
  max delay = 9459850
 
[root at gfs-test1 root]# gulm_tool nodeinfo gfs-test{3,2}.------
 Name: gfs-test2.------
  ip    = 160.75.100.23
  state = Logged in
  mode = Slave
  missed beats = 0
  last beat = 1120235802888768
  delay avg = 6678723
  max delay = 6880217



 
Aydin SASMAZ
System Support Engineer
ITU BIDB
Phone: +90 212 2853930
Fax    : +90 212 2856936

 
 
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/linux-cluster/attachments/20050704/79726010/attachment.htm>


More information about the Linux-cluster mailing list