[Linux-cluster] scsi reservation issue

Christopher Barry Christopher.Barry at qlogic.com
Wed Oct 31 19:16:59 UTC 2007


On Wed, 2007-10-31 at 11:54 -0600, Josh Gray wrote:
> Chris - How many other VM's do you have too?   Are the RH servers local disk
> or shared storage,  what about all the others?
> 
> Josh
> 
> 
> On 10/31/07 9:27 AM, "Christopher Barry" <Christopher.Barry at qlogic.com>
> wrote:
> 
> > Greetings all,
> > 
> > I have 2 vmware esx servers, each hitting a NetApp over FS, and each
> > with 3 RHCS cluster nodes trying to mount a gfs volume.
> > 
> > All of the nodes (1,2,& 3) on esx-01 can mount the volume fine, but none
> > of the nodes in the second esx box can mount the gfs volume at all, and
> > I get the following error in dmesg:
> > 
> > Lock_Harness 2.6.9-72.2 (built Apr 24 2007 12:45:38) installed
> > GFS 2.6.9-72.2 (built Apr 24 2007 12:45:54) installed
> > GFS: Trying to join cluster "lock_dlm", "kop-sds:gfs_home"
> > Lock_DLM (built Apr 24 2007 12:45:40) installed
> > GFS: fsid=kop-sds:gfs_home.2: Joined cluster. Now mounting FS...
> > GFS: fsid=kop-sds:gfs_home.2: jid=2: Trying to acquire journal lock...
> > GFS: fsid=kop-sds:gfs_home.2: jid=2: Looking at journal...
> > GFS: fsid=kop-sds:gfs_home.2: jid=2: Done
> > scsi2 (0,0,0) : reservation conflict
> > SCSI error : <2 0 0 0> return code = 0x18
> > end_request: I/O error, dev sdc, sector 523720263
> > scsi2 (0,0,0) : reservation conflict
> > SCSI error : <2 0 0 0> return code = 0x18
> > end_request: I/O error, dev sdc, sector 523720271
> > scsi2 (0,0,0) : reservation conflict
> > SCSI error : <2 0 0 0> return code = 0x18
> > end_request: I/O error, dev sdc, sector 523720279
> > GFS: fsid=kop-sds:gfs_home.2: fatal: I/O error
> > GFS: fsid=kop-sds:gfs_home.2:   block = 65464979
> > GFS: fsid=kop-sds:gfs_home.2:   function = gfs_logbh_wait
> > GFS: fsid=kop-sds:gfs_home.2:   file
> > = /builddir/build/BUILD/gfs-kernel-2.6.9-72/smp/src/gfs/dio.c, line =
> > 923
> > GFS: fsid=kop-sds:gfs_home.2:   time = 1193838678
> > GFS: fsid=kop-sds:gfs_home.2: about to withdraw from the cluster
> > GFS: fsid=kop-sds:gfs_home.2: waiting for outstanding I/O
> > GFS: fsid=kop-sds:gfs_home.2: telling LM to withdraw
> > lock_dlm: withdraw abandoned memory
> > GFS: fsid=kop-sds:gfs_home.2: withdrawn
> > GFS: fsid=kop-sds:gfs_home.2: can't get resource index inode: -5
> > 
> > 
> > Does anyone have a clue as to where I should start looking?
> > 
> > 
> > Thanks,
> > -C
> > 
> > --
> > Linux-cluster mailing list
> > Linux-cluster at redhat.com
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> > 
> 


Hi Josh,

My cluster nodes are on the local mirrored disks, I have several other
VMs that do not participate in the cluster, and they are located in a
VMFS on the shared storage.

Each cluster node has three disks:
sda* : the operating sys and it is local to the host
sdb1 : a 20MB quorum disk on netapp (which mounts fine from all nodes)
sdc1 : a 600GB gfs volume on netapp (only mounts on 3 nodes from first
host)

both the quorum and the gfs use RDMs (one each per host)
the rdms were created on nodes 1 and 4, and then nodes 2 & 3 mount the
existing rdms from node 1, and 5 & 6 mount the existing rdms from node
4.

Each disk has it's own virtual scsi controller.
sda has scsi controller set to none.
sdb and sdc have scsi controller set to physical.


The error I get when attempting to mount from nodes 4,5, & 6 is:

mount: /dev/sdc1: can't read superblock



Any help at all would be great.


Thanks,
-C




More information about the Linux-cluster mailing list