[Linux-cluster] CLVMD hangs on 2nd node startup and hangs all gfs nodes.

Jonathan Brassow jbrassow at redhat.com
Mon Apr 28 13:55:32 UTC 2008


I just got a blank page for you cluster.conf.  It must be /really/  
slimmed down. ;)

  brassow

On Apr 28, 2008, at 7:15 AM, Tracey Flanders wrote:

>
>
> I tested creating  a GFS disk with 2 nodes started in the cluster  
> without using LVMs and CLVMD stop.
> I mounted the disk on the first node but when I mounted the 2nd node  
> it did the same thing. So it seems its something
> other than CLVMD. I've attached my cluster.conf. It's kind of dumbed  
> down because I was troubleshooting. So I removed the GFS mount for a  
> services, etc.
> This the config I used.
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>> If you suspect a problem with clvmd, you could simply remove it from
>> the equation and retest, right?
>>
>> You could just use the underlying iSCSI device and mkfs.gfs on
>> that.... at least for testing if clvmd is the problem.
>>
>> I suppose you could also test if clvmd is the problem by testing the
>> logical volumes without GFS in the mix.  IOW, create some LVs and  
>> read/
>> write to them at the same time from different machines.  If this is
>> working, the file system should work.  If the file system doesn't,
>> then the problem is probably higher up than clvmd.
>>
>>  brassow
>>
>> On Apr 25, 2008, at 11:34 AM, Tracey Flanders wrote:
>>
>>> I've been trying to setup a 3 server cluster with GFS mounted over
>>> iSCSI on Qemu Virtual Machines. A 4th server acts as a iSCSI Target.
>>> I found and article that explains my issue, but I can't seem to
>>> figure out what the solution is. QUOTED from :http://kbase.redhat.com/faq/FAQ_51_10923.shtm
>>> After successfully setting up a cluster, cman_tool shows the
>>> cluster is healthy. Mounting the gfs mount on the first node works
>>> successfully. However, when mounting gfs on the second node, the
>>> mount command hangs. Writing to a file on the first node also hangs.
>>> On the second node, the following error is seen in /var/log/
>>> messages: Jul 18 14:49:27 blade3 kernel: Lock_Harness 2.6.9-72.2
>>> (built Apr 24 2007 12:45:55) installed Jul 18 14:49:27 blade3
>>> kernel: GFS 2.6.9-72.2 (built Apr 24 2007 12:46:12) installed Jul 18
>>> 14:52:53 blade3 kernel: GFS: Trying to join cluster "lock_dlm",
>>> "vcomcluster:testgfs" Jul 18 14:52:53 blade3 kernel: Lock_DLM (built
>>> Apr 24 2007 12:45:57) installed Jul 18 14:52:53 blade3 kernel: dlm:
>>> connect from non cluster node Jul 18 14:52:53 blade3 kernel: dlm:
>>> connect from non cluster node END QUOTE My Virtual Machines only
>>> have one interface so I still can't figure out why this is
>>> happening. I can successfully mount the GFS partition on any one
>>> node but as soon as I try to start the clvmd on a 2nd node it hangs
>>> the whole cluster. I'm wondering if its a Qemu VM network issue?
>>> Each host can ping each other by name and ip. The cluster works fine
>>> but I cant get GFS to work on th VMs. Is it possible to debug the
>>> clvmd to see what IP Address it is sending? Thanks, Tracey Flanders
>>> In a rush? Get real-time answers with Windows Live Messenger. --
>>> Linux-cluster mailing list
>>> Linux-cluster at redhat.com
>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>
> _________________________________________________________________
> Spell a grand slam in this game where word skill meets World Series.  
> Get in the game.
> http://club.live.com/word_slugger.aspx?icid=word_slugger_wlhm_admod_april08
>
> --
> Linux-cluster mailing list
> Linux-cluster at redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster




More information about the Linux-cluster mailing list