[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]

[Linux-cluster] Restart lvm with out dismounting LUN/FS



Restart LVM with out dismounting LUN/FS. Here is what u have to do.

#killall clvmd
#/usr/sbin/clvmd 

-----Original Message-----
From: linux-cluster-bounces redhat com
[mailto:linux-cluster-bounces redhat com] On Behalf Of
linux-cluster-request redhat com
Sent: Thursday, January 15, 2009 12:00 PM
To: linux-cluster redhat com
Subject: Linux-cluster Digest, Vol 57, Issue 14

Send Linux-cluster mailing list submissions to
	linux-cluster redhat com

To subscribe or unsubscribe via the World Wide Web, visit
	https://www.redhat.com/mailman/listinfo/linux-cluster
or, via email, send a message with subject or body 'help' to
	linux-cluster-request redhat com

You can reach the person managing the list at
	linux-cluster-owner redhat com

When replying, please edit your Subject line so it is more specific
than "Re: Contents of Linux-cluster digest..."


Today's Topics:

   1. Create Logical Volume  (Cluster Management)
   2. Re: Re: Fencing test (Paras pradhan)
   3. Documentation enhancement requests/patches? (denis)
   4. Re: List Cluster Resources (denis)
   5. "Simple" Managed NFS setup (denis)
   6. cman-2.0.98-1.el5 / question about a problem when	launching
      cman (Alain.Moulle)
   7. Re: Documentation enhancement requests/patches? (Bob Peterson)
   8. Re: cman-2.0.98-1.el5 / question about a problem	when
      launching cman (Chrissie Caulfield)
   9. Re: [Openais] cman in RHEL 5 cluster suite and	Openais
      (Chrissie Caulfield)
  10. GFS/clvmd question (Gary Romo)


----------------------------------------------------------------------

Message: 1
Date: Wed, 14 Jan 2009 19:29:23 +0100
From: "Cluster Management" <cluster xinet it>
Subject: [Linux-cluster] Create Logical Volume 
To: <linux-cluster redhat com>
Message-ID: <008201c97676$05c36600$114a3200$ it>
Content-Type: text/plain; charset="us-ascii"

Hi all,

 

i have a two_node cluster RHEL 5 and an external ISCSI storage. I use
XEN
for virtualizzation purpose. When i create a new LUN in my storage i use
hot_add command to discover it from nodes.

The problem is that i have to restart clvmd to be able to create a new
Logical Volume. This operation is very critical because i have to stop
or
migrate each VM running on the node ed i have to umout their own LUN.

Is there a way to update clvmd without restarting?

 

Thanks a lot,

--

Francesco Gallo

XiNet S.r.L.

gallo (at) xinet (dot) it

 

 

-------------- next part --------------
An HTML attachment was scrubbed...
URL:
https://www.redhat.com/archives/linux-cluster/attachments/20090114/e827f
01f/attachment.html

------------------------------

Message: 2
Date: Wed, 14 Jan 2009 13:48:58 -0600
From: "Paras pradhan" <pradhanparas gmail com>
Subject: Re: [Linux-cluster] Re: Fencing test
To: "linux clustering" <linux-cluster redhat com>
Message-ID:
	<8b711df40901141148k740cf738ha9e43f0222b0a4ce mail gmail com>
Content-Type: text/plain; charset=ISO-8859-1

On Thu, Jan 8, 2009 at 10:57 PM, Rajagopal Swaminathan
<raju rajsand gmail com> wrote:
> Greetings,
>
> On Fri, Jan 9, 2009 at 12:09 AM, Paras pradhan
<pradhanparas gmail com> wrote:
>>
>>
>> In an act to solve my fencing issue in my 2 node cluster, i tried to
>> run fence_ipmi to check if fencing is working or not. I need to know
>> what is my problem
>>
>> -
>> [root ha1lx ~]# fence_ipmilan -a 10.42.21.28 -o off -l admin -p admin
>> Powering off machine @ IPMI:10.42.21.28...ipmilan: Failed to connect
>> after 30 seconds
>> Failed
>> [root ha1lx ~]#
>> ---------------
>>
>>
>> Here 10.42.21.28 is an IP address assigned to IPMI interface and I am
>> running this command in the same host.
>>
>
> Sorry couldn't respond earlier as I do this on personal time (which as
> useual limited for us IT guys and gals ;-) ) and not during work per
> se..
>
> Do not run fence script from the node that you want to fence.
>
> Let us say you want to fence node 3.
> 1. Try pinging the node 3's IPMI from node 4. I should be successful
> 2. Issue the fence command from Node 4 with IP of Node 3 IPMI as
argument .
>
>
> HTH
>
> With warm regards
>
> Rajagopal
>
> --
> Linux-cluster mailing list
> Linux-cluster redhat com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>

Yes as you said, I am able to power down node4 using node3, so it
seems ipmi is working fine. But I dunno what is going on with my two
node cluster. Can a red hat cluster operates fine in a two nodes mode?
Do i need qdisk or it is optional. Which area do i need to focus to
run my 2 nodes red hat cluster using ipmi as fencing device.

Thanks
Paras.



------------------------------

Message: 3
Date: Thu, 15 Jan 2009 10:49:11 +0100
From: denis <denisb+gmane gmail com>
Subject: [Linux-cluster] Documentation enhancement requests/patches?
To: linux-cluster redhat com
Message-ID: <gkn0qn$n57$1 ger gmane org>
Content-Type: text/plain; charset=ISO-8859-1

Hi,

After getting to know the "Configuring and Managing a Red Hat Cluster"
documentation [1] fairly well, I have a few enhancement suggestions.
What is the best way to submit these?

[1]
http://www.redhat.com/docs/en-US/Red_Hat_Enterprise_Linux/5.2/html/Clust
er_Administration/

Regards
--
Denis Braekhus



------------------------------

Message: 4
Date: Thu, 15 Jan 2009 10:52:00 +0100
From: denis <denisb+gmane gmail com>
Subject: [Linux-cluster] Re: List Cluster Resources
To: linux-cluster redhat com
Message-ID: <gkn100$n57$2 ger gmane org>
Content-Type: text/plain; charset=ISO-8859-1

Chaitanya Kulkarni wrote:
> Hi All,
> 
> I am new to the RHEL Clusters. Is there any way, (other than the
> cluster.conf file) using which we can view / list all the Cluster
> Resources that are used under the Cluster Service (Resource Group)?
Some
> command which might give some output as -
> 
> Service Name = Service1
> 
> Resources -
> IP Address = <Value>
> File System = <Value>
> Script = <Value>

Hi Chaitanya,

I recently discovered the rg_test tool, it might be of help to you. It
does currently not have a man page, but check the "Configuring and
Managing a Red Hat Cluster" chapter "Debugging and Testing Services and
Resource Ordering" [1] for usage.

Hope this is of some help to you.

[1]
http://www.redhat.com/docs/en-US/Red_Hat_Enterprise_Linux/5.2/html/Clust
er_Administration/s1-clust-rsc-testing-config-CA.html

Regards
-- 
Denis Braekhus



------------------------------

Message: 5
Date: Thu, 15 Jan 2009 12:00:03 +0100
From: denis <denisb+gmane gmail com>
Subject: [Linux-cluster] "Simple" Managed NFS setup
To: linux-cluster redhat com
Message-ID: <gkn4vj$5jp$1 ger gmane org>
Content-Type: text/plain; charset=ISO-8859-1

Hi,

I have begun a setup with a pretty simple 3-node cluster and a couple of
services. One of these is NFS, and I have setup the basics as laid out
in the included cluster.conf below.

A couple of questions :

1. Do I need to keep the nfs-state information on the NFS_homes volume
so as to keep it in sync between clusternodes?

2. The nfsclient name="nfs" is added to enable the current NFS serving
node to mount its own export, otherwise I got

Jan 15 11:41:03 node03 mountd[14229]: mount request from unknown host
XX.XX.XX.174 for /mnt/nfshome (/mnt/nfshome)

This is obviously caused by the mount connecting as the NFS service
address instead of the hostaddress, what is the best way to resolve
this? Mounting with the serviceaddress is not a good solution it seems,
as failing the service over is problematic when that address is in use
locally.

3. I read "The Red Hat Cluster Suite NFS Cookbook" [1], as the reference
Red Hat documentation was a bit thin regarding best practices. Is there
more documentation available to read?


Any tips/pointers/help highly appreciated.


<rm>
        <failoverdomains>
                <failoverdomain name="failover_nfshome" ordered="1"
restricted="1">
                        <failoverdomainnode name="node01.domain"
priority="30"/>
                        <failoverdomainnode name="node02.domain"
priority="30"/>
                        <failoverdomainnode name="node03.domain"
priority="10"/>
                </failoverdomain>
        </failoverdomains>
        <resources>
                <ip address="XX.XX.XX.174" monitor_link="1"/>
                <nfsexport name="NFShome"/>
                <fs device="/dev/mapper/NFS_homes" fsid="2"
force_fsck="1" force_unmount="1" fstype="ext3" mountpoint="/mnt/nfshome"
name="nfs_homes" self_fence="0"/>
                <nfsclient name="node01" options="rw"
target="node01.domain"/>
                <nfsclient name="node02" options="rw"
target="node02.domain"/>
                <nfsclient name="node03" options="rw"
target="node03.domain"/>
                <nfsclient name="nfs" options="rw" target="nfs.domain"/>
        </resources>
        <service autostart="0" domain="failover_nfshome" exclusive="0"
name="client_nfshome" recovery="restart">
                <ip ref="XX.XX.XX.174"/>
                <fs ref="nfs_homes">
                        <nfsexport name="nfshome">
                                <nfsclient ref="node01"/>
                                <nfsclient ref="node02"/>
                                <nfsclient ref="node03"/>
				<nfsclient ref="nfs"/>
                        </nfsexport>
                </fs>
        </service>
</rm>


[1] http://sources.redhat.com/cluster/doc/nfscookbook.pdf

Best Regards
-- 
Denis Braekhus



------------------------------

Message: 6
Date: Thu, 15 Jan 2009 13:48:42 +0100
From: "Alain.Moulle" <Alain Moulle bull net>
Subject: [Linux-cluster] cman-2.0.98-1.el5 / question about a problem
	when	launching cman
To: linux-cluster redhat com
Message-ID: <496F30AA 3000106 bull net>
Content-Type: text/plain; charset="iso-8859-1"

Hi ,
About this problem, I wonder if it is a definitive behavior considered 
as normal ?
or if this will work differently in a next release of cman or openais ?
(in previous versions with cman-2.0.73, we did not had this problem)
Thanks if someone could give an answer...
Regards,
Alain
> Release : cman-2.0.98-1.el5
> (but same problem with 2.0.95)
>
> I face a problem when launching cman on a two-node cluster :
>
> 1. Launching cman on node 1 : OK
> 2. When launching cman on node 2, the log on node1 gives :
>     cman killed by node 2 because we rejoined the cluster without a
full 
> restart
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
https://www.redhat.com/archives/linux-cluster/attachments/20090115/dd62c
325/attachment.html

------------------------------

Message: 7
Date: Thu, 15 Jan 2009 08:59:56 -0500 (EST)
From: Bob Peterson <rpeterso redhat com>
Subject: Re: [Linux-cluster] Documentation enhancement
	requests/patches?
To: linux clustering <linux-cluster redhat com>
Message-ID:
	
<40063497 1429461232027996695 JavaMail root zmail02 collab prod int phx2
.redhat.com>
	
Content-Type: text/plain; charset=utf-8

----- "denis" <denisb+gmane gmail com> wrote:
| Hi,
| 
| After getting to know the "Configuring and Managing a Red Hat
| Cluster"
| documentation [1] fairly well, I have a few enhancement suggestions.
| What is the best way to submit these?
| 
| [1]
|
http://www.redhat.com/docs/en-US/Red_Hat_Enterprise_Linux/5.2/html/Clust
er_Administration/
| 
| Regards
| --
| Denis Braekhus

Hi Denis,

Probably the best way to do this is to open a new bugzilla record
against product Red Hat Enterprise Linux 5, component
"Documentation--cluster"

If you have permission to look at it, you can follow this example:
https://bugzilla.redhat.com/show_bug.cgi?id=471364

You can assign it to slevine redhat com or pkennedy redhat com 

Regards,

Bob Peterson
Red Hat GFS



------------------------------

Message: 8
Date: Thu, 15 Jan 2009 15:01:17 +0000
From: Chrissie Caulfield <ccaulfie redhat com>
Subject: Re: [Linux-cluster] cman-2.0.98-1.el5 / question about a
	problem	when	launching cman
To: linux clustering <linux-cluster redhat com>
Message-ID: <496F4FBD 4040607 redhat com>
Content-Type: text/plain; charset=ISO-8859-1

Alain.Moulle wrote:
> Hi ,
> About this problem, I wonder if it is a definitive behavior considered
> as normal ?
> or if this will work differently in a next release of cman or openais
?
> (in previous versions with cman-2.0.73, we did not had this problem)
> Thanks if someone could give an answer...
> Regards,
> Alain
>> Release : cman-2.0.98-1.el5
>> (but same problem with 2.0.95)
>>
>> I face a problem when launching cman on a two-node cluster :
>>
>> 1. Launching cman on node 1 : OK
>> 2. When launching cman on node 2, the log on node1 gives :
>>     cman killed by node 2 because we rejoined the cluster without a
full 
>> restart
> 

Alain,

I'm sure this question has been answer many times on IRC and on the
mailing list, as well as in the FAQ.


Chrissie



------------------------------

Message: 9
Date: Thu, 15 Jan 2009 15:05:33 +0000
From: Chrissie Caulfield <ccaulfie redhat com>
Subject: [Linux-cluster] Re: [Openais] cman in RHEL 5 cluster suite
	and	Openais
To: unleashing_vivek007 yahoo co in
Cc: linux clustering <linux-cluster redhat com>
Message-ID: <496F50BD 4020409 redhat com>
Content-Type: text/plain; charset=UTF-8

Vivek Purohit wrote:
> Hi Steve,
> Thanks for the previous reply.
> 
> I was able to run the checkpointing tests in the tarball Openais
> on RHEL 5.
> 
> I explored and came to know that the CMAN service of RHEL 5's
> clustersuite runs as aisexec; thus the tests could be run directly.
> 
> Can you please tell how the Openais is being used by RHEL 5's
> CMAN service.
> 

Hi,

You might like to read these two documents:

http://people.redhat.com/ccaulfie/docs/aiscman.pdf
http://people.redhat.com/ccaulfie/docs/CSNetworking.pdf


-- 

Chrissie



------------------------------

Message: 10
Date: Thu, 15 Jan 2009 09:08:08 -0700
From: Gary Romo <garromo us ibm com>
Subject: [Linux-cluster] GFS/clvmd question
To: linux-cluster redhat com
Message-ID:
	
<OF48AE2273 1EEE35B0-ON8725753F 0057EC20-8725753F 0058A29F us ibm com>
Content-Type: text/plain; charset="us-ascii"


Why can't I mount my gfs logical volume on the second node in  the
cluster?
I am creating a new GFS file system on an existing cluster.  Here is
what I
did;

1.  I determined I had space in an existing volume group (both nodes)
2.  I created my logical volume (node 1)
3.  I ran my gfs_mkfs (node 1)
4.  I mounted my new lv on node 1 only

Here is the error I get on node 2

# mount /gfs/new_mount
/sbin/mount.gfs: invalid device path "/dev/vggfs/new_lv"

I see that the logical volume is "inactive" on node2 and "ACTIVE" on
node 1

inactive          '/dev/vgclgfs/new_lv' [25.00 GB] inherit

ACTIVE            '/dev/vgclgfs/new_lv' [25.00 GB] inherit

What do I need to do in order to make this logical volume active on node
2 ?
I thought that this would have happened automagically via clvmd, and not
have to be done manually.


Gary Romo
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
https://www.redhat.com/archives/linux-cluster/attachments/20090115/1be5b
3b7/attachment.html

------------------------------

--
Linux-cluster mailing list
Linux-cluster redhat com
https://www.redhat.com/mailman/listinfo/linux-cluster

End of Linux-cluster Digest, Vol 57, Issue 14
*********************************************


[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]