[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]

Re: [Linux-cluster] Linux-cluster Digest, Vol 71, Issue 37



Hi all,

A> How many partitions can be mount using cluster server RHEL 5.4 (GFS2)
B> And how many VIP can use with the same
Regards,

Rajat J Patel

FIRST THEY IGNORE YOU...
THEN THEY LAUGH AT YOU...
THEN THEY FIGHT YOU...
THEN YOU WIN...


On Mon, Mar 22, 2010 at 9:30 PM, <linux-cluster-request redhat com> wrote:
Send Linux-cluster mailing list submissions to
       linux-cluster redhat com

To subscribe or unsubscribe via the World Wide Web, visit
       https://www.redhat.com/mailman/listinfo/linux-cluster
or, via email, send a message with subject or body 'help' to
       linux-cluster-request redhat com

You can reach the person managing the list at
       linux-cluster-owner redhat com

When replying, please edit your Subject line so it is more specific
than "Re: Contents of Linux-cluster digest..."


Today's Topics:

  1. Re: RHCS: How to display with command line HA     resources
     attached to service (Paul Morgan)
  2. Re: RHCS: How to display with command line        HA      resources
     attached to service (Moralejo, Alfredo)
  3. Re: falure during gfs2_grow caused node crash & data loss
     (Bob Peterson)
  4. Re: GFS create file performance (Jeff Sturm)


----------------------------------------------------------------------

Message: 1
Date: Sun, 21 Mar 2010 16:39:26 -0400
From: Paul Morgan <jumanjiman gmail com>
To: linux clustering <linux-cluster redhat com>
Subject: Re: [Linux-cluster] RHCS: How to display with command line HA
       resources attached to service
Message-ID:
       <1b6fc7dd1003211339s4e0e901ax9fc1a539ca8af646 mail gmail com>
Content-Type: text/plain; charset="utf-8"

Try cman_tool services

On Mar 21, 2010 11:32 AM, "Hoang, Alain" <Alain Hoang hp com> wrote:

 Hello,

With the command, clustat ?l ?s <Service>, I could get some information, but
I can not display:

  - List of HA resources attached
  - Failover Domain with its members


Is there any other command that could give me the information?

Best Regards,
Ki?n L?m Alain Hoang,
(???  ??? ,  ??  ?)
Technical Consultant, Factory consulting and training
HP Software and Solutions
Communications & Media Solutions
NGOSS Practice Delivery Management




--
Linux-cluster mailing list
Linux-cluster redhat com
https://www.redhat.com/mailman/listinfo/linux-cluster
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://www.redhat.com/archives/linux-cluster/attachments/20100321/6a566c57/attachment.html>

------------------------------

Message: 2
Date: Mon, 22 Mar 2010 10:48:38 +0100
From: "Moralejo, Alfredo" <alfredo moralejo roche com>
To: linux clustering <linux-cluster redhat com>
Subject: Re: [Linux-cluster] RHCS: How to display with command line     HA
       resources attached to service
Message-ID:
       <C64734E4E1C80E49955AD539DB2FBC3A69330DBF rkamsem703 emea roche com>
Content-Type: text/plain; charset="utf-8"

You can try rg_test command too. This is the one that provides more information

________________________________
From: linux-cluster-bounces redhat com [mailto:linux-cluster-bounces redhat com] On Behalf Of Paul Morgan
Sent: Sunday, March 21, 2010 9:39 PM
To: linux clustering
Subject: Re: [Linux-cluster] RHCS: How to display with command line HA resources attached to service


Try cman_tool services
On Mar 21, 2010 11:32 AM, "Hoang, Alain" <Alain Hoang hp com<mailto:Alain Hoang hp com>> wrote:
Hello,

With the command, clustat ?l ?s <Service>, I could get some information, but I can not display:

 *   List of HA resources attached
 *   Failover Domain with its members

Is there any other command that could give me the information?

Best Regards,
Ki?n L?m Alain Hoang,
(???  ??? ,  ??  ?)
Technical Consultant, Factory consulting and training
HP Software and Solutions
Communications & Media Solutions
NGOSS Practice Delivery Management




--
Linux-cluster mailing list
Linux-cluster redhat com<mailto:Linux-cluster redhat com>
https://www.redhat.com/mailman/listinfo/linux-cluster
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://www.redhat.com/archives/linux-cluster/attachments/20100322/678245f8/attachment.html>

------------------------------

Message: 3
Date: Mon, 22 Mar 2010 09:52:21 -0400 (EDT)
From: Bob Peterson <rpeterso redhat com>
To: bergman merctech com, linux clustering <linux-cluster redhat com>
Subject: Re: [Linux-cluster] falure during gfs2_grow caused node crash
       & data loss
Message-ID:
       <1296793451 662271269265941932 JavaMail root zmail06 collab prod int phx2 redhat com>

Content-Type: text/plain; charset=utf-8

----- bergman merctech com wrote:
| I just had a serious problem with gfs2_grow which caused a loss of
| data and a
| cluster node reboot.
|
| I was attempting to grow a gfs2 volume from 50GB => 145GB. The volume
| was
| mounted on both cluster nodes at the start of running "gfs2_grow".
| When I
| umounted the volume from _one_ node (not where gfs2_grow was running),
| the
| macine running gfs2_grow rebooted and the filesystem is damaged.
|
| The sequence of commands was as follows. Each command was successful
| until the
| "umount".
(snip)
| Mark

Hi Mark,

There's a good chance this was caused by bugzilla bug #546683 which
is scheduled to be released in 5.5.  However, I've also seen some
problems like this when a logical volume in LVM isn't marked as
clustered.  Make sure it is with the "vgs" command (check if the flags
end with a "c") and if not, do vgchange -cy <volgrp>

As for fsck.gfs2, it should never segfault.  IMHO, this is a bug
so please open a bugzilla record: Product: "Red Hat Enterprise Linux 5"
and component "gfs2-utils".  Assign it to me.

As for recovering your volume, you can try this but it's not guaranteed
to work:
(1) Reduce the volume to its size from before the gfs2_grow.
(2) Mount it from one node only, if you can (it may crash).
(3) If it lets you mount it, run gfs2_grow again.
(4) Unmount the volume.
(5) Mount the volume from both nodes.

If that doesn't work or if the system can't properly mount the volume
your choices are either (1) reformat the volume and restore from
backup, (2) Use gfs2_edit to patch the i_size field of the rindex file
to be a fairly small multiple of 96 then repeat steps 1 through 4.

Regards,

Bob Peterson
Red Hat File Systems



------------------------------

Message: 4
Date: Mon, 22 Mar 2010 10:08:17 -0400
From: Jeff Sturm <jeff sturm eprize com>
To: "linux clustering" <linux-cluster redhat com>
Subject: Re: [Linux-cluster] GFS create file performance
Message-ID:
       <64D0546C5EBBD147B75DE133D798665F055D8C5C hugo eprize local>
Content-Type: text/plain; charset="us-ascii"

> -----Original Message-----
> From: linux-cluster-bounces redhat com
[mailto:linux-cluster-bounces redhat com]
> On Behalf Of C. Handel
> Sent: Friday, March 19, 2010 5:43 PM
> To: linux-cluster redhat com
> Subject: Re: [Linux-cluster] GFS create file performance
>
> Is your session data valuable? what happens if you loose it? For web
> application this normally means, that users need to login again.

It varies.  Our "session" mechanism is used for a variety of purposes,
some very short lived, others that may persist for weeks.

In some cases the loss of this data will force the user to login again,
as you say.  In other examples a link that we send in an email may
become invalid.

We may decide eventually to adopt different storage backends for
short-lived session data, or transient vs. persistent data.

> How big is your data? What is the read/write ratio?

We have a 50GB GFS filesystem right now.  Reads/writes are close to 1:1.

> You could go for a memcache. Try two dedicated machines with lots of
> memory. Write your session storage to always write to both and read
> from one. Handle failure in software. Unbeatable performance. will
> saturate gigbit links with ease.

Yup, we're aware of this and other storage alternatives.  I wanted to
ask about it on the linux-cluster list to make sure we didn't overlook
anything regarding GFS.  I'm also curious to know what the present
limitations of GFS are.

We actually use GFS for several purposes.  One of those is to
synchronize web content--we used to run an elaborate system of rsync
processes to keep all content distributed over all nodes.  We've
replaced the use of rsync with a GFS filesystem (two master nodes, many
spectator nodes).  This is working well.

We also use GFS to distribute certain user-contributed content, such as
images or video.  This is a read-write filesystem mounted on all cluster
nodes.  GFS works well for this too.

Our only controversial use of GFS at the moment is the session data due
to the frequency of create/write/read/unlink that we need to support.
Following Steven Whitehouse's great explanation last week of inode
creation, resource groups and extended attributes, we tried disabling
selinux on certain cluster nodes.  Surprisingly, I've seen a reduction
of block I/O as high as 30-40% resulting from this.

-Jeff





------------------------------

--
Linux-cluster mailing list
Linux-cluster redhat com
https://www.redhat.com/mailman/listinfo/linux-cluster

End of Linux-cluster Digest, Vol 71, Issue 37
*********************************************


[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]