Bug 333181 - Add option to not fail-back service
Summary: Add option to not fail-back service
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Cluster Suite
Classification: Retired
Component: conga
Version: 4
Hardware: All
OS: Linux
low
low
Target Milestone: ---
Assignee: Ryan McCabe
QA Contact: Cluster QE
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2007-10-15 21:44 UTC by Anthony Green
Modified: 2018-10-20 01:01 UTC (History)
3 users (show)

Fixed In Version: RHBA-2008-0798
Doc Type: Bug Fix
Doc Text:
Clone Of:
Environment:
Last Closed: 2008-07-25 19:14:18 UTC
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Red Hat Product Errata RHBA-2008:0798 0 normal SHIPPED_LIVE conga bug fix and enhancement update 2008-07-25 19:14:11 UTC

Description Anthony Green 2007-10-15 21:44:09 UTC
Description of problem:
We use the priority option with our active/active cluster to insure that all of
our databases don’t start on the same node. What we’ve noticed is that when a
node is fenced and then returns to the cluster the databases are automatically
redistributed based on the priority in the cluster.conf.

The issue with this behavior is that if a node crashes, comes back up, crashes
again and comes back up again the database will be needlessly failed over
several times until someone logs into the cluster and stops it.

We request an option to have the databases stay put when a failed node returns
to the cluster.    


Version-Release number of selected component (if applicable):


How reproducible:


Steps to Reproduce:
1.
2.
3.
  
Actual results:


Expected results:


Additional info:

Comment 2 Lon Hohberger 2007-10-16 19:21:11 UTC
This can be manually configured currently by specifying nofailback="1" as part
of the failover domain configuration in cluster.conf, for example:

                <failoverdomains>
                        <failoverdomain name="all" ordered="1" nofailback="1">
                                <failoverdomainnode name="molly"
                                 priority="2"/>
                                <failoverdomainnode name="frederick" 
                                 priority="1"/>
                        </failoverdomain>
                </failoverdomains>

Note that this option does not work when dealing with a service bound to a
domain but currently running outside of its failover domain - for example, if a
service was running on a node named "buster", it would currently move to either
"molly" or "frederick" if one of those nodes came online irrespective of the
nofailback option.

The nofailback option is available for RHEL4 and RHEL5


Comment 3 Anthony Green 2007-10-16 19:52:22 UTC
(In reply to comment #2)
> The nofailback option is available for RHEL4 and RHEL5

Thanks.  Just to be clear -- this feature exists but has been undocumented to
date.  I assume we will updating the documentation to cover this feature, correct?





Comment 4 Rob Kenna 2007-10-17 22:00:18 UTC
re: comment #3.  Yes, we will be adding this to the documentation.

- Rob

Comment 7 Ryan McCabe 2008-01-18 18:53:46 UTC
Fix will be backported from RHEL5.

Comment 8 Lon Hohberger 2008-03-04 22:46:27 UTC
Removing dependency from 182423.

Comment 11 Kiersten (Kerri) Anderson 2008-04-18 20:47:21 UTC
Adding the feature described in bug #333181 introduced a regression that causes
an exception to be thrown when updating or adding a failover domain. This makes
it impossible to create or update failover domains via conga.

Comment 13 Brian Brock 2008-05-23 18:47:07 UTC
fix verified in 0.11.1-4.el4, GUI option in the Failover Domain subsection of
the cluster tab creates the cluster.conf option described above.


Comment 14 Paul Kennedy 2008-06-10 22:19:08 UTC
This needs to be documented in RHEL 4 Cluster_Administration. Likewise for RHEL
5 Cluster_Administration. Because the RHEL 5 bug for this is closed, I have
opened a doc bug for it: bug #450777. Will port change to RHEL 4 doc.

Comment 16 errata-xmlrpc 2008-07-25 19:14:18 UTC
An advisory has been issued which should help the problem
described in this bug report. This report is therefore being
closed with a resolution of ERRATA. For more information
on the solution and/or where to find the updated files,
please follow the link below. You may reopen this bug report
if the solution does not work for you.

http://rhn.redhat.com/errata/RHBA-2008-0798.html



Note You need to log in before you can comment on or make changes to this bug.