Red Hat Bugzilla – Bug 400211
RFE: Provide a method to fail over a service to another node after X number of restarts.
Last modified: 2010-10-22 16:42:28 EDT
For RH clustering the default recovery policy (when no recovery policy is
specified) on RHEL4 cluster is restart - "Restart the service in the node the
service is currently located.".
The desired behavior is to retry a restart on the same node, if that fails again
a failover to another node should be done.
Current default behavior appears endless attempts to restart.
RHEL3 had relocate after X restarts as well as a notion of "false"
starts, which is a special kind of restart.
-- Additional comment from email@example.com on 2007-11-20 18:06 EST --
Following patch allows a user to set max_restarts="x" and
restart_expire_time="y" in cluster.conf for services and virtual machines.
Basically, restart_expire_time allows one to throttle restarts - that is, if X
restarts occur within Y seconds, relocate the service instead of restarting it.
Requires recovery_policy="restart" (which is the default).
Also makes the parsing of time values more robust; you can now enter things
like : 1h30m as part of resource metadata or in cluster.conf.
-- Additional comment from firstname.lastname@example.org on 2007-11-20 18:55 EST --
Created an attachment (id=265531)
Patches in CVS.
*** Bug 247139 has been marked as a duplicate of this bug. ***
This request was evaluated by Red Hat Product Management for inclusion in a Red
Hat Enterprise Linux maintenance release. Product Management has requested
further review of this request by Red Hat Engineering, for potential
inclusion in a Red Hat Enterprise Linux Update release for currently deployed
products. This request is not yet committed for inclusion in an Update
*** Bug 431130 has been marked as a duplicate of this bug. ***
An advisory has been issued which should help the problem
described in this bug report. This report is therefore being
closed with a resolution of ERRATA. For more information
on the solution and/or where to find the updated files,
please follow the link below. You may reopen this bug report
if the solution does not work for you.