Bug 632319
| Summary: | dlm_controld daemon cpg_dispatch error 2 | ||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Product: | [Fedora] Fedora | Reporter: | Michael Hagmann <michael.hagmann> | ||||||||||
| Component: | cluster | Assignee: | Fabio Massimo Di Nitto <fdinitto> | ||||||||||
| Status: | CLOSED CURRENTRELEASE | QA Contact: | Fedora Extras Quality Assurance <extras-qa> | ||||||||||
| Severity: | medium | Docs Contact: | |||||||||||
| Priority: | low | ||||||||||||
| Version: | 13 | CC: | agk, ccaulfie, cfeist, extras-orphan, fdinitto, lhh, mbroz, swhiteho | ||||||||||
| Target Milestone: | --- | ||||||||||||
| Target Release: | --- | ||||||||||||
| Hardware: | All | ||||||||||||
| OS: | Linux | ||||||||||||
| Whiteboard: | |||||||||||||
| Fixed In Version: | Doc Type: | Bug Fix | |||||||||||
| Doc Text: | Story Points: | --- | |||||||||||
| Clone Of: | Environment: | ||||||||||||
| Last Closed: | 2011-02-08 11:06:45 UTC | Type: | --- | ||||||||||
| Regression: | --- | Mount Type: | --- | ||||||||||
| Documentation: | --- | CRM: | |||||||||||
| Verified Versions: | Category: | --- | |||||||||||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||||||||
| Cloudforms Team: | --- | Target Upstream Version: | |||||||||||
| Embargoed: | |||||||||||||
| Bug Depends On: | |||||||||||||
| Bug Blocks: | 631496 | ||||||||||||
| Attachments: |
|
||||||||||||
|
Description
Michael Hagmann
2010-09-09 17:01:42 UTC
Created attachment 446303 [details]
Sosreport from scecond node Leo
Created attachment 446304 [details]
Sosreport from first Node Scheat
I also not able to start cman again:
[root@scheat cluster]# service cman start
Starting cluster:
Checking Network Manager... [ OK ]
Global setup... [ OK ]
Loading kernel modules... [ OK ]
Mounting configfs... [ OK ]
Starting cman... [ OK ]
Waiting for quorum... [ OK ]
Starting fenced... [ OK ]
Starting dlm_controld... [ OK ]
Starting gfs_controld... [ OK ]
Unfencing self... [ OK ]
Joining fence domain...
[FAILED]
[root@scheat cluster]#
How could I configure manual fencing over luci ? I Think that's the problem.
Michael
After add the manual fencing the cluster works fine: [root@leo tmp]# cat /etc/cluster/cluster.conf <?xml version="1.0"?> <cluster config_version="18" name="gecco"> <fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/> <clusternodes> <clusternode name="scheat" nodeid="1" votes="1"> <fence> <method name="single"> <device name="human" nodename="scheat"/> </method> </fence> </clusternode> <clusternode name="leo" nodeid="2" votes="1"> <fence> <method name="single"> <device name="human" nodename="leo"/> </method> </fence> </clusternode> </clusternodes> <cman expected_votes="1" two_node="1"/> <fencedevices> <fencedevice name="human" agent="fence_manual"/> </fencedevices> <rm> <failoverdomains> <failoverdomain name="scheat" nofailback="0" ordered="0" restricted="1"> <failoverdomainnode name="scheat" priority="1"/> </failoverdomain> <failoverdomain name="all" nofailback="0" ordered="1" restricted="0"> <failoverdomainnode name="scheat" priority="1"/> <failoverdomainnode name="leo" priority="1"/> </failoverdomain> <failoverdomain name="leo" nofailback="0" ordered="0" restricted="1"> <failoverdomainnode name="leo" priority="1"/> </failoverdomain> </failoverdomains> <resources> <ip address="192.168.1.111" sleeptime="10"/> </resources> <service autostart="1" domain="all" exclusive="1" name="Webserver" recovery="relocate"> <ip ref="192.168.1.111"/> </service> </rm> </cluster> [root@leo tmp]# But luci don't show the Cluster anymore! https://bugzilla.redhat.com/show_bug.cgi?id=631496 thanks for Help Michael Manual override is built-in; there is no need to configure it. Also, there is no fence_manual agent. I do not think fencing configuration was the problem -- I think somehow the two config files got out of sync. The two sosreports have cluster conf versions 15 and 16. I am not sure why. Updating the cluster config by adding manual fencing brought the config files back into sync, causing things to work again. ok,
so I could deconfigure it again, and it should work ?
I asume the fencing was the problem because if i want to start cman the Joining fence domain... failed.
so this config version should also work?
<?xml version="1.0"?>
<cluster config_version="19" name="gecco">
<fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/>
<clusternodes>
<clusternode name="scheat" nodeid="1" votes="1">
<fence/>
</clusternode>
<clusternode name="leo" nodeid="2" votes="1">
<fence/>
</clusternode>
</clusternodes>
<cman expected_votes="1" two_node="1"/>
<fencedevices/>
<rm>
<failoverdomains>
<failoverdomain name="scheat" nofailback="0" ordered="0" restricted="1">
<failoverdomainnode name="scheat" priority="1"/>
</failoverdomain>
<failoverdomain name="all" nofailback="0" ordered="1" restricted="0">
<failoverdomainnode name="scheat" priority="1"/>
<failoverdomainnode name="leo" priority="1"/>
</failoverdomain>
<failoverdomain name="leo" nofailback="0" ordered="0" restricted="1">
<failoverdomainnode name="leo" priority="1"/>
</failoverdomain>
</failoverdomains>
<resources>
<ip address="192.168.1.111" sleeptime="10"/>
</resources>
<service autostart="1" domain="all" exclusive="1" name="Webserver"
recovery="relocate">
<ip ref="192.168.1.111"/>
</service>
</rm>
</cluster>
but why then luci don't allow me to admin the cluster ?
--> https://bugzilla.redhat.com/show_bug.cgi?id=631496
something is complete wrong
Michael
Lon you are right I update to no fencedevice and it works too Michael and now also luci works again ! is that normal that with a small configerror luci give a 500? Michael Created attachment 446354 [details]
sosreport after cluster.conf change to 19
Created attachment 446355 [details]
sosreport after cluster.conf change to 19
(In reply to comment #8) > and now also luci works again ! > > is that normal that with a small configerror luci give a 500? > > Michael This issue has been addressed in luci and now luci has a much wider understanding of the configuration. The configuration version problem could have been caused by the other issue you had of luci unable to talk to one of the ricci session (reported in another bug). |