Bug 894232 - Cluster configuration fails using cluster manager pcs
Summary: Cluster configuration fails using cluster manager pcs
Keywords:
Status: CLOSED WONTFIX
Alias: None
Product: Fedora
Classification: Fedora
Component: pcs
Version: 18
Hardware: ppc64
OS: All
unspecified
urgent
Target Milestone: ---
Assignee: Chris Feist
QA Contact: Fedora Extras Quality Assurance
URL:
Whiteboard:
Depends On:
Blocks: 906529
TreeView+ depends on / blocked
 
Reported: 2013-01-11 06:10 UTC by IBM Bug Proxy
Modified: 2014-02-05 22:57 UTC (History)
3 users (show)

Fixed In Version:
Clone Of:
: 906529 (view as bug list)
Environment:
Last Closed: 2014-02-05 22:57:13 UTC
Type: ---
Embargoed:


Attachments (Terms of Use)
dmesg.txt (16.71 KB, text/plain)
2013-01-11 06:11 UTC, IBM Bug Proxy
no flags Details
var-log-messages.txt (335.70 KB, text/plain)
2013-01-11 06:11 UTC, IBM Bug Proxy
no flags Details
strace-pcs-property (83 bytes, text/plain)
2013-01-11 06:11 UTC, IBM Bug Proxy
no flags Details
dmesg-F19.txt (27.92 KB, text/plain)
2013-06-15 08:41 UTC, IBM Bug Proxy
no flags Details
var-log-messages.-F19.txt.tgz (431.87 KB, application/x-compressed-tar)
2013-06-15 08:41 UTC, IBM Bug Proxy
no flags Details
var-log-cluster-corosync.log.tgz (161.25 KB, application/x-compressed-tar)
2013-06-15 08:41 UTC, IBM Bug Proxy
no flags Details


Links
System ID Private Priority Status Summary Last Updated
IBM Linux Technology Center 87779 0 None None None 2019-07-02 06:53:06 UTC

Description IBM Bug Proxy 2013-01-11 06:10:41 UTC
Problem Description
------------------------------
Configuring two node cluster on P7 LPARs using cluster manager pcs , noticed both nodes are not getting online and accessing pcs properties gives error as below.

[root@c57f1ju0203 ~]# pcs status
Last updated: Thu Jan 10 11:38:42 2013
Last change: Thu Jan 10 11:03:51 2013 via crmd on c57f1ju0203
Current DC: NONE
2 Nodes configured, unknown expected votes
0 Resources configured.


Node c57f1ju0203 (1): UNCLEAN (offline)
Node c57f1ju0204 (2): UNCLEAN (offline)
Full list of resources:


[root@c57f1ju0203 ~]# pcs property
ERROR: Unable to get crm_config
Call cib_query failed (-62): Timer expired
<null>

I also get error as below while trying to stup property using pcs.
[root@c57f1ju0203 ~]# pcs property set no-quorum-policy=ignore
Unable to get crm_config, is pacemaker running?


[root@c57f1ju0203 ~]# ps axf | egrep "corosync|pacemaker" | grep -v egrep
15825 ?        Ssl    0:15 corosync
15832 ?        Ssl    0:00 /usr/sbin/pacemakerd -f
15833 ?        Ssl    0:00  \_ /usr/libexec/pacemaker/cib
15835 ?        Ss     0:00  \_ /usr/libexec/pacemaker/stonithd
15836 ?        Ss     0:00  \_ /usr/libexec/pacemaker/lrmd
15837 ?        Ss     0:00  \_ /usr/libexec/pacemaker/attrd
15838 ?        Ss     0:00  \_ /usr/libexec/pacemaker/pengine

You can note here /usr/libexec/pacemaker/crmd is not running as mentioned in document "Cluster from scratch using pcs".

[root@c57f1ju0203 ~]# pcs status corosync

Membership information
----------------------
    Nodeid      Votes    Qdevice Name
         1          1         NR c57f1ju0203
         2          1         NR c57f1ju0204


--- corosync.conf---
[root@c57f1ju0203 ~]# cat /etc/corosync/corosync.conf
totem {
version: 2
secauth: off
cluster_name: mycluster
transport: udpu
}

nodelist {
  node {
        ring0_addr: c57f1ju0203
        nodeid: 1
       }
  node {
        ring0_addr: c57f1ju0204
        nodeid: 2
       }
}

quorum {
provider: corosync_votequorum
}

logging {
to_syslog: yes
}



[root@c57f1ju0203 ~]# systemctl status pacemaker.service
pacemaker.service - Pacemaker High Availability Cluster Manager
	  Loaded: loaded (/usr/lib/systemd/system/pacemaker.service; disabled)
	  Active: active (running) since Thu, 2013-01-10 11:03:50 EST; 38min ago
	Main PID: 15832 (pacemakerd)
	  CGroup: name=systemd:/system/pacemaker.service
		  ? 15832 /usr/sbin/pacemakerd -f
		  ? 15833 /usr/libexec/pacemaker/cib
		  ? 15835 /usr/libexec/pacemaker/stonithd
		  ? 15836 /usr/libexec/pacemaker/lrmd
		  ? 15837 /usr/libexec/pacemaker/attrd
		  ? 15838 /usr/libexec/pacemaker/pengine

Jan 10 11:03:53 c57f1ju0203 cib[15833]: warning: qb_ipcs_event_sendv: new_event_notification (15833-15941-10): Broken pipe (32)
Jan 10 11:03:53 c57f1ju0203 cib[15833]: warning: do_local_notify: A-Sync reply to crmd failed: No message of desired type
Jan 10 11:03:53 c57f1ju0203 pacemakerd[15832]: error: pcmk_child_exit: Child process crmd exited (pid=15941, rc=2)
Jan 10 11:03:53 c57f1ju0203 crmd[15942]: error: check_dead_member: We're not part of the cluster anymore
Jan 10 11:03:53 c57f1ju0203 crmd[15942]: error: do_log: FSA: Input I_ERROR from check_dead_member() received in state S_STARTING
Jan 10 11:03:53 c57f1ju0203 crmd[15942]: warning: do_state_transition: State transition S_STARTING -> S_RECOVERY [ input=I_ERROR cause=C_FSA_INTERNAL origin=check_dead_member ]
Jan 10 11:03:53 c57f1ju0203 cib[15833]: warning: qb_ipcs_event_sendv: new_event_notification (15833-15942-10): Broken pipe (32)
Jan 10 11:03:53 c57f1ju0203 cib[15833]: warning: do_local_notify: A-Sync reply to crmd failed: No message of desired type
Jan 10 11:03:53 c57f1ju0203 pacemakerd[15832]: error: pcmk_child_exit: Child process crmd exited (pid=15942, rc=2)
Jan 10 11:03:53 c57f1ju0203 pacemakerd[15832]: error: pcmk_child_exit: Child respawn count exceeded by crmd


[root@c57f1ju0203 ~]# systemctl status corosync.service
corosync.service - Corosync Cluster Engine
	  Loaded: loaded (/usr/lib/systemd/system/corosync.service; disabled)
	  Active: active (running) since Thu, 2013-01-10 11:03:50 EST; 39min ago
	 Process: 15818 ExecStart=/usr/share/corosync/corosync start (code=exited, status=0/SUCCESS)
	Main PID: 15825 (corosync)
	  CGroup: name=systemd:/system/corosync.service
		  ? 15825 corosync

Jan 10 11:03:49 c57f1ju0203 corosync[15825]: [QUORUM] Members[1]: 1
Jan 10 11:03:49 c57f1ju0203 corosync[15825]: [TOTEM ] A processor joined or left the membership and a new membership (9.114.32.144:4) was formed.
Jan 10 11:03:49 c57f1ju0203 corosync[15825]: [MAIN  ] Completed service synchronization, ready to provide service.
Jan 10 11:03:50 c57f1ju0203 corosync[15818]: Starting Corosync Cluster Engine (corosync): [  OK  ]
Jan 10 11:03:50 c57f1ju0203 systemd[1]: Started Corosync Cluster Engine.
Jan 10 11:03:50 c57f1ju0203 corosync[15825]: [YKD   ] Members[2]: 1 2
Jan 10 11:03:50 c57f1ju0203 corosync[15825]: [TOTEM ] A processor joined or left the membership and a new membership (9.114.32.144:8) was formed.
Jan 10 11:03:50 c57f1ju0203 corosync[15825]: [QUORUM] This node is within the primary component and will provide service.
Jan 10 11:03:50 c57f1ju0203 corosync[15825]: [YKD   ] Members[2]: 1 2
Jan 10 11:03:50 c57f1ju0203 corosync[15825]: [YKD   ] Completed service synchronization, ready to provide service.


[root@c57f1ju0203 ~]# uname -a
Linux c57f1ju0203 3.6.10-4.fc18.ppc64p7 #1 SMP Wed Dec 12 16:08:02 MST 2012 ppc64 ppc64 ppc64 GNU/Linux

--- Attached log files ---
var-log-messages.txt
dmesg.txt
strace-pcs-property.txt

Here is the package versions installed on system :
----------------------------------------------------------------------
[root@c57f1ju0203 ~]# rpm -qa | grep -i pcs
pcsc-lite-doc-1.8.7-1.fc18.noarch
pcsc-lite-libs-1.8.7-1.fc18.ppc64
pcs-0.9.27-3.fc18.ppc64
pcsc-lite-1.8.7-1.fc18.ppc64
pcsc-lite-openct-0.6.20-5.fc18.ppc64
pcsc-tools-1.4.17-4.fc18.ppc64
pcsc-lite-devel-1.8.7-1.fc18.ppc64
pcsc-perl-1.4.12-5.fc18.ppc64
pcsc-lite-ccid-1.4.8-1.fc18.ppc64

[root@c57f1ju0203 ~]# rpm -qa | grep -i corosync
corosync-2.1.0-1.fc18.ppc64
corosynclib-devel-2.1.0-1.fc18.ppc64
corosynclib-2.1.0-1.fc18.ppc64

[root@c57f1ju0203 ~]# rpm -qa | grep -i pacemaker
pacemaker-1.1.8-3.fc18.ppc64
pacemaker-cluster-libs-1.1.8-3.fc18.ppc64
pacemaker-libs-devel-1.1.8-3.fc18.ppc64
pacemaker-cts-1.1.8-3.fc18.ppc64
pacemaker-cli-1.1.8-3.fc18.ppc64
pacemaker-libs-1.1.8-3.fc18.ppc64
pacemaker-doc-1.1.8-3.fc18.ppc64

[root@c57f1ju0203 ~]# pacemakerd --features
Pacemaker 1.1.8-3.fc18 (Build: 394e906)
 Supporting:  generated-manpages agent-manpages ncurses libqb-logging libqb-ipc upstart systemd  corosync-native

Comment 1 IBM Bug Proxy 2013-01-11 06:11:00 UTC
Created attachment 676666 [details]
dmesg.txt

Comment 2 IBM Bug Proxy 2013-01-11 06:11:08 UTC
Created attachment 676667 [details]
var-log-messages.txt

Comment 3 IBM Bug Proxy 2013-01-11 06:11:18 UTC
Created attachment 676668 [details]
strace-pcs-property

Comment 4 Chris Feist 2013-01-11 20:09:35 UTC
Can you check and see if there are any errors in /var/log/audit/audit.log?  It appears as though pacemaker is not connecting properly (which is why you're getting the pcs errors).

If there are errors, can you try temporarily disabling selinux and the firewall to verify that they're not causing issues?

To disable selinux, edit /etc/sysconfig/selinux and change the 'SELINUX=' line to 'SELINUX=disabled' and restart.

To disable the firewall run:
systemctl stop iptables.service
systemctl disable iptables.service

Comment 5 IBM Bug Proxy 2013-01-14 04:10:58 UTC
------- Comment From maknayak.com 2013-01-14 04:06 EDT-------
(In reply to comment #10)

Hello Chris,

> Can you check and see if there are any errors in /var/log/audit/audit.log?

In  /var/log/audit/audit.log ,no errors but there are many failiures reported as below

[root@c57f1ju0203 ~]# cat /var/log/audit/audit.log | grep -i fail
type=SERVICE_STOP msg=audit(1357827937.860:17): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg=' comm="rngd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed'
type=USER_LOGIN msg=audit(1357828127.823:317): pid=1003 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login acct=28756E6B6E6F776E207573657229 exe="/usr/sbin/sshd" hostname=? addr=9.79.215.218 terminal=ssh res=failed'
type=USER_AUTH msg=audit(1357828169.503:322): pid=1005 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=PAM:authentication acct="root" exe="/usr/sbin/sshd" hostname=9.79.215.218 addr=9.79.215.218 terminal=ssh res=failed'
type=USER_AUTH msg=audit(1357828169.503:323): pid=1005 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=password acct="root" exe="/usr/sbin/sshd" hostname=? addr=9.79.215.218 terminal=ssh res=failed'
type=USER_LOGIN msg=audit(1357832401.047:405): pid=15557 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login acct=28756E6B6E6F776E207573657229 exe="/usr/sbin/sshd" hostname=? addr=9.114.32.145 terminal=ssh res=failed'
type=USER_LOGIN msg=audit(1357836968.293:518): pid=16288 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login acct=28756E6B6E6F776E207573657229 exe="/usr/sbin/sshd" hostname=? addr=9.79.215.218 terminal=ssh res=failed'
type=USER_AUTH msg=audit(1357836974.523:519): pid=16288 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=PAM:authentication acct="?" exe="/usr/sbin/sshd" hostname=9.79.215.218 addr=9.79.215.218 terminal=ssh res=failed'
type=USER_AUTH msg=audit(1357836974.523:520): pid=16288 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=password acct=28696E76616C6964207573657229 exe="/usr/sbin/sshd" hostname=? addr=9.79.215.218 terminal=ssh res=failed'
type=USER_AUTH msg=audit(1357836980.233:521): pid=16288 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=PAM:authentication acct="?" exe="/usr/sbin/sshd" hostname=9.79.215.218 addr=9.79.215.218 terminal=ssh res=failed'
type=USER_AUTH msg=audit(1357836980.233:522): pid=16288 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=password acct=28696E76616C6964207573657229 exe="/usr/sbin/sshd" hostname=? addr=9.79.215.218 terminal=ssh res=failed'
type=USER_LOGIN msg=audit(1357836980.233:526): pid=16288 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login acct=28696E76616C6964207573657229 exe="/usr/sbin/sshd" hostname=? addr=9.79.215.218 terminal=ssh res=failed'

Please see for details log attached file audit.log.

> It appears as though pacemaker is not connecting properly (which is why
> you're getting the pcs errors).
>
> If there are errors, can you try temporarily disabling selinux and the
> firewall to verify that they're not causing issues?
>
> To disable selinux, edit /etc/sysconfig/selinux and change the 'SELINUX='
> line to 'SELINUX=disabled' and restart.
>
> To disable the firewall run:
> systemctl stop iptables.service
> systemctl disable iptables.service

Selinux & firewalls are already disabled.

[root@c57f1ju0203 ~]# iptables -L
Chain INPUT (policy ACCEPT)
target     prot opt source               destination

Chain FORWARD (policy ACCEPT)
target     prot opt source               destination

Chain OUTPUT (policy ACCEPT)
target     prot opt source               destination

[root@c57f1ju0203 ~]# systemctl status firewalld.service
firewalld.service - firewalld - dynamic firewall daemon
Loaded: loaded (/usr/lib/systemd/system/firewalld.service; disabled)
Active: inactive (dead)
CGroup: name=systemd:/system/firewalld.service

Jan 10 09:25:39 c57f1ju0203.ppd.pok.ibm.com systemd[1]: Started firewalld - dynamic firewall daemon.
Jan 10 09:55:30 c57f1ju0203.ppd.pok.ibm.com systemd[1]: Stopping firewalld - dynamic firewall daemon...
Jan 10 09:55:36 c57f1ju0203.ppd.pok.ibm.com systemd[1]: Stopped firewalld - dynamic firewall daemon.

[root@c57f1ju0203 ~]# getenforce
Permissive

Thanks...
Manas

Comment 6 Chris Feist 2013-01-31 23:26:57 UTC
Can you give me the output of 'hostname' on all of your machines?

Comment 7 IBM Bug Proxy 2013-02-01 04:22:22 UTC
------- Comment From maknayak.com 2013-02-01 04:12 EDT-------
(In reply to comment #12)
> Can you give me the output of 'hostname' on all of your machines?

Hello Chris,

I have configured Short Node Names for convenience on both the nodes:

NODE-1:
[root@c57f1ju0204 ~]# hostname
c57f1ju0204

[root@c57f1ju0204 ~]# uname -a
Linux c57f1ju0204 3.6.10-4.fc18.ppc64p7 #1 SMP Wed Dec 12 16:08:02 MST 2012 ppc64 ppc64 ppc64 GNU/Linux

NODE-2:
[root@c57f1ju0203 ~]# hostname
c57f1ju0203

Some more info:

[root@c57f1ju0203 ~]# env | grep ais_
ais_mcast=239.255.1.1
ais_port=4000
ais_addr=9.114.32.0

[root@c57f1ju0204 ~]# env | grep -i ais
ais_mcast=239.255.1.1
ais_port=4000
ais_addr=9.114.32.0

Thanks...
Manas

Comment 8 IBM Bug Proxy 2013-03-13 09:22:31 UTC
------- Comment From maknayak.com 2013-03-13 09:17 EDT-------
Chris,
Any update on this issue ?

Thanks...
Manas

Comment 9 Chris Feist 2013-03-15 21:03:41 UTC
Manas,

Sorry for the delay, I don't see anything wrong with your configuration.  I've talked with one of the pacemaker developers and it sounds like several bugs are fixed with the latest 1.1.9 pacemaker release.  Can you try a 'pcs cluster stop' on all your nodes and yum update to the latest pacemaker & corosync and then run 'pcs cluster start'?

The latest version of pacemaker in fedora should be 1.1.9-0.1.70ad9fa.git.fc18

Let me know if you're getting the same issues.

Thanks,
Chris

Comment 10 IBM Bug Proxy 2013-06-15 08:41:10 UTC
------- Comment From maknayak.com 2013-06-15 08:31 EDT-------
(In reply to comment #15)
> Manas,
>
> Sorry for the delay, I don't see anything wrong with your configuration.
> I've talked with one of the pacemaker developers and it sounds like several
> bugs are fixed with the latest 1.1.9 pacemaker release.  Can you try a 'pcs
> cluster stop' on all your nodes and yum update to the latest pacemaker &
> corosync and then run 'pcs cluster start'?
>
> The latest version of pacemaker in fedora should be
> 1.1.9-0.1.70ad9fa.git.fc18
>
> Let me know if you're getting the same issues.
>
> Thanks,
> Chris

Hi Chris,
Sorry forthe late response.
I am using Pacemaker 1.1.9-0.1.70ad9fa.git.fc19 , and verified on F19 Beta ppc64 kernel. Looks like still it has issue. Even if pacemaker gtets started , it shows several failures. Also I noticed crash in '/usr/libexec/pacemaker/crmd' gets produced in /var/log/messages as soon as pacemaker gets started.

[root@miz03 cluster]# systemctl status corosync.service
Loaded: loaded (/usr/lib/systemd/system/corosync.service; enabled)
Active: active (running) since Sat 2013-06-15 04:05:54 EDT; 18min ago
Process: 2991 ExecStop=/usr/share/corosync/corosync stop (code=exited, status=0/SUCCESS)
Process: 3009 ExecStart=/usr/share/corosync/corosync start (code=exited, status=0/SUCCESS)
Main PID: 3016 (corosync)
??3016 corosync

Jun 15 04:05:54 miz03.austin.ibm.com corosync[3016]: [QB    ] server name: quorum
Jun 15 04:05:54 miz03.austin.ibm.com corosync[3016]: [QUORUM] Members[1]: -1351744759
Jun 15 04:05:54 miz03.austin.ibm.com corosync[3016]: [TOTEM ] A processor joined or left the membership and a new membership (9.3.110.175:16) was formed.
Jun 15 04:05:54 miz03.austin.ibm.com corosync[3016]: [MAIN  ] Completed service synchronization, ready to provide service.
Jun 15 04:05:54 miz03.austin.ibm.com systemd[1]: Started Corosync Cluster Engine.
Jun 15 04:05:57 miz03.austin.ibm.com corosync[3016]: [YKD   ] Members[2]: 1299055369 -1351744759
Jun 15 04:05:57 miz03.austin.ibm.com corosync[3016]: [TOTEM ] A processor joined or left the membership and a new membership (9.3.110.77:24) was formed.
Jun 15 04:05:57 miz03.austin.ibm.com corosync[3016]: [QUORUM] This node is within the primary component and will provide service.
Jun 15 04:05:57 miz03.austin.ibm.com corosync[3016]: [YKD   ] Members[2]: 1299055369 -1351744759
Jun 15 04:05:57 miz03.austin.ibm.com corosync[3016]: [YKD   ] Completed service synchronization, ready to provide service.

[root@miz03 cluster]# systemctl status pacemaker.service
Loaded: loaded (/usr/lib/systemd/system/pacemaker.service; enabled)
Active: active (running) since Sat 2013-06-15 04:07:57 EDT; 16min ago
Main PID: 3031 (pacemakerd)
??3031 /usr/sbin/pacemakerd -f
??3032 /usr/libexec/pacemaker/cib
??3034 /usr/libexec/pacemaker/stonithd
??3035 /usr/libexec/pacemaker/lrmd
??3036 /usr/libexec/pacemaker/attrd
??3037 /usr/libexec/pacemaker/pengine

Jun 15 04:10:34 miz03.austin.ibm.com crmd[3801]: error: do_log: FSA: Input I_TERMINATE from do_recover() received in state S_RECOVERY
Jun 15 04:10:34 miz03.austin.ibm.com crmd[3801]: notice: do_lrm_control: Disconnected from the LRM
Jun 15 04:10:34 miz03.austin.ibm.com crmd[3801]: notice: terminate_cs_connection: Disconnecting from Corosync
Jun 15 04:10:34 miz03.austin.ibm.com cib[3032]: warning: qb_ipcs_event_sendv: new_event_notification (3032-3801-13): Broken pipe (32)
Jun 15 04:10:34 miz03.austin.ibm.com crmd[3801]: error: do_exit: Could not recover from internal error
Jun 15 04:10:34 miz03.austin.ibm.com cib[3032]: warning: do_local_notify: A-Sync reply to crmd failed: No message of desired type
Jun 15 04:10:34 miz03.austin.ibm.com crmd[3801]: error: crm_abort: crm_glib_handler: Forked child 3802 to record non-fatal assert at logging.c:62 : g_hash_table_size: assertion `has...NULL' failed
Jun 15 04:10:35 miz03.austin.ibm.com crmd[3801]: error: crm_abort: crm_glib_handler: Forked child 3804 to record non-fatal assert at logging.c:62 : g_hash_table_destroy: assertion `...NULL' failed
Jun 15 04:10:36 miz03.austin.ibm.com pacemakerd[3031]: error: pcmk_child_exit: Child process crmd exited (pid=3801, rc=2)
Jun 15 04:10:36 miz03.austin.ibm.com pacemakerd[3031]: error: pcmk_process_exit: Child respawn count exceeded by crmd

You can see status for pacemaker service says lots of errors.

--- Other Details ---

[root@miz03 cluster]# uname -a
Linux miz03.austin.ibm.com 3.9.4-300.fc19.ppc64p7 #1 SMP Sun May 26 22:14:56 MST 2013 ppc64 ppc64 ppc64 GNU/Linux

[root@miz03 cluster]# rpm -qa | grep -i pacemaker
pacemaker-libs-1.1.9-0.1.70ad9fa.git.fc19.ppc64
pacemaker-1.1.9-0.1.70ad9fa.git.fc19.ppc64
pacemaker-cluster-libs-1.1.9-0.1.70ad9fa.git.fc19.ppc64
pacemaker-libs-devel-1.1.9-0.1.70ad9fa.git.fc19.ppc64
pacemaker-cts-1.1.9-0.1.70ad9fa.git.fc19.ppc64
pacemaker-cli-1.1.9-0.1.70ad9fa.git.fc19.ppc64
pacemaker-doc-1.1.9-0.1.70ad9fa.git.fc19.ppc64

[root@miz03 cluster]# rpm -qa | grep -i corosync
corosync-2.3.0-3.fc19.ppc64
corosynclib-2.3.0-3.fc19.ppc64
corosynclib-devel-2.3.0-3.fc19.ppc64

[root@miz03 cluster]# rpm -qa | grep -i crm
crm114-emacs-0-5.14.20100106.fc19.ppc64
crm114-0-5.14.20100106.fc19.ppc64

[root@miz03 cluster]# ps axf | egrep "corosync|pacemaker" | grep -v egrep
3016 ?        Ssl    0:08 corosync
3031 ?        Ssl    0:00 /usr/sbin/pacemakerd -f
3032 ?        Ssl    0:00  \_ /usr/libexec/pacemaker/cib
3034 ?        Ss     0:00  \_ /usr/libexec/pacemaker/stonithd
3035 ?        Ss     0:00  \_ /usr/libexec/pacemaker/lrmd
3036 ?        Ss     0:00  \_ /usr/libexec/pacemaker/attrd
3037 ?        Ss     0:00  \_ /usr/libexec/pacemaker/pengine

[root@miz03 cluster]# corosync-cmapctl | grep members
runtime.totem.pg.mrp.srp.members.1299055369.config_version (u64) = 0
runtime.totem.pg.mrp.srp.members.1299055369.ip (str) = r(0) ip(9.3.110.77)
runtime.totem.pg.mrp.srp.members.1299055369.join_count (u32) = 1
runtime.totem.pg.mrp.srp.members.1299055369.status (str) = joined
runtime.totem.pg.mrp.srp.members.2943222537.config_version (u64) = 0
runtime.totem.pg.mrp.srp.members.2943222537.ip (str) = r(0) ip(9.3.110.175)
runtime.totem.pg.mrp.srp.members.2943222537.join_count (u32) = 1
runtime.totem.pg.mrp.srp.members.2943222537.status (str) = joined

[root@miz03 cluster]# corosync-quorumtool -l
Nodeid      Votes Name
1299055369          1 miz05.austin.ibm.com
2943222537          1 miz03.austin.ibm.com (local)

[root@miz03 cluster]# crm_mon -1
Last updated: Sat Jun 15 04:27:45 2013
Last change: Sat Jun 15 03:58:32 2013
0 Nodes configured, unknown expected votes

Attached three files var-log-messages-F19.txt.tgz , dmesg-F19.txt  & var-log-cluster-corosync.log.tgz for reference.

Thanks...
Manas

Comment 11 IBM Bug Proxy 2013-06-15 08:41:26 UTC
Created attachment 761526 [details]
dmesg-F19.txt


------- Comment (attachment only) From maknayak.com 2013-06-15 08:32 EDT-------

Comment 12 IBM Bug Proxy 2013-06-15 08:41:38 UTC
Created attachment 761527 [details]
var-log-messages.-F19.txt.tgz


------- Comment (attachment only) From maknayak.com 2013-06-15 08:33 EDT-------

Comment 13 IBM Bug Proxy 2013-06-15 08:41:50 UTC
Created attachment 761528 [details]
var-log-cluster-corosync.log.tgz


------- Comment (attachment only) From maknayak.com 2013-06-15 08:34 EDT-------

Comment 14 Fedora End Of Life 2013-12-21 15:16:55 UTC
This message is a reminder that Fedora 18 is nearing its end of life.
Approximately 4 (four) weeks from now Fedora will stop maintaining
and issuing updates for Fedora 18. It is Fedora's policy to close all
bug reports from releases that are no longer maintained. At that time
this bug will be closed as WONTFIX if it remains open with a Fedora 
'version' of '18'.

Package Maintainer: If you wish for this bug to remain open because you
plan to fix it in a currently maintained version, simply change the 'version' 
to a later Fedora version prior to Fedora 18's end of life.

Thank you for reporting this issue and we are sorry that we may not be 
able to fix it before Fedora 18 is end of life. If you would still like 
to see this bug fixed and are able to reproduce it against a later version 
of Fedora, you are encouraged  change the 'version' to a later Fedora 
version prior to Fedora 18's end of life.

Although we aim to fix as many bugs as possible during every release's 
lifetime, sometimes those efforts are overtaken by events. Often a 
more recent Fedora release includes newer upstream software that fixes 
bugs or makes them obsolete.

Comment 15 Fedora End Of Life 2014-02-05 22:57:13 UTC
Fedora 18 changed to end-of-life (EOL) status on 2014-01-14. Fedora 18 is
no longer maintained, which means that it will not receive any further
security or bug fix updates. As a result we are closing this bug.

If you can reproduce this bug against a currently maintained version of
Fedora please feel free to reopen this bug against that version. If you
are unable to reopen this bug, please file a new report against the
current release. If you experience problems, please add a comment to this
bug.

Thank you for reporting this bug and we are sorry it could not be fixed.


Note You need to log in before you can comment on or make changes to this bug.