Problem Description ------------------------------ Configuring two node cluster on P7 LPARs using cluster manager pcs , noticed both nodes are not getting online and accessing pcs properties gives error as below. [root@c57f1ju0203 ~]# pcs status Last updated: Thu Jan 10 11:38:42 2013 Last change: Thu Jan 10 11:03:51 2013 via crmd on c57f1ju0203 Current DC: NONE 2 Nodes configured, unknown expected votes 0 Resources configured. Node c57f1ju0203 (1): UNCLEAN (offline) Node c57f1ju0204 (2): UNCLEAN (offline) Full list of resources: [root@c57f1ju0203 ~]# pcs property ERROR: Unable to get crm_config Call cib_query failed (-62): Timer expired <null> I also get error as below while trying to stup property using pcs. [root@c57f1ju0203 ~]# pcs property set no-quorum-policy=ignore Unable to get crm_config, is pacemaker running? [root@c57f1ju0203 ~]# ps axf | egrep "corosync|pacemaker" | grep -v egrep 15825 ? Ssl 0:15 corosync 15832 ? Ssl 0:00 /usr/sbin/pacemakerd -f 15833 ? Ssl 0:00 \_ /usr/libexec/pacemaker/cib 15835 ? Ss 0:00 \_ /usr/libexec/pacemaker/stonithd 15836 ? Ss 0:00 \_ /usr/libexec/pacemaker/lrmd 15837 ? Ss 0:00 \_ /usr/libexec/pacemaker/attrd 15838 ? Ss 0:00 \_ /usr/libexec/pacemaker/pengine You can note here /usr/libexec/pacemaker/crmd is not running as mentioned in document "Cluster from scratch using pcs". [root@c57f1ju0203 ~]# pcs status corosync Membership information ---------------------- Nodeid Votes Qdevice Name 1 1 NR c57f1ju0203 2 1 NR c57f1ju0204 --- corosync.conf--- [root@c57f1ju0203 ~]# cat /etc/corosync/corosync.conf totem { version: 2 secauth: off cluster_name: mycluster transport: udpu } nodelist { node { ring0_addr: c57f1ju0203 nodeid: 1 } node { ring0_addr: c57f1ju0204 nodeid: 2 } } quorum { provider: corosync_votequorum } logging { to_syslog: yes } [root@c57f1ju0203 ~]# systemctl status pacemaker.service pacemaker.service - Pacemaker High Availability Cluster Manager Loaded: loaded (/usr/lib/systemd/system/pacemaker.service; disabled) Active: active (running) since Thu, 2013-01-10 11:03:50 EST; 38min ago Main PID: 15832 (pacemakerd) CGroup: name=systemd:/system/pacemaker.service ? 15832 /usr/sbin/pacemakerd -f ? 15833 /usr/libexec/pacemaker/cib ? 15835 /usr/libexec/pacemaker/stonithd ? 15836 /usr/libexec/pacemaker/lrmd ? 15837 /usr/libexec/pacemaker/attrd ? 15838 /usr/libexec/pacemaker/pengine Jan 10 11:03:53 c57f1ju0203 cib[15833]: warning: qb_ipcs_event_sendv: new_event_notification (15833-15941-10): Broken pipe (32) Jan 10 11:03:53 c57f1ju0203 cib[15833]: warning: do_local_notify: A-Sync reply to crmd failed: No message of desired type Jan 10 11:03:53 c57f1ju0203 pacemakerd[15832]: error: pcmk_child_exit: Child process crmd exited (pid=15941, rc=2) Jan 10 11:03:53 c57f1ju0203 crmd[15942]: error: check_dead_member: We're not part of the cluster anymore Jan 10 11:03:53 c57f1ju0203 crmd[15942]: error: do_log: FSA: Input I_ERROR from check_dead_member() received in state S_STARTING Jan 10 11:03:53 c57f1ju0203 crmd[15942]: warning: do_state_transition: State transition S_STARTING -> S_RECOVERY [ input=I_ERROR cause=C_FSA_INTERNAL origin=check_dead_member ] Jan 10 11:03:53 c57f1ju0203 cib[15833]: warning: qb_ipcs_event_sendv: new_event_notification (15833-15942-10): Broken pipe (32) Jan 10 11:03:53 c57f1ju0203 cib[15833]: warning: do_local_notify: A-Sync reply to crmd failed: No message of desired type Jan 10 11:03:53 c57f1ju0203 pacemakerd[15832]: error: pcmk_child_exit: Child process crmd exited (pid=15942, rc=2) Jan 10 11:03:53 c57f1ju0203 pacemakerd[15832]: error: pcmk_child_exit: Child respawn count exceeded by crmd [root@c57f1ju0203 ~]# systemctl status corosync.service corosync.service - Corosync Cluster Engine Loaded: loaded (/usr/lib/systemd/system/corosync.service; disabled) Active: active (running) since Thu, 2013-01-10 11:03:50 EST; 39min ago Process: 15818 ExecStart=/usr/share/corosync/corosync start (code=exited, status=0/SUCCESS) Main PID: 15825 (corosync) CGroup: name=systemd:/system/corosync.service ? 15825 corosync Jan 10 11:03:49 c57f1ju0203 corosync[15825]: [QUORUM] Members[1]: 1 Jan 10 11:03:49 c57f1ju0203 corosync[15825]: [TOTEM ] A processor joined or left the membership and a new membership (9.114.32.144:4) was formed. Jan 10 11:03:49 c57f1ju0203 corosync[15825]: [MAIN ] Completed service synchronization, ready to provide service. Jan 10 11:03:50 c57f1ju0203 corosync[15818]: Starting Corosync Cluster Engine (corosync): [ OK ] Jan 10 11:03:50 c57f1ju0203 systemd[1]: Started Corosync Cluster Engine. Jan 10 11:03:50 c57f1ju0203 corosync[15825]: [YKD ] Members[2]: 1 2 Jan 10 11:03:50 c57f1ju0203 corosync[15825]: [TOTEM ] A processor joined or left the membership and a new membership (9.114.32.144:8) was formed. Jan 10 11:03:50 c57f1ju0203 corosync[15825]: [QUORUM] This node is within the primary component and will provide service. Jan 10 11:03:50 c57f1ju0203 corosync[15825]: [YKD ] Members[2]: 1 2 Jan 10 11:03:50 c57f1ju0203 corosync[15825]: [YKD ] Completed service synchronization, ready to provide service. [root@c57f1ju0203 ~]# uname -a Linux c57f1ju0203 3.6.10-4.fc18.ppc64p7 #1 SMP Wed Dec 12 16:08:02 MST 2012 ppc64 ppc64 ppc64 GNU/Linux --- Attached log files --- var-log-messages.txt dmesg.txt strace-pcs-property.txt Here is the package versions installed on system : ---------------------------------------------------------------------- [root@c57f1ju0203 ~]# rpm -qa | grep -i pcs pcsc-lite-doc-1.8.7-1.fc18.noarch pcsc-lite-libs-1.8.7-1.fc18.ppc64 pcs-0.9.27-3.fc18.ppc64 pcsc-lite-1.8.7-1.fc18.ppc64 pcsc-lite-openct-0.6.20-5.fc18.ppc64 pcsc-tools-1.4.17-4.fc18.ppc64 pcsc-lite-devel-1.8.7-1.fc18.ppc64 pcsc-perl-1.4.12-5.fc18.ppc64 pcsc-lite-ccid-1.4.8-1.fc18.ppc64 [root@c57f1ju0203 ~]# rpm -qa | grep -i corosync corosync-2.1.0-1.fc18.ppc64 corosynclib-devel-2.1.0-1.fc18.ppc64 corosynclib-2.1.0-1.fc18.ppc64 [root@c57f1ju0203 ~]# rpm -qa | grep -i pacemaker pacemaker-1.1.8-3.fc18.ppc64 pacemaker-cluster-libs-1.1.8-3.fc18.ppc64 pacemaker-libs-devel-1.1.8-3.fc18.ppc64 pacemaker-cts-1.1.8-3.fc18.ppc64 pacemaker-cli-1.1.8-3.fc18.ppc64 pacemaker-libs-1.1.8-3.fc18.ppc64 pacemaker-doc-1.1.8-3.fc18.ppc64 [root@c57f1ju0203 ~]# pacemakerd --features Pacemaker 1.1.8-3.fc18 (Build: 394e906) Supporting: generated-manpages agent-manpages ncurses libqb-logging libqb-ipc upstart systemd corosync-native
Created attachment 676666 [details] dmesg.txt
Created attachment 676667 [details] var-log-messages.txt
Created attachment 676668 [details] strace-pcs-property
Can you check and see if there are any errors in /var/log/audit/audit.log? It appears as though pacemaker is not connecting properly (which is why you're getting the pcs errors). If there are errors, can you try temporarily disabling selinux and the firewall to verify that they're not causing issues? To disable selinux, edit /etc/sysconfig/selinux and change the 'SELINUX=' line to 'SELINUX=disabled' and restart. To disable the firewall run: systemctl stop iptables.service systemctl disable iptables.service
------- Comment From maknayak.com 2013-01-14 04:06 EDT------- (In reply to comment #10) Hello Chris, > Can you check and see if there are any errors in /var/log/audit/audit.log? In /var/log/audit/audit.log ,no errors but there are many failiures reported as below [root@c57f1ju0203 ~]# cat /var/log/audit/audit.log | grep -i fail type=SERVICE_STOP msg=audit(1357827937.860:17): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg=' comm="rngd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' type=USER_LOGIN msg=audit(1357828127.823:317): pid=1003 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login acct=28756E6B6E6F776E207573657229 exe="/usr/sbin/sshd" hostname=? addr=9.79.215.218 terminal=ssh res=failed' type=USER_AUTH msg=audit(1357828169.503:322): pid=1005 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=PAM:authentication acct="root" exe="/usr/sbin/sshd" hostname=9.79.215.218 addr=9.79.215.218 terminal=ssh res=failed' type=USER_AUTH msg=audit(1357828169.503:323): pid=1005 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=password acct="root" exe="/usr/sbin/sshd" hostname=? addr=9.79.215.218 terminal=ssh res=failed' type=USER_LOGIN msg=audit(1357832401.047:405): pid=15557 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login acct=28756E6B6E6F776E207573657229 exe="/usr/sbin/sshd" hostname=? addr=9.114.32.145 terminal=ssh res=failed' type=USER_LOGIN msg=audit(1357836968.293:518): pid=16288 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login acct=28756E6B6E6F776E207573657229 exe="/usr/sbin/sshd" hostname=? addr=9.79.215.218 terminal=ssh res=failed' type=USER_AUTH msg=audit(1357836974.523:519): pid=16288 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=PAM:authentication acct="?" exe="/usr/sbin/sshd" hostname=9.79.215.218 addr=9.79.215.218 terminal=ssh res=failed' type=USER_AUTH msg=audit(1357836974.523:520): pid=16288 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=password acct=28696E76616C6964207573657229 exe="/usr/sbin/sshd" hostname=? addr=9.79.215.218 terminal=ssh res=failed' type=USER_AUTH msg=audit(1357836980.233:521): pid=16288 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=PAM:authentication acct="?" exe="/usr/sbin/sshd" hostname=9.79.215.218 addr=9.79.215.218 terminal=ssh res=failed' type=USER_AUTH msg=audit(1357836980.233:522): pid=16288 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=password acct=28696E76616C6964207573657229 exe="/usr/sbin/sshd" hostname=? addr=9.79.215.218 terminal=ssh res=failed' type=USER_LOGIN msg=audit(1357836980.233:526): pid=16288 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login acct=28696E76616C6964207573657229 exe="/usr/sbin/sshd" hostname=? addr=9.79.215.218 terminal=ssh res=failed' Please see for details log attached file audit.log. > It appears as though pacemaker is not connecting properly (which is why > you're getting the pcs errors). > > If there are errors, can you try temporarily disabling selinux and the > firewall to verify that they're not causing issues? > > To disable selinux, edit /etc/sysconfig/selinux and change the 'SELINUX=' > line to 'SELINUX=disabled' and restart. > > To disable the firewall run: > systemctl stop iptables.service > systemctl disable iptables.service Selinux & firewalls are already disabled. [root@c57f1ju0203 ~]# iptables -L Chain INPUT (policy ACCEPT) target prot opt source destination Chain FORWARD (policy ACCEPT) target prot opt source destination Chain OUTPUT (policy ACCEPT) target prot opt source destination [root@c57f1ju0203 ~]# systemctl status firewalld.service firewalld.service - firewalld - dynamic firewall daemon Loaded: loaded (/usr/lib/systemd/system/firewalld.service; disabled) Active: inactive (dead) CGroup: name=systemd:/system/firewalld.service Jan 10 09:25:39 c57f1ju0203.ppd.pok.ibm.com systemd[1]: Started firewalld - dynamic firewall daemon. Jan 10 09:55:30 c57f1ju0203.ppd.pok.ibm.com systemd[1]: Stopping firewalld - dynamic firewall daemon... Jan 10 09:55:36 c57f1ju0203.ppd.pok.ibm.com systemd[1]: Stopped firewalld - dynamic firewall daemon. [root@c57f1ju0203 ~]# getenforce Permissive Thanks... Manas
Can you give me the output of 'hostname' on all of your machines?
------- Comment From maknayak.com 2013-02-01 04:12 EDT------- (In reply to comment #12) > Can you give me the output of 'hostname' on all of your machines? Hello Chris, I have configured Short Node Names for convenience on both the nodes: NODE-1: [root@c57f1ju0204 ~]# hostname c57f1ju0204 [root@c57f1ju0204 ~]# uname -a Linux c57f1ju0204 3.6.10-4.fc18.ppc64p7 #1 SMP Wed Dec 12 16:08:02 MST 2012 ppc64 ppc64 ppc64 GNU/Linux NODE-2: [root@c57f1ju0203 ~]# hostname c57f1ju0203 Some more info: [root@c57f1ju0203 ~]# env | grep ais_ ais_mcast=239.255.1.1 ais_port=4000 ais_addr=9.114.32.0 [root@c57f1ju0204 ~]# env | grep -i ais ais_mcast=239.255.1.1 ais_port=4000 ais_addr=9.114.32.0 Thanks... Manas
------- Comment From maknayak.com 2013-03-13 09:17 EDT------- Chris, Any update on this issue ? Thanks... Manas
Manas, Sorry for the delay, I don't see anything wrong with your configuration. I've talked with one of the pacemaker developers and it sounds like several bugs are fixed with the latest 1.1.9 pacemaker release. Can you try a 'pcs cluster stop' on all your nodes and yum update to the latest pacemaker & corosync and then run 'pcs cluster start'? The latest version of pacemaker in fedora should be 1.1.9-0.1.70ad9fa.git.fc18 Let me know if you're getting the same issues. Thanks, Chris
------- Comment From maknayak.com 2013-06-15 08:31 EDT------- (In reply to comment #15) > Manas, > > Sorry for the delay, I don't see anything wrong with your configuration. > I've talked with one of the pacemaker developers and it sounds like several > bugs are fixed with the latest 1.1.9 pacemaker release. Can you try a 'pcs > cluster stop' on all your nodes and yum update to the latest pacemaker & > corosync and then run 'pcs cluster start'? > > The latest version of pacemaker in fedora should be > 1.1.9-0.1.70ad9fa.git.fc18 > > Let me know if you're getting the same issues. > > Thanks, > Chris Hi Chris, Sorry forthe late response. I am using Pacemaker 1.1.9-0.1.70ad9fa.git.fc19 , and verified on F19 Beta ppc64 kernel. Looks like still it has issue. Even if pacemaker gtets started , it shows several failures. Also I noticed crash in '/usr/libexec/pacemaker/crmd' gets produced in /var/log/messages as soon as pacemaker gets started. [root@miz03 cluster]# systemctl status corosync.service Loaded: loaded (/usr/lib/systemd/system/corosync.service; enabled) Active: active (running) since Sat 2013-06-15 04:05:54 EDT; 18min ago Process: 2991 ExecStop=/usr/share/corosync/corosync stop (code=exited, status=0/SUCCESS) Process: 3009 ExecStart=/usr/share/corosync/corosync start (code=exited, status=0/SUCCESS) Main PID: 3016 (corosync) ??3016 corosync Jun 15 04:05:54 miz03.austin.ibm.com corosync[3016]: [QB ] server name: quorum Jun 15 04:05:54 miz03.austin.ibm.com corosync[3016]: [QUORUM] Members[1]: -1351744759 Jun 15 04:05:54 miz03.austin.ibm.com corosync[3016]: [TOTEM ] A processor joined or left the membership and a new membership (9.3.110.175:16) was formed. Jun 15 04:05:54 miz03.austin.ibm.com corosync[3016]: [MAIN ] Completed service synchronization, ready to provide service. Jun 15 04:05:54 miz03.austin.ibm.com systemd[1]: Started Corosync Cluster Engine. Jun 15 04:05:57 miz03.austin.ibm.com corosync[3016]: [YKD ] Members[2]: 1299055369 -1351744759 Jun 15 04:05:57 miz03.austin.ibm.com corosync[3016]: [TOTEM ] A processor joined or left the membership and a new membership (9.3.110.77:24) was formed. Jun 15 04:05:57 miz03.austin.ibm.com corosync[3016]: [QUORUM] This node is within the primary component and will provide service. Jun 15 04:05:57 miz03.austin.ibm.com corosync[3016]: [YKD ] Members[2]: 1299055369 -1351744759 Jun 15 04:05:57 miz03.austin.ibm.com corosync[3016]: [YKD ] Completed service synchronization, ready to provide service. [root@miz03 cluster]# systemctl status pacemaker.service Loaded: loaded (/usr/lib/systemd/system/pacemaker.service; enabled) Active: active (running) since Sat 2013-06-15 04:07:57 EDT; 16min ago Main PID: 3031 (pacemakerd) ??3031 /usr/sbin/pacemakerd -f ??3032 /usr/libexec/pacemaker/cib ??3034 /usr/libexec/pacemaker/stonithd ??3035 /usr/libexec/pacemaker/lrmd ??3036 /usr/libexec/pacemaker/attrd ??3037 /usr/libexec/pacemaker/pengine Jun 15 04:10:34 miz03.austin.ibm.com crmd[3801]: error: do_log: FSA: Input I_TERMINATE from do_recover() received in state S_RECOVERY Jun 15 04:10:34 miz03.austin.ibm.com crmd[3801]: notice: do_lrm_control: Disconnected from the LRM Jun 15 04:10:34 miz03.austin.ibm.com crmd[3801]: notice: terminate_cs_connection: Disconnecting from Corosync Jun 15 04:10:34 miz03.austin.ibm.com cib[3032]: warning: qb_ipcs_event_sendv: new_event_notification (3032-3801-13): Broken pipe (32) Jun 15 04:10:34 miz03.austin.ibm.com crmd[3801]: error: do_exit: Could not recover from internal error Jun 15 04:10:34 miz03.austin.ibm.com cib[3032]: warning: do_local_notify: A-Sync reply to crmd failed: No message of desired type Jun 15 04:10:34 miz03.austin.ibm.com crmd[3801]: error: crm_abort: crm_glib_handler: Forked child 3802 to record non-fatal assert at logging.c:62 : g_hash_table_size: assertion `has...NULL' failed Jun 15 04:10:35 miz03.austin.ibm.com crmd[3801]: error: crm_abort: crm_glib_handler: Forked child 3804 to record non-fatal assert at logging.c:62 : g_hash_table_destroy: assertion `...NULL' failed Jun 15 04:10:36 miz03.austin.ibm.com pacemakerd[3031]: error: pcmk_child_exit: Child process crmd exited (pid=3801, rc=2) Jun 15 04:10:36 miz03.austin.ibm.com pacemakerd[3031]: error: pcmk_process_exit: Child respawn count exceeded by crmd You can see status for pacemaker service says lots of errors. --- Other Details --- [root@miz03 cluster]# uname -a Linux miz03.austin.ibm.com 3.9.4-300.fc19.ppc64p7 #1 SMP Sun May 26 22:14:56 MST 2013 ppc64 ppc64 ppc64 GNU/Linux [root@miz03 cluster]# rpm -qa | grep -i pacemaker pacemaker-libs-1.1.9-0.1.70ad9fa.git.fc19.ppc64 pacemaker-1.1.9-0.1.70ad9fa.git.fc19.ppc64 pacemaker-cluster-libs-1.1.9-0.1.70ad9fa.git.fc19.ppc64 pacemaker-libs-devel-1.1.9-0.1.70ad9fa.git.fc19.ppc64 pacemaker-cts-1.1.9-0.1.70ad9fa.git.fc19.ppc64 pacemaker-cli-1.1.9-0.1.70ad9fa.git.fc19.ppc64 pacemaker-doc-1.1.9-0.1.70ad9fa.git.fc19.ppc64 [root@miz03 cluster]# rpm -qa | grep -i corosync corosync-2.3.0-3.fc19.ppc64 corosynclib-2.3.0-3.fc19.ppc64 corosynclib-devel-2.3.0-3.fc19.ppc64 [root@miz03 cluster]# rpm -qa | grep -i crm crm114-emacs-0-5.14.20100106.fc19.ppc64 crm114-0-5.14.20100106.fc19.ppc64 [root@miz03 cluster]# ps axf | egrep "corosync|pacemaker" | grep -v egrep 3016 ? Ssl 0:08 corosync 3031 ? Ssl 0:00 /usr/sbin/pacemakerd -f 3032 ? Ssl 0:00 \_ /usr/libexec/pacemaker/cib 3034 ? Ss 0:00 \_ /usr/libexec/pacemaker/stonithd 3035 ? Ss 0:00 \_ /usr/libexec/pacemaker/lrmd 3036 ? Ss 0:00 \_ /usr/libexec/pacemaker/attrd 3037 ? Ss 0:00 \_ /usr/libexec/pacemaker/pengine [root@miz03 cluster]# corosync-cmapctl | grep members runtime.totem.pg.mrp.srp.members.1299055369.config_version (u64) = 0 runtime.totem.pg.mrp.srp.members.1299055369.ip (str) = r(0) ip(9.3.110.77) runtime.totem.pg.mrp.srp.members.1299055369.join_count (u32) = 1 runtime.totem.pg.mrp.srp.members.1299055369.status (str) = joined runtime.totem.pg.mrp.srp.members.2943222537.config_version (u64) = 0 runtime.totem.pg.mrp.srp.members.2943222537.ip (str) = r(0) ip(9.3.110.175) runtime.totem.pg.mrp.srp.members.2943222537.join_count (u32) = 1 runtime.totem.pg.mrp.srp.members.2943222537.status (str) = joined [root@miz03 cluster]# corosync-quorumtool -l Nodeid Votes Name 1299055369 1 miz05.austin.ibm.com 2943222537 1 miz03.austin.ibm.com (local) [root@miz03 cluster]# crm_mon -1 Last updated: Sat Jun 15 04:27:45 2013 Last change: Sat Jun 15 03:58:32 2013 0 Nodes configured, unknown expected votes Attached three files var-log-messages-F19.txt.tgz , dmesg-F19.txt & var-log-cluster-corosync.log.tgz for reference. Thanks... Manas
Created attachment 761526 [details] dmesg-F19.txt ------- Comment (attachment only) From maknayak.com 2013-06-15 08:32 EDT-------
Created attachment 761527 [details] var-log-messages.-F19.txt.tgz ------- Comment (attachment only) From maknayak.com 2013-06-15 08:33 EDT-------
Created attachment 761528 [details] var-log-cluster-corosync.log.tgz ------- Comment (attachment only) From maknayak.com 2013-06-15 08:34 EDT-------
This message is a reminder that Fedora 18 is nearing its end of life. Approximately 4 (four) weeks from now Fedora will stop maintaining and issuing updates for Fedora 18. It is Fedora's policy to close all bug reports from releases that are no longer maintained. At that time this bug will be closed as WONTFIX if it remains open with a Fedora 'version' of '18'. Package Maintainer: If you wish for this bug to remain open because you plan to fix it in a currently maintained version, simply change the 'version' to a later Fedora version prior to Fedora 18's end of life. Thank you for reporting this issue and we are sorry that we may not be able to fix it before Fedora 18 is end of life. If you would still like to see this bug fixed and are able to reproduce it against a later version of Fedora, you are encouraged change the 'version' to a later Fedora version prior to Fedora 18's end of life. Although we aim to fix as many bugs as possible during every release's lifetime, sometimes those efforts are overtaken by events. Often a more recent Fedora release includes newer upstream software that fixes bugs or makes them obsolete.
Fedora 18 changed to end-of-life (EOL) status on 2014-01-14. Fedora 18 is no longer maintained, which means that it will not receive any further security or bug fix updates. As a result we are closing this bug. If you can reproduce this bug against a currently maintained version of Fedora please feel free to reopen this bug against that version. If you are unable to reopen this bug, please file a new report against the current release. If you experience problems, please add a comment to this bug. Thank you for reporting this bug and we are sorry it could not be fixed.