Description of problem: Foreman installs should allow RHS or NFS to be used to hold images, and also allow for direct file access to those images. Version-Release number of selected component (if applicable): How reproducible: Steps to Reproduce: 1. 2. 3. Actual results: Expected results: Additional info:
glance and cinder should be separate. for glance, let separate into 2 categories: File: I think the only difference between nfs and gfs is the mount type, options, required packages. This is the command I use on my current controller. ssh ${CONFIG_GLANCE_HOST} "mkdir -p /var/lib/glance;chown 161.161 /var/lib/glance" ssh ${CONFIG_GLANCE_HOST} "echo 'ra-rhs-srv1-10g.storage.se.priv:/OSTACKglance /var/lib/glance glusterfs _netdev,selinux,backup-volfile-servers=ra-rhs-srv2-10g.storage.se.priv 0 0' >> /etc/fstab" ssh ${CONFIG_GLANCE_HOST} "mount /var/lib/glance" direct file Enabling this mean the images will not be transferred over http using the API endpoint but copying the file 'locally' This requires this on the controller: ssh ${CONFIG_GLANCE_HOST} "openstack-config --set /etc/glance/glance-api.conf DEFAULT show_image_direct_url True" ssh ${CONFIG_NOVA_API_HOST} "openstack-config --set /etc/nova/nova.conf DEFAULT allowed_direct_url_schemes \[file\]" An this on the computes: ssh $i "mkdir -p /var/lib/glance;chown 161.161 /var/lib/glance" ssh $i "echo 'ra-rhs-srv1-10g.storage.se.priv:/OSTACKglance /var/lib/glance glusterfs _netdev,selinux,backup-volfile-servers=ra-r hs-srv2-10g.storage.se.priv 0 0' >> /etc/fstab" ssh $i "mount /var/lib/glance" ssh $i "openstack-config --set /etc/nova/nova.conf DEFAULT allowed_direct_url_schemes \[file\]"
I just remembered, this is how I do the mount using pcs: ssh $HA_FIRST "pcs resource create glance-fs Filesystem device=\"ra-rhs-srv1-10g.storage.se.priv:/OSTACKglance\" directory=\"/var/lib/glance/\" fstype=\"glusterfs\" options=\"selinux,backup-volfile-servers=ra-rhs-srv2-10g.storage.se.priv\" --clone"
This BZ was cloned to split the 2 requests separately. One request was for nfs/rhs enablement for glance. That half is staying with this BZ and should stay with RHOS 4 A4. I believe this will be needed for HA. The direct file enablement is something that is moved to BZ 1082785 which is now targeted at RHOS 5.
configured Foreman: change this parameters 1) pcmk_fs_device = 10.35.64.106://nlevinki_glance 2) pcmk_fs_type = glusterfs 3) backend = file (this is the default config) 4) pcmk_fs_manage = true (this is the default config) installation of openstack failed Notice: /Stage[main]/Pacemaker::Corosync/Exec[Start Cluster openstack]/returns: [FAILED] Notice: /Stage[main]/Pacemaker::Corosync/Exec[Start Cluster openstack]/returns: Stopping cluster: Notice: /Stage[main]/Pacemaker::Corosync/Exec[Start Cluster openstack]/returns: Leaving fence domain... [ OK ] Notice: /Stage[main]/Pacemaker::Corosync/Exec[Start Cluster openstack]/returns: Stopping gfs_controld... [ OK ] Notice: /Stage[main]/Pacemaker::Corosync/Exec[Start Cluster openstack]/returns: Stopping dlm_controld... [ OK ] Notice: /Stage[main]/Pacemaker::Corosync/Exec[Start Cluster openstack]/returns: Stopping fenced... [ OK ] Notice: /Stage[main]/Pacemaker::Corosync/Exec[Start Cluster openstack]/returns: Stopping cman... [ OK ] Notice: /Stage[main]/Pacemaker::Corosync/Exec[Start Cluster openstack]/returns: Unloading kernel modules... [ OK ] Notice: /Stage[main]/Pacemaker::Corosync/Exec[Start Cluster openstack]/returns: Unmounting configfs... [ OK ] Error: /usr/sbin/pcs cluster start returned 1 instead of one of [0] Error: /Stage[main]/Pacemaker::Corosync/Exec[Start Cluster openstack]/returns: change from notrun to 0 failed: /usr/sbin/pcs cluster start returned 1 instead of one of [0] Notice: /Stage[main]/Pacemaker::Corosync/Exec[wait-for-settle]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Pacemaker::Corosync/Exec[wait-for-settle]: Skipping because of failed dependencies Notice: /Stage[main]/Pacemaker::Corosync/Notify[pacemaker settled]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Pacemaker::Corosync/Notify[pacemaker settled]: Skipping because of failed dependencies Notice: /Stage[main]/Pacemaker::Stonith/Exec[Disable STONITH]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Pacemaker::Stonith/Exec[Disable STONITH]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Pacemaker::Common/Exec[stonith-setup-complete]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Pacemaker::Common/Exec[stonith-setup-complete]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Hamysql::Node/Exec[wait-for-fs-to-be-active]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Hamysql::Node/Exec[wait-for-fs-to-be-active]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Hamysql::Node/Exec[sleep-so-really-sure-fs-is-mounted]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Hamysql::Node/Exec[sleep-so-really-sure-fs-is-mounted]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Hamysql::Node/Exec[create-socket-symlink-if-we-own-the-mount]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Hamysql::Node/Exec[create-socket-symlink-if-we-own-the-mount]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Hamysql::Node/Exec[wait-for-mysql-to-start]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Hamysql::Node/Exec[wait-for-mysql-to-start]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Hamysql::Node/File[are-we-running-mysql-script]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Hamysql::Node/File[are-we-running-mysql-script]: Skipping because of failed dependencies Notice: /File[/var/lib/puppet/concat/_etc_haproxy_haproxy.cfg/fragments/20-mysql-00_mysql_listen_block]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /File[/var/lib/puppet/concat/_etc_haproxy_haproxy.cfg/fragments/20-mysql-00_mysql_listen_block]: Skipping because of failed dependencies Notice: /File[/var/lib/puppet/concat/_etc_haproxy_haproxy.cfg/fragments/20-mysql-mysql_mysql_balancermember_mysql]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /File[/var/lib/puppet/concat/_etc_haproxy_haproxy.cfg/fragments/20-mysql-mysql_mysql_balancermember_mysql]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Hamysql::Mysql::Rootpw/Exec[set_mysql_rootpw]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Hamysql::Mysql::Rootpw/Exec[set_mysql_rootpw]: Skipping because of failed dependencies Notice: /File[/root/.my.cnf]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /File[/root/.my.cnf]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Pacemaker::Common/File[ha-all-in-one-util-bash-tests]/ensure: defined content as '{md5}73541c24158939763c0bddefb71a52e8' Notice: /Stage[main]/Quickstack::Pacemaker::Memcached/Exec[pcs-memcached-server-set-up-on-this-node]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Pacemaker::Memcached/Exec[pcs-memcached-server-set-up-on-this-node]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Pacemaker::Memcached/Exec[all-memcached-nodes-are-up]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Pacemaker::Memcached/Exec[all-memcached-nodes-are-up]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Load_balancer::Common/Sysctl::Value[net.ipv4.ip_nonlocal_bind]/Sysctl[net.ipv4.ip_nonlocal_bind]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Load_balancer::Common/Sysctl::Value[net.ipv4.ip_nonlocal_bind]/Sysctl[net.ipv4.ip_nonlocal_bind]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Load_balancer::Common/Sysctl::Value[net.ipv4.ip_nonlocal_bind]/Exec[exec_sysctl_net.ipv4.ip_nonlocal_bind]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Load_balancer::Common/Sysctl::Value[net.ipv4.ip_nonlocal_bind]/Exec[exec_sysctl_net.ipv4.ip_nonlocal_bind]: Skipping because of failed dependencies Notice: /File[/var/lib/puppet/concat/_etc_haproxy_haproxy.cfg/fragments/20-stats-00_stats_listen_block]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /File[/var/lib/puppet/concat/_etc_haproxy_haproxy.cfg/fragments/20-stats-00_stats_listen_block]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Pacemaker::Qpid/Qpid_user[openstack]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Pacemaker::Qpid/Qpid_user[openstack]: Skipping because of failed dependencies Notice: /File[/var/lib/puppet/concat/_etc_haproxy_haproxy.cfg/fragments/20-qpid-00_qpid_listen_block]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /File[/var/lib/puppet/concat/_etc_haproxy_haproxy.cfg/fragments/20-qpid-00_qpid_listen_block]: Skipping because of failed dependencies Notice: /File[/var/lib/puppet/concat/_etc_haproxy_haproxy.cfg/fragments/20-qpid-qpid_qpid_balancermember_qpid]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /File[/var/lib/puppet/concat/_etc_haproxy_haproxy.cfg/fragments/20-qpid-qpid_qpid_balancermember_qpid]: Skipping because of failed dependencies Notice: /Stage[main]/Haproxy/Concat[/etc/haproxy/haproxy.cfg]/Exec[concat_/etc/haproxy/haproxy.cfg]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Haproxy/Concat[/etc/haproxy/haproxy.cfg]/Exec[concat_/etc/haproxy/haproxy.cfg]: Skipping because of failed dependencies Notice: /Stage[main]/Haproxy/Concat[/etc/haproxy/haproxy.cfg]/Exec[concat_/etc/haproxy/haproxy.cfg]: Triggered 'refresh' from 4 events Notice: /File[/etc/haproxy/haproxy.cfg]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /File[/etc/haproxy/haproxy.cfg]: Skipping because of failed dependencies Info: Concat[/etc/haproxy/haproxy.cfg]: Scheduling refresh of Service[haproxy] Notice: /Stage[main]/Quickstack::Pacemaker::Qpid/Exec[pcs-qpid-server-set-up-on-this-node]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Pacemaker::Qpid/Exec[pcs-qpid-server-set-up-on-this-node]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Pacemaker::Qpid/Exec[all-qpid-nodes-are-up]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Pacemaker::Qpid/Exec[all-qpid-nodes-are-up]: Skipping because of failed dependencies Notice: /Stage[main]/Haproxy/Service[haproxy]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Haproxy/Service[haproxy]: Skipping because of failed dependencies Notice: /Stage[main]/Haproxy/Service[haproxy]: Triggered 'refresh' from 1 events Notice: /Stage[main]/Quickstack::Pacemaker::Load_balancer/Exec[pcs-haproxy-server-set-up-on-this-node]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Pacemaker::Load_balancer/Exec[pcs-haproxy-server-set-up-on-this-node]: Skipping because of failed dependencies Notice: /Stage[main]/Quickstack::Pacemaker::Load_balancer/Exec[all-haproxy-nodes-are-up]: Dependency Exec[Start Cluster openstack] has failures: true Warning: /Stage[main]/Quickstack::Pacemaker::Load_balancer/Exec[all-haproxy-nodes-are-up]: Skipping because of failed dependencies Notice: Finished catalog run in 731.18 seconds [root@dhcp160-217 ~]#
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. http://rhn.redhat.com/errata/RHSA-2014-0517.html