Hide Forgot
Description of problem: While Installing Mon on a ceph Node using Ansible, all other packages are getting Installed rpm -qa |grep ceph ceph-osd-10.0.4-2.el7cp.x86_64 ceph-ansible-1.0.3-1.el7.noarch python-cephfs-10.0.4-2.el7cp.x86_64 ceph-common-10.0.4-2.el7cp.x86_64 ceph-base-10.0.4-2.el7cp.x86_64 ceph-mds-10.0.4-2.el7cp.x86_64 ceph-10.0.4-2.el7cp.x86_64 libcephfs1-10.0.4-2.el7cp.x86_64 ceph-selinux-10.0.4-2.el7cp.x86_64 ceph-mon-10.0.4-2.el7cp.x86_64 Version-Release number of selected component (if applicable): ceph -v ceph version 10.0.4-2.el7cp (61adb020219fbad4508050b5f0a792246ba74dae) How reproducible: Always Steps to Reproduce: 1.Running ceph-ansible installation command i am seeing all the other packages are also getting installed. e.g. MDS, OSD.. Command: ansible-playbook site.yml -vv -i /etc/ansible/hosts --extra-vars '{"ceph_stable": true, "ceph_origin": "distro", "ceph_stable_rh_storage": true, "monitor_interface": "em3", "journal_collocation": true, "devices": ["/dev/sdi", "/dev/sdh"], "journal_size": 10000, "public_network": "10.70.44.0/22"}' -u roo Actual results: All packages are getting installed Expected results: In Mon node, other packages should not get installed Additional info: [mons] 10.70.44.40 [osds] 10.70.44.42 10.70.44.44 10.70.44.46
This is certainly not something we have seen before. You will need to investigate a bit more where these packages are coming from (what repo, from what repo file configuration etc...)
If ceph-ansible runs "yum install ceph" anywhere, then that could be the cause. The "ceph" package is a now a meta-package that will pull in all other packages. This is not causing yum errors in testing *yet* because we are not using errata-type puddles, so the ceph-10.0.4-2.el7cp.x86_64 RPM is still present in the puddle repositories. When we switch to delivering errata-type puddles, "yum install ceph" is going to fail.
*** Bug 1324229 has been marked as a duplicate of this bug. ***
@Alfredo,@Ken We enable the CEPH.repo on both mons and osd nodes, and like Ken mentioned the ceph ansible installs ceph package on all nodes. So could we differentiate if either a mon or osd is being installed, and install the necessary packages only. Thanks, Tejas
Verified on an ubuntu cluster: Moving to verified state ceph version 10.2.1-8redhat1xenial (0abd18aa5da0dde0f01404d8ac10876cb3691bb3) ceph-ansible-1.0.5-19.el7scon.noarch root@magna009:~# dpkg -l | grep ceph ii ceph-base 10.2.1-8redhat1xenial amd64 common ceph daemon libraries and management tools ii ceph-common 10.2.1-8redhat1xenial amd64 common utilities to mount and interact with a ceph storage cluster ii ceph-fs-common 10.2.1-8redhat1xenial amd64 common utilities to mount and interact with a ceph file system ii ceph-fuse 10.2.1-8redhat1xenial amd64 FUSE-based client for the Ceph distributed file system ii ceph-mds 10.2.1-8redhat1xenial amd64 metadata server for the ceph distributed file system ii ceph-mon 10.2.1-8redhat1xenial amd64 monitor server for the ceph storage system ii libcephfs1 10.2.1-8redhat1xenial amd64 Ceph distributed file system client library ii python-cephfs 10.2.1-8redhat1xenial amd64 Python libraries for the Ceph libcephfs library root@magna009:~# root@magna031:~# dpkg -l | grep ceph ii ceph-base 10.2.1-8redhat1xenial amd64 common ceph daemon libraries and management tools ii ceph-common 10.2.1-8redhat1xenial amd64 common utilities to mount and interact with a ceph storage cluster ii ceph-fs-common 10.2.1-8redhat1xenial amd64 common utilities to mount and interact with a ceph file system ii ceph-fuse 10.2.1-8redhat1xenial amd64 FUSE-based client for the Ceph distributed file system ii ceph-mds 10.2.1-8redhat1xenial amd64 metadata server for the ceph distributed file system ii ceph-osd 10.2.1-8redhat1xenial amd64 OSD server for the ceph storage system ii libcephfs1 10.2.1-8redhat1xenial amd64 Ceph distributed file system client library ii python-cephfs 10.2.1-8redhat1xenial amd64 Python libraries for the Ceph libcephfs library root@magna031:~#
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2016:1754