Bug 1750005

Summary: check_running_containers does not consider pacemaker managed ceph-nfs
Product: [Red Hat Storage] Red Hat Ceph Storage Reporter: Giulio Fidente <gfidente>
Component: Ceph-AnsibleAssignee: Giulio Fidente <gfidente>
Status: CLOSED ERRATA QA Contact: Yogev Rabl <yrabl>
Severity: high Docs Contact: Bara Ancincova <bancinco>
Priority: high    
Version: 3.3CC: agunn, aschoen, ceph-eng-bugs, ceph-qe-bugs, dsavinea, gmeno, nthomas, nweinber, pasik, tbarron, tchandra, tserlin, vashastr, ykaul, yrabl
Target Milestone: z1   
Target Release: 3.3   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: RHEL: ceph-ansible-3.2.27-1.el7cp Ubuntu: ceph-ansible_3.2.27-2redhat1 Doc Type: Bug Fix
Doc Text:
.Ceph Ansible can now successfully updates and restarts NFS Ganesha container when a custom suffix is used for the container name Previously, the value set for the `ceph_nfs_service_suffix` variable was not considered when checking the status and version of the Ceph NFS Ganesha (`ceph-nfs`) container for restart or update. Consequently, the `ceph-nfs` container was not updated or restarted because the `ceph-ansible` utility could not determine that the container was running. With this update, `ceph-ansible` uses the value of `ceph_nfs_service_suffix` to determine the status of the `ceph-nfs` container. As a result, `ceph-nfs` container is successfully updated or restarted as expected.
Story Points: ---
Clone Of: Environment:
Last Closed: 2019-10-22 13:29:30 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On:    
Bug Blocks: 1578730, 1726135    

Description Giulio Fidente 2019-09-07 07:44:14 UTC
The action which gathers status of existing containers in ceph-ansible is failing when the container is managed by pacemaker

2019-09-06 23:32:05,058 p=29950 u=mistral |  TASK [ceph-handler : check for a nfs container] ********************************
2019-09-06 23:32:05,058 p=29950 u=mistral |  task path: /usr/share/ceph-ansible/roles/ceph-handler/tasks/check_running_containers.yml:56
2019-09-06 23:32:05,059 p=29950 u=mistral |  Friday 06 September 2019  23:32:05 -0400 (0:00:00.219)       0:05:38.596 ****** 
2019-09-06 23:32:05,152 p=29950 u=mistral |  Using module file /usr/lib/python2.7/site-packages/ansible/modules/commands/command.py
2019-09-06 23:32:05,220 p=29950 u=mistral |  Escalation succeeded
2019-09-06 23:32:05,396 p=29950 u=mistral |  ok: [192.168.81.52] => {
    "changed": false, 
    "cmd": [
        "docker", 
        "ps", 
        "-q", 
        "--filter=name=ceph-nfs-overcloud-controller-0"
    ], 
    "delta": "0:00:00.030883", 
    "end": "2019-09-06 23:32:05.366332", 
    "failed_when_result": false, 
    "invocation": {
        "module_args": {
            "_raw_params": "docker ps -q --filter='name=ceph-nfs-overcloud-controller-0'", 
            "_uses_shell": false, 
            "argv": null, 
            "chdir": null, 
            "creates": null, 
            "executable": null, 
            "removes": null, 
            "stdin": null, 
            "warn": true
        }
    }, 
    "rc": 0, 
    "start": "2019-09-06 23:32:05.335449", 
    "stderr": "", 
    "stderr_lines": [], 
    "stdout": "", 
    "stdout_lines": []
}

the filter should also consider ceph-nfs-pacemaker to find a match.

Comment 1 RHEL Program Management 2019-09-07 07:44:21 UTC
Please specify the severity of this bug. Severity is defined here:
https://bugzilla.redhat.com/page.cgi?id=fields.html#bug_severity.

Comment 2 Giulio Fidente 2019-09-07 07:52:20 UTC
Needs to be backported in RHCS3 as well

Comment 17 Yogev Rabl 2019-10-17 14:07:31 UTC
Verified on rhceph:ceph-3.3-rhel-7-containers-candidate-71606-20191015210325 
and ceph-ansible-3.2.30-1.el7cp.noarch

Comment 19 errata-xmlrpc 2019-10-22 13:29:30 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2019:3173