Bug 1613038
| Summary: | [RFE][Cinder] RBD driver support to get manageable snapshots | ||
|---|---|---|---|
| Product: | Red Hat OpenStack | Reporter: | Sean Cohen <scohen> |
| Component: | openstack-cinder | Assignee: | Eric Harney <eharney> |
| Status: | CLOSED ERRATA | QA Contact: | Tzach Shefi <tshefi> |
| Severity: | medium | Docs Contact: | Kim Nylander <knylande> |
| Priority: | medium | ||
| Version: | 14.0 (Rocky) | CC: | gcharot, gregraka, mabrams, srevivo, tshefi |
| Target Milestone: | Upstream M1 | Keywords: | FutureFeature, TestOnly, Triaged |
| Target Release: | 15.0 (Stein) | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| URL: | https://blueprints.launchpad.net/cinder/+spec/ceph-list-manageable-volumes-and-snapshots | ||
| Whiteboard: | |||
| Fixed In Version: | openstack-cinder-14.0.1-0.20190420004000.84d3d12.el8ost | Doc Type: | Release Note |
| Doc Text: |
The Block Storage service (cinder) command, "snapshot-manageable-list," now lists the snapshots on the back end for Red Hat Ceph RADOS block devices (RBD).
|
Story Points: | --- |
| Clone Of: | Environment: | ||
| Last Closed: | 2019-09-21 11:16:49 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Sean Cohen
2018-08-06 20:14:39 UTC
Verified on:
openstack-cinder-14.0.1-0.20190712060430.0996f0a.el8ost.noarch
1. Create a volume (rbd backed)
(overcloud) [stack@undercloud-0 ~]$ cinder create 1 --image cirros
+--------------------------------+--------------------------------------+
| Property | Value |
+--------------------------------+--------------------------------------+
| attachments | [] |
| availability_zone | nova |
| bootable | false |
| consistencygroup_id | None |
| created_at | 2019-08-07T08:58:48.000000 |
| description | None |
| encrypted | False |
| id | 715a5c7c-b5ae-4a6e-a6b7-514329b97aa6 |
| metadata | {} |
| migration_status | None |
| multiattach | False |
| name | None |
| os-vol-host-attr:host | hostgroup@tripleo_ceph#tripleo_ceph |
| os-vol-mig-status-attr:migstat | None |
| os-vol-mig-status-attr:name_id | None |
| os-vol-tenant-attr:tenant_id | 5d7e49c661ae42e498dcbbfa8098a9f8 |
| replication_status | None |
| size | 1 |
| snapshot_id | None |
| source_volid | None |
| status | creating |
| updated_at | 2019-08-07T08:58:48.000000 |
| user_id | f32508264e86488abd194ef9ded8a4b2 |
| volume_type | tripleo |
+--------------------------------+--------------------------------------+
2. Create a snapshot of said volume:
(overcloud) [stack@undercloud-0 ~]$ cinder snapshot-create 715a5c7c-b5ae-4a6e-a6b7-514329b97aa6 --name snap1
+-------------+--------------------------------------+
| Property | Value |
+-------------+--------------------------------------+
| created_at | 2019-08-07T08:59:38.236502 |
| description | None |
| id | 69deb14a-36f5-48e0-804c-c1b908cf0711 |
| metadata | {} |
| name | snap1 |
| size | 1 |
| status | creating |
| updated_at | None |
| volume_id | 715a5c7c-b5ae-4a6e-a6b7-514329b97aa6 |
+-------------+--------------------------------------+
3. Check snapshot status:
(overcloud) [stack@undercloud-0 ~]$ cinder snapshot-list
+--------------------------------------+--------------------------------------+-----------+-------+------+
| ID | Volume ID | Status | Name | Size |
+--------------------------------------+--------------------------------------+-----------+-------+------+
| 69deb14a-36f5-48e0-804c-c1b908cf0711 | 715a5c7c-b5ae-4a6e-a6b7-514329b97aa6 | available | snap1 | 1 |
+--------------------------------------+--------------------------------------+-----------+-------+------+
4. Note this is only supported on/from micro version 3.8, but we see that now rbd snapshots also showup as managed/manageable.
(overcloud) [stack@undercloud-0 ~]$ cinder --os-volume-api-version 3.8 snapshot-manageable-list hostgroup@tripleo_ceph
+------------------------------------------------------------------+------+----------------+----------------------------------------------------------------+-----------------+--------------------------------------+------------+
| reference | size | safe_to_manage | source_reference | reason_not_safe | cinder_id | extra_info |
+------------------------------------------------------------------+------+----------------+----------------------------------------------------------------+-----------------+--------------------------------------+------------+
| {'source-name': 'snapshot-69deb14a-36f5-48e0-804c-c1b908cf0711'} | 1 | False | {'source-name': 'volume-715a5c7c-b5ae-4a6e-a6b7-514329b97aa6'} | already managed | 69deb14a-36f5-48e0-804c-c1b908cf0711 | - |
+------------------------------------------------------------------+------+----------------+----------------------------------------------------------------+-----------------+--------------------------------------+------------+
Looks good.
Added another test which skipped my mind before
1. Unmanage the snapshot from before:
(overcloud) [stack@undercloud-0 ~]$ cinder snapshot-unmanage 69deb14a-36f5-48e0-804c-c1b908cf0711
We see no more snapshots on openstack:
(overcloud) [stack@undercloud-0 ~]$ cinder snapshot-list
+----+-----------+--------+------+------+
| ID | Volume ID | Status | Name | Size |
+----+-----------+--------+------+------+
+----+-----------+--------+------+------+
2. Recheck snapshot-manageable-list
(overcloud) [stack@undercloud-0 ~]$ cinder --os-volume-api-version 3.8 snapshot-manageable-list hostgroup@tripleo_ceph
+------------------------------------------------------------------+------+----------------+----------------------------------------------------------------+-----------------+-----------+------------+
| reference | size | safe_to_manage | source_reference | reason_not_safe | cinder_id | extra_info |
+------------------------------------------------------------------+------+----------------+----------------------------------------------------------------+-----------------+-----------+------------+
| {'source-name': 'snapshot-69deb14a-36f5-48e0-804c-c1b908cf0711'} | 1 | True | {'source-name': 'volume-715a5c7c-b5ae-4a6e-a6b7-514329b97aa6'} | - | - | - |
+------------------------------------------------------------------+------+----------------+----------------------------------------------------------------+-----------------+-----------+------------+
Meaning we're able to see snapshots on the backend which have been removed are now longer a part of openstack at the moment.
Another test this time snap initiated by ceph.
1. Find the volume
bash-4.4# rbd -p volumes ls
volume-715a5c7c-b5ae-4a6e-a6b7-514329b97aa6
2, Find the snapshot we created before:
bash-4.4# rbd snap ls volumes/volume-715a5c7c-b5ae-4a6e-a6b7-514329b97aa6
SNAPID NAME SIZE PROTECTED TIMESTAMP
4 snapshot-69deb14a-36f5-48e0-804c-c1b908cf0711 1 GiB Wed Aug 7 08:59:38 2019
3. Create a snapshot of said volume this time from ceph's side:
rbd snap create volumes/volume-715a5c7c-b5ae-4a6e-a6b7-514329b97aa6@snap2
4. Show again snapshots:
bash-4.4# rbd snap ls volumes/volume-715a5c7c-b5ae-4a6e-a6b7-514329b97aa6
SNAPID NAME SIZE PROTECTED TIMESTAMP
4 snapshot-69deb14a-36f5-48e0-804c-c1b908cf0711 1 GiB Wed Aug 7 08:59:38 2019
5 snap2 1 GiB Thu Aug 8 05:49:29 2019
We see snap2 added/created on ceph's side.
5. Returning to openstack we should see these two snapshots, as expected we now see both of them:
(overcloud) [stack@undercloud-0 ~]$ cinder --os-volume-api-version 3.8 snapshot-manageable-list hostgroup@tripleo_ceph
+------------------------------------------------------------------+------+----------------+----------------------------------------------------------------+-----------------+-----------+------------+
| reference | size | safe_to_manage | source_reference | reason_not_safe | cinder_id | extra_info |
+------------------------------------------------------------------+------+----------------+----------------------------------------------------------------+-----------------+-----------+------------+
| {'source-name': 'snapshot-69deb14a-36f5-48e0-804c-c1b908cf0711'} | 1 | True | {'source-name': 'volume-715a5c7c-b5ae-4a6e-a6b7-514329b97aa6'} | - | - | - |
| {'source-name': 'snap2'} | 1 | True | {'source-name': 'volume-715a5c7c-b5ae-4a6e-a6b7-514329b97aa6'} | - | - | - |
+------------------------------------------------------------------+------+----------------+----------------------------------------------------------------+-----------------+-----------+------------+
Verified now even better :)
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2019:2811 |