Bug 1275522 - ceph-fuse should be installed on openshift node
ceph-fuse should be installed on openshift node
Status: CLOSED NOTABUG
Product: OpenShift Container Platform
Classification: Red Hat
Component: Installer (Show other bugs)
3.0.0
Unspecified Unspecified
medium Severity medium
: ---
: ---
Assigned To: Jason DeTiberus
Ma xiaoqiang
:
Depends On:
Blocks:
  Show dependency treegraph
 
Reported: 2015-10-27 03:11 EDT by Anping Li
Modified: 2016-07-03 20:45 EDT (History)
4 users (show)

See Also:
Fixed In Version:
Doc Type: Bug Fix
Doc Text:
Story Points: ---
Clone Of:
Environment:
Last Closed: 2015-10-27 10:36:30 EDT
Type: Bug
Regression: ---
Mount Type: ---
Documentation: ---
CRM:
Verified Versions:
Category: ---
oVirt Team: ---
RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: ---


Attachments (Terms of Use)

  None (edit)
Description Anping Li 2015-10-27 03:11:28 EDT
Description of problem:
Ceph-common is ceph the server package. ceph-fuse packages must be installed, without this package, the ceph volume can be mounted. 

If ansible can't install ceph-fuse, it should report a message at the end.


Version-Release number of selected component (if applicable):
ansible pull request 738

How reproducible:
always

Steps to Reproduce:
1. install openshift by ansible
2. check the ceph packages after installation
3. mount ceph volume, it report wrong fs type.

Actual results:

1. ceph-common-0.80.7-3.el7.x86_64 was installed

2. Mount failed:
# mount -t ceph 10.66.79.110:6789:/ /mnt -o name=admin,secretfile=ceph.key
mount: wrong fs type, bad option, bad superblock on 10.66.79.110:6789:/,
       missing codepage or helper program, or other error

       In some cases useful info is found in syslog - try
       dmesg | tail or so.


Expect results:
The package "ceph-fuse.x86_64 : Ceph fuse-based client" should be installed.
Comment 1 Jason DeTiberus 2015-10-27 10:36:30 EDT
The storage plugin does not use the ceph fuse client, rather it uses the ceph kernel module, libraries and tools provided by ceph-common.

kubernetes docs for ceph rbd volume plugin: https://github.com/kubernetes/kubernetes/tree/master/examples/rbd

PR for the origin repo with examples for the ceph volume plugin: https://github.com/openshift/origin/pull/4772

To test mounting a volume, you will need to use the rbd map command:
rbd map <volume> --pool <pool> -m <ceph monitor host> --key=<admin user's key>

The rbd command will return the name of a block device that can then be mounted (generally /dev/rbd0 for the first device and incrementing for additional mounts).

Note You need to log in before you can comment on or make changes to this bug.