[ceph-users] mount failed since failed to load ceph kernel module

Linh Vu vul at unimelb.edu.au
Mon Nov 13 18:24:06 PST 2017

Your kernel is way too old for CephFS Luminous. I'd use one of the newer kernels from elrepo.org. :) We're on 4.12 here on RHEL 7.4.

From: ceph-users <ceph-users-bounces at lists.ceph.com> on behalf of xiang.dai at sky-data.cn <xiang.dai at sky-data.cn>
Sent: Tuesday, 14 November 2017 1:13:47 PM
To: ceph-users at lists.ceph.com
Subject: [ceph-users] mount failed since failed to load ceph kernel module


I got a confused issue in docker as below:

After install ceph successfully, i want to mount cephfs but failed:

[root at dbffa72704e4 ~]$ /bin/mount<> /cephfs -t ceph -o name=admin,secretfile=/etc/ceph/admin.secret -v
failed to load ceph kernel module (1)
parsing options: rw,name=admin,secretfile=/etc/ceph/admin.secret
mount error 5 = Input/output error

But ceph related kernel modules have existed:

[root at dbffa72704e4 ~]$ lsmod | grep ceph
ceph                  327687  0
libceph               287066  1 ceph
dns_resolver           13140  2 nfsv4,libceph
libcrc32c              12644  3 xfs,libceph,dm_persistent_data

Check the ceph state(i only set data disk for osd):

[root at dbffa72704e4 ~]$ ceph -s
    id:     20f51975-303e-446f-903f-04e1feaff7d0
    health: HEALTH_WARN
            Reduced data availability: 128 pgs inactive
            Degraded data redundancy: 128 pgs unclean

    mon: 2 daemons, quorum dbffa72704e4,5807d12f920e
    mgr: dbffa72704e4(active), standbys: 5807d12f920e
    mds: cephfs-1/1/1 up  {0=5807d12f920e=up:creating}, 1 up:standby
    osd: 0 osds: 0 up, 0 in

    pools:   2 pools, 128 pgs
    objects: 0 objects, 0 bytes
    usage:   0 kB used, 0 kB / 0 kB avail
    pgs:     100.000% pgs unknown
             128 unknown

[root at dbffa72704e4 ~]$ ceph version
ceph version 12.2.1 (3e7492b9ada8bdc9a5cd0feafd42fbca27f9c38e) luminous (stable)

My container is based on centos:centos7.2.1511, kernel is 3e0728877e22 3.10.0-514.el7.x86_64.

I saw some ceph related images on docker hub so that i think above
operation is ok, did i miss something important?

Best Regards
Dai Xiang
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20171114/d632093c/attachment.html>

More information about the ceph-users mailing list