I am going to attempt to answer my own question here and someone can correct me 
if I am wrong.

Looking at a few of the other OSDs that we have replaced over the last year or 
so it looks like they are mounted using tmpfs as well and that this is just a 
result of switching from filestore to bluestore and that this is really nothing 
to worry about.

Thanks,
Shain



On 9/9/20, 11:16 AM, "Shain Miley" <smi...@npr.org> wrote:

    Hi,
    I recently added 3 new servers to Ceph cluster.  These servers use the 
H740p mini raid card and I had to install the HWE kernel in Ubuntu 16.04 in 
order to get the drives recognized.


    We have a 23 node cluster and normally when we add OSDs they end up 
mounting like this:

    /dev/sde1       3.7T  2.0T  1.8T  54% /var/lib/ceph/osd/ceph-15

    /dev/sdj1       3.7T  2.0T  1.7T  55% /var/lib/ceph/osd/ceph-20

    /dev/sdd1       3.7T  2.1T  1.6T  58% /var/lib/ceph/osd/ceph-14

    /dev/sdc1       3.7T  1.8T  1.9T  49% /var/lib/ceph/osd/ceph-13



    However I noticed this morning that the 3 new servers have the OSDs mounted 
like this:

    tmpfs            47G   28K   47G   1% /var/lib/ceph/osd/ceph-246

    tmpfs            47G   28K   47G   1% /var/lib/ceph/osd/ceph-240

    tmpfs            47G   28K   47G   1% /var/lib/ceph/osd/ceph-248

    tmpfs            47G   28K   47G   1% /var/lib/ceph/osd/ceph-237


    Is this normal for deployments going forward…or did something go wrong?  
These are 12TB drives but they are showing up as 47G here instead.


    We are using ceph version 12.2.13 and I installed this using ceph-deply 
version 2.0.1.



    Thanks in advance,



    Shain

    Shain Miley | Director of Platform and Infrastructure | Digital Media | 
smi...@npr.org
    _______________________________________________
    ceph-users mailing list -- ceph-users@ceph.io
    To unsubscribe send an email to ceph-users-le...@ceph.io

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to