According to dpkg, all of the cluster is running 12.2.5, so it is strange that 
one would report as 12.2.4.  I am certain that everything is on Luminous as its 
a new environment.

The problem seems to have been resolved by changing the trimming settings and 
rolling restarting the mds daemons.

— 
Doug Bell
Electric Alchemy Limited
Office: +1 303.578.6796 — Direct: +1 720.689.1776
fprint: 5DD2 B48A BC78 C1E6 BB17 C588 C790 72A2 7841 220E

> On May 30, 2018, at 5:39 PM, Gregory Farnum <gfar...@redhat.com> wrote:
> 
> Given what you've shown here, it's probably one of the odder cases CephFS is 
> subject to, rather than an actual "there's no disk space" error. How far is 
> the script actually getting? Is it possible your client doesn't have 
> permission to write to the RADOS pool and isn't finding that out until too 
> late?
> 
> ...actually, hrm, is your cluster running a consistent version? I see the 
> MDSes are different point releases, and if you've got some Jewel daemons in 
> the mix, there are a bunch more cases that could be applying under the old 
> behavior.
> -Greg
> 
> On Wed, May 30, 2018 at 5:36 AM Doug Bell <d...@electricalchemy.net 
> <mailto:d...@electricalchemy.net>> wrote:
> I am new to Ceph and have built a small Ceph instance on 3 servers.  I 
> realize the configuration is probably not ideal but I’d like to understand an 
> error I’m getting.
> 
> Ceph hosts are cm1, cm2, cm3.  Cephfs is mounted with ceph.fuse on a server 
> c1.  I am attempting to perform a simple cp-rp from one directory tree 
> already in cephfs to another directory also inside of cephfs.  The directory 
> tree is 2740 files totaling 93G.  Approximately 3/4 of the way through the 
> copy, the following error occurs:  "cp: failed to close ‘<filename>': No 
> space left on device”  The odd thing is that it seems to finish the copy, as 
> the final directory sizes are the same.  But scripts attached to the process 
> see an error so it is causing a problem.
> 
> Any idea what is happening?  I have watched all of the ceph logs on one of 
> the ceph servers and haven’t seen anything.
> 
> Here is some of the configuration.  The names actually aren’t obfuscated, 
> they really are that generic.  IP Addresses are altered though.
> 
> # ceph fs ls
> name: cephfs, metadata pool: cephfs_metadata, data pools: [cephfs_data ]
> 
> # ceph status
>  cluster:
>    id:     c14e77f1-9898-48d8-8a52-cd1f1c5bf689
>    health: HEALTH_WARN
>            1 MDSs behind on trimming
> 
>  services:
>    mon: 3 daemons, quorum cm1,cm3,cm2
>    mgr: cm3(active), standbys: cm2, cm1
>    mds: cephfs-1/1/1 up  {0=cm1=up:active}, 1 up:standby-replay, 1 up:standby
>    osd: 7 osds: 7 up, 7 in
> 
>  data:
>    pools:   2 pools, 256 pgs
>    objects: 377k objects, 401 GB
>    usage:   1228 GB used, 902 GB / 2131 GB avail
>    pgs:     256 active+clean
> 
>  io:
>    client:   852 B/s rd, 2 op/s rd, 0 op/s wr
> 
> # ceph osd status
> +----+------+-------+-------+--------+---------+--------+---------+-----------+
> | id | host |  used | avail | wr ops | wr data | rd ops | rd data |   state   
> |
> +----+------+-------+-------+--------+---------+--------+---------+-----------+
> | 0  | cm1  |  134G |  165G |    0   |     0   |    0   |     0   | exists,up 
> |
> | 1  | cm1  |  121G |  178G |    0   |     0   |    0   |     0   | exists,up 
> |
> | 2  | cm2  |  201G | 98.3G |    0   |     0   |    1   |    90   | exists,up 
> |
> | 3  | cm2  |  207G | 92.1G |    0   |     0   |    0   |     0   | exists,up 
> |
> | 4  | cm3  |  217G | 82.8G |    0   |     0   |    0   |     0   | exists,up 
> |
> | 5  | cm3  |  192G |  107G |    0   |     0   |    0   |     0   | exists,up 
> |
> | 6  | cm1  |  153G |  177G |    0   |     0   |    1   |    16   | exists,up 
> |
> +----+------+-------+-------+--------+---------+--------+---------+—————+
> 
> # ceph osd df
> ID CLASS WEIGHT  REWEIGHT SIZE  USE   AVAIL   %USE  VAR  PGS
> 0   ssd 0.29300  1.00000  299G  134G    165G 44.74 0.78  79
> 1   ssd 0.29300  1.00000  299G  121G    178G 40.64 0.70  75
> 6   ssd 0.32370  1.00000  331G  153G    177G 46.36 0.80 102
> 2   ssd 0.29300  1.00000  299G  201G 100754M 67.20 1.17 129
> 3   ssd 0.29300  1.00000  299G  207G  94366M 69.28 1.20 127
> 4   ssd 0.29300  1.00000  299G  217G  84810M 72.39 1.26 131
> 5   ssd 0.29300  1.00000  299G  192G    107G 64.15 1.11 125
>                    TOTAL 2131G 1228G    902G 57.65
> MIN/MAX VAR: 0.70/1.26  STDDEV: 12.36
> 
> # ceph fs get cephfs
> Filesystem 'cephfs' (1)
> fs_name cephfs
> epoch   1047
> flags   c
> created 2018-03-20 13:58:51.860813
> modified        2018-03-20 13:58:51.860813
> tableserver     0
> root    0
> session_timeout 60
> session_autoclose       300
> max_file_size   1099511627776
> last_failure    0
> last_failure_osd_epoch  98
> compat  compat={},rocompat={},incompat={1=base v0.20,2=client writeable 
> ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds 
> uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file 
> layout v2}
> max_mds 1
> in      0
> up      {0=74127}
> failed
> damaged
> stopped
> data_pools      [1]
> metadata_pool   2
> inline_data     disabled
> balancer
> standby_count_wanted    1
> 74127:  10.1.2.157:6800/3141645279 <http://10.1.2.157:6800/3141645279> 'cm1' 
> mds.0.36 up:active seq 5 (standby for rank 0)
> 64318:  10.1.2.194:6803/2623342769 <http://10.1.2.194:6803/2623342769> 'cm2' 
> mds.0.0 up:standby-replay seq 497658 (standby for rank 0)
> 
> # ceph fs status
> cephfs - 9 clients
> ======
> +------+----------------+-----+---------------+-------+-------+
> | Rank |     State      | MDS |    Activity   |  dns  |  inos |
> +------+----------------+-----+---------------+-------+-------+
> |  0   |     active     | cm1 | Reqs:    0 /s |  295k |  292k |
> | 0-s  | standby-replay | cm2 | Evts:    0 /s |    0  |    0  |
> +------+----------------+-----+---------------+-------+-------+
> +-----------------+----------+-------+-------+
> |       Pool      |   type   |  used | avail |
> +-----------------+----------+-------+-------+
> | cephfs_metadata | metadata |  167M |  160G |
> |   cephfs_data   |   data   |  401G |  160G |
> +-----------------+----------+-------+-------+
> 
> +-------------+
> | Standby MDS |
> +-------------+
> |     cm3     |
> +-------------+
> +----------------------------------------------------------------------------------+---------+
> |                                     version                                 
>      | daemons |
> +----------------------------------------------------------------------------------+---------+
> | ceph version 12.2.5 (cad919881333ac92274171586c827e01f554a70a) luminous 
> (stable) |   cm1   |
> | ceph version 12.2.4 (52085d5249a80c5f5121a76d6288429f35e4e77b) luminous 
> (stable) |   cm3   |
> +----------------------------------------------------------------------------------+---------+
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to