Hello,

I have a Nautilus cluster with a cephfs volume, on grafana, it shows that 
cephfs_data pool is almost full[1] but if I give a look to the pool
usage, it looks like I have plenty of space. Which metrics are used by grafana?

1. https://framapic.org/5r7J86s55x6k/jGSIsjEUPYMU.png

pool usage:

> artemis@icitsrv5:~$ ceph df detail
> RAW STORAGE:
>     CLASS     SIZE        AVAIL       USED        RAW USED     %RAW USED 
>     hdd       662 TiB     296 TiB     366 TiB      366 TiB         55.32 
>     TOTAL     662 TiB     296 TiB     366 TiB      366 TiB         55.32 
>  
> POOLS:
>     POOL                           ID     STORED      OBJECTS     USED        
> %USED     MAX AVAIL     QUOTA OBJECTS     QUOTA BYTES     DIRTY       USED 
> COMPR     UNDER COMPR 
>     .rgw.root                       3     8.1 KiB          15     2.8 MiB     
>     0        63 TiB     N/A               N/A                  15            
> 0 B             0 B 
>     default.rgw.control             4         0 B           8         0 B     
>     0        63 TiB     N/A               N/A                   8            
> 0 B             0 B 
>     default.rgw.meta                5      26 KiB          85      16 MiB     
>     0        63 TiB     N/A               N/A                  85            
> 0 B             0 B 
>     default.rgw.log                 6         0 B         207         0 B     
>     0        63 TiB     N/A               N/A                 207            
> 0 B             0 B 
>     cephfs_data                     7     113 TiB     139.34M     186 TiB     
> 49.47       138 TiB     N/A               N/A             139.34M            
> 0 B             0 B 
>     cephfs_metadata                 8      54 GiB      10.21M      57 GiB     
>  0.03        63 TiB     N/A               N/A              10.21M            
> 0 B             0 B 
>     default.rgw.buckets.data        9     122 TiB      54.57M     173 TiB     
> 47.70       138 TiB     N/A               N/A              54.57M            
> 0 B             0 B 
>     default.rgw.buckets.index      10     2.6 GiB      19.97k     2.6 GiB     
>     0        63 TiB     N/A               N/A              19.97k            
> 0 B             0 B 
>     default.rgw.buckets.non-ec     11      67 MiB         186     102 MiB     
>     0        63 TiB     N/A               N/A                 186            
> 0 B             0 B 
>     device_health_metrics          12     1.2 MiB         145     1.2 MiB     
>     0        63 TiB     N/A               N/A                 145            
> 0 B             0 B 

Best,

-- 
Yoann Moulin
EPFL IC-IT
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to