On Tue, Jan 28, 2020 at 4:26 PM CASS Philip <p.c...@epcc.ed.ac.uk> wrote:

> I have a query about https://docs.ceph.com/docs/master/cephfs/createfs/:
>
>
>
> “The data pool used to create the file system is the “default” data pool
> and the location for storing all inode backtrace information, used for hard
> link management and disaster recovery. For this reason, all inodes created
> in CephFS have at least one object in the default data pool.”
>
>
>
> This does not match my experience (nautilus servers, nautlius FUSE client
> or Centos 7 kernel client). I have a cephfs with a replicated top-level
> pool and a directory set to use erasure coding with setfattr, though I also
> did the same test using the subvolume commands with the same result.  "Ceph
> df detail" shows no objects used in the top level pool, as shown in
> https://gist.github.com/pcass-epcc/af24081cf014a66809e801f33bcb535b (also
> displayed in-line below)
>

Hmm I think this is tripping over the longstanding issue that omap data is
not reflected in the pool stats (although I would expect it to still show
up as objects, but perhaps the "ceph df" view has a different reporting
chain? Or else I'm confused somehow.)
But anyway...


>
>
> It would be useful if indeed clients didn’t have to write to the top-level
> pool, since that would mean we could give different clients permission only
> to pool-associated subdirectories without giving everyone write access to a
> pool with data structures shared between all users of the filesystem.
>

*Clients* don't need write permission to the default data pool unless you
want them to write files there. The backtraces are maintained by the MDS. :)
-Greg


>
>
> [root@hdr-admon01 ec]# ceph df detail; ceph fs ls; ceph fs status
>
> RAW STORAGE:
>
>     CLASS     SIZE        AVAIL       USED        RAW USED     %RAW USED
>
>     hdd       3.3 PiB     3.3 PiB      32 TiB       32 TiB          0.95
>
>     nvme      2.9 TiB     2.9 TiB     504 MiB      2.5 GiB          0.08
>
>     TOTAL     3.3 PiB     3.3 PiB      32 TiB       32 TiB          0.95
>
>
>
> POOLS:
>
>     POOL                           ID     STORED      OBJECTS
> USED        %USED     MAX AVAIL     QUOTA OBJECTS     QUOTA BYTES
> DIRTY     USED COMPR     UNDER COMPR
>
>     cephfs.fs1.metadata             5     162 MiB          63     324
> MiB      0.01       1.4 TiB     N/A               N/A
> 63            0 B             0 B
>
>     cephfs.fs1-replicated.data      6         0 B           0         0
> B         0       1.0 PiB     N/A               N/A
> 0            0 B             0 B
>
>     cephfs.fs1-ec.data              7     8.0 GiB       2.05k      11
> GiB         0       2.4 PiB     N/A               N/A
> 2.05k            0 B             0 B
>
> name: fs1, metadata pool: cephfs.fs1.metadata, data pools:
> [cephfs.fs1-replicated.data cephfs.fs1-ec.data ]
>
> fs1 - 4 clients
>
> ===
>
> +------+--------+------------+---------------+-------+-------+
>
> | Rank | State  |    MDS     |    Activity   |  dns  |  inos |
>
> +------+--------+------------+---------------+-------+-------+
>
> |  0   | active | hdr-meta02 | Reqs:    0 /s |   29  |   16  |
>
> +------+--------+------------+---------------+-------+-------+
>
> +----------------------------+----------+-------+-------+
>
> |            Pool            |   type   |  used | avail |
>
> +----------------------------+----------+-------+-------+
>
> |    cephfs.fs1.metadata     | metadata |  324M | 1414G |
>
> | cephfs.fs1-replicated.data |   data   |    0  | 1063T |
>
> |     cephfs.fs1-ec.data     |   data   | 11.4G | 2505T |
>
> +----------------------------+----------+-------+-------+
>
> +-------------+
>
> | Standby MDS |
>
> +-------------+
>
> |  hdr-meta01 |
>
> +-------------+
>
> MDS version: ceph version 14.2.5
> (ad5bd132e1492173c85fda2cc863152730b16a92) nautilus (stable)
>
>
>
> [root@hdr-admon01 ec]# ll /test-fs/ec/
>
> total 12582912
>
> -rw-r--r--. 1 root root 4294967296 Jan 27 22:26 new-file
>
> -rw-r--r--. 2 root root 4294967296 Jan 28 14:06 new-file2
>
> -rw-r--r--. 2 root root 4294967296 Jan 28 14:06
> new-file-same-inode-as-newfile2
>
>
>
> Regards,
> Phil
> _________________________________________
> *Philip Cass*
>
> *HPC Systems Specialist – Senior Systems Administrator *
>
> *EPCC*
>
>
>
> Advanced Computing Facility
>
> Bush Estate
>
> Penicuik
>
>
>
> *Tel:*                                        +44 (0)131 4457815
>
> *Email:*                                   p.c...@epcc.ed.ac.uk
>
>
> _________________________________________
>
>
>
> The University of Edinburgh is a charitable body, registered in Scotland,
> with registration number SC005336.
>
> The information contained in this e-mail (including any attachments) is
> confidential and is intended for the use of the addressee only.  If you
> have received this message in error, please delete it and notify the
> originator immediately.
>
> Please consider the environment before printing this email.
>
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to