(I think John knows the answer, but sending to ceph-users for archival purposes)

Hi John,

A Ceph cluster lost a PG with CephFS metadata in there and it is currently 
doing a CephFS disaster recovery as described here: 
http://docs.ceph.com/docs/master/cephfs/disaster-recovery/

This data pool has 1.4B objects and currently has 16 concurrent scan_extents 
scans running:

# cephfs-data-scan --debug-rados=10 scan_extents --worker_n 0 --worker_m 16 
cephfs_metadata
# cephfs-data-scan --debug-rados=10 scan_extents --worker_n 1 --worker_m 16 
cephfs_metadata
..
..
# cephfs-data-scan --debug-rados=10 scan_extents --worker_n 15 --worker_m 16 
cephfs_metadata

According to the source in DataScan.cc:
* worker_n: Worker number
* worker_m: Worker count

So with the commands above I have 16 workers running, correct? For the 
scan_inodes I want to scale out to 32 workers to speed up the process even more.

Just to double-check before I send a new PR to update the docs, this is the 
right way to run the tool, correct?

If not, before sending the PR and starting scan_inodes on this cluster, what is 
the correct way to invoke the tool?

Thanks!

Wido
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to