Hi, Or you can query the MDS(s) with: ceph tell mds.* dump inode <inode> 2>/dev/null | grep path
for example: user@server:~$ ceph tell mds.* dump inode 1099836155033 2>/dev/null | grep path "path": "/ec42/default/joliot/gipsi/gpu_burn.sif", "stray_prior_path": "", Arnaud Le 01/05/2023 15:07, « Loic Tortay » <tor...@cc.in2p3.fr <mailto:tor...@cc.in2p3.fr>> a écrit : On 01/05/2023 11:35, Frank Schilder wrote: > Hi all, > > I think we might be hitting a known problem > (https://tracker.ceph.com/issues/57244 > <https://tracker.ceph.com/issues/57244>). I don't want to fail the mds yet, > because we have troubles with older kclients that miss the mds restart and > hold on to cache entries referring to the killed instance, leading to hanging > jobs on our HPC cluster. > > I have seen this issue before and there was a process in D-state that > dead-locked itself. Usually, killing this process succeeded and resolved the > issue. However, this time I can't find such a process. > > The tracker mentions that one can delete the file/folder. I have the inode > number, but really don't want to start a find on a 1.5PB file system. Is > there a better way to find what path is causing the issue (ask the MDS > directly, look at a cache dump, or similar)? Is there an alternative to > deletion or MDS fail? > Hello, If you have the inode number, you can retrieve the name with something like: rados getxattr -p $POOL ${ino}.00000000 parent | \ ceph-dencoder type inode_backtrace_t import - decode dump_json | \ jq -M '[.ancestors[].dname]' | tr -d '[[",\]]' | \ awk 't!=""{t=$1 "/" t;}t==""{t=$1;}END{print t}' Where $POOL is the "default pool" name (for files) or the metadata pool name (for directories) and $ino is the inode number (in hexadecimal). Loïc. -- | Loīc Tortay <tor...@cc.in2p3.fr <mailto:tor...@cc.in2p3.fr>> - IN2P3 Computing Centre | _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io <mailto:ceph-users@ceph.io> To unsubscribe send an email to ceph-users-le...@ceph.io <mailto:ceph-users-le...@ceph.io> _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io