Hello,
I have had a very bad thing happen. 10 of my OSD's were attached to
two hosts at the same time. They were automatically mounted at system
boot.
When I realised I stopped them, one node at a time. However they now
wont mount (with the SAS pathing fixed)
I'm getting rocksDB corruption issues (Unsurprisingly) that cannot be
resolved by ceph-bluestore-tool or ceph-kvstore-tool as they cannot
even deal with the rocksdb's.
I've lost enough disks that entire PG's are wiped out.
I'm hoping there is a way to recover some of the data. I have another
6 4TB disks that are blank and can be used if it will help.
Some of the errors I'm seeing are:
rocksdb: Corruption: Mismatch in unique ID on table file 41122.
Expected: {4769373877066773223,17385751700056157561} Actual:
{17358529186729819771,7012539495376980102} in file db/MANIFEST-0411055
rocksdb: verify_sharding unable to list column families: Corruption:
checksum mismatch in file db/MANIFEST-062921
And a different one I can't find right now, but it says something
about missing a newline....
Any advice or help would be appreciated,
Thanks in advance,
Mal
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]