This problem may also be related to the below unsolved issue, which
specifically mentions 'unfound' objects. Sadly, there is probably
nothing in the report which will help with your troubleshooting.
https://tracker.ceph.com/issues/44286
C.
___
ceph-
Hi all,
We had a situation where 1 drive failed at the same time as a node. This
caused files in cephfs not not be readable and 'ceph status' to display
the error message "pgs not active".
Our cluster is either 3 replicas or equivalent EC (k2m2). Eventually
all the PGs became active and not
Hi all,
The documentation for "min_size" says "Sets the minimum number of
replicas required for I/O".
https://docs.ceph.com/en/latest/rados/operations/pools/
Can anyone confirm that a PG below "min_size" but still online can still
be read?
If someone says "the PG can be read" I will open a
Thanks!
There is wishful thinking on the web:
"Min_size is the number of replicas, when Ceph stops writing data. Read
operations are still possible."
e.g.
https://forum.proxmox.com/threads/ceph-pool-of-size-4-with-min_size-of-2-safe.65108/post-293976
I'll open a wishlist/enhancement issue.
You may have run into this bug:
https://tracker.ceph.com/issues/44286
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
Do you mean unfound instead of undersized? There is an as yet
unreproducible bug:
https://tracker.ceph.com/issues/44286
(Please follow this bug if it affects you! I've experienced it and am
leery of doing any drive swaps or upgrades until it is fixed.)
Chad.
___
Hi Jake and all,
We're having what looks to be the exact same problem. In our case it
happened when I was "draining" an OSD for removal. (ceph crush
remove...) Adding the OSD back doesn't help workaround the bug.
Everything is either triply replicated or EC k3m2, either of which
should st
Hi Jake,
In case it's relevant we are on Nautilus 14.2.6, not Mimic.
Yeah, my guess is that it is multiversion. Also, my scenario simply
should not have lost any data, so don't kick yourself too hard.
This command supposedly check cephfs for damaged files. It came back
with nothing for me
Hi Simone,
Maybe you've hit this bug:
https://tracker.ceph.com/issues/44286
?
Chad.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
Maybe I've marked the object as "lost" and removed the failed
OSD.
The cluster now is healthy, but I'd like to understand if it's likely
to bother me again in the future.
Yeah, I don't know.
Within the last month there are 4 separate instances of people
mentioning "unfound" object in their
No, I don't have cache tiering enabled. I also found strange that the PG
was marked unfound: the cluster was perfectly healthy before the kernel
panic and a single OSD failure shouldn't create mush hassle.
Yes, it is a bug unless using a singly replicated pool!
C.
__
Awhile back I thought there were some limitations which prevented us
from trying this, but I cannot remember...
What does the ceph vfs gain you over exporting by cephfs kernel module
(kernel 4.19). What does it lose you?
(I.e. pros and cons versus kernel module?)
Thanks!
C.
It's based on v
12 matches
Mail list logo