There are 30 osds.
Thu, 22 Aug 2019 14:38:10 +0700
wahyu.muqs...@gmail.com ==> ceph-users@lists.ceph.com, Lars Täuber
:
> how many osd do you use ?
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph
All osd are up.
I manually mark one out of 30 "out" not "down".
The primary osd of the stuck pgs are neither marked as out nor as down.
Thanks
Lars
Thu, 22 Aug 2019 15:01:12 +0700
wahyu.muqs...@gmail.com ==> wahyu.muqs...@gmail.com, Lars Täuber
:
> I think you use too few osd. when you use era
Hi there!
We also experience this behaviour of our cluster while it is moving pgs.
# ceph health detail
HEALTH_ERR 1 MDSs report slow metadata IOs; Reduced data availability: 2 pgs
inactive; Degraded data redundancy (low space): 1 pg backfill_toofull
MDS_SLOW_METADATA_IO 1 MDSs report slow metad
https://tracker.ceph.com/issues/41255 is probably reporting the same issue.
On Thu, Aug 22, 2019 at 6:31 PM Lars Täuber wrote:
>
> Hi there!
>
> We also experience this behaviour of our cluster while it is moving pgs.
>
> # ceph health detail
> HEALTH_ERR 1 MDSs report slow metadata IOs; Reduced
Hi,
In a couple of situations I have encountered that Virtual Machines
running on RBD had a high I/O-wait, nearly 100%, on their vdX (VirtIO)
or sdX (Virtio-SCSI) devices while they were performing CPU intensive tasks.
These servers would be running a very CPU intensive application while
*not* do
Hello,
I am considering enabling optimal crush tunables in our Jewel cluster (4
nodes, 52 OSD, used as OpenStack Cinder+Nova backend = RBD images). I've
got two questions:
1. Do I understand right that having the optimal tunables on can be
considered best practice and should be applied in most sce
On Thu, Aug 22, 2019 at 9:23 AM Wido den Hollander wrote:
>
> Hi,
>
> In a couple of situations I have encountered that Virtual Machines
> running on RBD had a high I/O-wait, nearly 100%, on their vdX (VirtIO)
> or sdX (Virtio-SCSI) devices while they were performing CPU intensive tasks.
>
> These
Hi Folks,
I've updated hsbench (new S3 benchmark) to 0.2
Notable changes since 0.1:
- Can now output CSV results
- Can now output JSON results
- Fix for poor read performance with low thread counts
- New bucket listing benchmark with a new "mk" flag that lets you
control the number of ke
We just had metadata damage show up on our Jewel cluster. I tried a few
things like renaming directories and scanning, but the damage would just
show up again in less than 24 hours. I finally just copied the directories
with the damage to a tmp location on CephFS, then swapped it with the
damaged o
Hello,
yesterday I've added 4th OSD node (increase from 39 to 52 OSDs) into our
Jewel cluster. Backfilling of remapped pgs is still running and seems it
will run for another day until complete.
I know the pg_num of largest is undersized and I should increase it from
512 to 2048.
The question is -
On 8/22/19 3:59 PM, Jason Dillaman wrote:
> On Thu, Aug 22, 2019 at 9:23 AM Wido den Hollander wrote:
>>
>> Hi,
>>
>> In a couple of situations I have encountered that Virtual Machines
>> running on RBD had a high I/O-wait, nearly 100%, on their vdX (VirtIO)
>> or sdX (Virtio-SCSI) devices whil
On Thu, Aug 22, 2019 at 11:29 AM Wido den Hollander wrote:
>
>
>
> On 8/22/19 3:59 PM, Jason Dillaman wrote:
> > On Thu, Aug 22, 2019 at 9:23 AM Wido den Hollander wrote:
> >>
> >> Hi,
> >>
> >> In a couple of situations I have encountered that Virtual Machines
> >> running on RBD had a high I/O-
I am interested in keeping a revision history of ceph-iscsi's gateway.conf
object for any and all changes. It seems to me this may come in handy to revert
the environment to a previous state. My question is are there any existing
tools which do similar or could someone please suggest, if they ex
On 8/22/19 9:38 PM, Wesley Dillingham wrote:
> I am interested in keeping a revision history of ceph-iscsi's
> gateway.conf object for any and all changes. It seems to me this may
> come in handy to revert the environment to a previous state. My question
> is are there any existing tools which do
Thank you for providing the profiling data, Vladimir. There are 5078 threads
and most of them are waiting. Here is a list of the deepest call of each thread
with duplicates removed.
+ 100.00% epoll_wait
+ 100.00%
get_obj_data::flush(rgw::OwningList&&)
Root affected got more than 70TB free. The only solution is manual reweight
the OSD. But in this situacion balancer in unmap mode should move data to
get all HEALTHY
Hope some fix come in the next 14.2.X to fix that issue.
Ceph 14.2.2 Centos 7.6
cluster:
id: e1ee8086-7cce-43fd
16 matches
Mail list logo