Hi,
We are running a cluster with ceph luminous 12.2.0. Some of the OSDs are
getting full and we are running ceph osd reweight-by-utilization to
re-balance the OSDs. We have also set
mon_osd_backfillfull_ratio 0.8 (This is to prevent moving data to an
overloaded OSD when re-weighting)
mon_osd_nea
Hello,
Is it possible to get directory/file layout information (size, pool)
of a CephFS directory directly from a metadata server without the need
to mount the fs? Or better through the restful plugin...
When mounted I can get infos about the directory/file layout using the
getfattr command...
B
Hi Monis,
The settings you mention do not prevent data movement to overloaded OSD's,
they are a threshold when CEPH warns when an OSD is nearfull or
backfillfull.
No expert on this but setting backfillfull lower then nearfull is not
recommended, the nearfull state should be reached first in stead
Hi all,
We recently had an osd breakdown. After that i have manually added osd's
thinking that ceph repairs by itself.
I am running ceph 11 version
root@node16:~# ceph -v
ceph version 11.2.1 (e0354f9d3b1eea1d75a7dd487ba8098311be38a7)
root@node16:~# ceph -s
cluster 7c75f6e9-b858-4ac4-aa26-48
Hi Dilip,
Looking at the output of ceph -s it's still recovering (there are still pgs
in recovery_wait, backfill_wait, recovering state) so you will have to be
patient to let ceph recover.
The output of ceph osd dump doesn't mention osd.7 (it's referring to pool 7)
Kind regards,
Caspar Smit
201
Hi All,
We are having 5 node cluster with EC 4+1 . Each node has 68 HDD . Now we
are trying to add new node with 68 disks to the cluster .
We tried to add new node and created all OSDs in one go , the cluster
stopped all client traffic and does only backfilling .
Any procedure to add the new nod
You can add new OSDs with 0 weight and edit below script to increase
the osd weights instead of decreasing.
https://github.com/cernceph/ceph-scripts/blob/master/tools/ceph-gentle-reweight
On Wed, Apr 18, 2018 at 2:16 PM, nokia ceph wrote:
> Hi All,
>
> We are having 5 node cluster with EC 4+1 .
Hi ,
Thank you for the quick response .. We will try to adapt the script to
increase the osd weight.
How to create osd with weight 0 using ceph-volume tool? Or we have to
create OSD and later modify the OSD weight to 0.
Thanks,
Muthu
On Wed, Apr 18, 2018 at 5:02 PM, Serkan Çoban wrote:
> You
You can modify ceph.conf and add line below for newly added osd weight of 0
[osd]
osd_crush_initial_weight = 0
Also you can read this thread:
http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-February/007591.html
On Wed, Apr 18, 2018 at 3:01 PM, nokia ceph wrote:
> Hi ,
>
> Thank you fo
I keep seeing these threads where adding nodes has such an impact on the
cluster as a whole, that I wonder what the rest of the cluster looks like.
Normally I’d just advise someone to put a limit on the concurrent backfills
that can be done, and `osd max backfills` by default already is 1. Could
On Tue, Apr 17, 2018 at 3:52 PM, Steven Vacaroaia wrote:
> Hi,
>
> I can see many slow requests in the logs but no clue which OSD is the
> culprit
>
> If I remember correctly, that info was provided in the logs in previous CEPH
> version
>
> How can I find the culprit ?
>
> Previous log entry
>
>
>68 OSDs per node sounds an order of magnitude above what you should be doing,
>unless you have vast experience with Ceph and its memory requirements under
>stress.
I don't think so. We are also evaluating 90 OSDs per node. In order to
know it works you need to test all the scenarios. Redhat supp
You'll find it said time and time agin on the ML... avoid disks of
different sizes in the same cluster. It's a headache that sucks. It's
not impossible, it's not even overly hard to pull off... but it's
very easy to cause a mess and a lot of headaches. It will also make
it harder to diagnose perf
>>I don't find any clue. Next time it happens, could you please try
>>"ceph tell mds.xxx heap release"
don't seem to work
USER PID %CPU %MEMVSZ RSS TTY STAT START TIME COMMAND
ceph 1211357 13.1 18.6 12676452 12286508 ? Ssl avril05 2567:28
/usr/bin/ceph-mds -f --c
>> There is no way to fill up all disks evenly with the same number of
>> Bytes and then stop filling the small disks when they're full and
>> only continue filling the larger disks.
>This is possible with adjusting crush weights. Initially the smaller
>drives are weighted more highly than l
15 matches
Mail list logo