[ceph-users] Cluster Re-balancing

2018-04-18 Thread Monis Monther
Hi, We are running a cluster with ceph luminous 12.2.0. Some of the OSDs are getting full and we are running ceph osd reweight-by-utilization to re-balance the OSDs. We have also set mon_osd_backfillfull_ratio 0.8 (This is to prevent moving data to an overloaded OSD when re-weighting) mon_osd_nea

[ceph-users] CephFS get directory size without mounting the fs

2018-04-18 Thread Martin Palma
Hello, Is it possible to get directory/file layout information (size, pool) of a CephFS directory directly from a metadata server without the need to mount the fs? Or better through the restful plugin... When mounted I can get infos about the directory/file layout using the getfattr command... B

Re: [ceph-users] Cluster Re-balancing

2018-04-18 Thread Caspar Smit
Hi Monis, The settings you mention do not prevent data movement to overloaded OSD's, they are a threshold when CEPH warns when an OSD is nearfull or backfillfull. No expert on this but setting backfillfull lower then nearfull is not recommended, the nearfull state should be reached first in stead

[ceph-users] pg's are stuck in active+undersized+degraded+remapped+backfill_wait even after introducing new osd's to cluster

2018-04-18 Thread Dilip Renkila
Hi all, We recently had an osd breakdown. After that i have manually added osd's thinking that ceph repairs by itself. I am running ceph 11 version root@node16:~# ceph -v ceph version 11.2.1 (e0354f9d3b1eea1d75a7dd487ba8098311be38a7) root@node16:~# ceph -s cluster 7c75f6e9-b858-4ac4-aa26-48

Re: [ceph-users] pg's are stuck in active+undersized+degraded+remapped+backfill_wait even after introducing new osd's to cluster

2018-04-18 Thread Caspar Smit
Hi Dilip, Looking at the output of ceph -s it's still recovering (there are still pgs in recovery_wait, backfill_wait, recovering state) so you will have to be patient to let ceph recover. The output of ceph osd dump doesn't mention osd.7 (it's referring to pool 7) Kind regards, Caspar Smit 201

[ceph-users] scalability new node to the existing cluster

2018-04-18 Thread nokia ceph
Hi All, We are having 5 node cluster with EC 4+1 . Each node has 68 HDD . Now we are trying to add new node with 68 disks to the cluster . We tried to add new node and created all OSDs in one go , the cluster stopped all client traffic and does only backfilling . Any procedure to add the new nod

Re: [ceph-users] scalability new node to the existing cluster

2018-04-18 Thread Serkan Çoban
You can add new OSDs with 0 weight and edit below script to increase the osd weights instead of decreasing. https://github.com/cernceph/ceph-scripts/blob/master/tools/ceph-gentle-reweight On Wed, Apr 18, 2018 at 2:16 PM, nokia ceph wrote: > Hi All, > > We are having 5 node cluster with EC 4+1 .

Re: [ceph-users] scalability new node to the existing cluster

2018-04-18 Thread nokia ceph
Hi , Thank you for the quick response .. We will try to adapt the script to increase the osd weight. How to create osd with weight 0 using ceph-volume tool? Or we have to create OSD and later modify the OSD weight to 0. Thanks, Muthu On Wed, Apr 18, 2018 at 5:02 PM, Serkan Çoban wrote: > You

Re: [ceph-users] scalability new node to the existing cluster

2018-04-18 Thread Serkan Çoban
You can modify ceph.conf and add line below for newly added osd weight of 0 [osd] osd_crush_initial_weight = 0 Also you can read this thread: http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-February/007591.html On Wed, Apr 18, 2018 at 3:01 PM, nokia ceph wrote: > Hi , > > Thank you fo

Re: [ceph-users] scalability new node to the existing cluster

2018-04-18 Thread Hans van den Bogert
I keep seeing these threads where adding nodes has such an impact on the cluster as a whole, that I wonder what the rest of the cluster looks like. Normally I’d just advise someone to put a limit on the concurrent backfills that can be done, and `osd max backfills` by default already is 1. Could

Re: [ceph-users] ceph 12.2.4 - which OSD has slow requests ?

2018-04-18 Thread Alex Gorbachev
On Tue, Apr 17, 2018 at 3:52 PM, Steven Vacaroaia wrote: > Hi, > > I can see many slow requests in the logs but no clue which OSD is the > culprit > > If I remember correctly, that info was provided in the logs in previous CEPH > version > > How can I find the culprit ? > > Previous log entry > >

Re: [ceph-users] scalability new node to the existing cluster

2018-04-18 Thread Serkan Çoban
>68 OSDs per node sounds an order of magnitude above what you should be doing, >unless you have vast experience with Ceph and its memory requirements under >stress. I don't think so. We are also evaluating 90 OSDs per node. In order to know it works you need to test all the scenarios. Redhat supp

[ceph-users] osds with different disk sizes may killing, > performance (?? ?)

2018-04-18 Thread Chad William Seys
You'll find it said time and time agin on the ML... avoid disks of different sizes in the same cluster. It's a headache that sucks. It's not impossible, it's not even overly hard to pull off... but it's very easy to cause a mess and a lot of headaches. It will also make it harder to diagnose perf

Re: [ceph-users] ceph mds memory usage 20GB : is it normal ?

2018-04-18 Thread Alexandre DERUMIER
>>I don't find any clue. Next time it happens, could you please try >>"ceph tell mds.xxx heap release" don't seem to work USER PID %CPU %MEMVSZ RSS TTY STAT START TIME COMMAND ceph 1211357 13.1 18.6 12676452 12286508 ? Ssl avril05 2567:28 /usr/bin/ceph-mds -f --c

Re: [ceph-users] osds with different disk sizes may killing, > performance (?? ?)

2018-04-18 Thread Van Leeuwen, Robert
>> There is no way to fill up all disks evenly with the same number of >> Bytes and then stop filling the small disks when they're full and >> only continue filling the larger disks. >This is possible with adjusting crush weights. Initially the smaller >drives are weighted more highly than l