Ok, thank you guys

The version is 10.2.10
Matteo

> Il giorno 20 nov 2017, alle ore 23:15, Christian Balzer <ch...@gol.com> ha 
> scritto:
> 
> On Mon, 20 Nov 2017 10:35:36 -0800 Chris Taylor wrote:
> 
>> On 2017-11-20 3:39 am, Matteo Dacrema wrote:
>>> Yes I mean the existing Cluster.
>>> SSDs are on a fully separate pool.
>>> Cluster is not busy during recovery and deep scrubs but I think it’s
>>> better to limit replication in some way when switching to replica 3.
>>> 
>>> My question is to understand if I need to set some options parameters
>>> to limit the impact of the creation of new objects.I’m also concerned
>>> about disk filling up during recovery because of inefficient data
>>> balancing.  
>> 
>> You can try using osd_recovery_sleep to slow down the backfilling so it 
>> does not cause the client io to hang.
>> 
>> ceph tell osd.* injectargs "--osd_recovery_sleep 0.1"
>> 
> 
> Which is one of the things that is version specific and we don't know the
> version yet.
> 
> The above will work with Hammer and should again with Luminous, but not so
> much with the unified queue bits inbetween. 
> 
> Christian
> 
>> 
>>> 
>>> Here osd tree
>>> 
>>> ID  WEIGHT    TYPE NAME        UP/DOWN REWEIGHT PRIMARY-AFFINITY
>>> -10  19.69994 root ssd
>>> -11   5.06998     host ceph101
>>> 166   0.98999         osd.166       up  1.00000          1.00000
>>> 167   1.00000         osd.167       up  1.00000          1.00000
>>> 168   1.00000         osd.168       up  1.00000          1.00000
>>> 169   1.07999         osd.169       up  1.00000          1.00000
>>> 170   1.00000         osd.170       up  1.00000          1.00000
>>> -12   4.92998     host ceph102
>>> 171   0.98000         osd.171       up  1.00000          1.00000
>>> 172   0.92999         osd.172       up  1.00000          1.00000
>>> 173   0.98000         osd.173       up  1.00000          1.00000
>>> 174   1.00000         osd.174       up  1.00000          1.00000
>>> 175   1.03999         osd.175       up  1.00000          1.00000
>>> -13   4.69998     host ceph103
>>> 176   0.84999         osd.176       up  1.00000          1.00000
>>> 177   0.84999         osd.177       up  1.00000          1.00000
>>> 178   1.00000         osd.178       up  1.00000          1.00000
>>> 179   1.00000         osd.179       up  1.00000          1.00000
>>> 180   1.00000         osd.180       up  1.00000          1.00000
>>> -14   5.00000     host ceph104
>>> 181   1.00000         osd.181       up  1.00000          1.00000
>>> 182   1.00000         osd.182       up  1.00000          1.00000
>>> 183   1.00000         osd.183       up  1.00000          1.00000
>>> 184   1.00000         osd.184       up  1.00000          1.00000
>>> 185   1.00000         osd.185       up  1.00000          1.00000
>>> -1 185.19835 root default
>>> -2  18.39980     host ceph001
>>> 63   0.79999         osd.63        up  1.00000          1.00000
>>> 64   0.79999         osd.64        up  1.00000          1.00000
>>> 65   0.79999         osd.65        up  1.00000          1.00000
>>> 146   0.70000         osd.146       up  1.00000          1.00000
>>> 147   0.70000         osd.147       up  1.00000          1.00000
>>> 148   0.90999         osd.148       up  1.00000          1.00000
>>> 149   0.79999         osd.149       up  1.00000          1.00000
>>> 150   0.79999         osd.150       up  1.00000          1.00000
>>> 151   0.79999         osd.151       up  1.00000          1.00000
>>> 152   0.79999         osd.152       up  1.00000          1.00000
>>> 153   0.70000         osd.153       up  1.00000          1.00000
>>> 154   0.70000         osd.154       up  1.00000          1.00000
>>> 155   0.89999         osd.155       up  1.00000          1.00000
>>> 156   0.84999         osd.156       up  1.00000          1.00000
>>> 157   0.70000         osd.157       up  1.00000          1.00000
>>> 158   0.70000         osd.158       up  1.00000          1.00000
>>> 159   0.84999         osd.159       up  1.00000          1.00000
>>> 160   0.90999         osd.160       up  1.00000          1.00000
>>> 161   0.90999         osd.161       up  1.00000          1.00000
>>> 162   0.90999         osd.162       up  1.00000          1.00000
>>> 163   0.79999         osd.163       up  1.00000          1.00000
>>> 164   0.90999         osd.164       up  1.00000          1.00000
>>> 165   0.64999         osd.165       up  1.00000          1.00000
>>> -3  19.41982     host ceph002
>>> 23   0.79999         osd.23        up  1.00000          1.00000
>>> 24   0.70000         osd.24        up  1.00000          1.00000
>>> 25   0.90999         osd.25        up  1.00000          1.00000
>>> 26   0.59999         osd.26        up  1.00000          1.00000
>>> 27   0.95000         osd.27        up  1.00000          1.00000
>>> 28   0.64999         osd.28        up  1.00000          1.00000
>>> 29   0.75000         osd.29        up  1.00000          1.00000
>>> 30   0.89999         osd.30        up  1.00000          1.00000
>>> 31   0.90999         osd.31        up  1.00000          1.00000
>>> 32   0.90999         osd.32        up  1.00000          1.00000
>>> 33   0.89999         osd.33        up  1.00000          1.00000
>>> 34   0.90999         osd.34        up  1.00000          1.00000
>>> 35   0.90999         osd.35        up  1.00000          1.00000
>>> 36   0.84999         osd.36        up  1.00000          1.00000
>>> 37   0.89999         osd.37        up  1.00000          1.00000
>>> 38   1.00000         osd.38        up  1.00000          1.00000
>>> 39   0.79999         osd.39        up  1.00000          1.00000
>>> 40   0.90999         osd.40        up  1.00000          1.00000
>>> 41   0.84999         osd.41        up  1.00000          1.00000
>>> 42   0.84999         osd.42        up  1.00000          1.00000
>>> 43   0.90999         osd.43        up  1.00000          1.00000
>>> 44   0.75000         osd.44        up  1.00000          1.00000
>>> 45   0.79999         osd.45        up  1.00000          1.00000
>>> -4  16.78984     host ceph003
>>> 20   1.00000         osd.20        up  1.00000          1.00000
>>> 21   0.90999         osd.21        up  1.00000          1.00000
>>> 22   0.90999         osd.22        up  1.00000          1.00000
>>> 46   0.79999         osd.46        up  1.00000          1.00000
>>> 47   0.84999         osd.47        up  1.00000          1.00000
>>> 48   0.75000         osd.48        up  1.00000          1.00000
>>> 49   0.90999         osd.49        up  1.00000          1.00000
>>> 50   0.90999         osd.50        up  1.00000          1.00000
>>> 51   0.90999         osd.51        up  1.00000          1.00000
>>> 52   0.90999         osd.52        up  1.00000          1.00000
>>> 53   0.70000         osd.53        up  1.00000          1.00000
>>> 54   0.70000         osd.54        up  1.00000          1.00000
>>> 55   0.90999         osd.55        up  1.00000          1.00000
>>> 56   0.90999         osd.56        up  1.00000          1.00000
>>> 57   0.75000         osd.57        up  1.00000          1.00000
>>> 58   0.79999         osd.58        up  1.00000          1.00000
>>> 59   0.70000         osd.59        up  1.00000          1.00000
>>> 60   0.90999         osd.60        up  1.00000          1.00000
>>> 61   0.75000         osd.61        up  1.00000          1.00000
>>> 62   0.79999         osd.62        up  1.00000          1.00000
>>> -5  16.30986     host ceph004
>>> 66   0.95000         osd.66        up  1.00000          1.00000
>>> 67   0.79999         osd.67        up  1.00000          1.00000
>>> 68   0.84999         osd.68        up  1.00000          1.00000
>>> 69   0.84999         osd.69        up  1.00000          1.00000
>>> 70   0.70000         osd.70        up  1.00000          1.00000
>>> 71   0.75000         osd.71        up  1.00000          1.00000
>>> 72   0.64999         osd.72        up  1.00000          1.00000
>>> 73   0.90999         osd.73        up  1.00000          1.00000
>>> 74   0.75000         osd.74        up  1.00000          1.00000
>>> 75   0.81999         osd.75        up  1.00000          1.00000
>>> 76   0.90999         osd.76        up  1.00000          1.00000
>>> 77   0.79999         osd.77        up  1.00000          1.00000
>>> 78   0.59999         osd.78        up  1.00000          1.00000
>>> 79   1.00000         osd.79        up  1.00000          1.00000
>>> 80   0.90999         osd.80        up  1.00000          1.00000
>>> 81   0.79999         osd.81        up  1.00000          1.00000
>>> 82   0.89999         osd.82        up  1.00000          1.00000
>>> 83   0.70000         osd.83        up  1.00000          1.00000
>>> 84   0.75000         osd.84        up  1.00000          1.00000
>>> 85   0.90999         osd.85        up  1.00000          1.00000
>>> -6  16.71983     host ceph005
>>> 86   0.75000         osd.86        up  1.00000          1.00000
>>> 87   0.90999         osd.87        up  1.00000          1.00000
>>> 88   0.89999         osd.88        up  1.00000          1.00000
>>> 89   0.90999         osd.89        up  1.00000          1.00000
>>> 90   0.90999         osd.90        up  1.00000          1.00000
>>> 91   0.89999         osd.91        up  1.00000          1.00000
>>> 92   0.89999         osd.92        up  1.00000          1.00000
>>> 93   0.90999         osd.93        up  1.00000          1.00000
>>> 94   0.84999         osd.94        up  1.00000          1.00000
>>> 95   0.90999         osd.95        up  1.00000          1.00000
>>> 96   0.50000         osd.96        up  1.00000          1.00000
>>> 97   0.79999         osd.97        up  1.00000          1.00000
>>> 98   0.59999         osd.98        up  1.00000          1.00000
>>> 99   0.75000         osd.99        up  1.00000          1.00000
>>> 100   1.00000         osd.100       up  1.00000          1.00000
>>> 101   0.79999         osd.101       up  1.00000          1.00000
>>> 102   0.90999         osd.102       up  1.00000          1.00000
>>> 103   0.79999         osd.103       up  1.00000          1.00000
>>> 104   0.90999         osd.104       up  1.00000          1.00000
>>> 105   0.79999         osd.105       up  1.00000          1.00000
>>> -7  17.19980     host ceph006
>>> 106   0.90999         osd.106       up  1.00000          1.00000
>>> 107   0.89999         osd.107       up  1.00000          1.00000
>>> 108   0.70000         osd.108       up  1.00000          1.00000
>>> 109   0.84999         osd.109       up  1.00000          1.00000
>>> 110   0.79999         osd.110       up  1.00000          1.00000
>>> 111   0.79999         osd.111       up  1.00000          1.00000
>>> 112   0.79999         osd.112       up  1.00000          1.00000
>>> 113   0.90999         osd.113       up  1.00000          1.00000
>>> 115   0.90999         osd.115       up  1.00000          1.00000
>>> 116   0.90999         osd.116       up  1.00000          1.00000
>>> 117   0.90999         osd.117       up  1.00000          1.00000
>>> 118   0.90999         osd.118       up  1.00000          1.00000
>>> 119   0.59999         osd.119       up  1.00000          1.00000
>>> 120   0.90999         osd.120       up  1.00000          1.00000
>>> 121   0.90999         osd.121       up  1.00000          1.00000
>>> 122   0.90999         osd.122       up  1.00000          1.00000
>>> 123   0.79999         osd.123       up  1.00000          1.00000
>>> 124   0.75000         osd.124       up  1.00000          1.00000
>>> 125   1.09999         osd.125       up  1.00000          1.00000
>>> 114   0.90999         osd.114       up  1.00000          1.00000
>>> -8  16.61984     host ceph007
>>> 126   0.90999         osd.126       up  1.00000          1.00000
>>> 127   0.75000         osd.127       up  1.00000          1.00000
>>> 128   0.84999         osd.128       up  1.00000          1.00000
>>> 129   0.90999         osd.129       up  1.00000          1.00000
>>> 130   0.70000         osd.130       up  1.00000          1.00000
>>> 131   0.78000         osd.131       up  1.00000          1.00000
>>> 132   0.79999         osd.132       up  1.00000          1.00000
>>> 133   0.54999         osd.133       up  1.00000          1.00000
>>> 134   0.79999         osd.134       up  1.00000          1.00000
>>> 135   0.87000         osd.135       up  1.00000          1.00000
>>> 136   0.90999         osd.136       up  1.00000          1.00000
>>> 137   0.90999         osd.137       up  1.00000          1.00000
>>> 138   0.90999         osd.138       up  1.00000          1.00000
>>> 139   0.89999         osd.139       up  1.00000          1.00000
>>> 140   0.75000         osd.140       up  1.00000          1.00000
>>> 141   0.75000         osd.141       up  1.00000          1.00000
>>> 142   0.90999         osd.142       up  1.00000          1.00000
>>> 143   0.84999         osd.143       up  1.00000          1.00000
>>> 145   0.90999         osd.145       up  1.00000          1.00000
>>> 144   0.89999         osd.144       up  1.00000          1.00000
>>> -9  16.93982     host ceph008
>>>  0   0.90999         osd.0         up  1.00000          1.00000
>>>  1   0.70000         osd.1         up  1.00000          1.00000
>>>  2   0.90999         osd.2         up  1.00000          1.00000
>>>  3   0.90999         osd.3         up  1.00000          1.00000
>>>  4   0.79999         osd.4         up  1.00000          1.00000
>>>  5   0.64999         osd.5         up  1.00000          1.00000
>>>  6   0.64999         osd.6         up  1.00000          1.00000
>>>  7   0.90999         osd.7         up  1.00000          1.00000
>>>  8   0.79999         osd.8         up  1.00000          1.00000
>>>  9   0.90999         osd.9         up  1.00000          1.00000
>>> 10   0.75000         osd.10        up  1.00000          1.00000
>>> 11   0.90999         osd.11        up  1.00000          1.00000
>>> 12   0.89999         osd.12        up  1.00000          1.00000
>>> 13   0.75000         osd.13        up  1.00000          1.00000
>>> 14   1.09999         osd.14        up  1.00000          1.00000
>>> 15   0.84999         osd.15        up  1.00000          1.00000
>>> 16   0.90999         osd.16        up  1.00000          1.00000
>>> 17   0.90999         osd.17        up  1.00000          1.00000
>>> 18   0.79999         osd.18        up  1.00000          1.00000
>>> 19   0.90999         osd.19        up  1.00000          1.00000
>>> -15  28.79987     host ceph009
>>> 186   1.00000         osd.186       up  1.00000          1.00000
>>> 187   1.00000         osd.187       up  1.00000          1.00000
>>> 188   1.00000         osd.188       up  1.00000          1.00000
>>> 189   1.00000         osd.189       up  1.00000          1.00000
>>> 190   1.00000         osd.190       up  1.00000          1.00000
>>> 191   1.00000         osd.191       up  1.00000          1.00000
>>> 192   1.00000         osd.192       up  1.00000          1.00000
>>> 193   1.00000         osd.193       up  1.00000          1.00000
>>> 194   1.00000         osd.194       up  1.00000          1.00000
>>> 195   1.79999         osd.195       up  1.00000          1.00000
>>> 196   1.79999         osd.196       up  1.00000          1.00000
>>> 197   1.79999         osd.197       up  1.00000          1.00000
>>> 198   1.79999         osd.198       up  1.00000          1.00000
>>> 199   1.79999         osd.199       up  1.00000          1.00000
>>> 200   1.79999         osd.200       up  1.00000          1.00000
>>> 201   1.79999         osd.201       up  1.00000          1.00000
>>> 202   1.79999         osd.202       up  1.00000          1.00000
>>> 203   1.79999         osd.203       up  1.00000          1.00000
>>> 204   1.79999         osd.204       up  1.00000          1.00000
>>> 205   1.79999         osd.205       up  1.00000          1.00000
>>> -16  17.99988     host ceph010
>>> 206   1.79999         osd.206       up  1.00000          1.00000
>>> 207   1.79999         osd.207       up  1.00000          1.00000
>>> 208   1.79999         osd.208       up  1.00000          1.00000
>>> 209   1.79999         osd.209       up  1.00000          1.00000
>>> 210   1.79999         osd.210       up  1.00000          1.00000
>>> 211   1.79999         osd.211       up  1.00000          1.00000
>>> 212   1.79999         osd.212       up  1.00000          1.00000
>>> 213   1.79999         osd.213       up  1.00000          1.00000
>>> 214   1.79999         osd.214       up  1.00000          1.00000
>>> 215   1.79999         osd.215       up  1.00000          1.00000
>>> 
>>> 
>>> 
>>>> Il giorno 20 nov 2017, alle ore 12:17, Christian Balzer 
>>>> <ch...@gol.com> ha scritto:
>>>> 
>>>> 
>>>> Hello,
>>>> 
>>>> On Mon, 20 Nov 2017 11:56:31 +0100 Matteo Dacrema wrote:
>>>> 
>>>>> Hi,
>>>>> 
>>>>> I need to switch a cluster of over 200 OSDs from replica 2 to replica 
>>>>> 3  
>>>> I presume this means the existing cluster and not adding 100 OSDs...
>>>> 
>>>>> There are two different crush maps for HDD and SSDs also mapped to 
>>>>> two different pools.
>>>>> 
>>>>> Is there a best practice to use? Can this provoke troubles?
>>>>> 
>>>> Are your SSDs a cache-tier or are they a fully separate pool?
>>>> 
>>>> As for troubles, how busy is your cluster during the recovery of 
>>>> failed
>>>> OSDs or deep scrubs?
>>>> 
>>>> There are 2 things to consider here:
>>>> 
>>>> 1. The re-balancing and additional replication of all the data, which 
>>>> you
>>>> can control/ease by the various knobs present. Ceph version matters to
>>>> which are relevant/useful. It shouldn't impact things too much, unless
>>>> your cluster was at the very edge of it's capacity anyway.
>>>> 
>>>> 2. The little detail that after 1) is done, your cluster will be
>>>> noticeably slower than before, especially in the latency department.
>>>> In short, you don't just need to have the disk space to go 3x, but 
>>>> also
>>>> enough IOPS/bandwidth reserves.
>>>> 
>>>> Christian
>>>> 
>>>>> Thank you
>>>>> Matteo
>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> ceph-users@lists.ceph.com
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>> 
>>>> 
>>>> 
>>>> --
>>>> Christian Balzer        Network/Systems Engineer
>>>> ch...@gol.com      Rakuten Communications
>>>> 
>>>> --
>>>> Questo messaggio e' stato analizzato con Libra ESVA ed e' risultato 
>>>> non infetto.
>>>> Seguire il link qui sotto per segnalarlo come spam:
>>>> http://mx01.enter.it/cgi-bin/learn-msg.cgi?id=524464756E.A33EC
>>>> 
>>>> 
>>> 
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com  
>> 
> 
> 
> -- 
> Christian Balzer        Network/Systems Engineer                
> ch...@gol.com <mailto:ch...@gol.com>          Rakuten Communications
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
> 
> --
> Questo messaggio e' stato analizzato con Libra ESVA ed e' risultato non 
> infetto.
> Seguire il link qui sotto per segnalarlo come spam: 
> http://mx01.enter.it/cgi-bin/learn-msg.cgi?id=78F40403AD.A839B 
> <http://mx01.enter.it/cgi-bin/learn-msg.cgi?id=78F40403AD.A839B>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to