Hi,

I’m trying to setup one-way rbd-mirroring for a ceph-cluster used by an
openstack cloud, but the rbd-mirror is unable to “catch up” with the
changes. However it appears to me as if it's not due to the ceph-clusters
or the network but due to the server running the rbd-mirror process running
out of cpu?

Is a high cpu load to be expected or is it a symptom of something else?
Or in other words, what can I check/do to get the mirroring working? 😊

# rbd mirror pool status nova
health: WARNING
images: 596 total
    572 starting_replay
    24 replaying

top - 13:31:36 up 79 days,  5:31,  1 user,  load average: 32.27, 26.82,
25.33
Tasks: 360 total,  17 running, 182 sleeping,   0 stopped,   0 zombie
%Cpu(s):  8.9 us, 70.0 sy,  0.0 ni, 18.5 id,  0.0 wa,  0.0 hi,  2.7 si,
0.0 st
KiB Mem : 13205185+total, 12862490+free,   579508 used,  2847444 buff/cache
KiB Swap:        0 total,        0 free,        0 used. 12948856+avail Mem
    PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM
 TIME+ COMMAND
2336553 ceph      20   0   17.1g 178160  20344 S 417.2  0.1  21:50.61
rbd-mirror
2312698 root      20   0       0      0      0 I  70.2  0.0  70:11.51
kworker/12:2
2312851 root      20   0       0      0      0 R  69.2  0.0  62:29.69
kworker/24:1
2324627 root      20   0       0      0      0 I  68.4  0.0  40:36.77
kworker/14:1
2235817 root      20   0       0      0      0 I  68.0  0.0 469:14.08
kworker/8:0
2241720 root      20   0       0      0      0 R  67.3  0.0 437:46.51
kworker/9:1
2306648 root      20   0       0      0      0 R  66.9  0.0 109:27.44
kworker/25:0
2324625 root      20   0       0      0      0 R  66.9  0.0  40:37.53
kworker/13:1
2336318 root      20   0       0      0      0 R  66.7  0.0  14:51.96
kworker/27:3
2324643 root      20   0       0      0      0 I  66.5  0.0  36:21.46
kworker/15:2
2294989 root      20   0       0      0      0 I  66.3  0.0 134:09.89
kworker/11:1
2324626 root      20   0       0      0      0 I  66.3  0.0  39:44.14
kworker/28:2
2324019 root      20   0       0      0      0 I  65.3  0.0  44:51.80
kworker/26:1
2235814 root      20   0       0      0      0 R  65.1  0.0 459:14.70
kworker/29:2
2294174 root      20   0       0      0      0 I  64.5  0.0 220:58.50
kworker/30:1
2324355 root      20   0       0      0      0 R  63.3  0.0  45:04.29
kworker/10:1
2263800 root      20   0       0      0      0 R  62.9  0.0 353:38.48
kworker/31:1
2270765 root      20   0       0      0      0 R  60.2  0.0 294:46.34
kworker/0:0
2294798 root      20   0       0      0      0 R  59.8  0.0 148:48.23
kworker/1:2
2307128 root      20   0       0      0      0 R  59.8  0.0  86:15.45
kworker/6:2
2307129 root      20   0       0      0      0 I  59.6  0.0  85:29.66
kworker/5:0
2294826 root      20   0       0      0      0 R  58.2  0.0 138:53.56
kworker/7:3
2294575 root      20   0       0      0      0 I  57.8  0.0 155:03.74
kworker/2:3
2294310 root      20   0       0      0      0 I  57.2  0.0 176:10.92
kworker/4:2
2295000 root      20   0       0      0      0 I  57.2  0.0 132:47.28
kworker/3:2
2307060 root      20   0       0      0      0 I  56.6  0.0  87:46.59
kworker/23:2
2294931 root      20   0       0      0      0 I  56.4  0.0 133:31.47
kworker/17:2
2318659 root      20   0       0      0      0 I  56.2  0.0  55:01.78
kworker/16:2
2336304 root      20   0       0      0      0 I  56.0  0.0  11:45.92
kworker/21:2
2306947 root      20   0       0      0      0 R  55.6  0.0  90:45.31
kworker/22:2
2270628 root      20   0       0      0      0 I  53.8  0.0 273:43.31
kworker/19:3
2294797 root      20   0       0      0      0 R  52.3  0.0 141:13.67
kworker/18:0
2330537 root      20   0       0      0      0 R  52.3  0.0  25:33.25
kworker/20:2

The main cluster has 12 nodes with 120 OSDs and the backup cluster has 6
nodes with 60 OSDs (but roughly the same amount of storage), the rbd-mirror
runs on a separate server with 2* E5-2650v2 cpus and 128GB memory.

Best regards
/Magnus
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to