Hi, Muthusamy Muthiah

I'm not totally sure that this is a memory leak.
We had same problems with bluestore on ceph v11.2.0.
Reduce bluestore cache helped us to solve it and stabilize OSD memory
consumption on the 3GB level.

Perhaps this will help you:

bluestore_cache_size = 104857600

On Wed, Feb 15, 2017 at 2:33 PM, Ilya Letkowski <mj12.svetz...@gmail.com>
wrote:

> Hi, Muthusamy Muthiah
>
> I'm not totally sure that this is a memory leak.
> We had same problems with bluestore on ceph v11.2.0.
> Reduce bluestore cache helped us to solve it and stabilize OSD memory
> consumption on the 3GB level.
>
> Perhaps this will help you:
>
> bluestore_cache_size = 104857600
>
>
>
> On Tue, Feb 14, 2017 at 11:52 AM, Muthusamy Muthiah <
> muthiah.muthus...@gmail.com> wrote:
>
>> Hi All,
>>
>> On all our 5 node cluster with ceph 11.2.0 we encounter memory leak
>> issues.
>>
>> Cluster details : 5 node with 24/68 disk per node , EC : 4+1 , RHEL 7.2
>>
>> Some traces using sar are below and attached the memory utilisation graph
>> .
>>
>> (16:54:42)[cn2.c1 sa] # sar -r
>> 07:50:01 kbmemfree kbmemused %memused kbbuffers kbcached kbcommit %commit
>> kbactive kbinact kbdirty
>> 10:20:01 32077264 132754368 80.54 16176 3040244 77767024 47.18 51991692
>> 2676468 260
>>
>>
>>
>>
>>
>>
>>
>>
>> *10:30:01 32208384 132623248 80.46 16176 3048536 77832312 47.22 51851512
>> 2684552 1210:40:01 32067244 132764388 80.55 16176 3059076 77832316 47.22
>> 51983332 2694708 26410:50:01 30626144 134205488 81.42 16176 3064340
>> 78177232 47.43 53414144 2693712 411:00:01 28927656 135903976 82.45 16176
>> 3074064 78958568 47.90 55114284 2702892 1211:10:01 27158548 137673084 83.52
>> 16176 3080600 80553936 48.87 56873664 2708904 1211:20:01 26455556 138376076
>> 83.95 16176 3080436 81991036 49.74 57570280 2708500 811:30:01 26002252
>> 138829380 84.22 16176 3090556 82223840 49.88 58015048 2718036 1611:40:01
>> 25965924 138865708 84.25 16176 3089708 83734584 50.80 58049980 2716740
>> 1211:50:01 26142888 138688744 84.14 16176 3089544 83800100 50.84 57869628
>> 2715400 16*
>>
>> ...
>> ...
>>
>> In the attached graph, there is increase in memory utilisation by
>> ceph-osd during soak test. And when it reaches the system limit of 128GB
>> RAM , we could able to see the below dmesg logs related to memory out when
>> the system reaches close to 128GB RAM. OSD.3 killed due to Out of memory
>> and started again.
>>
>> [Tue Feb 14 03:51:02 2017] *tp_osd_tp invoked oom-killer:
>> gfp_mask=0x280da, order=0, oom_score_adj=0*
>> [Tue Feb 14 03:51:02 2017] tp_osd_tp cpuset=/ mems_allowed=0-1
>> [Tue Feb 14 03:51:02 2017] CPU: 20 PID: 11864 Comm: tp_osd_tp Not tainted
>> 3.10.0-327.13.1.el7.x86_64 #1
>> [Tue Feb 14 03:51:02 2017] Hardware name: HP ProLiant XL420 Gen9/ProLiant
>> XL420 Gen9, BIOS U19 09/12/2016
>> [Tue Feb 14 03:51:02 2017]  ffff8819ccd7a280 0000000030e84036
>> ffff881fa58f7528 ffffffff816356f4
>> [Tue Feb 14 03:51:02 2017]  ffff881fa58f75b8 ffffffff8163068f
>> ffff881fa3478360 ffff881fa3478378
>> [Tue Feb 14 03:51:02 2017]  ffff881fa58f75e8 ffff8819ccd7a280
>> 0000000000000001 000000000001f65f
>> [Tue Feb 14 03:51:02 2017] Call Trace:
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff816356f4>] dump_stack+0x19/0x1b
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff8163068f>] dump_header+0x8e/0x214
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff8116ce7e>]
>> oom_kill_process+0x24e/0x3b0
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff8116c9e6>] ?
>> find_lock_task_mm+0x56/0xc0
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff8116d6a6>]
>> *out_of_memory+0x4b6/0x4f0*
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff81173885>]
>> __alloc_pages_nodemask+0xa95/0xb90
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff811b792a>]
>> alloc_pages_vma+0x9a/0x140
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff811976c5>]
>> handle_mm_fault+0xb85/0xf50
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff811957fb>] ?
>> follow_page_mask+0xbb/0x5c0
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff81197c2b>]
>> __get_user_pages+0x19b/0x640
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff8119843d>]
>> get_user_pages_unlocked+0x15d/0x1f0
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff8106544f>]
>> get_user_pages_fast+0x9f/0x1a0
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff8121de78>]
>> do_blockdev_direct_IO+0x1a78/0x2610
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff81218c40>] ? I_BDEV+0x10/0x10
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff8121ea65>]
>> __blockdev_direct_IO+0x55/0x60
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff81218c40>] ? I_BDEV+0x10/0x10
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff81219297>]
>> blkdev_direct_IO+0x57/0x60
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff81218c40>] ? I_BDEV+0x10/0x10
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff8116af63>]
>> generic_file_aio_read+0x6d3/0x750
>> [Tue Feb 14 03:51:02 2017]  [<ffffffffa038ad5c>] ?
>> xfs_iunlock+0x11c/0x130 [xfs]
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff811690db>] ? unlock_page+0x2b/0x30
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff81192f21>] ? __do_fault+0x401/0x510
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff8121970c>] blkdev_aio_read+0x4c/0x70
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff811ddcfd>] do_sync_read+0x8d/0xd0
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff811de45c>] vfs_read+0x9c/0x170
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff811df182>] SyS_pread64+0x92/0xc0
>> [Tue Feb 14 03:51:02 2017]  [<ffffffff81645e89>]
>> system_call_fastpath+0x16/0x1b
>>
>>
>> Feb 14 03:51:40 fr-paris kernel: *Out of memory: Kill process 7657
>> (ceph-osd) score 45 or sacrifice child*
>> Feb 14 03:51:40 fr-paris kernel: Killed process 7657 (ceph-osd)
>> total-vm:8650208kB, anon-rss:6124660kB, file-rss:1560kB
>> Feb 14 03:51:41 fr-paris systemd:* ceph-osd@3.service: main process
>> exited, code=killed, status=9/KILL*
>> Feb 14 03:51:41 fr-paris systemd: Unit ceph-osd@3.service entered failed
>> state.
>> Feb 14 03:51:41 fr-paris systemd: *ceph-osd@3.service failed.*
>> Feb 14 03:51:41 fr-paris systemd: cassandra.service: main process exited,
>> code=killed, status=9/KILL
>> Feb 14 03:51:41 fr-paris systemd: Unit cassandra.service entered failed
>> state.
>> Feb 14 03:51:41 fr-paris systemd: cassandra.service failed.
>> Feb 14 03:51:41 fr-paris ceph-mgr: 2017-02-14 03:51:41.978878
>> 7f51a3154700 -1 mgr ms_dispatch osd_map(7517..7517 src has 6951..7517) v3
>> Feb 14 03:51:42 fr-paris systemd: Device 
>> dev-disk-by\x2dpartlabel-ceph\x5cx20block.device
>> appeared twice with different sysfs paths /sys/devices/pci0000:00/0000:0
>> 0:03.2/0000:03:00.0/host0/target0:0:0/0:0:0:9/block/sdj/sdj2 and
>> /sys/devices/pci0000:00/0000:00:03.2/0000:03:00.0/host0/targ
>> et0:0:0/0:0:0:4/block/sde/sde2
>> Feb 14 03:51:42 fr-paris ceph-mgr: 2017-02-14 03:51:42.992477
>> 7f51a3154700 -1 mgr ms_dispatch osd_map(7518..7518 src has 6951..7518) v3
>> Feb 14 03:51:43 fr-paris ceph-mgr: 2017-02-14 03:51:43.508990
>> 7f51a3154700 -1 mgr ms_dispatch mgrdigest v1
>> Feb 14 03:51:48 fr-paris ceph-mgr: 2017-02-14 03:51:48.508970
>> 7f51a3154700 -1 mgr ms_dispatch mgrdigest v1
>> Feb 14 03:51:53 fr-paris ceph-mgr: 2017-02-14 03:51:53.509592
>> 7f51a3154700 -1 mgr ms_dispatch mgrdigest v1
>> Feb 14 03:51:58 fr-paris ceph-mgr: 2017-02-14 03:51:58.509936
>> 7f51a3154700 -1 mgr ms_dispatch mgrdigest v1
>> Feb 14 03:52:01 fr-paris systemd: ceph-osd@3.service holdoff time over,
>> scheduling restart.
>> Feb 14 03:52:02 fr-paris systemd: *Starting Ceph object storage daemon
>> osd.3.*..
>> Feb 14 03:52:02 fr-paris systemd: Started Ceph object storage daemon
>> osd.3.
>> Feb 14 03:52:02 fr-paris numactl: 2017-02-14 03:52:02.307106 7f1e499bb940
>> -1 WARNING: the following dangerous and experimental features are enabled:
>> bluestore,rocksdb
>> Feb 14 03:52:02 fr-paris numactl: 2017-02-14 03:52:02.317687 7f1e499bb940
>> -1 WARNING: the following dangerous and experimental features are enabled:
>> bluestore,rocksdb
>> Feb 14 03:52:02 fr-paris numactl: starting osd.3 at - osd_data
>> /var/lib/ceph/osd/ceph-3 /var/lib/ceph/osd/ceph-3/journal
>> Feb 14 03:52:02 fr-paris numactl: 2017-02-14 03:52:02.333522 7f1e499bb940
>> -1 WARNING: experimental feature 'bluestore' is enabled
>> Feb 14 03:52:02 fr-paris numactl: Please be aware that this feature is
>> experimental, untested,
>> Feb 14 03:52:02 fr-paris numactl: unsupported, and may result in data
>> corruption, data loss,
>> Feb 14 03:52:02 fr-paris numactl: and/or irreparable damage to your
>> cluster.  Do not use
>> Feb 14 03:52:02 fr-paris numactl: feature with important data.
>>
>> This seems to happen only in 11.2.0 and not in 11.1.x . Could you please
>> help us in resolving this issue by means of any config change to limit the
>> memory use on ceph-osd or a bug in the current kraken release.
>>
>> Thanks,
>> Muthu
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>
>
> --
> С уважением / Best regards
>
> Илья Летковский / Ilya Letkouski
>
> Phone, Viber: +375 29 3237335
>
> Minsk, Belarus (GMT+3)
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to