On Tue, Feb 20, 2018 at 9:33 PM, nokia ceph <nokiacephus...@gmail.com> wrote:
> Hi Alfredo Deza, > > I understand the point between lvm and simple however we see issue , was > it issue in luminous because we use same ceph config and workload from > client. The graphs i attached in previous mail is from ceph-volume lvm osd. > If the issue is a performance regression in Luminous I wouldn't know :( I was trying to say that if you are seeing the same regression with previously deployed OSDs then it can't possibly be a thing we are doing incorrectly in ceph-volume > > In this case does it ococcupies 2 times only inside ceph. If we consider > only lvm based system does this high iops because of dm-cache created for > each osd?. > Not sure again. Maybe someone else might be able to chime in on this. > > Meanwhile i will update some graphs to show this once i have. > > Thanks, > Muthu > > On Tuesday, February 20, 2018, Alfredo Deza <ad...@redhat.com> wrote: > >> >> >> On Mon, Feb 19, 2018 at 9:29 PM, nokia ceph <nokiacephus...@gmail.com> >> wrote: >> >>> Hi Alfredo Deza, >>> >>> We have 5 node platforms with lvm osd created from scratch and another 5 >>> node platform migrated from kraken which is ceph volume simple. Both has >>> same issue . Both platform has only hdd for osd. >>> >>> We also noticed 2 times disk iops more compare to kraken , this causes >>> less read performance. During rocksdb compaction the situation is worse. >>> >>> >>> Meanwhile we are building another platform creating osd using ceph-disk >>> and analyse on this. >>> >> >> If you have two platforms, one with `simple` and the other one with `lvm` >> experiencing the same, then something else must be at fault here. >> >> The `simple` setup in ceph-volume basically keeps everything as it was >> before, it just captures details of what devices were being used so OSDs >> can be started. There is no interaction from ceph-volume >> in there that could cause something like this. >> >> >> >>> Thanks, >>> Muthu >>> >>> >>> >>> On Tuesday, February 20, 2018, Alfredo Deza <ad...@redhat.com> wrote: >>> >>>> >>>> >>>> On Mon, Feb 19, 2018 at 2:01 PM, nokia ceph <nokiacephus...@gmail.com> >>>> wrote: >>>> >>>>> Hi All, >>>>> >>>>> We have 5 node clusters with EC 4+1 and use bluestore since last year >>>>> from Kraken. >>>>> Recently we migrated all our platforms to luminous 12.2.2 and finally >>>>> all OSDs migrated to ceph-volume simple type and on few platforms >>>>> installed >>>>> ceph using ceph-volume . >>>>> >>>>> Now we see two times more traffic in read compare to client traffic on >>>>> migrated platform and newly created platforms . This was not the case in >>>>> older releases where ceph status read B/W will be same as client read >>>>> traffic. >>>>> >>>>> Some network graphs : >>>>> >>>>> *Client network interface* towards ceph public interface : shows >>>>> *4.3Gbps* read >>>>> >>>>> >>>>> [image: Inline image 2] >>>>> >>>>> *Ceph Node Public interface* : Each node around 960Mbps * 5 node =* >>>>> 4.6 Gbps *- this matches. >>>>> [image: Inline image 3] >>>>> >>>>> Ceph status output : show 1032 MB/s =* 8.06 Gbps* >>>>> >>>>> cn6.chn6us1c1.cdn ~# ceph status >>>>> cluster: >>>>> id: abda22db-3658-4d33-9681-e3ff10690f88 >>>>> health: HEALTH_OK >>>>> >>>>> services: >>>>> mon: 5 daemons, quorum cn6,cn7,cn8,cn9,cn10 >>>>> mgr: cn6(active), standbys: cn7, cn9, cn10, cn8 >>>>> osd: 340 osds: 340 up, 340 in >>>>> >>>>> data: >>>>> pools: 1 pools, 8192 pgs >>>>> objects: 270M objects, 426 TB >>>>> usage: 581 TB used, 655 TB / 1237 TB avail >>>>> pgs: 8160 active+clean >>>>> 32 active+clean+scrubbing >>>>> >>>>> io: >>>>> client: *1032 MB/s rd*, 168 MB/s wr, 1908 op/s rd, 1594 op/s wr >>>>> >>>>> >>>>> Write operation we don't see this issue. Client traffic and this >>>>> matches. >>>>> Is this expected behavior in Luminous and ceph-volume lvm or a bug ? >>>>> Wrong calculation in ceph status read B/W ? >>>>> >>>> >>>> You mentioned `ceph-volume simple` but here you say lvm. With LVM >>>> ceph-volume will create the OSDs from scratch, while "simple" will keep >>>> whatever OSD was created before. >>>> >>>> Have you created the OSDs from scratch with ceph-volume? or is it just >>>> using "simple" , managing a previously deployed OSD? >>>> >>>>> >>>>> Please provide your feedback. >>>>> >>>>> Thanks, >>>>> Muthu >>>>> >>>>> >>>>> >>>>> _______________________________________________ >>>>> ceph-users mailing list >>>>> ceph-users@lists.ceph.com >>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>> >>>>> >>>> >>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com