olcDbShmKey only applies to BDB and HDB backends but I'm using the new MDB 
backend.


Am 28.02.19 um 14:47 schrieb Marc Roos:
> If you have every second disk io with your current settings, which I 
> also had with 'default' settings. There are some optimizations you can 
> do, bringing it down to every 50 seconds or so. Adding the olcDbShmKey 
> will allow for slapd to access the db cache. 
> I am getting an error of sharedmemory settings when rebooting (centos7), 
> but maintainers of slapd said that I can ignore that. Dont have any 
> problems since using this also.
> 
> 
> 
> -----Original Message-----
> From: Uwe Sauter [mailto:uwe.sauter...@gmail.com] 
> Sent: 28 February 2019 14:34
> To: Marc Roos; ceph-users; vitalif
> Subject: Re: [ceph-users] Fwd: Re: Blocked ops after change from 
> filestore on HDD to bluestore on SDD
> 
> Do you have anything particular in mind? I'm using mdb backend with 
> maxsize = 1GB but currently the files are only about 23MB.
> 
> 
>>
>> I am having quite a few openldap servers (slaves) running also, make 
>> sure to use proper caching that saves a lot of disk io.
>>
>>
>>
>>
>> -----Original Message-----
>> Sent: 28 February 2019 13:56
>> To: uwe.sauter...@gmail.com; Uwe Sauter; Ceph Users
>> Subject: *****SPAM***** Re: [ceph-users] Fwd: Re: Blocked ops after 
>> change from filestore on HDD to bluestore on SDD
>>
>> "Advanced power loss protection" is in fact a performance feature, not 
> 
>> a safety one.
>>
>>
>> 28 февраля 2019 г. 13:03:51 GMT+03:00, Uwe Sauter 
>> <uwe.sauter...@gmail.com> пишет:
>>
>>      Hi all,
>>      
>>      thanks for your insights.
>>      
>>      Eneko,
>>      
>>
>>              We tried to use a Samsung 840 Pro SSD as OSD some time ago 
> and it 
>> was a no-go; it wasn't that performance was bad, it
>>              just didn't work for the kind of use of OSD. Any HDD was 
> better than 
>> it (the disk was healthy and have been used in a
>>              software raid-1 for a pair of years).
>>              
>>              I suggest you check first that your Samsung 860 Pro disks 
> work well 
>> for Ceph. Also, how is your host's RAM?
>>
>>
>>      As already mentioned the hosts each have 64GB RAM. Each host has 
> 3 
>> SSDs for OSD usage. Each OSD is using about 1.3GB virtual
>>      memory / 400MB residual memory.
>>      
>>      
>>      
>>      Joachim,
>>      
>>
>>              I can only recommend the use of enterprise SSDs. We've 
> tested many 
>> consumer SSDs in the past, including your SSDs. Many
>>              of them are not suitable for long-term use and some weard 
> out within 
>> 6 months.
>>
>>
>>      Unfortunately I couldn't afford enterprise grade SSDs. But I 
> suspect 
>> that my workload (about 20 VMs for our infrastructure, the
>>      most IO demanding is probably LDAP) is light enough that wearout 
>> won't be a problem.
>>      
>>      The issue I'm seeing then is probably related to direct IO if 
> using 
>> bluestore. But with filestore, the file system cache probably
>>      hides the latency issues.
>>      
>>      
>>      Igor,
>>      
>>
>>              AFAIR Samsung 860 Pro isn't for enterprise market, you 
> shouldn't use 
>> consumer SSDs for Ceph.
>>              
>>              I had some experience with Samsung 960 Pro a while ago and 
> it turned 
>> out that it handled fsync-ed writes very slowly
>>              (comparing to the original/advertised performance). Which 
> one can 
>> probably explain by the lack of power loss protection
>>              for these drives. I suppose it's the same in your case.
>>              
>>              Here are a couple links on the topic:
>>              
>>              
>> https://www.percona.com/blog/2018/02/08/fsync-performance-storage-devi
>> ces/
>>              
>>              
>> https://www.sebastien-han.fr/blog/2014/10/10/ceph-how-to-test-if-your-
>> ssd-is-suitable-as-a-journal-device/
>>
>>
>>      Power loss protection wasn't a criteria for me as the cluster 
> hosts 
>> are distributed in two buildings with separate battery backed
>>      UPSs. As mentioned above I suspect the main difference for my 
> case 
>> between filestore and bluestore is file system cache vs. direct
>>      IO. Which means I will keep using filestore.
>>      
>>      Regards,
>>      
>>              Uwe
>> ________________________________
>>
>>      ceph-users mailing list
>>      ceph-users@lists.ceph.com
>>      http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>> --
>> With best regards,
>> Vitaliy Filippov
>>
>>
> 
> 
> 

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to