Re: [ceph-users] slow requests are blocked > 32 sec. Implicated osds 0, 2, 3, 4, 5 (REQUEST_SLOW)

2019-06-07 Thread Robert LeBlanc
With the low number of OSDs, you are probably satuarting the disks. Check with `iostat -xd 2` and see what the utilization of your disks are. A lot of SSDs don't perform well with Ceph's heavy sync writes and performance is terrible. If some of your drives are 100% while others are lower utilizati

Re: [ceph-users] Any CEPH's iSCSI gateway users?

2019-06-07 Thread Heðin Ejdesgaard Møller
-BEGIN PGP SIGNED MESSAGE- Hash: SHA256 I recently upgraded a RHCS-3.0 cluster with 4 iGW's to RHCS-3.2 on top of RHEL- 7.6 Big block size performance went from ~350MB/s to about 1100MB/s on each lun, seen from a VM in vSphere-6.5 with data read from an ssd pool and written to a hdd pool,

[ceph-users] radosgw dying

2019-06-07 Thread DHilsbos
All; I have a test and demonstration cluster running (3 hosts, MON, MGR, 2x OSD per host), and I'm trying to add a 4th host for gateway purposes. The radosgw process keeps dying with: 2019-06-07 15:59:50.700 7fc4ef273780 0 ceph version 14.2.1 (d555a9489eb35f84f2e1ef49b77e19da9d113972) nautilus

Re: [ceph-users] Can I limit OSD memory usage?

2019-06-07 Thread EDH - Manuel Rios Fernandez
Hi Sergei, Please add to your host: For 64GB RAM, reserve 1GB. vm.min_free_kbytes = 1048576 For 128GB RAM, reserve 2GB. vm.min_free_kbytes = 2097152 For 256GB RAM, reserve 3GB. vm.min_free_kbytes = 3145728 This will prevent to your OSD to use ALL memory of host and OOM act. Regards -M

[ceph-users] Can I limit OSD memory usage?

2019-06-07 Thread Sergei Genchev
Hi, My OSD processes are constantly getting killed by OOM killer. My cluster has 5 servers, each with 18 spinning disks, running 18 OSD daemons in 48GB of memory. I was trying to limit OSD cache, according to http://docs.ceph.com/docs/mimic/rados/configuration/bluestore-config-ref/ [osd] bluest

Re: [ceph-users] PG stuck peering - OSD cephx: verify_authorizer key problem

2019-06-07 Thread Jan Pekař - Imatic
I think I found source of my problems. One of my monitors was on problematic disk. It was not responding for few seconds so mon was taken out and in the cluster frequently (few times a day). I expect, that key exchange was during that time taken with this problematic monitor and was not successf

Re: [ceph-users] OSD RAM recommendations

2019-06-07 Thread Mark Nelson
The truth of the matter is that folks try to boil this down to some kind of hard and fast rule but it's often not that simple. With our current default settings for pglog, rocksdb WAL buffers, etc, the OSD basically needs about 1GB of RAM for bare-bones operation (not under recovery or extreme

Re: [ceph-users] OSD RAM recommendations

2019-06-07 Thread EDH - Manuel Rios Fernandez
In nautilus min 4GB per disk . In case of ssd/nvme 6-12GB per disk. 8GB per disk is a good way to get performance +2/4GB for the OS. Regards, Manuel -Mensaje original- De: ceph-users En nombre de jes...@krogh.cc Enviado el: viernes, 7 de junio de 2019 19:36 Para: Jorge Garcia CC: cep

Re: [ceph-users] OSD RAM recommendations

2019-06-07 Thread jesper
> I'm a bit confused by the RAM recommendations for OSD servers. I have > also seen conflicting information in the lists (1 GB RAM per OSD, 1 GB > RAM per TB, 3-5 GB RAM per OSD, etc.). I guess I'm a lot better with a > concrete example: I think it depends on the usagepattern - the more the better

[ceph-users] OSD RAM recommendations

2019-06-07 Thread Jorge Garcia
I'm a bit confused by the RAM recommendations for OSD servers. I have also seen conflicting information in the lists (1 GB RAM per OSD, 1 GB RAM per TB, 3-5 GB RAM per OSD, etc.). I guess I'm a lot better with a concrete example: Say this is your cluster (using Bluestore): 8 Machines serving

Re: [ceph-users] Reweight OSD to 0, why doesn't report degraded if UP set under Pool Size

2019-06-07 Thread Tarek Zegar
Paul / All I'm not sure what warning your are referring to, I'm on Nautilus. The point I'm getting at is if you weight out all OSD on a host with a cluster of 3 OSD hosts with 3 OSD each, crush rule = host, then write to the cluster, it *should* imo not just say remapped but undersized / degrade

Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-07 Thread John Petrini
How's iowait look on your disks? How have you configured your disks and what are your cache settings? Did you disable cstates? On Friday, June 7, 2019, Stolte, Felix wrote: > Hi Sinan, > > thanks for the numbers. I am a little bit surprised that your SSD pool has nearly the same stats as you SA

Re: [ceph-users] Remove rbd image after interrupt of deletion command

2019-06-07 Thread Jason Dillaman
On Fri, Jun 7, 2019 at 7:22 AM Sakirnth Nagarasa wrote: > > On 6/6/19 5:09 PM, Jason Dillaman wrote: > > On Thu, Jun 6, 2019 at 10:13 AM Sakirnth Nagarasa > > wrote: > >> > >> On 6/6/19 3:46 PM, Jason Dillaman wrote: > >>> Can you run "rbd trash ls --all --long" and see if your image > >>> is lis

Re: [ceph-users] balancer module makes OSD distribution worse

2019-06-07 Thread Josh Haft
95% of usage is CephFS. Remaining is split between RGW and RBD. On Wed, Jun 5, 2019 at 3:05 PM Gregory Farnum wrote: > > I think the mimic balancer doesn't include omap data when trying to > balance the cluster. (Because it doesn't get usable omap stats from > the cluster anyway; in Nautilus I th

Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-07 Thread Stolte, Felix
Hi Sinan, thanks for the numbers. I am a little bit surprised that your SSD pool has nearly the same stats as you SAS pool. Nevertheless I would expect our pools to perform like your SAS pool, at least regarding to writes since all our write ops should be placed on our SSDs. But since I only

Re: [ceph-users] Understanding Cephfs / how to have fs in line with OSD pool ?

2019-06-07 Thread Vincent Pharabot
Ok it seems that OSD pools are finally to correctly restored (quite same amount of objects/datas than before restoration) so that should explain the situation I will have to dig why Vincent Le ven. 7 juin 2019 à 08:41, Vincent Pharabot a écrit : > Hello Cephers, > > > > I’m trying to understand

Re: [ceph-users] Remove rbd image after interrupt of deletion command

2019-06-07 Thread Sakirnth Nagarasa
On 6/6/19 5:09 PM, Jason Dillaman wrote: > On Thu, Jun 6, 2019 at 10:13 AM Sakirnth Nagarasa > wrote: >> >> On 6/6/19 3:46 PM, Jason Dillaman wrote: >>> Can you run "rbd trash ls --all --long" and see if your image >>> is listed? >> >> No, it is not listed. >> >> I did run: >> rbd trash ls --all -

Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-07 Thread Sinan Polat
Hi Felix, I have 2 Pools, a SSD only and a SAS only pool. SSD pool is spread over 12 OSD servers. SAS pool is spread over 6 OSD servers. See results (SSD Only Pool): # sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G --file-test-mode=rndrw --file-rw-ratio=2 run sysbench 1.

Re: [ceph-users] is rgw crypt default encryption key long term supported ?

2019-06-07 Thread Scheurer François
Hello Casey We found something weird during our testing of the rgw_crypt_default_encryption_key=""xxx" parameter. s3cms behaves like expected: s3cmd is then always writing encrypted objects s3cmd can read encrypted and unencrypted objects but swift does not support encryption: swift can read

Re: [ceph-users] performance in a small cluster

2019-06-07 Thread Paul Emmerich
On Tue, Jun 4, 2019 at 1:39 PM wrote: > >> Basically they max out at around 1000 IOPS and report 100% > >> utilization and feel slow. > >> > >> Haven't seen the 5200 yet. > > Micron 5100s performs wonderfully! > > You have to just turn its write cache off: > > hdparm -W 0 /dev/sdX > > 1000 IOPS m

Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-07 Thread Stolte, Felix
Hi Sinan, that would be great. The numbers should differ a lot, since you have an all flash pool, but it would be interesting, what we could expect from such a configuration. Regards Felix - ---

Re: [ceph-users] getting pg inconsistent periodly

2019-06-07 Thread Paul Emmerich
On Mon, Jun 3, 2019 at 9:02 AM Hervé Ballans wrote: > Hi all, > > For information, I updated my Luminous cluster to the latest version > 12.2.12 two weeks ago and, since then, I no longer encounter any problems > of inconsistent pgs :) > You probably were affected by https://tracker.ceph.com/iss

Re: [ceph-users] Any CEPH's iSCSI gateway users?

2019-06-07 Thread Paul Emmerich
Hi, ceph-iscsi 3.0 fixes a lot of problems and limitations of the older gateway. Best way to run it on Debian/Ubuntu is to build it yourself Paul -- Paul Emmerich Looking for help with your Ceph cluster? Contact us at https://croit.io croit GmbH Freseniusstr. 31h 81247 München www.croit.io

Re: [ceph-users] Reweight OSD to 0, why doesn't report degraded if UP set under Pool Size

2019-06-07 Thread Paul Emmerich
remapped no longer triggers a health warning in nautilus. Your data is still there, it's just on the wrong OSD if that OSD is still up and running. Paul -- Paul Emmerich Looking for help with your Ceph cluster? Contact us at https://croit.io croit GmbH Freseniusstr. 31h 81247 München www.cro

Re: [ceph-users] OSD hanging on 12.2.12 by message worker

2019-06-07 Thread Stefan Kooman
Quoting Max Vernimmen (vernim...@textkernel.nl): > Thank you for the suggestion to use the bitmap allocator. I looked at the > ceph documentation and could find no mention of this setting. This makes me > wonder how safe and production ready this setting really is. I'm hesitant > to apply that to o

Re: [ceph-users] typical snapmapper size

2019-06-07 Thread Dan van der Ster
On Thu, Jun 6, 2019 at 8:00 PM Sage Weil wrote: > > Hello RBD users, > > Would you mind running this command on a random OSD on your RBD-oriented > cluster? > > ceph-objectstore-tool \ > --data-path /var/lib/ceph/osd/ceph-NNN \ > > '["meta",{"oid":"snapmapper","key":"","snapid":0,"hash":2758339

Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-07 Thread Sinan Polat
Hi Felix, I can run your commands inside an OpenStack VM. Tthe storage cluster contains of 12 OSD servers, holding each 8x 960GB SSD. Luminous FileStore. Replicated 3. Would it help you to run your command on my cluster? Sinan > Op 7 jun. 2019 om 08:52 heeft Stolte, Felix het > volgende gesc

[ceph-users] v12.2.12 mds FAILED assert(session->get_nref() == 1)

2019-06-07 Thread Dan van der Ster
Hi all, Just a quick heads up, and maybe a check if anyone else is affected. After upgrading our MDS's from v12.2.11 to v12.2.12, we started getting crashes with /builddir/build/BUILD/ceph-12.2.12/src/mds/MDSRank.cc: 1304: FAILED assert(session->get_nref() == 1) I opened a ticket here with

Re: [ceph-users] OSD hanging on 12.2.12 by message worker

2019-06-07 Thread Igor Fedotov
Hi Max, I don't think this is allocator related issue. The symptoms that triggered us to start using bitmap allocator over stupid one were: - write op latency gradually increasing over time (days not hours) - perf showing significant amount of time spent in allocator related function - OSD

Re: [ceph-users] OSD hanging on 12.2.12 by message worker

2019-06-07 Thread Max Vernimmen
Thank you for the suggestion to use the bitmap allocator. I looked at the ceph documentation and could find no mention of this setting. This makes me wonder how safe and production ready this setting really is. I'm hesitant to apply that to our production environment. If the allocator setting helps

[ceph-users] 200 clusters vs 1 admin (Cephalocon 2019)

2019-06-07 Thread Bartosz Rabiega
Hi, I did the talk "200 clusters vs 1 admin" on latest Cephalocon (Barcelona 2019). There were some questions on the backstage about implementation details, opensourcing, etc. I just thought that if anyone has some questions then you may catch me here via email (bartosz.rabi...@corp.ovh.