Re: [ceph-users] "session established", "io error", "session lost, hunting for new mon" solution/fix

2019-07-16 Thread Ilya Dryomov
On Fri, Jul 12, 2019 at 5:38 PM Marc Roos wrote: > > > Thanks Ilya for explaining. Am I correct to understand from the link[0] > mentioned in the issue, that because eg. I have an unhealthy state for > some time (1 pg on a insignificant pool) I have larger osdmaps, > triggering this issue? Or is j

Re: [ceph-users] Returning to the performance in a small cluster topic

2019-07-16 Thread Lars Marowsky-Bree
On 2019-07-15T19:08:26, " Drobyshevskiy, Vladimir " wrote: Hi Vladimir, is this a replicated or EC pool? > The cluster itself: > nautilus > 6 nodes, 7 SSD with 2 OSDs per SSD (14 OSDs in overall). You mean 14 OSDs per node, right? > Each node: 2x Intel Xeon E5-2665 v1 (governor = perf

[ceph-users] HEALTH_WARN 1 MDSs report slow metadata IOs

2019-07-16 Thread Dietmar Rieder
Hi, We are running ceph version 14.1.2 with cephfs only. I just noticed that one of our pgs had scrub errors which I could repair # ceph health detail HEALTH_ERR 1 MDSs report slow metadata IOs; 1 MDSs report slow requests; 1 scrub errors; Possible data damage: 1 pg inconsistent MDS_SLOW_METADAT

Re: [ceph-users] "session established", "io error", "session lost, hunting for new mon" solution/fix

2019-07-16 Thread Marc Roos
Can you give me a little pointer on how and where to do the verbose logging? I am really interested in discovery why I am getting the large osdmap message. Maybe it is because of the snapshot being created? -Original Message- From: Ilya Dryomov [mailto:idryo...@gmail.com] Cc: paul.e

[ceph-users] pools limit

2019-07-16 Thread M Ranga Swami Reddy
Hello - I have created 10 nodes ceph cluster with 14.x version. Can you please confirm below: Q1 - Can I create 100+ pool (or more) on the cluster? (the reason is - creating a pool per project). Any limitation on pool creation? Q2 - In the above pool - I use 128 PG-NUM - to start with and enable

Re: [ceph-users] Returning to the performance in a small cluster topic

2019-07-16 Thread Drobyshevskiy, Vladimir
Hi, Lars! > is this a replicated or EC pool? > This is a replicated pool with size = 3. > > > The cluster itself: > > nautilus > > 6 nodes, 7 SSD with 2 OSDs per SSD (14 OSDs in overall). > > You mean 14 OSDs per node, right? > No, I mean it is 14 OSDs on 7 SSDs at all: 5 nodes with 1 SSD

Re: [ceph-users] Returning to the performance in a small cluster topic

2019-07-16 Thread Lars Marowsky-Bree
On 2019-07-16T19:27:32, " Drobyshevskiy, Vladimir " wrote: > This is a replicated pool with size = 3. > Is it possible to check this anyhow? ping shows average latency ~0.147 ms > which is pretty high for IB but might be reasonable (IPoIB). This means that, barring any other overhead, you're li

Re: [ceph-users] Returning to the performance in a small cluster topic

2019-07-16 Thread Drobyshevskiy, Vladimir
Hello, Paul! You are effectively measuring the latency with jobs=1 here (which is > appropriate considering that the WAL of a DB is effectively limited by > latency) and yeah, a networked file system will always be a little bit > slower than a local disk. > > But I think you should be able to get

Re: [ceph-users] pools limit

2019-07-16 Thread Janne Johansson
Den tis 16 juli 2019 kl 16:16 skrev M Ranga Swami Reddy < swamire...@gmail.com>: > Hello - I have created 10 nodes ceph cluster with 14.x version. Can you > please confirm below: > Q1 - Can I create 100+ pool (or more) on the cluster? (the reason is - > creating a pool per project). Any limitation

Re: [ceph-users] HEALTH_WARN 1 MDSs report slow metadata IOs

2019-07-16 Thread Dietmar Rieder
On 7/16/19 4:11 PM, Dietmar Rieder wrote: > Hi, > > We are running ceph version 14.1.2 with cephfs only. > > I just noticed that one of our pgs had scrub errors which I could repair > > # ceph health detail > HEALTH_ERR 1 MDSs report slow metadata IOs; 1 MDSs report slow requests; > 1 scrub erro

Re: [ceph-users] What if etcd is lost

2019-07-16 Thread Janne Johansson
Den mån 15 juli 2019 kl 23:05 skrev Oscar Segarra : > Hi Frank, > Thanks a lot for your quick response. > Yes, the use case that concerns me is the following: > 1.- I bootstrap a complete cluster mons, osds, mgr, mds, nfs, etc using > etcd as a key store > as a key store ... for what? Are you stu

[ceph-users] New best practices for osds???

2019-07-16 Thread Stolte, Felix
Hi guys, our ceph cluster is performing way less than it could, based on the disks we are using. We could narrow it down to the storage controller (LSI SAS3008 HBA) in combination with an SAS expander. Yesterday we had a meeting with our hardware reseller and sale representatives of the hardwar

Re: [ceph-users] New best practices for osds???

2019-07-16 Thread Marc Roos
With the sas expander you are putting more drives on 'one port'. Just make sure you do not create a bottle neck there, adding to many drives. I guess this depends on the speed of the drives. Then you should be fine not? -Original Message- From: Stolte, Felix [mailto:f.sto...@fz-ju

Re: [ceph-users] What if etcd is lost

2019-07-16 Thread Paul Emmerich
All that etcd stuff is specific to ceph-docker, it supports storing config options in there. Most of these features are obselete with the mon config store nowadays. To answer the original question: all you need to recover your Ceph cluster are the mon and osd disks, they have all the data you need

Re: [ceph-users] New best practices for osds???

2019-07-16 Thread Paul Emmerich
On Tue, Jul 16, 2019 at 5:43 PM Stolte, Felix wrote: > > They told us, that "best practices" for ceph would be to deploy disks as > Raid 0 consisting of one disk using a raid controller with a big writeback > cache. > > Since this "best practice" is new to me, I would like to hear your opinion >

Re: [ceph-users] What if etcd is lost

2019-07-16 Thread Oscar Segarra
Thanks a lot Janne, Well, maybe I'm missunderstanding how ceph stores keyrings in etcd... https://github.com/ceph/ceph-container/blob/master/src/daemon/config.kv.etcd.sh bootstrap="bootstrap${array[1]}Keyring" etcdctl "${ETCDCTL_OPTS[@]}" "${KV_TLS[@]}" set "${CLUSTER_PATH}"/" ${bootstrap}" < "$

Re: [ceph-users] pools limit

2019-07-16 Thread Paul Emmerich
100+ pools work fine if you can get the PG count right (auto-scaler helps, there are some options that you'll need to tune for small-ish pools). But it's not a "nice" setup. Have you considered using namespaces instead? Paul -- Paul Emmerich Looking for help with your Ceph cluster? Contact us

Re: [ceph-users] What if etcd is lost

2019-07-16 Thread Janne Johansson
Nah, it was me not running container versions. The bootstrap keys are used to get daemons up without giving out too much admin access to them, so my guess is that as soon as you have a cluster going, you can always read out or create new bootstrap keys if needed later, but they are not necessary f

Re: [ceph-users] What if etcd is lost

2019-07-16 Thread Oscar Segarra
Hi Paul, That is the initial question, is it possible to recover my ceph cluster (docker based) if I loose all information stored in the etcd... I don't know if anyone has a clear answer to these questions.. 1.- I bootstrap a complete cluster mons, osds, mgr, mds, nfs, etc using etcd as a key s

Re: [ceph-users] What if etcd is lost

2019-07-16 Thread Janne Johansson
Den tis 16 juli 2019 kl 18:15 skrev Oscar Segarra : > Hi Paul, > That is the initial question, is it possible to recover my ceph cluster > (docker based) if I loose all information stored in the etcd... > I don't know if anyone has a clear answer to these questions.. > 1.- I bootstrap a complete c

Re: [ceph-users] New best practices for osds???

2019-07-16 Thread Kaspar Bosma
Hi,I concur with Paul. Any kind of RAID for OSD devices would be ill advised.Make sure the 3008 SAS controller is flashed with the IT firmware so it will operate in JBOD mode.Regards, KasparOp 16 juli 2019 om 17:56 schreef Paul Emmerich : On Tue, Jul 16, 2019 at 5:43 PM Stolte, Felix < f.sto...@f

Re: [ceph-users] pools limit

2019-07-16 Thread M Ranga Swami Reddy
Thanks for your reply.. Here, new pool creations and pg auto scale may cause rebalance..which impact the ceph cluster performance.. Please share name space detail like how to use etc On Tue, 16 Jul, 2019, 9:30 PM Paul Emmerich, wrote: > 100+ pools work fine if you can get the PG count right (

[ceph-users] Multisite RGW - endpoints configuration

2019-07-16 Thread P. O.
Hi all, I have multisite RGW setup with one zonegroup and two zones. Each zone has one endpoint configured like below: "zonegroups": [ { ... "is_master": "true", "endpoints": ["http://192.168.100.1:80";], "zones": [ { "name": "primary_1", "endpoints": ["http://192.168.100.1:80";]

Re: [ceph-users] Multisite RGW - endpoints configuration

2019-07-16 Thread Casey Bodley
We used to have issues when a load balancer was in front of the sync endpoints, because our http client didn't time out stalled connections. Those are resolved in luminous, but we still recommend using the radosgw addresses directly to avoid shoveling data through an extra proxy. Internally, sy

Re: [ceph-users] What if etcd is lost

2019-07-16 Thread Oscar Segarra
Hi, The main question is that /etc/ceph is not longer needed when you use etcd... docker run -d --net=host \ -v /var/lib/ceph:/var/lib/ceph \ -e MON_IP=192.168.0.20 \ -e CEPH_PUBLIC_NETWORK=192.168.0.0/24 \ -e KV_TYPE=etcd \ -e KV_IP=192.168.0.20 \ ceph/daemon mon I'd like to know/predict the b

[ceph-users] cephfs snapshot scripting questions

2019-07-16 Thread Robert Ruge
Greetings. Before I reinvent the wheel has anyone written a script to maintain X number of snapshots on a cephfs file system that can be run through cron? I am aware of the cephfs-snap code but just wondering if there are any other options out there. On a related note which of these options wou

Re: [ceph-users] New best practices for osds???

2019-07-16 Thread Brent Kennedy
Went over this with a few clusters built last year, it seems best practices ( at least in the white papers I have read ) only use jbod configuration( and so have we for all clusters ). I agree regarding the writeback cache issues and not using it but that's more due to how we use hardware ( defini

Re: [ceph-users] New best practices for osds???

2019-07-16 Thread Mike O'Connor
On 17/7/19 1:12 am, Stolte, Felix wrote: > Hi guys, > > our ceph cluster is performing way less than it could, based on the disks we > are using. We could narrow it down to the storage controller (LSI SAS3008 > HBA) in combination with an SAS expander. Yesterday we had a meeting with our > hardw

Re: [ceph-users] HEALTH_WARN 1 MDSs report slow metadata IOs

2019-07-16 Thread Dietmar Rieder
On 7/16/19 5:34 PM, Dietmar Rieder wrote: > On 7/16/19 4:11 PM, Dietmar Rieder wrote: >> Hi, >> >> We are running ceph version 14.1.2 with cephfs only. >> >> I just noticed that one of our pgs had scrub errors which I could repair >> >> # ceph health detail >> HEALTH_ERR 1 MDSs report slow metadata