Re: [ceph-users] luminous vs jewel rbd performance

2017-09-19 Thread Rafael Lopez
Hi Alexandre, Yeah we are using filestore for the moment with luminous. With regards to client, I tried both jewel and luminous librbd versions against the luminous cluster - similar results. I am running fio on a physical machine with fio rbd engine. This is a snippet of the fio config for the r

Re: [ceph-users] v12.2.0 bluestore - OSD down/crash " internal heartbeat not healthy, dropping ping reques "

2017-09-19 Thread Henrik Korkuc
On 17-09-20 08:06, nokia ceph wrote: Hello, Env:- RHEL 7.2 , 3.10.0-327.el7.x86_64 , EC 4+1 , bluestore We are writing to ceph via librados C API  . Testing with rados no issues. The same we tested with Jewel/kraken without any issue. Need your view how to debug this issue? maybe similar to

Re: [ceph-users] luminous vs jewel rbd performance

2017-09-19 Thread Alexandre DERUMIER
Hi so, you use also filestore on luminous ? do you have also upgraded librbd on client ? (are you benching inside a qemu machine ? or directly with fio-rbd ?) (I'm going to do a lot of benchmarks in coming week, I'll post results on mailing soon.) - Mail original - De: "Rafael Lop

[ceph-users] luminous vs jewel rbd performance

2017-09-19 Thread Rafael Lopez
hey guys. wondering if anyone else has done some solid benchmarking of jewel vs luminous, in particular on the same cluster that has been upgraded (same cluster, client and config). we have recently upgraded a cluster from 10.2.9 to 12.2.0, and unfortunately i only captured results from a single

[ceph-users] v12.2.0 bluestore - OSD down/crash " internal heartbeat not healthy, dropping ping reques "

2017-09-19 Thread nokia ceph
Hello, Env:- RHEL 7.2 , 3.10.0-327.el7.x86_64 , EC 4+1 , bluestore We are writing to ceph via librados C API . Testing with rados no issues. The same we tested with Jewel/kraken without any issue. Need your view how to debug this issue? >> OSD.log == ~~~ 2017-09-18 14:51:59.895746 7f1e7

Re: [ceph-users] OSD assert hit suicide timeout

2017-09-19 Thread Jordan Share
We had suicide timeouts, but unfortunately I can't remember the specific root cause at this point. It was definitely one of two things: * too low of net.netfilter.nf_conntrack_max (preventing the osds from opening new connections to each other) * too low of kernel.pid_max or kernel.threads-m

Re: [ceph-users] OSD assert hit suicide timeout

2017-09-19 Thread Stanley Zhang
We don't use EC pools, but my experience with similar slow requests on RGW+replicated_pools is that in the logs you need to find out the first slow request and identify where it's from, for example, is it deep-scrub, or some client accessing corrupted objects, disk errors etc. On 20/09/17 8:1

Re: [ceph-users] ceph-osd restartd via systemd in case of disk error

2017-09-19 Thread Stanley Zhang
I like this, there is some similar ideas we probably can borrow from Cassandra on disk failure # policy for data disk failures: # die: shut down gossip and Thrift and kill the JVM for any fs errors or #  single-sstable errors, so the node can be replaced. # stop_paranoid: shut down gossip a

Re: [ceph-users] Ceph fails to recover

2017-09-19 Thread David Turner
Can you please provide the output of `ceph status`, `ceph osd tree`, and `ceph health detail`? Thank you. On Tue, Sep 19, 2017 at 2:59 PM Jonas Jaszkowic < jonasjaszkowic.w...@gmail.com> wrote: > Hi all, > > I have setup a Ceph cluster consisting of one monitor, 32 OSD hosts (1 OSD > of size 320

[ceph-users] OSD assert hit suicide timeout

2017-09-19 Thread David Turner
Just starting 3 nights ago we started seeing OSDs randomly going down in our cluster (Jewel 10.2.7). At first I saw that each OSD that was recently marked down in the cluster (`ceph osd dump | grep -E '^osd\.[0-9]+\s' | sort -nrk11` sorted list of OSDs by which OSDs have been marked down in the mo

[ceph-users] Ceph fails to recover

2017-09-19 Thread Jonas Jaszkowic
Hi all, I have setup a Ceph cluster consisting of one monitor, 32 OSD hosts (1 OSD of size 320GB per host) and 16 clients which are reading and writing to the cluster. I have one erasure coded pool (shec plugin) with k=8, m=4, c=3 and pg_num=256. Failure domain is host. I am able to reach a HEA

Re: [ceph-users] mds: failed to decode message of type 43 v7: buffer::end_of_buffer

2017-09-19 Thread Gregory Farnum
You've probably run in to http://tracker.ceph.com/issues/16010 — do you have very large directories? (Or perhaps just a whole bunch of unlinked files which the MDS hasn't managed to trim yet?) On Tue, Sep 19, 2017 at 11:51 AM Christian Salzmann-Jäckel < christian.salzm...@fu-berlin.de> wrote: > H

[ceph-users] mds: failed to decode message of type 43 v7: buffer::end_of_buffer

2017-09-19 Thread Christian Salzmann-Jäckel
Hi, we run cephfs (10.2.9 on Debian jessie; 108 OSDs on 9 nodes) as scratch filesystem for a HPC cluster using IPoIB interconnect with kernel client (Debian backports kernel version 4.9.30). Our clients started blocking on file system access. Logs show 'mds0: Behind on trimming' and slow reque

Re: [ceph-users] What HBA to choose? To expand or not to expand?

2017-09-19 Thread Jake Young
On Tue, Sep 19, 2017 at 9:38 AM Kees Meijs wrote: > Hi Jake, > > On 19-09-17 15:14, Jake Young wrote: > > Ideally you actually want fewer disks per server and more servers. > > This has been covered extensively in this mailing list. Rule of thumb > > is that each server should have 10% or less of

Re: [ceph-users] Ceph OSD crash starting up

2017-09-19 Thread David Turner
Adding the old OSD back in with its data shouldn't help you at all. Your cluster has finished backfilling and has the proper amount of copies of all of its data. The time you would want to add a removed OSD back to a cluster is when you have unfound objects. The scrub errors and inconsistent PGs

Re: [ceph-users] Ceph OSD crash starting up

2017-09-19 Thread Gonzalo Aguilar Delgado
Hi David, What I want is to add the OSD back with its data yes. But avoiding any troubles that can happen from the time it was out. Is it possible? I suppose that some pg has been updated after. Will ceph manage it gracefully? Ceph status is getting worse every day. ceph status cluster

Re: [ceph-users] Clarification on sequence of recovery and client ops after OSDs rejoin cluster (also, slow requests)

2017-09-19 Thread Florian Haas
On Mon, Sep 18, 2017 at 2:02 PM, Christian Theune wrote: > Hi, > > and here’s another update which others might find quite interesting. > > Florian and I spend some time discussing the issue further, face to face. I > had one switch that I brought up again (—osd-recovery-start-delay) which I > l

Re: [ceph-users] What HBA to choose? To expand or not to expand?

2017-09-19 Thread Kees Meijs
Hi Jake, On 19-09-17 15:14, Jake Young wrote: > Ideally you actually want fewer disks per server and more servers. > This has been covered extensively in this mailing list. Rule of thumb > is that each server should have 10% or less of the capacity of your > cluster. That's very true, but let's f

Re: [ceph-users] What HBA to choose? To expand or not to expand?

2017-09-19 Thread Jake Young
On Tue, Sep 19, 2017 at 7:34 AM Kees Meijs wrote: > Hi list, > > It's probably something to discuss over coffee in Ede tomorrow but I'll > ask anyway: what HBA is best suitable for Ceph nowadays? > > In an earlier thread I read some comments about some "dumb" HBAs running > in IT mode but still b

Re: [ceph-users] What HBA to choose? To expand or not to expand?

2017-09-19 Thread Wido den Hollander
> Op 19 september 2017 om 13:34 schreef Kees Meijs : > > > Hi list, > > It's probably something to discuss over coffee in Ede tomorrow but I'll > ask anyway: what HBA is best suitable for Ceph nowadays? > I still prefer LSI (Avago) in most systems. A 8-port or 16-port controller and a bunch

Re: [ceph-users] Ceph OSD crash starting up

2017-09-19 Thread David Turner
Are you asking to add the osd back with its data or add it back in as a fresh osd. What is your `ceph status`? On Tue, Sep 19, 2017, 5:23 AM Gonzalo Aguilar Delgado < gagui...@aguilardelgado.com> wrote: > Hi David, > > Thank you for the great explanation of the weights, I thought that ceph > was

[ceph-users] What HBA to choose? To expand or not to expand?

2017-09-19 Thread Kees Meijs
Hi list, It's probably something to discuss over coffee in Ede tomorrow but I'll ask anyway: what HBA is best suitable for Ceph nowadays? In an earlier thread I read some comments about some "dumb" HBAs running in IT mode but still being able to use cache on the HBA. Does it make sense? Or, is th

Re: [ceph-users] s3cmd not working with luminous radosgw

2017-09-19 Thread Sean Purdy
On Tue, 19 Sep 2017, Yoann Moulin said: > Hello, > > Does anyone have tested s3cmd or other tools to manage ACL on luminous > radosGW ? Don't know about ACL, but s3cmd for other things works for me. Version 1.6.1 My config file includes (but is not limited to): access_key = yourkey secret_ke

Re: [ceph-users] ceph-osd restartd via systemd in case of disk error

2017-09-19 Thread Wido den Hollander
> Op 19 september 2017 om 10:24 schreef Adrian Saul > : > > > > I understand what you mean and it's indeed dangerous, but see: > > https://github.com/ceph/ceph/blob/master/systemd/ceph-osd%40.service > > > > Looking at the systemd docs it's difficult though: > > https://www.freedesktop.org/soft

Re: [ceph-users] s3cmd not working with luminous radosgw

2017-09-19 Thread jwil...@stads.net
Op 19-9-2017 om 11:24 schreef Yoann Moulin: Hello, Does anyone have tested s3cmd or other tools to manage ACL on luminous radosGW ? I have opened an issue on s3cmd too https://github.com/s3tools/s3cmd/issues/919 Just an extra option. Have you tried --signature-v2  on s3cmd? Thanks for your h

Re: [ceph-users] s3cmd not working with luminous radosgw

2017-09-19 Thread Yoann Moulin
Hello, Does anyone have tested s3cmd or other tools to manage ACL on luminous radosGW ? I have opened an issue on s3cmd too https://github.com/s3tools/s3cmd/issues/919 Thanks for your help Yoann > I have a fresh luminous cluster in test and I made a copy of a bucket (4TB > 1.5M files) with r

Re: [ceph-users] Ceph OSD crash starting up

2017-09-19 Thread Gonzalo Aguilar Delgado
Hi David, Thank you for the great explanation of the weights, I thought that ceph was adjusting them based on disk. But it seems it's not. But the problem was not that I think the node was failing because a software bug because the disk was not full anymeans. /dev/sdb1 9

Re: [ceph-users] ceph-osd restartd via systemd in case of disk error

2017-09-19 Thread Manuel Lausch
Am Tue, 19 Sep 2017 08:24:48 + schrieb Adrian Saul : > > I understand what you mean and it's indeed dangerous, but see: > > https://github.com/ceph/ceph/blob/master/systemd/ceph-osd%40.service > > > > Looking at the systemd docs it's difficult though: > > https://www.freedesktop.org/software/s

Re: [ceph-users] ceph-osd restartd via systemd in case of disk error

2017-09-19 Thread Adrian Saul
> I understand what you mean and it's indeed dangerous, but see: > https://github.com/ceph/ceph/blob/master/systemd/ceph-osd%40.service > > Looking at the systemd docs it's difficult though: > https://www.freedesktop.org/software/systemd/man/systemd.service.ht > ml > > If the OSD crashes due to ano

Re: [ceph-users] ceph-osd restartd via systemd in case of disk error

2017-09-19 Thread Wido den Hollander
> Op 19 september 2017 om 10:02 schreef Manuel Lausch : > > > Hi, > > I see a issue with systemd's restart behaviour and disk IO-errors > If a disk fails with IO-errors ceph-osd stops running. Systemd detects > this and starts the daemon again. In our cluster I did see some loops > with osd cr

[ceph-users] ceph-osd restartd via systemd in case of disk error

2017-09-19 Thread Manuel Lausch
Hi, I see a issue with systemd's restart behaviour and disk IO-errors If a disk fails with IO-errors ceph-osd stops running. Systemd detects this and starts the daemon again. In our cluster I did see some loops with osd crashes caused by disk failure and restarts triggerd by systemd. Every time w