[ceph-users] rbd showmapped -p and --image options missing in rbd version 10.2.4, why?

2016-12-09 Thread Stéphane Klein
Hi, with: rbd version 0.80.7, `rbd showmapped` have this options: * -p, --pool source pool name * --imageimage name This options missing in rdb version 10.2.4 Why ? It is a regression ? Is there another command to list map by pool name and image name ? Best

Re: [ceph-users] rbd showmapped -p and --image options missing in rbd version 10.2.4, why?

2016-12-09 Thread Ilya Dryomov
On Fri, Dec 9, 2016 at 10:52 AM, Stéphane Klein wrote: > Hi, > > with: rbd version 0.80.7, `rbd showmapped` have this options: > > * -p, --pool source pool name > * --imageimage name > > This options missing in rdb version 10.2.4 > > Why ? It is a regression ?

Re: [ceph-users] problem after reinstalling system

2016-12-09 Thread Dan van der Ster
On Thu, Dec 8, 2016 at 5:51 PM, Jake Young wrote: > Hey Dan, > > I had the same issue that Jacek had after changing my OS and Ceph version > from Ubuntu 14 - Hammer to Centos 7 - Jewel. I was also able to recover from > the failure by renaming the .ldb files to .sst files. > > Do you know why thi

[ceph-users] Performance measurements CephFS vs. RBD

2016-12-09 Thread plataleas
Hi all We enabled CephFS on our Ceph Cluster consisting of: - 3 Monitor servers - 2 Metadata servers - 24 OSD (3 OSD / Server) - Spinning disks, OSD Journal is on SSD - Public and Cluster Network separated, all 1GB - Release: Jewel 10.2.3 With CephFS we reach roughly 1/3 of the write performance

[ceph-users] OSDs down after reboot

2016-12-09 Thread sandeep.cool...@gmail.com
Hi, Im using jewel (10.2.4) release on centos 7.2, after rebooting one of the OSD node, the osd doesn't start. Even after trying the 'systemctl start ceph-osd@.service'. Does we have to make entry for in fstab for our ceph osd's folder or ceph does it automatically? Then i mounted the correct pa

Re: [ceph-users] OSDs down after reboot

2016-12-09 Thread John Petrini
Try using systemctl start ceph-osd* I usually refer to this documentation for ceph + systemd https://www.suse.com/documentation/ses-1/book_storage_admin/data/ceph_operating_services.html ___ John Petrini NOC Systems Administrator // *CoreDial, LLC* // coredial.com // [image: Twitter]

Re: [ceph-users] 10.2.4 Jewel released

2016-12-09 Thread Graham Allan
On Thu, Dec 8, 2016 at 5:19 AM, Francois Lafont < francois.lafont.1...@gmail.com> wrote: > On 12/08/2016 11:24 AM, Ruben Kerkhof wrote: > > > I've been running this on one of my servers now for half an hour, and > > it fixes the issue. > > It's the same for me. ;) > > ~$ ceph -v > ceph version 10.

Re: [ceph-users] 10.2.4 Jewel released

2016-12-09 Thread Francois Lafont
Hi, Just for information, after the upgrade to the version 10.2.4-1-g5d3c76c (5d3c76c1c6e991649f0beedb80e6823606176d9e) of all my cluster (osd, mon and mds) since ~30 hours, I have no problem (my cluster is a small cluster with 5 nodes and 4 osds per nodes and 3 monitors and I just use cephfs). B

Re: [ceph-users] 10.2.4 Jewel released

2016-12-09 Thread Alex Evonosky
Francois- Sounds great. May I asked what procedure you did to upgrade? Thank you! On Fri, Dec 9, 2016 at 12:20 PM, Francois Lafont < francois.lafont.1...@gmail.com> wrote: > Hi, > > Just for information, after the upgrade to the version > 10.2.4-1-g5d3c76c (5d3c76c1c6e991649f0beedb80e682360

Re: [ceph-users] 10.2.4 Jewel released

2016-12-09 Thread Francois Lafont
On 12/09/2016 06:39 PM, Alex Evonosky wrote: > Sounds great. May I asked what procedure you did to upgrade? Of course. ;) It's here: https://shaman.ceph.com/repos/ceph/wip-msgr-jewel-fix2/ (I think this link was pointed by Greg Farnum or Sage Weil in a previous message). Personally I use Ubunt

Re: [ceph-users] 10.2.4 Jewel released

2016-12-09 Thread Alex Evonosky
Thank you sir. Ubuntu here as well. On Fri, Dec 9, 2016 at 12:54 PM, Francois Lafont < francois.lafont.1...@gmail.com> wrote: > On 12/09/2016 06:39 PM, Alex Evonosky wrote: > > > Sounds great. May I asked what procedure you did to upgrade? > > Of course. ;) > > It's here: https://shaman.cep

[ceph-users] Pgs stuck on undersized+degraded+peered

2016-12-09 Thread fridifree
Hi, 1 of 3 of my osd servers is down and I get this error And I do not have any access to rbds on the cluster Any suggestions? Thank you ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] filestore_split_multiple hardcoded maximum?

2016-12-09 Thread David Turner
Our 32k PGs each have about 25-30k objects (25-30GB per PG). When we first contracted with Redhat support, they recommended for us to have our setting at about 4000 files per directory before splitting into subfolders. When we split into subfolders with that setting, an osd_heartbeat_grace (ho

Re: [ceph-users] 10.2.4 Jewel released

2016-12-09 Thread Udo Lembke
Hi, unfortunately there are no Debian Jessie packages... Don't know that an recompile take such an long time for ceph... I think such an important fix should hit the repros faster. Udo On 09.12.2016 18:54, Francois Lafont wrote: > On 12/09/2016 06:39 PM, Alex Evonosky wrote: > >> Sounds grea

Re: [ceph-users] 2x replication: A BIG warning

2016-12-09 Thread Kees Meijs
Hi Wido, Since it's a Friday night, I decided to just go for it. ;-) It took a while to rebalance the cache tier but all went well. Thanks again for your valuable advice! Best regards, enjoy your weekend, Kees On 07-12-16 14:58, Wido den Hollander wrote: >> Anyway, any things to consider or cou

Re: [ceph-users] CephFS FAILED assert(dn->get_linkage()->is_null())

2016-12-09 Thread Goncalo Borges
Hi Sean, Rob. I saw on the tracker that you were able to resolve the mds assert by manually cleaning the corrupted metadata. Since I am also hitting that issue and I suspect that i will face an mds assert of the same type sooner or later, can you please explain a bit further what operations did

[ceph-users] Problems with multipart RGW uploads.

2016-12-09 Thread Martin Bureau
Hello, I am looking for help with a problem we have with our Jewel (10.2.4-1-g5d3c76c) cluster. Some files (which show up in the bucket listing) cannot be downloaded and return HTTP 404 and "ERROR: got unexpected error when trying to read object: -2" in the rgw log. Regards, Martin

[ceph-users] Kraken 11.x feedback

2016-12-09 Thread Ben Hines
Anyone have any good / bad experiences with Kraken? I haven't seen much discussion of it. Particularly from the RGW front. I'm still on Infernalis for our cluster, considering going up to K. thanks, -Ben ___ ceph-users mailing list ceph-users@lists.cep

Re: [ceph-users] Kraken 11.x feedback

2016-12-09 Thread Samuel Just
Is there a particular reason you are sticking to the versions with shorter support periods? -Sam On Fri, Dec 9, 2016 at 11:38 AM, Ben Hines wrote: > Anyone have any good / bad experiences with Kraken? I haven't seen much > discussion of it. Particularly from the RGW front. > > I'm still on Infern

Re: [ceph-users] Kraken 11.x feedback

2016-12-09 Thread Ben Hines
Not particularly, i just never did the Jewel upgrade. (normally like to stay relatively current) -Ben On Fri, Dec 9, 2016 at 11:40 AM, Samuel Just wrote: > Is there a particular reason you are sticking to the versions with > shorter support periods? > -Sam > > On Fri, Dec 9, 2016 at 11:38 AM, B

Re: [ceph-users] 10.2.4 Jewel released

2016-12-09 Thread Andrey Shevel
I did yum update and found out that ceph has version 10.2.4 and also after update I have the message "all OSDs are running jewel or later but the 'require_jewel_osds' osdmap flag is not set" === [ceph@ceph-swift-gateway ~]$ ceph -v ceph version 10.2.4 (9411351cc8ce9ee03fb

Re: [ceph-users] 10.2.4 Jewel released

2016-12-09 Thread David Turner
"all OSDs are running jewel or later but the 'require_jewel_osds' osdmap flag is not set" It's noted in the release notes that this will happen and that you then just set the flag and it goes away. [cid:image0cf66a.JPG@55847b03.4ab222e1]

Re: [ceph-users] Performance measurements CephFS vs. RBD

2016-12-09 Thread Gregory Farnum
On Fri, Dec 9, 2016 at 6:58 AM, plataleas wrote: > Hi all > > We enabled CephFS on our Ceph Cluster consisting of: > - 3 Monitor servers > - 2 Metadata servers > - 24 OSD (3 OSD / Server) > - Spinning disks, OSD Journal is on SSD > - Public and Cluster Network separated, all 1GB > - Release: Jewe

[ceph-users] High load on OSD processes

2016-12-09 Thread lewis.geo...@innoscale.net
Hello, I am testing out a new node setup for us and I have configured a node in a single node cluster. It has 24 OSDs. Everything looked okay during the initial build and I was able to run the 'rados bench' on it just fine. However, if I just let the cluster sit and run for a few minutes withou

[ceph-users] Server crashes on high mount volume

2016-12-09 Thread Diego Castro
Hello, my case is very specific but i think other may have this issue. I have a ceph cluster up and running hosting block storage for my openshift (kubernetes) cluster. Things goes bad when i "evacuate" a node, which is move all containers to other hosts, when this happens i can see a lot of map/m

Re: [ceph-users] 2x replication: A BIG warning

2016-12-09 Thread Oliver Humpage
> On 7 Dec 2016, at 15:01, Wido den Hollander wrote: > > I would always run with min_size = 2 and manually switch to min_size = 1 if > the situation really requires it at that moment. Thanks for this thread, it’s been really useful. I might have misunderstood, but does min_size=2 also mean th

Re: [ceph-users] filestore_split_multiple hardcoded maximum?

2016-12-09 Thread Dan van der Ster
Coincidentally, we've been suffering from split-induced slow requests on one of our clusters for the past week. I wanted to add that it isn't at all obvious when slow requests are being caused by filestore splitting. (When you increase the filestore/osd logs to 10, probably also 20, all you see is

Re: [ceph-users] High load on OSD processes

2016-12-09 Thread Reed Dier
Assuming you deployed within the last 48 hours, I’m going to bet you are using v10.2.4 which has an issue that causes high cpu utilization. Should see large ramp up in loadav after 15 minutes exactly. See mailing list thread here: https://www.mail-archive.com/ceph-users@lists.ceph.com/msg34390.

Re: [ceph-users] 2x replication: A BIG warning

2016-12-09 Thread David Turner
I'm pretty certain that the write returns as complete only after all active OSDs for a PG have completed the write regardless of min_size. [cid:image87d2ad.JPG@6e2c58b3.4d9df465] David Turner | Cloud Operations Engineer | StorageC

Re: [ceph-users] High load on OSD processes

2016-12-09 Thread Diego Castro
Same here, is there any ETA to publish CentOS packages? --- Diego Castro / The CloudFather GetupCloud.com - Eliminamos a Gravidade 2016-12-09 18:59 GMT-03:00 Reed Dier : > Assuming you deployed within the last 48 hours, I’m going to bet you are > using v10.2.4 which has an issue that causes hig

Re: [ceph-users] High load on OSD processes

2016-12-09 Thread lewis.geo...@innoscale.net
Hi Reed, Yes, this was just installed yesterday and that is the version. I just retested and it is exactly 15 minutes when the load starts to climb. So, just like Diego, do you know if there is a fix for this yet and when it might be available on the repo? Should I try to install the prior min

Re: [ceph-users] High load on OSD processes

2016-12-09 Thread Reed Dier
I don’t think there is a graceful path to downgrade. There is a hot fix upstream I believe. My understanding is the build is being tested for release. Francois Lafont posted in the other thread: > Begin forwarded message: > > From: Francois Lafont > Subject: Re: [ceph-users] 10.2.4 Jewel rele

Re: [ceph-users] CephFS FAILED assert(dn->get_linkage()->is_null())

2016-12-09 Thread Chris Sarginson
Hi Goncarlo, In the end we ascertained that the assert was coming from reading corrupt data in the mds journal. We have followed the sections at the following link (http://docs.ceph.com/docs/jewel/cephfs/disaster-recovery/) in order down to (and including) MDS Table wipes (only wiping the "sessio

Re: [ceph-users] Pgs stuck on undersized+degraded+peered

2016-12-09 Thread Christian Wuerdig
Hi, it's useful to generally provide some detail around the setup, like: What are your pool settings - size and min_size? What is your failure domain - osd or host? What version of ceph are you running on which OS? You can check which specific PGs are problematic by running "ceph health detail" a