Re: [ceph-users] [Solved] Oeps: lost cluster with: ceph osd require-osd-release luminous

2017-09-12 Thread Jan-Willem Michels
On 9/12/17 9:13 PM, Josh Durgin wrote: Could you post your crushmap? PGs mapping to no OSDs is a symptom of something wrong there. You can stop the osds from changing position at startup with 'osd crush update on start = false': Yes I had found that. Thanks. Seems be be by design, which

Re: [ceph-users] upgrade Hammer>Jewel>Luminous OSD fail to start

2017-09-12 Thread kevin parrikar
Thank you all for your suggestions: This is what i followed for the upgrade: Hammer to Jewel: apt-get dist-upgrade on each node seperately. stopped monitor process; stopped osd; changed permission to ceph:ceph recursively for /var/lib/ceph/ restarted monitor process; restarted osd; *ceph osd se

Re: [ceph-users] Power outages!!! help!

2017-09-12 Thread hjcho616
Ronny, Did bunch of ceph pg repair pg# and got the scrub errors down to 10... well was 9, trying to fix one became 10.. waiting for it to fix (I did that noout trick as I only have two copies).  8 of those scrub errors looks like it would need data from osd.0. HEALTH_ERR 22 pgs are stuck inactiv

[ceph-users] moving mons across networks

2017-09-12 Thread Blair Bethwaite
Hi all, We're looking at readdressing the mons (moving to a different subnet) on one of our clusters. Most of the existing clients are OpenStack guests on Libvirt+KVM and we have a major upgrade to do for those in coming weeks that will mean they have to go down briefly, that will give us an oppor

[ceph-users] debian-hammer wheezy Packages file incomplete?

2017-09-12 Thread David
Hi! Noticed tonight during maintenance that the hammer repo for debian wheezy only has 2 packages listed in the Packages file. Thought perhaps it's being moved to archive or something. However the files are still there: https://download.ceph.com/debian-hammer/pool/main/c/ceph/

Re: [ceph-users] Rgw install manual install luminous

2017-09-12 Thread Jean-Charles Lopez
Hi, see comment in line Regards JC > On Sep 12, 2017, at 13:31, Marc Roos wrote: > > > > I have been trying to setup the rados gateway (without deploy), but I am > missing some commands to enable the service I guess? How do I populate > the /var/lib/ceph/radosgw/ceph-gw1. I didn’t see any

[ceph-users] Rgw install manual install luminous

2017-09-12 Thread Marc Roos
I have been trying to setup the rados gateway (without deploy), but I am missing some commands to enable the service I guess? How do I populate the /var/lib/ceph/radosgw/ceph-gw1. I didn’t see any command like the ceph-mon. service ceph-radosgw@gw1 start Gives: 2017-09-12 22:26:06.390523 7fb9

Re: [ceph-users] upgrade Hammer>Jewel>Luminous OSD fail to start

2017-09-12 Thread Lincoln Bryant
Did you set the sortbitwise flag, fix OSD ownership (or use the "setuser match path" option) and such after upgrading from Hammer to Jewel? I am not sure if that matters here, but it might help if you elaborate on your upgrade process a bit. --Lincoln > On Sep 12, 2017, at 2:22 PM, kevin parri

Re: [ceph-users] upgrade Hammer>Jewel>Luminous OSD fail to start

2017-09-12 Thread Steve Taylor
It seems like I've seen similar behavior in the past with the changing of the osd user context between hammer and jewel. Hammer ran osds as root, and they switched to running as the ceph user in jewel. That doesn't really seem to match your scenario perfectly, but I think the errors you're seein

Re: [ceph-users] upgrade Hammer>Jewel>Luminous OSD fail to start

2017-09-12 Thread kevin parrikar
Can some one please help me on this.I have no idea how to bring up the cluster to operational state. Thanks, Kev On Tue, Sep 12, 2017 at 11:12 AM, kevin parrikar wrote: > hello All, > I am trying to upgrade a small test setup having one monitor and one osd > node which is in hammer release . >

Re: [ceph-users] Oeps: lost cluster with: ceph osd require-osd-release luminous

2017-09-12 Thread Josh Durgin
Could you post your crushmap? PGs mapping to no OSDs is a symptom of something wrong there. You can stop the osds from changing position at startup with 'osd crush update on start = false': http://docs.ceph.com/docs/master/rados/operations/crush-map/#crush-location Josh Sent from Nine

Re: [ceph-users] Ceph Developers Monthly - September

2017-09-12 Thread Joao Eduardo Luis
On 09/12/2017 04:59 PM, Leonardo Vaz wrote: Hey Cephers, In case you missed September's Ceph Developer Monthly, it is now up on our YouTube channel: https://youtu.be/xds1nsDoYqY Thanks Leonardo! Much appreciated ;) -Joao ___ ceph-users mailin

Re: [ceph-users] Luminous BlueStore EC performance

2017-09-12 Thread Mohamad Gebai
Sorry for the delay. We used the default k=2 and m=1. Mohamad On 09/07/2017 06:22 PM, Christian Wuerdig wrote: > What type of EC config (k+m) was used if I may ask? > > On Fri, Sep 8, 2017 at 1:34 AM, Mohamad Gebai wrote: >> Hi, >> >> These numbers are probably not as detailed as you'd like, bu

Re: [ceph-users] Ceph Developers Monthly - September

2017-09-12 Thread Leonardo Vaz
On Wed, Sep 06, 2017 at 02:06:24AM -0300, Leonardo Vaz wrote: > On Wed, Aug 30, 2017 at 01:04:51AM -0300, Leonardo Vaz wrote: > > Hey Cephers, > > > > This is just a friendly reminder that the next Ceph Developer Montly > > meeting is coming up: > > > > http://wiki.ceph.com/Planning > > > > If

[ceph-users] radosgw multi tenancy support with openstack newton

2017-09-12 Thread Kim-Norman Sahm
Hi, does swift api (radosgw on ceph jewel) multi tenancy working with openstack newton and keystone v3? http://docs.ceph.com/docs/jewel/radosgw/keystone/ I've tried to integrate ceph-radosgw as swift service in openstack newton with keystone v3 and the authentication is working but all buckets

Re: [ceph-users] Power outages!!! help!

2017-09-12 Thread hjcho616
Thank you for those references!  I'll have to go study some more.  Good portion of that inconsistent seems to be from missing data from osd.0. =P  There appears to be some from okay drives. =P  Kicked off "ceph pg repair pg#" few times, but doesn't seem to change much yet. =P  As far as smart ou

Re: [ceph-users] RBD I/O errors with QEMU [luminous upgrade/osd change]

2017-09-12 Thread Jason Dillaman
When the exclusive-lock feature is used, any and all Ceph users used for RBD purposes should be double-checked to ensure that they have permission to blacklist clients. This would effect both librbd and krbd, but only after a non-clean shutdown where the image is left in a locked state by a dead cl

Re: [ceph-users] RBD I/O errors with QEMU [luminous upgrade/osd change]

2017-09-12 Thread Blair Bethwaite
You're the OP, so for that, thanks! Our upgrade plan (for Thursday this week) was modified today to include prep work to double-check the caps. On 12 September 2017 at 21:26, Nico Schottelius wrote: > > Well, we basically needed to fix it, that's why did it :-) > > > Blair Bethwaite writes: > >>

Re: [ceph-users] RBD I/O errors with QEMU [luminous upgrade/osd change]

2017-09-12 Thread Nico Schottelius
Well, we basically needed to fix it, that's why did it :-) Blair Bethwaite writes: > Great to see this issue sorted. > > I have to say I am quite surprised anyone would implement the > export/import workaround mentioned here without *first* racing to this > ML or IRC and crying out for help. T

Re: [ceph-users] Bluestore "separate" WAL and DB (and WAL/DB size?)

2017-09-12 Thread Richard Hesketh
Thanks for the links. That does seem to largely confirm that what I haven't horribly misunderstood anything and I've not been doing anything obviously wrong while converting my disks; there's no point specifying separate WAL/DB partitions if they're going to go on the same device, throw as much

Re: [ceph-users] Power outages!!! help!

2017-09-12 Thread Ronny Aasen
you can start by posting more details. atleast "ceph osd tree" "cat ceph.conf" and "ceph osd df" so we can see what settings you are running, and how your cluster is balanced at the moment. generally: inconsistent pg's are pg's that have scrub errors. use rados list-inconsistent-pg [pool] and

Re: [ceph-users] Ceph cluster with SSDs

2017-09-12 Thread Christian Balzer
Please don't remove the ML. I'm not a support channel and if I reply to mails it is so that others hopefully will learn from that. ML re-added. On Mon, 11 Sep 2017 16:30:18 +0530 M Ranga Swami Reddy wrote: > >>> >> Here I have NVMes from Intel. but as the support of these NVMes not > >>> >> th

Re: [ceph-users] [SOLVED] output discards (queue drops) on switchport

2017-09-12 Thread Andreas Herrmann
Hi Bairo, On 12.09.2017 00:41, Blair Bethwaite wrote: > On 12 September 2017 at 01:15, Blair Bethwaite > wrote: >> Flow-control may well just mask the real problem. Did your throughput >> improve? Also, does that mean flow-control is on for all ports on the >> switch...? IIUC, then such "global