Re: [ceph-users] TCP failed connection attempts

2014-03-27 Thread Dan Van Der Ster
On 26 Mar 2014 at 21:33:06, Sergey Malinin (h...@newmail.com) wrote: This is typical (output from netstat -s): 50329019 active connections openings 15218590 passive connection openings 44167087 failed connection attempts Taking into account that presumably yo

Re: [ceph-users] ec pools and radosgw

2014-03-27 Thread Loic Dachary
Hi Michael, Could you please show the exact commands you've used to modify the k & m values ? Cheers On 27/03/2014 00:48, Michael Nelson wrote: > I am playing around with erasure coded pools on 0.78-348 (firefly) and am > attempting to enable EC on the .rgw.buckets pool for radosgw > (fresh in

Re: [ceph-users] TCP failed connection attempts

2014-03-27 Thread Sergey Malinin
On 27.03.14, 10:52, Dan Van Der Ster wrote: On 26 Mar 2014 at 21:33:06, Sergey Malinin (h...@newmail.com ) wrote: This is typical (output from netstat -s): 50329019 active connections openings 15218590 passive connection openings

Re: [ceph-users] TCP failed connection attempts

2014-03-27 Thread Dan Van Der Ster
On 27 Mar 2014 at 10:44:35, Sergey Malinin (h...@newmail.com) wrote: sysctl has nothing to do with that since those are just counters. You can debug failed connections by logging connection resets: iptables -I INPUT -p tcp -m tcp --tcp-flags RST RST -j LOG Thanks for tha

Re: [ceph-users] RGW hung, 2 OSDs using 100% CPU

2014-03-27 Thread Craig Lewis
The osd.8 log shows it doing some deep scrubbing here. Perhaps that is what caused your earlier issues with CPU usage? When I first noticed the CPU usage, I checked iotop and iostat. Both said there was no disk activity, on any OSD. At 14:17:25, I ran radosgw-admin --name=client.radosgw.c

Re: [ceph-users] if partition name changes, will ceph get corrupted?

2014-03-27 Thread Chris Kitzmiller
>> We use /dev/disk/by-path for this reason, but we confirmed that is stable >> for our HBAs. Maybe /dev/disk/by-something is consistent with your >> controller. > > The upstart/udev scripts will handle mounting and osd id detection, at > least on Ubuntu. I'll caution that while the OSD will be c

Re: [ceph-users] ec pools and radosgw

2014-03-27 Thread Michael Nelson
On Thu, 27 Mar 2014, Loic Dachary wrote: Hi Michael, Could you please show the exact commands you've used to modify the k & m values ? ceph osd crush rule create-erasure ecruleset ceph osd erasure-code-profile set myprofile ruleset-failure-domain=osd k=3 m=3 ceph osd pool create .rgw.buckets

Re: [ceph-users] ec pools and radosgw

2014-03-27 Thread Yehuda Sadeh
On Wed, Mar 26, 2014 at 4:48 PM, Michael Nelson wrote: > I am playing around with erasure coded pools on 0.78-348 (firefly) and am > attempting to enable EC on the .rgw.buckets pool for radosgw > (fresh install). > > If I use a plain EC profile (no settings changed), uploads of various sizes > wor

Re: [ceph-users] ec pools and radosgw

2014-03-27 Thread Yehuda Sadeh
On Thu, Mar 27, 2014 at 1:17 PM, Michael Nelson wrote: > > > On Thu, 27 Mar 2014, Yehuda Sadeh wrote: > >> On Wed, Mar 26, 2014 at 4:48 PM, Michael Nelson >> wrote: >>> >>> I am playing around with erasure coded pools on 0.78-348 (firefly) and am >>> attempting to enable EC on the .rgw.buckets po

[ceph-users] degraded objects after adding OSD?

2014-03-27 Thread Chad Seys
Hi all, Beginning with a cluster with only "active+clean" PGS, adding an OSD causes objects to be "degraded". Does this mean that ceph deletes replicas before copying them to the new OSD? Or does degraded also mean that there are not replicas on the target OSD, even though there are alrea

[ceph-users] Ceph meetup Amsterdam: April 24th 2014

2014-03-27 Thread Wido den Hollander
Hi all, I think it's time to organize a informal Ceph meetup in Amsterdam :-) I have some office space available in Amsterdam at a datacenter (with Ceph clusters running there!) and I think it would be fun to organize a Ceph meetup. No formal schedule or something, just some Ceph users (or p

[ceph-users] OSDs vanishing from Ceph cluster?

2014-03-27 Thread Dan Koren
Just ran into this problem: a week ago I set up a Ceph cluster on 4 systems, with one admin node and 3 mon+osd nodes, then ran a few casual IO tests. I returned to work after a few days out of town at a conference, and now my Ceph cluster appears to have no OSDs! root@rts24:/var/log/ceph# ceph sta

Re: [ceph-users] MDS crash when client goes to sleep

2014-03-27 Thread hjcho616
Looks like client is waking up ok now.  Thanks. Will those fixes be included in next release? Firefly? Regards, Hong From: hjcho616 To: Gregory Farnum Cc: "ceph-users@lists.ceph.com" Sent: Tuesday, March 25, 2014 11:56 AM Subject: Re: [ceph-users] MDS cras

Re: [ceph-users] rbd + qemu osd performance

2014-03-27 Thread Cédric Lemarchand
> Le 26 mars 2014 à 00:30, Andrei Mikhailovsky a écrit : > > The osd fragmentation level of zfs is at 8% at the moment, not sure if this > should impact the performance by this much. I will defrag it over night and > check tomorrow to see if it makes the difference. Sorry if this is a little

Re: [ceph-users] ec pools and radosgw

2014-03-27 Thread Michael Nelson
On Thu, 27 Mar 2014, Yehuda Sadeh wrote: On Wed, Mar 26, 2014 at 4:48 PM, Michael Nelson wrote: I am playing around with erasure coded pools on 0.78-348 (firefly) and am attempting to enable EC on the .rgw.buckets pool for radosgw (fresh install). If I use a plain EC profile (no settings ch

Re: [ceph-users] help, add mon failed lead to cluster failure

2014-03-27 Thread Joao Eduardo Luis
On 26/03/14 10:40, duan.xuf...@zte.com.cn wrote: Hi, I just add a new mon to a health cluster by following website manual "http://ceph.com/docs/master/rados/operations/add-or-rm-mons/"; "ADDING MONITORS" step by step, but when i execute step 6: ceph mon add [:] the command didn't ret

[ceph-users] radosgw-admin usage show Does not seem to work properly with start and end dates

2014-03-27 Thread Michael Baysek
I've noticed what seems to be a strange artifact in the radosgw-admin tool when I query for usage data in one hour intervals. For this exercise, I have had a script uploading and downloading files to the object store constantly, waiting 5 minutes in between runs. The user in this case is 'mik

Re: [ceph-users] RBD as backend for iSCSI SAN Targets

2014-03-27 Thread Karol Kozubal
Hi Jianing, Sorry for the late reply, I missed your contribution to the thread. Thank you for your response. I am still waiting for some of my hardware and will begin testing the new setup with firefly once it is available as a long term support release. I am looking forward to testing the new se

Re: [ceph-users] RGW hung, 2 OSDs using 100% CPU

2014-03-27 Thread Craig Lewis
In the interest of removing variables, I removed all snapshots on all pools, then restarted all ceph daemons at the same time. This brought up osd.8 as well. The cluster started recovering. Now osd.4 and osd.13 are doing this. Any suggestions for how I can see what the hung OSDs are doing?