[ceph-users] Hammer issues (rgw)

2015-06-29 Thread Gleb Borisov
Hi We've just upgraded our storage from 0.94 to 0.94.2 and realized that we have a lot of garbage and corrupted objects in our buckets. First of all, we found several corrupted objects (missing data in the middle of object) uploaded via S3 multipart upload with enabled retry policy. It seems that

Re: [ceph-users] 'pgs stuck unclean ' problem

2015-06-29 Thread jan.zeller
Hi, as I had the same issue in a little virtualized test environment (3 x 10g lvm volumes) I would like to understand the 'weight' thing. I did not find any "userfriendly explanation" for that kind of problem. The only explanation I found is on http://ceph.com/docs/master/rados/operations/crus

Re: [ceph-users] CephFS posix test performance

2015-06-29 Thread Gregory Farnum
Zheng, I don't have any idea what pieces have changed in that kernel range. Did we have to flip some switches that slowed things down and we expect to flip back, or did something more fundamental happen? Do these results make any sense? I'm a little surprised to find ceph-fuse that much faster than

Re: [ceph-users] radosgw backup

2015-06-29 Thread Konstantin Ivanov
Replication is the way to make cluster failover. I need to make failover for files (erroneous deletion, for example). Snapshot is good thing in this case. Is there implementation in ceph? 2015-06-11 20:50 GMT+03:00 Michael Kuriger : > You may be able to use replication. Here is a site showing a

Re: [ceph-users] Old vs New pool on same OSDs - Performance Difference

2015-06-29 Thread Nick Fisk
Hi Somnath, Sorry for the delay. Release is Hammer, so I can probably drop that setting then. I have hopefully managed to capture a couple of slow IO's on an idle cluster, does the below look about right to you? I can see there is a delay after this entry " get_object_context: obc NOT found in

Re: [ceph-users] 'pgs stuck unclean ' problem

2015-06-29 Thread Abhishek L
jan.zel...@id.unibe.ch writes: > Hi, > > as I had the same issue in a little virtualized test environment (3 x 10g lvm > volumes) I would like to understand the 'weight' thing. > I did not find any "userfriendly explanation" for that kind of problem. > > The only explanation I found is on > ht

[ceph-users] Removing empty placement groups / empty objects

2015-06-29 Thread Burkhard Linke
Hi, I've noticed that a number of placement groups in our setup contain objects, but no actual data (ceph pg dump | grep remapped during a hard disk replace operation): 7.616 26360 0 52720 4194304 3003 3003 active+remapped+wait_backfill 2015-06-29 13:43:28.716

[ceph-users] Where is what type if IO generated?

2015-06-29 Thread Steffen Tilsch
Hello Cephers, I got some questions regarding where what type of IO is generated. As far as I understand it looks like this (please see picture: http://imageshack.com/a/img673/4563/zctaGA.jpg ) : 1. Clients -> OSD (Journal): - Is it sequential write? - Is it parallel due to the many open soc

[ceph-users] backup RGW in federated gateway

2015-06-29 Thread Stéphane DUGRAVOT
Hi all I am interested in RGW and I would make a backup of differents pools (objects from a s3 connector) in a second Ceph cluster. From this doc: * http://ceph.com/docs/master/radosgw/federated-config/ If I activate the synchronization agent, copying data will synchronously the main ar

[ceph-users] bucket owner vs S3 ACL?

2015-06-29 Thread Valery Tschopp
Hi guys, We use the radosgw (v0.80.9) with the Openstack Keystone integration. One project have been deleted, so now I have to transfer the ownership of all the buckets to another user/project. Using radosgw-admin I have changed the owner: radosgw-admin bucket link --uid --bucket And the

Re: [ceph-users] straw to straw2 migration

2015-06-29 Thread Christoph Adomeit
Hello Dziannis, I am also planning to change our cluster from straw to straw2 because we habe different hdd sizes and changes in the HDD-Sizes always triggers a lot of reorganization load. Did you experience any issues ? Did you already change the other hosts ? Don't you think we will have less

Re: [ceph-users] Old vs New pool on same OSDs - Performance Difference

2015-06-29 Thread Somnath Roy
Nick, The line " get_object_context: obc NOT found in cache" should be fine, it will be rare (even after obc cache implementation) you will have a hit during random workload. Is the second IO on the pool you are having problem ? I saw huge wait time there during getattr completion. In the first

[ceph-users] How to use cgroup to bind ceph-osd to a specific cpu core?

2015-06-29 Thread Ray Sun
Cephers, I want to bind each of my ceph-osd to a specific cpu core, but I didn't find any document to explain that, could any one can provide me some detailed information. Thanks. Currently, my ceph is running like this: oot 28692 1 0 Jun23 ?00:37:26 /usr/bin/ceph-mon -i seed.e

[ceph-users] infiniband implementation

2015-06-29 Thread German Anders
hi cephers, Want to know if there's any 'best' practice or procedure to implement Ceph with Infiniband FDR 56gb/s for front and back end connectivity. Any crush tunning parameters, etc. The Ceph cluster has: - 8 OSD servers - 2x Intel Xeon E5 8C with HT - 128G RAM - 2x 200G Intel

Re: [ceph-users] infiniband implementation

2015-06-29 Thread Adam Boyhan
One thing that jumps out at me is using the S3700 for OS but the S3500 for journals. I would use the S3700 for journals and S3500 for the OS. Looks pretty good other than that! From: "German Anders" To: "ceph-users" Sent: Monday, June 29, 2015 12:24:41 PM Subject: [ceph-users] infiniban

Re: [ceph-users] How to use cgroup to bind ceph-osd to a specific cpu core?

2015-06-29 Thread Somnath Roy
Presently, you have to do it by using tool like ‘taskset’ or ‘numactl’… Thanks & Regards Somnath From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Ray Sun Sent: Monday, June 29, 2015 9:19 AM To: ceph-users@lists.ceph.com Subject: [ceph-users] How to use cgroup to bind ceph-

Re: [ceph-users] How to use cgroup to bind ceph-osd to a specific cpu core?

2015-06-29 Thread Jan Schermer
I promised you all our scripts for automatic cgroup assignment - they are in our production already and I just need to put them on github, stay tuned tomorrow :-) Jan > On 29 Jun 2015, at 19:41, Somnath Roy wrote: > > Presently, you have to do it by using tool like ‘taskset’ or ‘numactl’… >

Re: [ceph-users] infiniband implementation

2015-06-29 Thread German Anders
Thanks a lot Adam, it was an error typo, the 3700 are for the journals and the 3500 for the OS. Any special crush algorithm configuration for IB and for the mix of SSD and SATA OSD daemons? Thanks in advance, *German* 2015-06-29 14:05 GMT-03:00 Adam Boyhan : > One thing that jumps out at me is

Re: [ceph-users] RHEL 7.1 ceph-disk failures creating OSD with ver 0.94.2

2015-06-29 Thread Bruce McFarland
It doesn't appear to be related to using wwn's for the drive id. The verbose output shows ceph converting from wwn to sd letter. I ran with verbose on and used sd letters for the data drive and the journal and get the same failures. I'm attempting to create OSD's manually now. [root@ceph0 ceph

Re: [ceph-users] RHEL 7.1 ceph-disk failures creating OSD with ver 0.94.2

2015-06-29 Thread Bruce McFarland
Using the "manual" method of creating an OSD on RHEL 7.1 with Ceph 94.2 turns up an issue with the ondisk fsid of the journal device. From a quick web search I've found reference to this exact same issue from earlier this year. Is there a version of Ceph that works with RHEL 7.1??? [root@ceph0

Re: [ceph-users] RHEL 7.1 ceph-disk failures creating OSD with ver 0.94.2

2015-06-29 Thread Bruce McFarland
Do these issues occur in Centos 7 also? > -Original Message- > From: Bruce McFarland > Sent: Monday, June 29, 2015 12:06 PM > To: 'Loic Dachary'; 'ceph-users@lists.ceph.com' > Subject: RE: [ceph-users] RHEL 7.1 ceph-disk failures creating OSD with ver > 0.94.2 > > Using the "manual" metho

Re: [ceph-users] Old vs New pool on same OSDs - Performance Difference

2015-06-29 Thread Nick Fisk
Sorry, forgot to enable that. Here is another capture with it on and I think you are spot on as I can see a 100ms delay doing the getattr request. Any ideas how to debug further? Thanks for the help by the way, really appreciated. 2015-06-29 22:48:50.851645 7fd8a2a1e700 15 osd.1 26349 enqueue_op

Re: [ceph-users] Old vs New pool on same OSDs - Performance Difference

2015-06-29 Thread Somnath Roy
Nick, I think you are probably hitting the issue of crossing the xattr size limit that XFS can inline (255 bytes). In your case "_" xattr size is 267 bytes. Sage talked about that in one of his earlier mails..You can try to apply the following patch (not backported to hammer yet) and see if it is

Re: [ceph-users] How to use cgroup to bind ceph-osd to a specific cpu core?

2015-06-29 Thread Zhang, Jian
Ray, Just wondering, what’s the benefit for binding the ceph-osd to a specific CPU core? Thanks Jian From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Ray Sun Sent: Tuesday, June 30, 2015 12:19 AM To: ceph-users@lists.ceph.com Subject: [ceph-users] How to use cgroup to bin

Re: [ceph-users] CephFS posix test performance

2015-06-29 Thread Yan, Zheng
I tried 4.1 kernel and 0.94.2 ceph-fuse. their performance are about the same. fuse: Files=191, Tests=1964, 60 wallclock secs ( 0.43 usr 0.08 sys + 1.16 cusr 0.65 csys = 2.32 CPU) kernel: Files=191, Tests=2286, 61 wallclock secs ( 0.45 usr 0.08 sys + 1.21 cusr 0.72 csys = 2.46 CPU) > O

Re: [ceph-users] krbd splitting large IO's into smaller IO's

2015-06-29 Thread Z Zhang
Hi Ilya, Thanks for your explanation. This makes sense. Will you make max_segments to be configurable? Could you pls point me the fix you have made? We might help to test it. Thanks. David Zhang > Date: Fri, 26 Jun 2015 18:21:55 +0300 > Subject: Re: [ceph-users] krbd splitting large IO's into s

Re: [ceph-users] infiniband implementation

2015-06-29 Thread Haomai Wang
On Tue, Jun 30, 2015 at 12:24 AM, German Anders wrote: > hi cephers, > >Want to know if there's any 'best' practice or procedure to implement > Ceph with Infiniband FDR 56gb/s for front and back end connectivity. Any > crush tunning parameters, etc. > > The Ceph cluster has: > > - 8 OSD server

Re: [ceph-users] How to use cgroup to bind ceph-osd to a specific cpu core?

2015-06-29 Thread Ray Sun
Jian, As we put compute and storage together, we don't want them to bother each other during the runtime. Thanks. Best Regards -- Ray On Tue, Jun 30, 2015 at 8:50 AM, Zhang, Jian wrote: > Ray, > > Just wondering, what’s the benefit for binding the ceph-osd to a specific > CPU core? > > > > Tha

Re: [ceph-users] How to use cgroup to bind ceph-osd to a specific cpu core?

2015-06-29 Thread Ray Sun
​Sound great, any update please let me know.​ Best Regards -- Ray On Tue, Jun 30, 2015 at 1:46 AM, Jan Schermer wrote: > I promised you all our scripts for automatic cgroup assignment - they are > in our production already and I just need to put them on github, stay tuned > tomorrow :-) > > Jan