[ceph-users] Hammer patching on Wheezy?

2017-06-28 Thread Steffen Winther Sørensen
Ceph users, Got a Hammer cluster installed on old debian wheezy (7.11) boxes (I know :) root@node4:~# dpkg -l | grep -i ceph ii ceph 0.94.9-1~bpo70+1 amd64 distributed storage and file system ii ceph-common 0.94.9-1~bpo70

[ceph-users] pgs stuck unclean after removing OSDs

2017-06-28 Thread Jan Kasprzak
Hello, TL;DR: what to do when my cluster reports stuck unclean pgs? Detailed description: One of the nodes in my cluster died. CEPH correctly rebalanced itself, and reached the HEALTH_OK state. I have looked at the failed server, and decided to take it out of the cluster permanently, bec

Re: [ceph-users] Ceph and IPv4 -> IPv6

2017-06-28 Thread george.vasilakakos
I don't think you can do that, it would require running a mixed cluster which, going by the docs, doesn't seem to be supported. From: Jake Young [jak3...@gmail.com] Sent: 27 June 2017 22:42 To: Wido den Hollander; ceph-users@lists.ceph.com; Vasilakakos, George (ST

Re: [ceph-users] Ceph and IPv4 -> IPv6

2017-06-28 Thread george.vasilakakos
Hey Wido, Thanks for your suggestion. It sounds like the process might be feasible but I'd be looking for an "official" thing to do to a production cluster. Something that's documented ceph.com/docs, tested and "endorsed" if you will by the Ceph team. We could try this on a pre-prod environmen

[ceph-users] mon/osd cannot start with RDMA

2017-06-28 Thread 한승진
Hello Cephers! I am testing CEPH over RDMA now. I cloned the latest source code of ceph. I added below configs in ceph.conf ms_type = async+rdma ms_cluster_type = async+rdma ms_async_rdma_device_name = mlx4_0 However, I got same error message when I start ceph-mon and ceph-osd service. The me

Re: [ceph-users] mon/osd cannot start with RDMA

2017-06-28 Thread Haomai Wang
On Wed, Jun 28, 2017 at 6:02 PM, 한승진 wrote: > Hello Cephers! > > I am testing CEPH over RDMA now. > > I cloned the latest source code of ceph. > > I added below configs in ceph.conf > > ms_type = async+rdma > ms_cluster_type = async+rdma > ms_async_rdma_device_name = mlx4_0 > > However, I got same

Re: [ceph-users] mon/osd cannot start with RDMA

2017-06-28 Thread Haomai Wang
you can decrease "ms_async_rdma_send_buffers" and "ms_async_rdma_receive_buffers" to see whether help if the reason is system limitation On Wed, Jun 28, 2017 at 6:09 PM, Haomai Wang wrote: > On Wed, Jun 28, 2017 at 6:02 PM, 한승진 wrote: >> Hello Cephers! >> >> I am testing CEPH over RDMA now. >> >

[ceph-users] cephfs df with EC pool

2017-06-28 Thread Jake Grimmett
Dear All, Sorry is this has been covered before, but is it possible to configure cephfs to report free space based on what is available in the main storage tier? My "df" shows 76%, this gives a false sense of security, when the EC tier is 93% full... i.e. # df -h /ceph Filesystem Size Used

Re: [ceph-users] cephfs df with EC pool

2017-06-28 Thread John Spray
On Wed, Jun 28, 2017 at 12:19 PM, Jake Grimmett wrote: > Dear All, > > Sorry is this has been covered before, but is it possible to configure > cephfs to report free space based on what is available in the main > storage tier? There's a ticket for this here: http://tracker.ceph.com/issues/19109

Re: [ceph-users] cephfs df with EC pool

2017-06-28 Thread Burkhard Linke
Hi, On 06/28/2017 01:19 PM, Jake Grimmett wrote: Dear All, Sorry is this has been covered before, but is it possible to configure cephfs to report free space based on what is available in the main storage tier? My "df" shows 76%, this gives a false sense of security, when the EC tier is 93% f

Re: [ceph-users] Ceph and IPv4 -> IPv6

2017-06-28 Thread Wido den Hollander
> Op 28 juni 2017 om 11:24 schreef george.vasilaka...@stfc.ac.uk: > > > Hey Wido, > > Thanks for your suggestion. It sounds like the process might be feasible but > I'd be looking for an "official" thing to do to a production cluster. > Something that's documented ceph.com/docs, tested and "e

Re: [ceph-users] Ceph and IPv4 -> IPv6

2017-06-28 Thread ceph
On 28/06/2017 13:42, Wido den Hollander wrote: > Honestly I think there aren't that many IPv6 deployments with Ceph out there. > I for sure am a big fan a deployer of Ceph+IPv6, but I don't know many around > me. I got that ! Because IPv6 is so much better than IPv4 :dance: > Wido > >> Best re

Re: [ceph-users] LevelDB corruption

2017-06-28 Thread Mazzystr
The corruption is back in mons logs... 2017-06-28 08:16:53.078495 7f1a0b9da700 1 leveldb: Compaction error: Corruption: bad entry in block 2017-06-28 08:16:53.078499 7f1a0b9da700 1 leveldb: Waiting after background compaction error: Corruption: bad entry in block On Tue, Jun 27, 2017 at 10:42

Re: [ceph-users] Ceph and IPv4 -> IPv6

2017-06-28 Thread george.vasilakakos
> I don't think either. I don't think there is another way then just 'hacky' > changing the MONMaps. There have been talks of being able to make Ceph > dual-stack, but I don't think there is any code in the source right now. Yeah, that's what I'd like to know. What do the Ceph team think of prov

Re: [ceph-users] pgs stuck unclean after removing OSDs

2017-06-28 Thread David Turner
A couple things. You didn't `ceph osd crush remove osd.21` after doing the other bits. Also you will want to remove the bucket (re: host) from the crush map as it will now be empty. Right now you have a host in the crush map with a weight, but no osds to put that data on. It has a weight becaus

Re: [ceph-users] pgs stuck unclean after removing OSDs

2017-06-28 Thread Mazzystr
I've been using this procedure to remove OSDs... OSD_ID= ceph auth del osd.${OSD_ID} ceph osd down ${OSD_ID} ceph osd out ${OSD_ID} ceph osd rm ${OSD_ID} ceph osd crush remove osd.${OSD_ID} systemctl disable ceph-osd@${OSD_ID}.service systemctl stop ceph-osd@${OSD_ID}.service sed -i "/ceph-$OSD_ID

[ceph-users] Radosgw versioning S3 compatible?

2017-06-28 Thread Martin Emrich
Hi! Is the Object Gateway S3 API supposed to be compatible with Amazon S3 regarding versioning? Object Versioning is listed as supported in Ceph 12.1, but using the standard Node.js aws-sdk module (s3.putBucketVersioning()) results in "NotImplemented". Thanks Martin _

Re: [ceph-users] pgs stuck unclean after removing OSDs

2017-06-28 Thread David Turner
I would stop the service, down, out, rm, auth del, crush remove, disable service, fstab, umount. So you did remove it from your crush map, then? Could you post your `ceph osd tree`? On Wed, Jun 28, 2017, 10:12 AM Mazzystr wrote: > I've been using this procedure to remove OSDs... > > OSD_ID= >

Re: [ceph-users] cephfs df with EC pool

2017-06-28 Thread Jake Grimmett
Hi John, Many thanks for your reply. Glad to hear there is a ticket for this. But also glad that it's not a "show stopper", just an inconvenience :) best, Jake On 28/06/17 12:29, John Spray wrote: > On Wed, Jun 28, 2017 at 12:19 PM, Jake Grimmett > wrote: >> Dear All, >> >> Sorry is this h

Re: [ceph-users] RGW lifecycle not expiring objects

2017-06-28 Thread Daniel Gryniewicz
This is almost certainly because it's using v4 auth, which is not well supported in RGW yet. Can you try with v3 auth? Daniel On 06/27/2017 07:57 PM, Graham Allan wrote: I upgraded my test cluster to Luminous 12.1.0, and a separate problem made me realize a possible cause for the lifecycle fa

[ceph-users] Very HIGH Disk I/O latency on instances

2017-06-28 Thread Keynes_Lee
We were using HP Helion 2.1.5 ( OpenStack + Ceph ) The OpenStack version is Kilo and Ceph version is firefly The way we backup VMs is create a snapshot by Ceph commands (rbd snapshot) then download (rbd export) it. We found a very high Disk Read / Write latency during creating / deleting snapsh

Re: [ceph-users] num_caps

2017-06-28 Thread Henrik Korkuc
On 17-05-15 14:49, John Spray wrote: On Mon, May 15, 2017 at 1:36 PM, Henrik Korkuc wrote: On 17-05-15 13:40, John Spray wrote: On Mon, May 15, 2017 at 10:40 AM, Ranjan Ghosh wrote: Hi all, When I run "ceph daemon mds. session ls" I always get a fairly large number for num_caps (200.000). I

Re: [ceph-users] Radosgw versioning S3 compatible?

2017-06-28 Thread Martin Emrich
Correction: It's about the Version expiration, not the versioning itself. We send this rule: Rules: [ { Status: 'Enabled', Prefix: '', NoncurrentVersionExpiration: { NoncurrentDays: 60 }, Expiration: { ExpiredO

Re: [ceph-users] Very HIGH Disk I/O latency on instances

2017-06-28 Thread Peter Maloney
On 06/28/17 16:52, keynes_...@wistron.com wrote: > > We were using HP Helion 2.1.5 ( OpenStack + Ceph ) > > The OpenStack version is *Kilo* and Ceph version is *firefly* > > > > The way we backup VMs is create a snapshot by Ceph commands (rbd > snapshot) then download (rbd export) it. > > > > W

Re: [ceph-users] pgs stuck unclean after removing OSDs

2017-06-28 Thread Jan Kasprzak
David Turner wrote: : A couple things. You didn't `ceph osd crush remove osd.21` after doing the : other bits. Also you will want to remove the bucket (re: host) from the : crush map as it will now be empty. Right now you have a host in the crush : map with a weight, but no osds to put that data

Re: [ceph-users] rbd-fuse performance

2017-06-28 Thread Mykola Golub
On Tue, Jun 27, 2017 at 07:17:22PM -0400, Daniel K wrote: > rbd-nbd isn't good as it stops at 16 block devices (/dev/nbd0-15) modprobe nbd nbds_max=1024 Or, if nbd module is loaded by rbd-nbd, use --nbds_max command line option. -- Mykola Golub ___ c

Re: [ceph-users] Radosgw versioning S3 compatible?

2017-06-28 Thread Yehuda Sadeh-Weinraub
On Wed, Jun 28, 2017 at 8:13 AM, Martin Emrich wrote: > Correction: It’s about the Version expiration, not the versioning itself. > > We send this rule: > > > > Rules: [ > > { > > Status: 'Enabled', > > Prefix: '', > > NoncurrentVersionExpiration: { > >

Re: [ceph-users] Radosgw versioning S3 compatible?

2017-06-28 Thread Daniel Gryniewicz
On 06/28/2017 12:52 PM, Yehuda Sadeh-Weinraub wrote: On Wed, Jun 28, 2017 at 8:13 AM, Martin Emrich wrote: Correction: It’s about the Version expiration, not the versioning itself. We send this rule: Rules: [ { Status: 'Enabled', Prefix: '', N

[ceph-users] Mapping data and metadata between rados and cephfs

2017-06-28 Thread Lefman, Jonathan
Hi all, I would like to create a 1-to-1 mapping between rados and cephfs. Here's the usage scenario: 1. Upload file via rest api through rados compatible APIs 2. Run "local" operations on the file delivered via rados on the linked cephfs mount 3. Retrieve/download file via rados API on newly cr

Re: [ceph-users] RGW lifecycle not expiring objects

2017-06-28 Thread Graham Allan
That seems to be it! I couldn't see a way to specify the auth version with aws cli (is there a way?). However it did work with s3cmd and v2 auth: % s3cmd --signature-v2 setlifecycle lifecycle.xml s3://testgta s3://testgta/: Lifecycle Policy updated (I believe that with Kraken, this threw an err

Re: [ceph-users] Mapping data and metadata between rados and cephfs

2017-06-28 Thread David Turner
You want to access the same data via a rados API and via cephfs? Are you thinking RadosGW? On Wed, Jun 28, 2017 at 1:54 PM Lefman, Jonathan wrote: > Hi all, > > > > I would like to create a 1-to-1 mapping between rados and cephfs. Here's > the usage scenario: > > > > 1. Upload file via rest api

Re: [ceph-users] Mapping data and metadata between rados and cephfs

2017-06-28 Thread Lefman, Jonathan
Yes, sorry. I meant the RadosGW. I still do not know what the mechanism is to enable the mapping between data inserted by the rados component and the cephfs component. I hope that makes sense. -Jon From: David Turner [mailto:drakonst...@gmail.com] Sent: Wednesday, June 28, 2017 2:46 PM To: Lefm

Re: [ceph-users] Mapping data and metadata between rados and cephfs

2017-06-28 Thread David Turner
CephFS and RGW store data differently. I have never heard of, nor do I believe that it's possible, to have CephFS and RGW sharing the same data pool. On Wed, Jun 28, 2017 at 2:48 PM Lefman, Jonathan wrote: > Yes, sorry. I meant the RadosGW. I still do not know what the mechanism is > to enable

Re: [ceph-users] Obtaining perf counters/stats from krbd client

2017-06-28 Thread Gregory Farnum
On Thu, Jun 22, 2017 at 11:27 AM Prashant Murthy wrote: > Hi Ceph users, > > We are currently using the Ceph kernel client module (krbd) in our > deployment and we were looking to determine if there are ways by which we > can obtain perf counters, log dumps, etc from such a deployment. Has > anyb

Re: [ceph-users] Mapping data and metadata between rados and cephfs

2017-06-28 Thread Lefman, Jonathan
Thanks for the prompt reply. I was hoping that there would be an s3fs (https://github.com/s3fs-fuse/s3fs-fuse) equivalent for Ceph since there are numerous functional similarities. Ideally one would be able to upload data to a bucket and have the file synced to the local filesystem mount of that

Re: [ceph-users] Mapping data and metadata between rados and cephfs

2017-06-28 Thread David Turner
CephFS is very different from RGW. You may be able to utilize s3fs-fuse to interface with RGW, but I haven't heard of anyone using that on the ML before. On Wed, Jun 28, 2017 at 2:57 PM Lefman, Jonathan wrote: > Thanks for the prompt reply. I was hoping that there would be an s3fs ( > https://g

Re: [ceph-users] Mapping data and metadata between rados and cephfs

2017-06-28 Thread Gregory Farnum
Yeah, CephFS and RGW can't cross-communicate at all. Out of the original choices, "this will never happen". Somebody who was very, very dedicated could set something up. But it would basically be the same as running s3fs in the Ceph servers instead of on the clients (or probably the other way arou

Re: [ceph-users] Obtaining perf counters/stats from krbd client

2017-06-28 Thread Prashant Murthy
Thanks Greg. I thought as much :( We are looking into what stats are available and also for ways by which we can obtain stats from krbd, given this is going to be our current deployment in the near term. If this is useful to the community, we can share our findings and a proposal, once we have thi

Re: [ceph-users] Very HIGH Disk I/O latency on instances

2017-06-28 Thread Gregory Farnum
On Wed, Jun 28, 2017 at 9:17 AM Peter Maloney < peter.malo...@brockmann-consult.de> wrote: > On 06/28/17 16:52, keynes_...@wistron.com wrote: > > We were using HP Helion 2.1.5 ( OpenStack + Ceph ) > > The OpenStack version is *Kilo* and Ceph version is *firefly* > > > > The way we backup VMs is cr

Re: [ceph-users] Obtaining perf counters/stats from krbd client

2017-06-28 Thread Gregory Farnum
Please do! :) On Wed, Jun 28, 2017 at 1:51 PM Prashant Murthy wrote: > Thanks Greg. I thought as much :( > > We are looking into what stats are available and also for ways by which we > can obtain stats from krbd, given this is going to be our current > deployment in the near term. If this is us

Re: [ceph-users] Mapping data and metadata between rados and cephfs

2017-06-28 Thread Matt Benjamin
Hi, A supported way to access S3 objects from a filesystem mount is with RGW NFS. That is, RGW now exports the S3 namespace directly as files and directories, one consumer is an nfs-ganesha NFS driver. Regards, Matt - Original Message - > From: "David Turner" > To: "Jonathan Lefman"

Re: [ceph-users] Ceph and IPv4 -> IPv6

2017-06-28 Thread Gregory Farnum
On Wed, Jun 28, 2017 at 6:33 AM wrote: > > I don't think either. I don't think there is another way then just > 'hacky' changing the MONMaps. There have been talks of being able to make > Ceph dual-stack, but I don't think there is any code in the source right > now. > > Yeah, that's what I'd lik

Re: [ceph-users] Mapping data and metadata between rados and cephfs

2017-06-28 Thread Gregory Farnum
On Wed, Jun 28, 2017 at 2:10 PM Matt Benjamin wrote: > Hi, > > A supported way to access S3 objects from a filesystem mount is with RGW > NFS. That is, RGW now exports the S3 namespace directly as files and > directories, one consumer is an nfs-ganesha NFS driver. > This supports a very specifi

Re: [ceph-users] Mapping data and metadata between rados and cephfs

2017-06-28 Thread Matt Benjamin
Hi, That's true, sure. We hope to support async mounts and more normal workflows in future, but those are important caveats. Editing objects in place doesn't work with RGW NFS. Matt - Original Message - > From: "Gregory Farnum" > To: "Matt Benjamin" , "David Turner" > > Cc: ceph-u

Re: [ceph-users] Mapping data and metadata between rados and cephfs

2017-06-28 Thread Lefman, Jonathan
Thanks to everyone for the helpful feedback. I appreciate the responsiveness. -Jon -Original Message- From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Matt Benjamin Sent: Wednesday, June 28, 2017 4:20 PM To: Gregory Farnum Cc: ceph-users@lists.ceph.com Subject:

Re: [ceph-users] Performance issue with small files, and weird "workaround"

2017-06-28 Thread Ruben Rodriguez
On 06/27/2017 07:08 PM, Jason Dillaman wrote: > Have you tried blktrace to determine if there are differences in the > IO patterns to the rbd-backed virtio-scsi block device (direct vs > indirect through loop)? I tried today with the kernel tracing features, and I'll give blktrace a go if necessa

[ceph-users] Ceph mount rbd

2017-06-28 Thread lista
Hi People, I am testing the new enviroment, with ceph + rbd with ubuntu 16.04, and i have one question. I have my cluster ceph and mount the using the comands to ceph in my linux enviroment : rbd create veeamrepo --size 20480 rbd --image veeamrepo info modprobe rbd rbd map veeamrepo rbd featu

[ceph-users] qemu-img convert vs rbd import performance

2017-06-28 Thread Murali Balcha
Need some help resolving the performance issues on the my ceph cluster. We are running acute performance issues when we are using qemu-img convert. However rbd import operation works perfectly alright. Please ignore image format for a minute. I am trying to understand why rbd import performs wel

Re: [ceph-users] Ceph mount rbd

2017-06-28 Thread Maged Mokhtar
On 2017-06-28 22:55, li...@marcelofrota.info wrote: > Hi People, > > I am testing the new enviroment, with ceph + rbd with ubuntu 16.04, and i > have one question. > > I have my cluster ceph and mount the using the comands to ceph in my linux > enviroment : > > rbd create veeamrepo --size 204

Re: [ceph-users] rbd-fuse performance

2017-06-28 Thread Daniel K
thank you! On Wed, Jun 28, 2017 at 11:48 AM, Mykola Golub wrote: > On Tue, Jun 27, 2017 at 07:17:22PM -0400, Daniel K wrote: > > > rbd-nbd isn't good as it stops at 16 block devices (/dev/nbd0-15) > > modprobe nbd nbds_max=1024 > > Or, if nbd module is loaded by rbd-nbd, use --nbds_max command l

Re: [ceph-users] qemu-img convert vs rbd import performance

2017-06-28 Thread Murali Balcha
We will give it a try. I have another cluster of similar configuration and the converts are working fine. We have not changed any queue depth setting on that setup either. If it turns out to be queue depth how can we set queue setting for qemu-img convert operation? Thank you. Sent from my i

Re: [ceph-users] LevelDB corruption

2017-06-28 Thread Brad Hubbard
On Wed, Jun 28, 2017 at 10:18 PM, Mazzystr wrote: > The corruption is back in mons logs... > > 2017-06-28 08:16:53.078495 7f1a0b9da700 1 leveldb: Compaction error: > Corruption: bad entry in block > 2017-06-28 08:16:53.078499 7f1a0b9da700 1 leveldb: Waiting after background > compaction error: C

Re: [ceph-users] qemu-img convert vs rbd import performance

2017-06-28 Thread Jason Dillaman
Given that your time difference is roughly 10x, best guess is that qemu-img is sending the IO operations synchronously (queue depth = 1), whereas, by default, "rbd import" will send up to 10 write requests in parallel to the backing OSDs. Such an assumption assumes that you have really high latency

Re: [ceph-users] qemu-img convert vs rbd import performance

2017-06-28 Thread Jason Dillaman
Perhaps just one cluster has low latency and the other has excessively high latency? You can use "rbd bench-write" to verify. On Wed, Jun 28, 2017 at 8:04 PM, Murali Balcha wrote: > We will give it a try. I have another cluster of similar configuration and > the converts are working fine. We hav

Re: [ceph-users] Ceph mount rbd

2017-06-28 Thread Jason Dillaman
... additionally, the forthcoming 4.12 kernel release will support non-cooperative exclusive locking. By default, since 4.9, when the exclusive-lock feature is enabled, only a single client can write to the block device at a time -- but they will cooperatively pass the lock back and forth upon writ

Re: [ceph-users] LevelDB corruption

2017-06-28 Thread Mazzystr
just one MON On Wed, Jun 28, 2017 at 8:05 PM, Brad Hubbard wrote: > On Wed, Jun 28, 2017 at 10:18 PM, Mazzystr wrote: > > The corruption is back in mons logs... > > > > 2017-06-28 08:16:53.078495 7f1a0b9da700 1 leveldb: Compaction error: > > Corruption: bad entry in block > > 2017-06-28 08:16:

Re: [ceph-users] LevelDB corruption

2017-06-28 Thread Brad Hubbard
On Thu, Jun 29, 2017 at 11:58 AM, Mazzystr wrote: > just one MON Try just replacing that MON then? > > On Wed, Jun 28, 2017 at 8:05 PM, Brad Hubbard wrote: >> >> On Wed, Jun 28, 2017 at 10:18 PM, Mazzystr wrote: >> > The corruption is back in mons logs... >> > >> > 2017-06-28 08:16:53.078495 7

Re: [ceph-users] LevelDB corruption

2017-06-28 Thread Mazzystr
I should be able to try that tomorrow. I'll report back in afterward On Wed, Jun 28, 2017 at 10:09 PM, Brad Hubbard wrote: > On Thu, Jun 29, 2017 at 11:58 AM, Mazzystr wrote: > > just one MON > > Try just replacing that MON then? > > > > > On Wed, Jun 28, 2017 at 8:05 PM, Brad Hubbard > wrote

Re: [ceph-users] Ceph and IPv4 -> IPv6

2017-06-28 Thread Wido den Hollander
> Op 28 juni 2017 om 22:12 schreef Gregory Farnum : > > > On Wed, Jun 28, 2017 at 6:33 AM wrote: > > > > I don't think either. I don't think there is another way then just > > 'hacky' changing the MONMaps. There have been talks of being able to make > > Ceph dual-stack, but I don't think there

Re: [ceph-users] Radosgw versioning S3 compatible?

2017-06-28 Thread Martin Emrich
Thanks! We'll try that and I'll report back then... Regards, Martin -Ursprüngliche Nachricht- Von: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] Im Auftrag von Daniel Gryniewicz Gesendet: Mittwoch, 28. Juni 2017 19:12 An: ceph-users@lists.ceph.com Betreff: Re: [ceph-users] Rad