Re: [ceph-users] Using same name for rgw / beast web front end

2019-09-11 Thread Casey Bodley
Hi Eric, boost::beast is a low-level c++ http protocol library that's hosted at https://github.com/boostorg/beast. Radosgw uses this library, along with boost::asio, as the basis for its 'beast frontend'. The motivation behind this frontend is its flexible threading model and support for asyn

Re: [ceph-users] RGW configuration parameters

2019-07-30 Thread Casey Bodley
, Tom On Tue, 30 Jul 2019 at 16:57, Casey Bodley <mailto:cbod...@redhat.com>> wrote: Hi Thomas, I see that you're familiar with rgw_max_chunk_size, which is the most object data that radosgw will write in a single osd request. Each PutObj and GetObj request will

Re: [ceph-users] RGW configuration parameters

2019-07-30 Thread Casey Bodley
Hi Thomas, I see that you're familiar with rgw_max_chunk_size, which is the most object data that radosgw will write in a single osd request. Each PutObj and GetObj request will issue multiple osd requests in parallel, up to these configured window sizes. Raising these values can potentially

Re: [ceph-users] Large OMAP Objects in zone.rgw.log pool

2019-07-25 Thread Casey Bodley
he intention of one day doing multi site replication. That has never happened. The cluster has a single realm, which contains a single zonegroup, and that zonegroup contains a single zone. -Brett On Thu, Jul 25, 2019 at 2:16 PM Casey Bodley <mailto:cbod...@redhat.com>> wrote: Hi

Re: [ceph-users] Large OMAP Objects in zone.rgw.log pool

2019-07-25 Thread Casey Bodley
Hi Brett, These meta.log objects store the replication logs for metadata sync in multisite. Log entries are trimmed automatically once all other zones have processed them. Can you verify that all zones in the multisite configuration are reachable and syncing? Does 'radosgw-admin sync status'

Re: [ceph-users] RGW Admin REST metadata caps

2019-07-23 Thread Casey Bodley
the /admin/metadata apis require caps of type "metadata" source: https://github.com/ceph/ceph/blob/master/src/rgw/rgw_rest_metadata.h#L37 On 7/23/19 12:53 PM, Benjeman Meekhof wrote: Ceph Nautilus, 14.2.2, RGW civetweb. Trying to read from the RGW admin api /metadata/user with request URL lik

Re: [ceph-users] Multisite RGW - endpoints configuration

2019-07-17 Thread Casey Bodley
nchronization rgws = the faster replication)? W dniu wtorek, 16 lipca 2019 Casey Bodley mailto:cbod...@redhat.com>> napisał(a): We used to have issues when a load balancer was in front of the sync endpoints, because our http client didn't time out stalled conn

Re: [ceph-users] Multisite RGW - endpoints configuration

2019-07-16 Thread Casey Bodley
We used to have issues when a load balancer was in front of the sync endpoints, because our http client didn't time out stalled connections. Those are resolved in luminous, but we still recommend using the radosgw addresses directly to avoid shoveling data through an extra proxy. Internally, sy

Re: [ceph-users] RGW Beast crash 14.2.1

2019-07-11 Thread Casey Bodley
On 7/11/19 3:28 AM, EDH - Manuel Rios Fernandez wrote: Hi Folks, This night RGW crashed without sense using beast as fronted. We solved turning on civetweb again. Should be report to tracker? Please do. It looks like this crashed during startup. Can you please include the rgw_frontends co

Re: [ceph-users] Stop metadata sync in multi-site RGW

2019-06-19 Thread Casey Bodley
Right, the sync_from fields in the zone configuration only relate to data sync within the zonegroup. Can you clarify what your goal is? Are you just trying to pause the replication for a while, or disable it permanently? To pause replication, you can configure rgw_run_sync_thread=0 on all gat

Re: [ceph-users] Even more objects in a single bucket?

2019-06-17 Thread Casey Bodley
Hi Harry, When dynamic resharding was introduced for luminous, this limit on the number of bucket index shards was increased from 7877 to 65521. However, you're likely to have problems with bucket listing performance before you get to 7877 shards, because every listing request has to read from

Re: [ceph-users] RGW Multisite Q's

2019-06-14 Thread Casey Bodley
On 6/12/19 11:49 AM, Peter Eisch wrote: Hi, Could someone be able to point me to a blog or documentation page which helps me resolve the issues noted below? All nodes are Luminous, 12.2.12; one realm, one zonegroup (clustered haproxies fronting), two zones (three rgw in each); All endpoint re

Re: [ceph-users] RGW 405 Method Not Allowed on CreateBucket

2019-06-14 Thread Casey Bodley
Hi Drew, Judging by the "PUT /" in the request line, this request is using the virtual hosted bucket format [1]. This means the bucket name is part of the dns name and Host header, rather than in the path of the http request. Making this work in radosgw takes a little extra configuration [2].

Re: [ceph-users] rocksdb corruption, stale pg, rebuild bucket index

2019-06-12 Thread Casey Bodley
Hi Harald, If the bucket reshard didn't complete, it's most likely one of the new bucket index shards that got corrupted here and the original index shard should still be intact. Does $BAD_BUCKET_ID correspond to the new/resharded instance id? If so, once the rocksdb/osd issues are resolved,

Re: [ceph-users] [Ceph-large] Large Omap Warning on Log pool

2019-06-12 Thread Casey Bodley
Hi Aaron, The data_log objects are storing logs for multisite replication. Judging by the pool name '.us-phx2.log', this cluster was created before jewel. Are you (or were you) using multisite or radosgw-agent? If not, you'll want to turn off the logging (log_meta and log_data -> false) in y

Re: [ceph-users] is rgw crypt default encryption key long term supported ?

2019-06-11 Thread Casey Bodley
and verify that the object is still readable). Is that a bug? Thank you . Cheers Francois From: Scheurer François Sent: Wednesday, May 29, 2019 9:28 AM To: Casey Bodley; ceph-users@lists.ceph.com Subject: Re: is rgw crypt default encryption key long

Re: [ceph-users] is rgw crypt default encryption key long term supported ?

2019-05-28 Thread Casey Bodley
AES256 ________ From: Casey Bodley Sent: Tuesday, May 28, 2019 3:55 PM To: Scheurer François; ceph-users@lists.ceph.com Subject: Re: is rgw crypt default encryption key long term supported ? Hi François, Re

Re: [ceph-users] is rgw crypt default encryption key long term supported ?

2019-05-28 Thread Casey Bodley
Hi François, Removing support for either of rgw_crypt_default_encryption_key or rgw_crypt_s3_kms_encryption_keys would mean that objects encrypted with those keys would no longer be accessible. It's unlikely that we'll do that, but if we do it would be announced with a long deprecation period

Re: [ceph-users] large omap object in usage_log_pool

2019-05-24 Thread Casey Bodley
On 5/24/19 1:15 PM, shubjero wrote: Thanks for chiming in Konstantin! Wouldn't setting this value to 0 disable the sharding? Reference: http://docs.ceph.com/docs/mimic/radosgw/config-ref/ rgw override bucket index max shards Description:Represents the number of shards for the bucket index ob

Re: [ceph-users] Radosgw object size limit?

2019-05-10 Thread Casey Bodley
On 5/10/19 10:20 AM, Jan Kasprzak wrote: Hello Casey (and the ceph-users list), I am returning to my older problem to which you replied: Casey Bodley wrote: : There is a rgw_max_put_size which defaults to 5G, which limits the : size of a single PUT request. But in that case, the http

Re: [ceph-users] Ceph Bucket strange issues rgw.none + id and marker diferent.

2019-05-07 Thread Casey Bodley
slow down all bucket listings. Regards -Mensaje original- De: ceph-users En nombre de Casey Bodley Enviado el: martes, 7 de mayo de 2019 17:07 Para: ceph-users@lists.ceph.com Asunto: Re: [ceph-users] Ceph Bucket strange issues rgw.none + id and marker diferent. When the bucket id is

Re: [ceph-users] Ceph Bucket strange issues rgw.none + id and marker diferent.

2019-05-07 Thread Casey Bodley
When the bucket id is different than the bucket marker, that indicates the bucket has been resharded. Bucket stats shows 128 shards, which is reasonable for that object count. The rgw.none category in bucket stats is nothing to worry about. What ceph version is this? This reminds me of a fix i

Re: [ceph-users] Object Gateway - Server Side Encryption

2019-04-25 Thread Casey Bodley
On 4/25/19 11:33 AM, Francois Scheurer wrote: Hello Amardeep We are trying the same as you on luminous. s3cmd --access_key xxx  --secret_key xxx  --host-bucket '%(bucket)s.s3.xxx.ch' --host s3.xxx.ch --signature-v2 --no-preserve --server-side-encryption \ --server-side-encryption-kms-idhttps://

Re: [ceph-users] Multi-site replication speed

2019-04-16 Thread Casey Bodley
Hi Brian, On 4/16/19 1:57 AM, Brian Topping wrote: On Apr 15, 2019, at 5:18 PM, Brian Topping > wrote: If I am correct, how do I trigger the full sync? Apologies for the noise on this thread. I came to discover the `radosgw-admin [meta]data sync init` command.

Re: [ceph-users] RGW: Reshard index of non-master zones in multi-site

2019-04-05 Thread Casey Bodley
Hi Iain, Resharding is not supported in multisite. The issue is that the master zone needs to be authoritative for all metadata. If bucket reshard commands run on the secondary zone, they create new bucket instance metadata that the master zone never sees, so replication can't reconcile those chan

Re: [ceph-users] Looking up buckets in multi-site radosgw configuration

2019-03-19 Thread Casey Bodley
On 3/19/19 12:05 AM, David Coles wrote: I'm looking at setting up a multi-site radosgw configuration where data is sharded over multiple clusters in a single physical location; and would like to understand how Ceph handles requests in this configuration. Looking through the radosgw source[1] i

Re: [ceph-users] Rados Gateway using S3 Api does not store file correctly

2019-03-18 Thread Casey Bodley
Hi Dan, We just got a similar report about SSE-C in http://tracker.ceph.com/issues/38700 that seems to be related to multipart uploads. Could you please add some details there about your s3 client, its multipart chunk size, and your ceph version? On 3/18/19 2:38 PM, Dan Smith wrote: Hello,

Re: [ceph-users] Need clarification about RGW S3 Bucket Tagging

2019-03-14 Thread Casey Bodley
there instead and update the doc to clarify that it's not supported. If anyone's interested in working on this feature, the rgw team would happy to assist! Thanks, Casey On 3/14/19 4:05 AM, Konstantin Shalygin wrote: Hi. I CC'ed Casey Bodley as new RGW tech lead. Luminous d

Re: [ceph-users] radosgw sync falling behind regularly

2019-03-08 Thread Casey Bodley
a so a 'sync init' wouldn't refresh that info.   But a 'metadata sync init' might get rid of the stale bucket sync info and stop the sync errors.   Would that be the way to go? Thanks, Trey On Wed, Mar 6, 2019 at 11:47 AM Casey Bodley <mailto:cbod...@redhat.com&g

Re: [ceph-users] Radosgw object size limit?

2019-03-07 Thread Casey Bodley
There is a rgw_max_put_size which defaults to 5G, which limits the size of a single PUT request. But in that case, the http response would be 400 EntityTooLarge. For multipart uploads, there's also a rgw_multipart_part_upload_limit that defaults to 1 parts, which would cause a 416 InvalidRa

Re: [ceph-users] radosgw sync falling behind regularly

2019-03-06 Thread Casey Bodley
sync errors. I have found that these buckets do have entries in 'radosgw-admin datalog list', and my guess is this could be fixed by doing a 'radosgw-admin datalog trim' for each entry on the master zone. Does that sound right?  :-) Thanks again for the detailed expla

Re: [ceph-users] radosgw sync falling behind regularly

2019-03-05 Thread Casey Bodley
Hi Christian, I think you've correctly intuited that the issues are related to the use of 'bucket sync disable'. There was a bug fix for that feature in http://tracker.ceph.com/issues/26895, and I recently found that a block of code was missing from its luminous backport. That missing code is

Re: [ceph-users] Multisite Ceph setup sync issue

2019-01-29 Thread Casey Bodley
On Tue, Jan 29, 2019 at 12:24 PM Krishna Verma wrote: > > Hi Ceph Users, > > > > I need your to fix sync issue in multisite setup. > > > > I have 2 cluster in different datacenter that we want to use for > bidirectional data replication. By followed the documentation > http://docs.ceph.com/docs/

Re: [ceph-users] Bucket logging howto

2019-01-28 Thread Casey Bodley
On Sat, Jan 26, 2019 at 6:57 PM Marc Roos wrote: > > > > > From the owner account of the bucket I am trying to enable logging, but > I don't get how this should work. I see the s3:PutBucketLogging is > supported, so I guess this should work. How do you enable it? And how do > you access the log? >

Re: [ceph-users] rgw/s3: performance of range requests

2019-01-07 Thread Casey Bodley
On 1/7/19 3:15 PM, Giovani Rinaldi wrote: Hello! I've been wondering if range requests are more efficient than doing "whole" requests for relatively large objects (100MB-1GB). More precisely, my doubt is regarding the use of OSD/RGW resources, that is, does the entire object is retrieved from

Re: [ceph-users] any way to see enabled/disabled status of bucket sync?

2019-01-02 Thread Casey Bodley
Hi Christian, The easiest way to do that is probably the 'radosgw-admin bucket sync status' command, which will print "Sync is disabled for bucket ..." if disabled. Otherwise, you could use 'radosgw-admin metadata get' to inspect that flag in the bucket instance metadata. On 12/31/18 2:20 P

Re: [ceph-users] radosgw-admin unable to store user information

2019-01-02 Thread Casey Bodley
On 12/26/18 4:58 PM, Dilip Renkila wrote: Hi all, Some useful information >>/>> />>///What do the following return?/ >>/>> >> />>/>> >> $ radosgw-admin zone get/ /root@ctrl1:~# radosgw-admin zone get { "id": "8bfdf8a3-c165-44e9-9ed6-deff8a5d852f", "name": "default", "domain_root": "default.r

Re: [ceph-users] civitweb segfaults

2018-12-11 Thread Casey Bodley
Hi Leon, Are you running with a non-default value of rgw_gc_max_objs? I was able to reproduce this exact stack trace by setting rgw_gc_max_objs = 0; I can't think of any other way to get a 'Floating point exception' here. On 12/11/18 10:31 AM, Leon Robinson wrote: Hello, I have found a suref

Re: [ceph-users] rwg/civetweb log verbosity level

2018-11-28 Thread Casey Bodley
This stuff is logged under the 'civetweb' subsystem, so can be turned off with 'debug_civetweb = 0'. You can configure 'debug_rgw' separately. On 11/28/18 1:03 AM, zyn赵亚楠 wrote: Hi there, I have a question about rgw/civetweb log settings. Currently, rgw/civetweb prints 3 lines of logs with l

Re: [ceph-users] Disabling RGW Encryption support in Luminous

2018-10-16 Thread Casey Bodley
That's not currently possible, no. And I don't think it's a good idea to add such a feature; if the client requests that something be encrypted, the server should either encrypt it or reject the request. There is a config called rgw_crypt_s3_kms_encryption_keys that we use for testing, though,

Re: [ceph-users] can I define buckets in a multi-zone config that are exempted from replication?

2018-10-08 Thread Casey Bodley
On 10/08/2018 03:45 PM, Christian Rice wrote: Just getting started here, but I am setting up a three-zone realm, each with a pair of S3 object gateways, Luminous on Debian.  I’m wondering if there’s a straightforward way to exempt some buckets from replicating to other zones?  The idea bein

Re: [ceph-users] RADOS async client memory usage explodes when reading several objects in sequence

2018-09-12 Thread Casey Bodley
On 09/12/2018 05:29 AM, Daniel Goldbach wrote: Hi all, We're reading from a Ceph Luminous pool using the librados asychronous I/O API. We're seeing some concerning memory usage patterns when we read many objects in sequence. The expected behaviour is that our memory usage stabilises at a s

Re: [ceph-users] data_extra_pool for RGW Luminous still needed?

2018-09-04 Thread Casey Bodley
On 09/03/2018 10:07 PM, Nhat Ngo wrote: Hi all, I am new to Ceph and we are setting up a new RadosGW and Ceph storage cluster on Luminous. We are using only EC for our `buckets.data` pool at the moment. However, I just read the Red Hat Ceph object Gateway for Production article and it

Re: [ceph-users] Delay replicate for ceph radosgw multi-site v2

2018-08-28 Thread Casey Bodley
On 08/28/2018 09:24 AM, Jason Dillaman wrote: On Mon, Aug 27, 2018 at 11:19 PM đức phạm xuân wrote: Hello Jason Dillaman, I'm working with Ceph Object Storage Multi-Site v2, ceph's version is mimic. Now I want to delay replicate data from a master site to a slave site. I don't know whether

Re: [ceph-users] radosgw: need couple of blind (indexless) buckets, how-to?

2018-08-24 Thread Casey Bodley
On 08/24/2018 06:44 AM, Konstantin Shalygin wrote: Answer to myself. radosgw-admin realm create --rgw-realm=default --default radosgw-admin zonegroup modify --rgw-zonegroup=default --rgw-realm=default radosgw-admin period update --commit radosgw-admin zonegroup placement add --rgw-zonegroup="

Re: [ceph-users] RGW pools don't show up in luminous

2018-08-24 Thread Casey Bodley
On 08/23/2018 01:22 PM, Robert Stanford wrote:  I installed a new Ceph cluster with Luminous, after a long time working with Jewel.  I created my RGW pools the same as always (pool create default.rgw.buckets.data etc.), but they don't show up in ceph df with Luminous.  Has the command chang

Re: [ceph-users] Reset Object ACLs in RGW

2018-08-02 Thread Casey Bodley
On 08/02/2018 07:35 AM, Thomas White wrote: Hi all, At present I have a cluster with a user on the RGW who has lost access to many of his files. The bucket has the correct ACL to be accessed by the account and so with their access and secret key many items can be listed, but are unable to b

Re: [ceph-users] Why LZ4 isn't built with ceph?

2018-07-25 Thread Casey Bodley
On 07/25/2018 08:39 AM, Elias Abacioglu wrote: Hi I'm wondering why LZ4 isn't built by default for newer Linux distros like Ubuntu Xenial? I understand that it wasn't built for Trusty because of too old lz4 libraries. But why isn't built for the newer distros? Thanks, Elias _

Re: [ceph-users] multisite and link speed

2018-07-18 Thread Casey Bodley
On Tue, Jul 17, 2018 at 10:16 AM, Robert Stanford wrote: > > I have ceph clusters in a zone configured as active/passive, or > primary/backup. If the network link between the two clusters is slower than > the speed of data coming in to the active cluster, what will eventually > happen? Will dat

Re: [ceph-users] rgw non-ec pool and multipart uploads

2018-06-26 Thread Casey Bodley
Not quite. Only 'multipart meta' objects are stored in this non-ec pool - these objects just track a list of parts that have been written for a given multipart upload. This list is stored in the omap database, which isn't supported for ec pools. The actual object data for these parts are writte

Re: [ceph-users] radosgw bucket listing (s3 ls s3://$bucketname) slow with ~2 billion objects

2018-05-01 Thread Casey Bodley
The main problem with efficiently listing many-sharded buckets is the requirement to provide entries in sorted order. This means that each http request has to fetch ~1000 entries from every shard, combine them into a sorted order, and throw out the leftovers. The next request to continue the li

Re: [ceph-users] RGW bucket lifecycle policy vs versioning

2018-04-26 Thread Casey Bodley
On 04/26/2018 07:22 AM, Sean Purdy wrote: Hi, Both versioned buckets and lifecycle policies are implemented in ceph, and look useful. But are lifecycle policies implemented for versioned buckets? i.e. can I set a policy that will properly expunge all "deleted" objects after a certain time?

Re: [ceph-users] Fixing bad radosgw index

2018-04-16 Thread Casey Bodley
On 04/14/2018 12:54 PM, Robert Stanford wrote:  I deleted my default.rgw.buckets.data and default.rgw.buckets.index pools in an attempt to clean them out.  I brought this up on the list and received replies telling me essentially, "You shouldn't do that." There was however no helpful advice

Re: [ceph-users] RGW multisite sync issues

2018-04-06 Thread Casey Bodley
On 04/06/2018 10:57 AM, Josef Zelenka wrote: Hi everyone, i'm currently setting up RGW multisite(one cluster is jewel(primary), the other is luminous - this is only for testing, on prod we will have the same version - jewel on both), but i can't get bucket synchronization to work. Data gets

Re: [ceph-users] Object lifecycle and indexless buckets

2018-03-20 Thread Casey Bodley
On 03/20/2018 01:33 PM, Robert Stanford wrote:  Hello,  Does object expiration work on indexless (blind) buckets?  Thank you ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com No. Lifecy

Re: [ceph-users] Object Gateway - Server Side Encryption

2018-03-13 Thread Casey Bodley
On 03/10/2018 12:58 AM, Amardeep Singh wrote: On Saturday 10 March 2018 02:01 AM, Casey Bodley wrote: On 03/08/2018 07:16 AM, Amardeep Singh wrote: Hi, I am trying to configure server side encryption using Key Management Service as per documentation http://docs.ceph.com/docs/master

Re: [ceph-users] Object Gateway - Server Side Encryption

2018-03-09 Thread Casey Bodley
On 03/08/2018 07:16 AM, Amardeep Singh wrote: Hi, I am trying to configure server side encryption using Key Management Service as per documentation http://docs.ceph.com/docs/master/radosgw/encryption/ Configured Keystone/Barbican integration and its working, tested using curl commands. Aft

Re: [ceph-users] Significance of the us-east-1 region when using S3 clients to talk to RGW

2018-02-27 Thread Casey Bodley
s3cmd does have special handling for 'US' and 'us-east-1' that skips the LocationConstraint on bucket creation: https://github.com/s3tools/s3cmd/blob/master/S3/S3.py#L380 On 02/26/2018 05:16 PM, David Turner wrote: I just realized the difference between the internal realm, local realm, and lo

Re: [ceph-users] Is the minimum length of a part in a RGW multipart upload configurable?

2018-02-16 Thread Casey Bodley
On 02/16/2018 12:39 AM, F21 wrote: I am uploading parts to RGW using the S3 multipart upload functionality. I tried uploading a part sized at 500 KB and received a EntityTooSmall error from the server. I am assuming that it expects each part to have a minimum size of 5MB like S3. I found `r

Re: [ceph-users] Error message in the logs: "meta sync: ERROR: failed to read mdlog info with (2) No such file or directory"

2018-01-17 Thread Casey Bodley
On 01/15/2018 09:57 AM, Victor Flávio wrote: Hello, We've have a radosgw cluster(verion 12.2.2) in multisite mode. Our cluster is formed by one master realm, with one master zonegroup and two zones(which one is the master zone). We've followed the instructions of Ceph documentation to insta

Re: [ceph-users] How to "reset" rgw?

2018-01-10 Thread Casey Bodley
On 01/10/2018 04:34 AM, Martin Emrich wrote: Hi! As I cannot find any solution for my broken rgw pools, the only way out is to give up and "reset". How do I throw away all rgw data from a ceph cluster? Just delete all rgw pools? Or are some parts stored elsewhere (monitor, ...)? Thanks,

Re: [ceph-users] Luminous rgw hangs after sighup

2017-12-11 Thread Casey Bodley
There have been other issues related to hangs during realm reconfiguration, ex http://tracker.ceph.com/issues/20937. We decided to revert the use of SIGHUP to trigger realm reconfiguration in https://github.com/ceph/ceph/pull/16807. I just started a backport of that for luminous. On 12/11/20

Re: [ceph-users] 10.2.10: "default" zonegroup in custom root pool not found

2017-11-15 Thread Casey Bodley
On 11/15/2017 12:11 AM, Richard Chan wrote: After creating a non-default root pool rgw_realm_root_pool = gold.rgw.root rgw_zonegroup_root_pool = gold.rgw.root rgw_period_root_pool = gold.rgw.root rgw_zone_root_pool = gold.rgw.root rgw_region = gold.rgw.root You probably meant to set rgw_regio

Re: [ceph-users] radosgw multi site different period

2017-11-15 Thread Casey Bodley
Your period configuration is indeed consistent between zones. This "master is on a different period" error is specific to the metadata sync status. It's saying that zone b is unable to finish syncing the metadata changes from zone a that occurred during the previous period. Even though zone b w

Re: [ceph-users] Luminous 12.2.1 - RadosGW Multisite doesnt replicate multipart uploads

2017-10-12 Thread Casey Bodley
Thanks Enrico. I wrote a test case that reproduces the issue, and opened http://tracker.ceph.com/issues/21772 to track the bug. It sounds like this is a regression in luminous. On 10/11/2017 06:41 PM, Enrico Kern wrote: or this: { "shard_id": 22, "entries": [ {

Re: [ceph-users] RGW flush_read_list error

2017-10-11 Thread Casey Bodley
Hi Travis, This is reporting an error when sending data back to the client. Generally it means that the client timed out and closed the connection. Are you also seeing failures on the client side? Casey On 10/10/2017 06:45 PM, Travis Nielsen wrote: In Luminous 12.2.1, when running a GET on

Re: [ceph-users] RGW Multisite metadata sync init

2017-08-25 Thread Casey Bodley
Hi David, The 'data sync init' command won't touch any actual object data, no. Resetting the data sync status will just cause a zone to restart a full sync of the --source-zone's data changes log. This log only lists which buckets/shards have changes in them, which causes radosgw to consider

Re: [ceph-users] RGW Multisite metadata sync init

2017-08-25 Thread Casey Bodley
Hi David, The 'radosgw-admin sync error list' command may be useful in debugging sync failures for specific entries. For users, we've seen some sync failures caused by conflicting user metadata that was only present on the secondary site. For example, a user that had the same access key or em

Re: [ceph-users] RGW Multisite Sync Memory Usage

2017-07-26 Thread Casey Bodley
Hi Ryan, Sorry to hear about the crashes. Based on the fact that it's happening on the source zone, I'm guessing that you're hitting this infinite loop that leads to OOM: http://tracker.ceph.com/issues/20386. The jewel backport for this one is still pending, so I raised its priority to Urgent

Re: [ceph-users] radosgw: scrub causing slow requests in the md log

2017-06-22 Thread Casey Bodley
On 06/22/2017 10:40 AM, Dan van der Ster wrote: On Thu, Jun 22, 2017 at 4:25 PM, Casey Bodley wrote: On 06/22/2017 04:00 AM, Dan van der Ster wrote: I'm now running the three relevant OSDs with that patch. (Recompiled, replaced /usr/lib64/rados-classes/libcls_log.so with the new ve

Re: [ceph-users] radosgw: scrub causing slow requests in the md log

2017-06-22 Thread Casey Bodley
s the entire range inside of the suicide timeout. Casey On Wed, Jun 21, 2017 at 3:51 PM, Casey Bodley wrote: That patch looks reasonable. You could also try raising the values of osd_op_thread_suicide_timeout and filestore_op_thread_suicide_timeout on that osd in order to trim more at a ti

Re: [ceph-users] radosgw: scrub causing slow requests in the md log

2017-06-21 Thread Casey Bodley
_map_get_vals(hctx, from_index, log_index_prefix, max_entries, &keys); What do you think? -- Dan On Mon, Jun 19, 2017 at 5:32 PM, Casey Bodley wrote: Hi Dan, That's good news that it can remove 1000 keys at a time without hitting timeouts. The output of 'du' will depend on when

Re: [ceph-users] radosgw: scrub causing slow requests in the md log

2017-06-19 Thread Casey Bodley
put in the log), consider running offline compaction by adding 'leveldb compact on mount = true' to the osd config and restarting. Casey On 06/19/2017 11:01 AM, Dan van der Ster wrote: On Thu, Jun 15, 2017 at 7:56 PM, Casey Bodley wrote: On 06/14/2017 05:59 AM, Dan van der Ster wrote

Re: [ceph-users] radosgw: scrub causing slow requests in the md log

2017-06-15 Thread Casey Bodley
On 06/14/2017 05:59 AM, Dan van der Ster wrote: Dear ceph users, Today we had O(100) slow requests which were caused by deep-scrubbing of the metadata log: 2017-06-14 11:07:55.373184 osd.155 [2001:1458:301:24::100:d]:6837/3817268 7387 : cluster [INF] 24.1d deep-scrub starts ... 2017-06-14 11:2

Re: [ceph-users] radosgw global quotas - how to set in jewel?

2017-04-05 Thread Casey Bodley
asey On 04/03/2017 03:23 PM, Graham Allan wrote: Ah, thanks, I thought I was going crazy for a bit there! The global quota would be useful for us (now wanting to retroactively impose quotas on pre-existing users), but we can script a workaround instead. Thanks, Graham On 03/29/2017 10:17 AM, Cas

Re: [ceph-users] radosgw global quotas - how to set in jewel?

2017-03-29 Thread Casey Bodley
be the right place to look for jewel. Am I missing something, or should I open a bug? Graham On 03/21/2017 03:18 PM, Graham Allan wrote: On 03/17/2017 11:47 AM, Casey Bodley wrote: On 03/16/2017 03:47 PM, Graham Allan wrote: This might be a dumb question, but I'm not at all sure what t

Re: [ceph-users] radosgw global quotas

2017-03-17 Thread Casey Bodley
On 03/16/2017 03:47 PM, Graham Allan wrote: This might be a dumb question, but I'm not at all sure what the "global quotas" in the radosgw region map actually do. It is like a default quota which is applied to all users or buckets, without having to set them individually, or is it a blanket/a

Re: [ceph-users] radosgw. Strange behavior in 2 zone configuration

2017-03-06 Thread Casey Bodley
On 03/03/2017 07:40 AM, K K wrote: Hello, all! I have successfully create 2 zone cluster(se and se2). But my radosgw machines are sending many GET /admin/log requests to each other after put 10k items to cluster via radosgw. It's look like: 2017-03-03 17:31:17.897872 7f21b9083700 1 civetw

Re: [ceph-users] radosgw-admin bucket link: empty bucket instance id

2017-02-21 Thread Casey Bodley
When it complains about a missing bucket instance id, that's what it's expecting to get from the --bucket-id argument. That's the "id" field shown in bucket stats. Try this? $ radosgw-admin bucket link --bucket=XXX --bucket-id=YYY --uid=ZZZ Casey On 02/21/2017 08:30 AM, Valery Tschopp wrote:

Re: [ceph-users] RADOSGW S3 api ACLs

2017-02-16 Thread Casey Bodley
On 02/16/2017 07:17 AM, Josef Zelenka wrote: Hello everyone, i've been struggling for the past few days with setting up ACLs for buckets on my radosgw. I want to use the buckets with the s3 API and i want them to have the ACL set up like this: every file that gets pushed into the bucket is au

Re: [ceph-users] ceph rados gw, select objects by metadata

2017-01-30 Thread Casey Bodley
On 01/30/2017 06:11 AM, Johann Schwarzmeier wrote: Hello Wido, That is not good news, but it's what i expected. Thanks for your qick answer. Jonny Am 2017-01-30 11:57, schrieb Wido den Hollander: Op 30 januari 2017 om 10:29 schreef Johann Schwarzmeier : Hello, I’m quite new to ceph and ra

Re: [ceph-users] High CPU load with radosgw instances

2016-09-16 Thread Casey Bodley
In the meantime, we've made changes to radosgw so that it can detect and work around this libcurl bug. You can track the progress of this workaround (currently in master and pending backport to jewel) at http://tracker.ceph.com/issues/16695. Casey On 09/16/2016 01:38 PM, Ken Dreyer wrote: H

Re: [ceph-users] rgw meta pool

2016-09-09 Thread Casey Bodley
Hi, My (limited) understanding of this metadata heap pool is that it's an archive of metadata entries and their versions. According to Yehuda, this was intended to support recovery operations by reverting specific metadata objects to a previous version. But nothing has been implemented so far

Re: [ceph-users] Excluding buckets in RGW Multi-Site Sync

2016-09-08 Thread Casey Bodley
On 09/08/2016 08:35 AM, Wido den Hollander wrote: Hi, I've been setting up a RGW Multi-Site [0] configuration in 6 VMs. 3 VMs per cluster and one RGW per cluster. Works just fine, I can create a user in the master zone, create buckets and upload data using s3cmd (S3). What I see is that ALL