Hello Orit, thanks.
I will do all 6 just in case. Also as an FYI I originally had all 6 as
endpoint (3 in each zone) but have it down to just the two "1" servers
talking to each other until I can get it working. Eventually I would like
to have all 6 cross connecting again.
*rgw-primary-1:*
radosgw-admin zonegroup get
{
"id": "235b010c-22e2-4b43-8fcc-8ae01939273e",
"name": "us",
"api_name": "us",
"is_master": "true",
"endpoints": [
"http:\/\/LB_FQDN:80"
],
"hostnames": [],
"hostnames_s3website": [],
"master_zone": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"zones": [
{
"id": "58aa3eef-fc1f-492c-a08e-9c6019e7c266",
"name": "us-phx",
"endpoints": [
"http:\/\/PHX-RGW-1:80"
],
"log_meta": "false",
"log_data": "true",
"bucket_index_max_shards": 0,
"read_only": "false"
},
{
"id": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"name": "us-dfw",
"endpoints": [
"http:\/\/DFW-RGW-1:80"
],
"log_meta": "true",
"log_data": "true",
"bucket_index_max_shards": 0,
"read_only": "false"
}
],
"placement_targets": [
{
"name": "default-placement",
"tags": []
}
],
"default_placement": "default-placement",
"realm_id": "3af93a86-916a-490f-b38f-17922b472b19"
}
radosgw-admin zone get
{
"id": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"name": "us-dfw",
"domain_root": "us-dfw.rgw.data.root",
"control_pool": "us-dfw.rgw.control",
"gc_pool": "us-dfw.rgw.gc",
"log_pool": "us-dfw.rgw.log",
"intent_log_pool": "us-dfw.rgw.intent-log",
"usage_log_pool": "us-dfw.rgw.usage",
"user_keys_pool": "us-dfw.rgw.users.keys",
"user_email_pool": "us-dfw.rgw.users.email",
"user_swift_pool": "us-dfw.rgw.users.swift",
"user_uid_pool": "us-dfw.rgw.users.uid",
"system_key": {
"access_key": "SYSTEM_ACCESS_KEY",
"secret_key": "SYSTEM_SECRET_KEY"
},
"placement_pools": [
{
"key": "default-placement",
"val": {
"index_pool": "us-dfw.rgw.buckets.index",
"data_pool": "us-dfw.rgw.buckets.data",
"data_extra_pool": "us-dfw.rgw.buckets.non-ec",
"index_type": 0
}
}
],
"metadata_heap": "us-dfw.rgw.meta",
"realm_id": "3af93a86-916a-490f-b38f-17922b472b19"
}
----
*rgw-primary-2:*
radosgw-admin zonegroup get
{
"id": "235b010c-22e2-4b43-8fcc-8ae01939273e",
"name": "us",
"api_name": "us",
"is_master": "true",
"endpoints": [
"http:\/\/LB_FQDN:80"
],
"hostnames": [],
"hostnames_s3website": [],
"master_zone": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"zones": [
{
"id": "58aa3eef-fc1f-492c-a08e-9c6019e7c266",
"name": "us-phx",
"endpoints": [
"http:\/\/PHX-RGW-1:80"
],
"log_meta": "false",
"log_data": "true",
"bucket_index_max_shards": 0,
"read_only": "false"
},
{
"id": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"name": "us-dfw",
"endpoints": [
"http:\/\/DFW-RGW-1:80"
],
"log_meta": "true",
"log_data": "true",
"bucket_index_max_shards": 0,
"read_only": "false"
}
],
"placement_targets": [
{
"name": "default-placement",
"tags": []
}
],
"default_placement": "default-placement",
"realm_id": "3af93a86-916a-490f-b38f-17922b472b19"
}
radosgw-admin zone get
{
"id": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"name": "us-dfw",
"domain_root": "us-dfw.rgw.data.root",
"control_pool": "us-dfw.rgw.control",
"gc_pool": "us-dfw.rgw.gc",
"log_pool": "us-dfw.rgw.log",
"intent_log_pool": "us-dfw.rgw.intent-log",
"usage_log_pool": "us-dfw.rgw.usage",
"user_keys_pool": "us-dfw.rgw.users.keys",
"user_email_pool": "us-dfw.rgw.users.email",
"user_swift_pool": "us-dfw.rgw.users.swift",
"user_uid_pool": "us-dfw.rgw.users.uid",
"system_key": {
"access_key": "SYSTEM_ACCESS_KEY",
"secret_key": "SYSTEM_SECRET_KEY"
},
"placement_pools": [
{
"key": "default-placement",
"val": {
"index_pool": "us-dfw.rgw.buckets.index",
"data_pool": "us-dfw.rgw.buckets.data",
"data_extra_pool": "us-dfw.rgw.buckets.non-ec",
"index_type": 0
}
}
],
"metadata_heap": "us-dfw.rgw.meta",
"realm_id": "3af93a86-916a-490f-b38f-17922b472b19"
}
----
*rgw-primary-3:*
radosgw-admin zonegroup get
{
"id": "235b010c-22e2-4b43-8fcc-8ae01939273e",
"name": "us",
"api_name": "us",
"is_master": "true",
"endpoints": [
"http:\/\/LB_FQDN:80"
],
"hostnames": [],
"hostnames_s3website": [],
"master_zone": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"zones": [
{
"id": "58aa3eef-fc1f-492c-a08e-9c6019e7c266",
"name": "us-phx",
"endpoints": [
"http:\/\/PHX-RGW-1:80"
],
"log_meta": "false",
"log_data": "true",
"bucket_index_max_shards": 0,
"read_only": "false"
},
{
"id": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"name": "us-dfw",
"endpoints": [
"http:\/\/DFW-RGW-1:80"
],
"log_meta": "true",
"log_data": "true",
"bucket_index_max_shards": 0,
"read_only": "false"
}
],
"placement_targets": [
{
"name": "default-placement",
"tags": []
}
],
"default_placement": "default-placement",
"realm_id": "3af93a86-916a-490f-b38f-17922b472b19"
}
sudo radosgw-admin zone get
{
"id": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"name": "us-dfw",
"domain_root": "us-dfw.rgw.data.root",
"control_pool": "us-dfw.rgw.control",
"gc_pool": "us-dfw.rgw.gc",
"log_pool": "us-dfw.rgw.log",
"intent_log_pool": "us-dfw.rgw.intent-log",
"usage_log_pool": "us-dfw.rgw.usage",
"user_keys_pool": "us-dfw.rgw.users.keys",
"user_email_pool": "us-dfw.rgw.users.email",
"user_swift_pool": "us-dfw.rgw.users.swift",
"user_uid_pool": "us-dfw.rgw.users.uid",
"system_key": {
"access_key": "SYSTEM_ACCESS_KEY",
"secret_key": "SYSTEM_SECRET_KEY"
},
"placement_pools": [
{
"key": "default-placement",
"val": {
"index_pool": "us-dfw.rgw.buckets.index",
"data_pool": "us-dfw.rgw.buckets.data",
"data_extra_pool": "us-dfw.rgw.buckets.non-ec",
"index_type": 0
}
}
],
"metadata_heap": "us-dfw.rgw.meta",
"realm_id": "3af93a86-916a-490f-b38f-17922b472b19"
}
---
*rgw-secondary-1:*
radosgw-admin zonegroup get
{
"id": "235b010c-22e2-4b43-8fcc-8ae01939273e",
"name": "us",
"api_name": "us",
"is_master": "true",
"endpoints": [
"http:\/\/LB_FQDN:80"
],
"hostnames": [],
"hostnames_s3website": [],
"master_zone": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"zones": [
{
"id": "58aa3eef-fc1f-492c-a08e-9c6019e7c266",
"name": "us-phx",
"endpoints": [
"http:\/\/PHX-RGW-1:80"
],
"log_meta": "false",
"log_data": "true",
"bucket_index_max_shards": 0,
"read_only": "false"
},
{
"id": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"name": "us-dfw",
"endpoints": [
"http:\/\/DFW-RGW-1:80"
],
"log_meta": "true",
"log_data": "true",
"bucket_index_max_shards": 0,
"read_only": "false"
}
],
"placement_targets": [
{
"name": "default-placement",
"tags": []
}
],
"default_placement": "default-placement",
"realm_id": "3af93a86-916a-490f-b38f-17922b472b19"
}
radosgw-admin zone get
{
"id": "58aa3eef-fc1f-492c-a08e-9c6019e7c266",
"name": "us-phx",
"domain_root": "us-phx.rgw.data.root",
"control_pool": "us-phx.rgw.control",
"gc_pool": "us-phx.rgw.gc",
"log_pool": "us-phx.rgw.log",
"intent_log_pool": "us-phx.rgw.intent-log",
"usage_log_pool": "us-phx.rgw.usage",
"user_keys_pool": "us-phx.rgw.users.keys",
"user_email_pool": "us-phx.rgw.users.email",
"user_swift_pool": "us-phx.rgw.users.swift",
"user_uid_pool": "us-phx.rgw.users.uid",
"system_key": {
"access_key": "SYSTEM_ACCESS_KEY",
"secret_key": "SYSTEM_SECRET_KEY"
},
"placement_pools": [
{
"key": "default-placement",
"val": {
"index_pool": "us-phx.rgw.buckets.index",
"data_pool": "us-phx.rgw.buckets.data",
"data_extra_pool": "us-phx.rgw.buckets.non-ec",
"index_type": 0
}
}
],
"metadata_heap": "us-phx.rgw.meta",
"realm_id": "3af93a86-916a-490f-b38f-17922b472b19"
}
---
*rgw-secondary-2:*
radosgw-admin zonegroup get
{
"id": "235b010c-22e2-4b43-8fcc-8ae01939273e",
"name": "us",
"api_name": "us",
"is_master": "true",
"endpoints": [
"http:\/\/LB_FQDN:80"
],
"hostnames": [],
"hostnames_s3website": [],
"master_zone": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"zones": [
{
"id": "58aa3eef-fc1f-492c-a08e-9c6019e7c266",
"name": "us-phx",
"endpoints": [
"http:\/\/PHX-RGW-1:80"
],
"log_meta": "false",
"log_data": "true",
"bucket_index_max_shards": 0,
"read_only": "false"
},
{
"id": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"name": "us-dfw",
"endpoints": [
"http:\/\/DFW-RGW-1:80"
],
"log_meta": "true",
"log_data": "true",
"bucket_index_max_shards": 0,
"read_only": "false"
}
],
"placement_targets": [
{
"name": "default-placement",
"tags": []
}
],
"default_placement": "default-placement",
"realm_id": "3af93a86-916a-490f-b38f-17922b472b19"
}
radosgw-admin zone get
{
"id": "58aa3eef-fc1f-492c-a08e-9c6019e7c266",
"name": "us-phx",
"domain_root": "us-phx.rgw.data.root",
"control_pool": "us-phx.rgw.control",
"gc_pool": "us-phx.rgw.gc",
"log_pool": "us-phx.rgw.log",
"intent_log_pool": "us-phx.rgw.intent-log",
"usage_log_pool": "us-phx.rgw.usage",
"user_keys_pool": "us-phx.rgw.users.keys",
"user_email_pool": "us-phx.rgw.users.email",
"user_swift_pool": "us-phx.rgw.users.swift",
"user_uid_pool": "us-phx.rgw.users.uid",
"system_key": {
"access_key": "SYSTEM_ACCESS_KEY",
"secret_key": "SYSTEM_SECRET_KEY"
},
"placement_pools": [
{
"key": "default-placement",
"val": {
"index_pool": "us-phx.rgw.buckets.index",
"data_pool": "us-phx.rgw.buckets.data",
"data_extra_pool": "us-phx.rgw.buckets.non-ec",
"index_type": 0
}
}
],
"metadata_heap": "us-phx.rgw.meta",
"realm_id": "3af93a86-916a-490f-b38f-17922b472b19"
}
---
*rgw-secondary-3*
radosgw-admin zonegroup get
{
"id": "235b010c-22e2-4b43-8fcc-8ae01939273e",
"name": "us",
"api_name": "us",
"is_master": "true",
"endpoints": [
"http:\/\/LB_FQDN:80"
],
"hostnames": [],
"hostnames_s3website": [],
"master_zone": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"zones": [
{
"id": "58aa3eef-fc1f-492c-a08e-9c6019e7c266",
"name": "us-phx",
"endpoints": [
"http:\/\/PHX-RGW-1:80"
],
"log_meta": "false",
"log_data": "true",
"bucket_index_max_shards": 0,
"read_only": "false"
},
{
"id": "6c830b44-4e39-4e19-9bd8-03c37c2021f2",
"name": "us-dfw",
"endpoints": [
"http:\/\/DFW-RGW-1:80"
],
"log_meta": "true",
"log_data": "true",
"bucket_index_max_shards": 0,
"read_only": "false"
}
],
"placement_targets": [
{
"name": "default-placement",
"tags": []
}
],
"default_placement": "default-placement",
"realm_id": "3af93a86-916a-490f-b38f-17922b472b19"
}
radosgw-admin zone get
{
"id": "58aa3eef-fc1f-492c-a08e-9c6019e7c266",
"name": "us-phx",
"domain_root": "us-phx.rgw.data.root",
"control_pool": "us-phx.rgw.control",
"gc_pool": "us-phx.rgw.gc",
"log_pool": "us-phx.rgw.log",
"intent_log_pool": "us-phx.rgw.intent-log",
"usage_log_pool": "us-phx.rgw.usage",
"user_keys_pool": "us-phx.rgw.users.keys",
"user_email_pool": "us-phx.rgw.users.email",
"user_swift_pool": "us-phx.rgw.users.swift",
"user_uid_pool": "us-phx.rgw.users.uid",
"system_key": {
"access_key": "SYSTEM_ACCESS_KEY",
"secret_key": "SYSTEM_SECRET_KEY"
},
"placement_pools": [
{
"key": "default-placement",
"val": {
"index_pool": "us-phx.rgw.buckets.index",
"data_pool": "us-phx.rgw.buckets.data",
"data_extra_pool": "us-phx.rgw.buckets.non-ec",
"index_type": 0
}
}
],
"metadata_heap": "us-phx.rgw.meta",
"realm_id": "3af93a86-916a-490f-b38f-17922b472b19"
}
On Thu, Sep 22, 2016 at 5:18 AM Orit Wasserman <[email protected]> wrote:
> Hi John,
> Can you provide your zonegroup and zones configurations on all 3 rgw?
> (run the commands on each rgw)
>
> Thanks,
> Orit
>
> On Wed, Sep 21, 2016 at 11:14 PM, John Rowe <[email protected]>
> wrote:
> > Hello,
> >
> > We have 2 Ceph clusters running in two separate data centers, each one
> with
> > 3 mons, 3 rgws, and 5 osds. I am attempting to get bi-directional
> multi-site
> > replication setup as described in the ceph documentation here:
> > http://docs.ceph.com/docs/jewel/radosgw/multisite/
> >
> > We are running Jewel v 10.2.2:
> > rpm -qa | grep ceph
> > ceph-base-10.2.2-0.el7.x86_64
> > ceph-10.2.2-0.el7.x86_64
> > ceph-radosgw-10.2.2-0.el7.x86_64
> > libcephfs1-10.2.2-0.el7.x86_64
> > python-cephfs-10.2.2-0.el7.x86_64
> > ceph-selinux-10.2.2-0.el7.x86_64
> > ceph-mon-10.2.2-0.el7.x86_64
> > ceph-osd-10.2.2-0.el7.x86_64
> > ceph-release-1-1.el7.noarch
> > ceph-common-10.2.2-0.el7.x86_64
> > ceph-mds-10.2.2-0.el7.x86_64
> >
> > It appears syncing is happening, however it is not able to sync the
> > metadata, and therefore no users/buckets from the primary are going to
> the
> > secondary.
> > Primary sync status:
> > radosgw-admin sync status
> > realm 3af93a86-916a-490f-b38f-17922b472b19 (my_realm)
> > zonegroup 235b010c-22e2-4b43-8fcc-8ae01939273e (us)
> > zone 6c830b44-4e39-4e19-9bd8-03c37c2021f2 (us-dfw)
> > metadata sync no sync (zone is master)
> > data sync source: 58aa3eef-fc1f-492c-a08e-9c6019e7c266 (us-phx)
> > syncing
> > full sync: 0/128 shards
> > incremental sync: 128/128 shards
> > data is caught up with source
> >
> > radosgw-admin data sync status --source-zone=us-phx
> > {
> > "sync_status": {
> > "info": {
> > "status": "sync",
> > "num_shards": 128
> > },
> > "markers": [
> > ...
> > }
> >
> > radosgw-admin metadata sync status
> > {
> > "sync_status": {
> > "info": {
> > "status": "init",
> > "num_shards": 0,
> > "period": "",
> > "realm_epoch": 0
> > },
> > "markers": []
> > },
> > "full_sync": {
> > "total": 0,
> > "complete": 0
> > }
> > }
> > Secondary sync status:
> > radosgw-admin sync status
> > realm 3af93a86-916a-490f-b38f-17922b472b19 (pardot)
> > zonegroup 235b010c-22e2-4b43-8fcc-8ae01939273e (us)
> > zone 58aa3eef-fc1f-492c-a08e-9c6019e7c266 (us-phx)
> > metadata sync failed to read sync status: (2) No such file or directory
> > data sync source: 6c830b44-4e39-4e19-9bd8-03c37c2021f2 (us-dfw)
> > syncing
> > full sync: 0/128 shards
> > incremental sync: 128/128 shards
> > data is behind on 10 shards
> > oldest incremental change not applied: 2016-09-20
> > 15:00:17.0.330225s
> >
> > radosgw-admin data sync status --source-zone=us-dfw
> > {
> > "sync_status": {
> > "info": {
> > "status": "building-full-sync-maps",
> > "num_shards": 128
> > },
> > ....
> > }
> >
> > radosgw-admin metadata sync status --source-zone=us-dfw
> > ERROR: sync.read_sync_status() returned ret=-2
> >
> >
> >
> > In the logs I am seeing (date stamps not in order to pick out non-dupes):
> > Primary logs:
> > 2016-09-20 15:02:44.313204 7f2a2dffb700 0 ERROR:
> > client_io->complete_request() returned -5
> > 2016-09-20 10:31:57.501247 7faf4bfff700 0 ERROR: failed to wait for op,
> > ret=-11: POST
> >
> http://pardot0-cephrgw1-3-phx.ops.sfdc.net:80/admin/realm/period?period=385c44c7-0506-4204-90d7-9d26a6cbaad2&epoch=12&rgwx-zonegroup=f46ce11b-ee5d-489b-aa30-752fc5353931
> > 2016-09-20 10:32:03.391118 7fb12affd700 0 ERROR: failed to fetch datalog
> > info
> > 2016-09-20 10:32:03.491520 7fb12affd700 0 ERROR: lease cr failed, done
> > early
> >
> > Secondary logs;
> > 2016-09-20 10:28:15.290050 7faab2fed700 0 ERROR: failed to get bucket
> > instance info for bucket
> > id=BUCKET1:78301214-35bb-41df-a77f-24968ee4b3ff.104293.1
> > 2016-09-20 10:28:15.290108 7faab5ff3700 0 ERROR: failed to get bucket
> > instance info for bucket
> > id=BUCKET1:78301214-35bb-41df-a77f-24968ee4b3ff.104293.1
> > 2016-09-20 10:28:15.290571 7faab77f6700 0 ERROR: failed to get bucket
> > instance info for bucket
> > id=BUCKET1:78301214-35bb-41df-a77f-24968ee4b3ff.104293.1
> > 2016-09-20 10:28:15.304619 7faaad7e2700 0 ERROR: failed to get bucket
> > instance info for bucket
> > id=BUCKET1:78301214-35bb-41df-a77f-24968ee4b3ff.104293.1
> > 2016-09-20 10:28:38.169629 7fa98bfff700 0 ERROR: failed to distribute
> cache
> > for .rgw.root:periods.385c44c7-0506-4204-90d7-9d26a6cbaad2.12
> > 2016-09-20 10:28:38.169642 7fa98bfff700 -1 period epoch 12 is not newer
> than
> > current epoch 12, discarding update
> > 2016-09-21 03:19:01.550808 7fe10bfff700 0 rgw meta sync: ERROR: failed
> to
> > fetch mdlog info
> > 2016-09-21 15:45:09.799195 7fcd677fe700 0 ERROR: failed to fetch remote
> > data log info: ret=-11
> >
> > Each of those logs are repeated several times each, and constantly.
> >
> > Any help would be greatly appreciated. Thanks!
> >
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > [email protected]
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com