[root@OSKVM1 ~]# grep -v "^#" /etc/nova/nova.conf|grep -v ^$ [DEFAULT]
instance_usage_audit = True instance_usage_audit_period = hour notify_on_state_change = vm_and_task_state notification_driver = messagingv2 rbd_user=cinder rbd_secret_uuid=1989f7a6-4ecb-4738-abbf-2962c29b2bbb rpc_backend = rabbit auth_strategy = keystone my_ip = 10.1.0.4 network_api_class = nova.network.neutronv2.api.API security_group_api = neutron linuxnet_interface_driver = nova.network.linux_net.NeutronLinuxBridgeInterfaceDriver firewall_driver = nova.virt.firewall.NoopFirewallDriver enabled_apis=osapi_compute,metadata [api_database] connection = mysql://nova:nova@controller/nova [barbican] [cells] [cinder] os_region_name = RegionOne [conductor] [cors] [cors.subdomain] [database] [ephemeral_storage_encryption] [glance] host = controller [guestfs] [hyperv] [image_file_url] [ironic] [keymgr] [keystone_authtoken] auth_uri = http://controller:5000 auth_url = http://controller:35357 auth_plugin = password project_domain_id = default user_domain_id = default project_name = service username = nova password = nova [libvirt] inject_password=false inject_key=false inject_partition=-2 live_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER, VIR_MIGRATE_LIVE, VIR_MIGRATE_PERSIST_DEST, VIR_MIGRATE_TUNNELLED disk_cachemodes ="network=writeback" images_type=rbd images_rbd_pool=vms images_rbd_ceph_conf =/etc/ceph/ceph.conf rbd_user=cinder rbd_secret_uuid=1989f7a6-4ecb-4738-abbf-2962c29b2bbb hw_disk_discard=unmap [matchmaker_redis] [matchmaker_ring] [metrics] [neutron] url = http://controller:9696 auth_url = http://controller:35357 auth_plugin = password project_domain_id = default user_domain_id = default region_name = RegionOne project_name = service username = neutron password = neutron service_metadata_proxy = True metadata_proxy_shared_secret = XXXXX [osapi_v21] [oslo_concurrency] lock_path = /var/lib/nova/tmp [oslo_messaging_amqp] [oslo_messaging_qpid] [oslo_messaging_rabbit] rabbit_host = controller rabbit_userid = openstack rabbit_password = XXXXX [oslo_middleware] [rdp] [serial_console] [spice] [ssl] [trusted_computing] [upgrade_levels] [vmware] [vnc] enabled = True vncserver_listen = 0.0.0.0 novncproxy_base_url = http://controller:6080/vnc_auto.html vncserver_proxyclient_address = $my_ip [workarounds] [xenserver] [zookeeper] [root@OSKVM1 ceph]# ls -ltr total 24 -rwxr-xr-x 1 root root 92 May 10 12:58 rbdmap -rw------- 1 root root 0 Jun 28 11:05 tmpfDt6jw -rw-r--r-- 1 root root 63 Jul 5 12:59 ceph.client.admin.keyring -rw-r--r-- 1 glance glance 64 Jul 5 14:51 ceph.client.glance.keyring -rw-r--r-- 1 cinder cinder 64 Jul 5 14:53 ceph.client.cinder.keyring -rw-r--r-- 1 cinder cinder 71 Jul 5 14:54 ceph.client.cinder-backup.keyring -rwxrwxrwx 1 root root 438 Jul 7 14:19 ceph.conf [root@OSKVM1 ceph]# more ceph.client.cinder.keyring [client.cinder] key = AQCIAHxX9ga8LxAAU+S3Vybdu+Cm2bP3lplGnA== [root@OSKVM1 ~]# rados lspools rbd volumes images backups vms [root@OSKVM1 ~]# rbd -p rbd ls [root@OSKVM1 ~]# rbd -p volumes ls volume-27717a88-3c80-420f-8887-4ca5c5b94023 volume-3bd22868-cb2a-4881-b9fb-ae91a6f79cb9 volume-b9cf7b94-cfb6-4b55-816c-10c442b23519 [root@OSKVM1 ~]# rbd -p images ls 9aee6c4e-3b60-49d5-8e17-33953e384a00 a8b45c8a-a5c8-49d8-a529-1e4088bdbf3f [root@OSKVM1 ~]# rbd -p vms ls [root@OSKVM1 ~]# rbd -p backup *I could create cinder and attach it to one of already built nova instance.* [root@OSKVM1 ceph]# nova volume-list WARNING: Command volume-list is deprecated and will be removed after Nova 13.0.0 is released. Use python-cinderclient or openstackclient instead. +--------------------------------------+-----------+------------------+------+-------------+--------------------------------------+ | ID | Status | Display Name | Size | Volume Type | Attached to | +--------------------------------------+-----------+------------------+------+-------------+--------------------------------------+ | 14a572d0-2834-40d6-9650-cb3e18271963 | available | nova-vol_gg | 10 | - | | | 3bd22868-cb2a-4881-b9fb-ae91a6f79cb9 | in-use | nova-vol_1 | 2 | - | d06f7c3b-5bbd-4597-99ce-fa981d2e10db | | 27717a88-3c80-420f-8887-4ca5c5b94023 | available | cinder-ceph-vol1 | 10 | - | | +--------------------------------------+-----------+------------------+------+-------------+--------------------------------------+ On Fri, Jul 8, 2016 at 11:33 AM, Gaurav Goyal <er.gauravgo...@gmail.com> wrote: > Hi Kees, > > I regenerated the UUID as per your suggestion. > Now i have same UUID in host1 and host2. > I could create volumes and attach them to existing VMs. > > I could create new glance images. > > But still finding the same error while instance launch via GUI. > > > 2016-07-08 11:23:25.067 86007 INFO nova.compute.resource_tracker > [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Auditing locally > available compute resources for node controller > > 2016-07-08 11:23:25.527 86007 INFO nova.compute.resource_tracker > [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Total usable vcpus: > 40, total allocated vcpus: 0 > > 2016-07-08 11:23:25.527 86007 INFO nova.compute.resource_tracker > [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Final resource view: > name=controller phys_ram=193168MB used_ram=1024MB phys_disk=8168GB > used_disk=1GB total_vcpus=40 used_vcpus=0 pci_stats=None > > 2016-07-08 11:23:25.560 86007 INFO nova.compute.resource_tracker > [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Compute_service record > updated for OSKVM1:controller > > 2016-07-08 11:24:25.065 86007 INFO nova.compute.resource_tracker > [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Auditing locally > available compute resources for node controller > > 2016-07-08 11:24:25.561 86007 INFO nova.compute.resource_tracker > [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Total usable vcpus: > 40, total allocated vcpus: 0 > > 2016-07-08 11:24:25.562 86007 INFO nova.compute.resource_tracker > [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Final resource view: > name=controller phys_ram=193168MB used_ram=1024MB phys_disk=8168GB > used_disk=1GB total_vcpus=40 used_vcpus=0 pci_stats=None > > 2016-07-08 11:24:25.603 86007 INFO nova.compute.resource_tracker > [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Compute_service record > updated for OSKVM1:controller > > 2016-07-08 11:25:18.138 86007 INFO nova.compute.manager > [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 > 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] Starting instance... > > 2016-07-08 11:25:18.255 86007 INFO nova.compute.claims > [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 > 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] Attempting claim: memory 512 MB, disk > 1 GB > > 2016-07-08 11:25:18.255 86007 INFO nova.compute.claims > [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 > 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] Total memory: 193168 MB, used: > 1024.00 MB > > 2016-07-08 11:25:18.256 86007 INFO nova.compute.claims > [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 > 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] memory limit: 289752.00 MB, free: > 288728.00 MB > > 2016-07-08 11:25:18.256 86007 INFO nova.compute.claims > [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 > 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] Total disk: 8168 GB, used: 1.00 GB > > 2016-07-08 11:25:18.257 86007 INFO nova.compute.claims > [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 > 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] disk limit: 8168.00 GB, free: 8167.00 > GB > > 2016-07-08 11:25:18.271 86007 INFO nova.compute.claims > [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 > 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] Claim successful > > 2016-07-08 11:25:18.747 86007 INFO nova.virt.libvirt.driver > [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 > 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] Creating image > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager > [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 > 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] Instance failed to spawn > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] Traceback (most recent call last): > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2156, in > _build_resources > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] yield resources > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2009, in > _build_and_run_instance > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] > block_device_info=block_device_info) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 2527, > in spawn > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] admin_pass=admin_password) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 2953, > in _create_image > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] instance, size, > fallback_from_host) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 6406, > in _try_fetch_image_cache > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] size=size) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/imagebackend.py", line > 240, in cache > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] *args, **kwargs) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/imagebackend.py", line > 811, in create_image > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] prepare_template(target=base, > max_size=size, *args, **kwargs) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/oslo_concurrency/lockutils.py", line 254, > in inner > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] return f(*args, **kwargs) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/imagebackend.py", line > 230, in fetch_func_sync > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] fetch_func(target=target, *args, > **kwargs) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 2947, > in clone_fallback_to_fetch > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] libvirt_utils.fetch_image(*args, > **kwargs) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/utils.py", line 408, in > fetch_image > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] max_size=max_size) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/virt/images.py", line 123, in > fetch_to_raw > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] max_size=max_size) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/virt/images.py", line 113, in fetch > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] IMAGE_API.download(context, > image_href, dest_path=path) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/image/api.py", line 182, in download > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] dst_path=dest_path) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/image/glance.py", line 367, in > download > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] > _reraise_translated_image_exception(image_id) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/image/glance.py", line 613, in > _reraise_translated_image_exception > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] six.reraise(new_exc, None, > exc_trace) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/image/glance.py", line 365, in > download > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] image_chunks = > self._client.call(context, 1, 'data', image_id) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/nova/image/glance.py", line 231, in call > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] result = getattr(client.images, > method)(*args, **kwargs) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/glanceclient/v1/images.py", line 148, in > data > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] % urlparse.quote(str(image_id))) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/glanceclient/common/http.py", line 280, > in get > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] return self._request('GET', url, > **kwargs) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/glanceclient/common/http.py", line 272, > in _request > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] resp, body_iter = > self._handle_response(resp) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] File > "/usr/lib/python2.7/site-packages/glanceclient/common/http.py", line 93, in > _handle_response > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] raise exc.from_response(resp, > resp.content) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] Invalid: 400 Bad Request: Unknown > scheme 'file' found in URI (HTTP 400) > > 2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] > > 2016-07-08 11:25:19.575 86007 INFO nova.compute.manager > [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 > 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] Terminating instance > > 2016-07-08 11:25:19.583 86007 INFO nova.virt.libvirt.driver [-] [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] During wait destroy, instance > disappeared. > > 2016-07-08 11:25:19.665 86007 INFO nova.virt.libvirt.driver > [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 > 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] Deleting instance files > /var/lib/nova/instances/bf4839c8-2af6-4959-9158-fe411e1cfae7_del > > 2016-07-08 11:25:19.666 86007 INFO nova.virt.libvirt.driver > [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 > 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: > bf4839c8-2af6-4959-9158-fe411e1cfae7] Deletion of > /var/lib/nova/instances/bf4839c8-2af6-4959-9158-fe411e1cfae7_del complete > > 2016-07-08 11:25:26.073 86007 INFO nova.compute.resource_tracker > [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Auditing locally > available compute resources for node controller > > 2016-07-08 11:25:26.477 86007 INFO nova.compute.resource_tracker > [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Total usable vcpus: > 40, total allocated vcpus: 0 > > 2016-07-08 11:25:26.478 86007 INFO nova.compute.resource_tracker > [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Final resource view: > name=controller phys_ram=193168MB used_ram=1024MB phys_disk=8168GB > used_disk=1GB total_vcpus=40 used_vcpus=0 pci_stats=None > > > Regards > Gaurav Goyal > > On Fri, Jul 8, 2016 at 10:17 AM, Kees Meijs <k...@nefos.nl> wrote: > >> Hi, >> >> I'd recommend generating an UUID and use it for all your compute nodes. >> This way, you can keep your configuration in libvirt constant. >> >> Regards, >> Kees >> >> On 08-07-16 16:15, Gaurav Goyal wrote: >> > >> > For below section, should i generate separate UUID for both compte >> hosts? >> > >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com