Hello, Are you configured these two paremeters in cinder.conf?
rbd_user rbd_secret_uuid Regards. 2016-07-07 15:39 GMT+02:00 Gaurav Goyal <er.gauravgo...@gmail.com>: > Hello Mr. Kees, > > Thanks for your response! > > My setup is > > Openstack Node 1 -> controller + network + compute1 (Liberty Version) > Openstack node 2 --> Compute2 > > Ceph version Hammer > > I am using dell storage with following status > > DELL SAN storage is attached to both hosts as > > [root@OSKVM1 ~]# iscsiadm -m node > > 10.35.0.3:3260,1 > iqn.2001-05.com.equallogic:0-1cb196-07a83c107-47700000018575af-vol1 > > 10.35.0.8:3260,1 > iqn.2001-05.com.equallogic:0-1cb196-07a83c107-47700000018575af-vol1 > > 10.35.0.*:3260,-1 > iqn.2001-05.com.equallogic:0-1cb196-20d83c107-7290000002157606-vol2 > > 10.35.0.8:3260,1 > iqn.2001-05.com.equallogic:0-1cb196-20d83c107-7290000002157606-vol2 > > 10.35.0.*:3260,-1 > iqn.2001-05.com.equallogic:0-1cb196-f0783c107-70a000000245761a-vol3 > > 10.35.0.8:3260,1 > iqn.2001-05.com.equallogic:0-1cb196-f0783c107-70a000000245761a-vol3 > > 10.35.0.*:3260,-1 > iqn.2001-05.com.equallogic:0-1cb196-fda83c107-927000000275761a-vol4 > 10.35.0.8:3260,1 > iqn.2001-05.com.equallogic:0-1cb196-fda83c107-927000000275761a-vol4 > > > Since in my setup same LUNs are MAPPED to both hosts > > i choose 2 LUNS on Openstack Node 1 and 2 on Openstack Node 2 > > > *Node1 has * > > /dev/sdc1 2.0T 3.1G 2.0T 1% /var/lib/ceph/osd/ceph-0 > > /dev/sdd1 2.0T 3.8G 2.0T 1% /var/lib/ceph/osd/ceph-1 > > *Node 2 has * > > /dev/sdd1 2.0T 3.4G 2.0T 1% /var/lib/ceph/osd/ceph-2 > > /dev/sde1 2.0T 3.5G 2.0T 1% /var/lib/ceph/osd/ceph-3 > > [root@OSKVM1 ~]# ceph status > > cluster 9f923089-a6c0-4169-ace8-ad8cc4cca116 > > health HEALTH_WARN > > mon.OSKVM1 low disk space > > monmap e1: 1 mons at {OSKVM1=10.24.0.4:6789/0} > > election epoch 1, quorum 0 OSKVM1 > > osdmap e40: 4 osds: 4 up, 4 in > > pgmap v1154: 576 pgs, 5 pools, 6849 MB data, 860 objects > > 13857 MB used, 8154 GB / 8168 GB avail > > 576 active+clean > > *Can you please help me to know if it is correct configuration as per my > setup?* > > After this setup, i am trying to configure Cinder and Glance to use RBD > for a backend. > Glance image is already stored in RBD. > Following this link http://docs.ceph.com/docs/master/rbd/rbd-openstack/ > > I have managed to install glance image in rbd. But i am finding some issue > in cinder configuration. Can you please help me on this? > As per link, i need to configure these parameters under [ceph] but i do > not have different section for [ceph]. infact i could find all these > parameters under [DEFAULT]. Is it ok to configure them under [DEFAULT]. > CONFIGURING CINDER > <http://docs.ceph.com/docs/master/rbd/rbd-openstack/#configuring-cinder> > > OpenStack requires a driver to interact with Ceph block devices. You must > also specify the pool name for the block device. On your OpenStack node, > edit/etc/cinder/cinder.conf by adding: > > [DEFAULT] > ... > enabled_backends = ceph > ... > [ceph] > volume_driver = cinder.volume.drivers.rbd.RBDDriver > rbd_pool = volumes > rbd_ceph_conf = /etc/ceph/ceph.conf > rbd_flatten_volume_from_snapshot = false > rbd_max_clone_depth = 5 > rbd_store_chunk_size = 4 > rados_connect_timeout = -1 > glance_api_version = 2 > > I find following error in cinder service status > > systemctl status openstack-cinder-volume.service > > Jul 07 09:37:01 OSKVM1 cinder-volume[136247]: 2016-07-07 09:37:01.058 > 136259 ERROR cinder.service [-] Manager for service cinder-volume > OSKVM1@ceph is reporting problems, not sending heartbeat. Service will > appear "down". > > Jul 07 09:37:02 OSKVM1 cinder-volume[136247]: 2016-07-07 09:37:02.040 > 136259 WARNING cinder.volume.manager > [req-561ddd3c-9560-4374-a958-7a2c103af7ee - - - - -] Update driver status > failed: (config name ceph) is uninitialized. > > Jul 07 09:37:11 OSKVM1 cinder-volume[136247]: 2016-07-07 09:37:11.059 > 136259 ERROR cinder.service [-] Manager for service cinder-volume > OSKVM1@ceph is reporting problems, not sending heartbeat. Service will > appear "down". > > > > [root@OSKVM2 ~]# rbd -p images ls > > a8b45c8a-a5c8-49d8-a529-1e4088bdbf3f > > [root@OSKVM2 ~]# rados df > > pool name KB objects clones degraded > unfound rd rd KB wr wr KB > > backups 0 0 0 0 > 0 0 0 0 0 > > images 7013377 860 0 0 > 0 9486 7758 2580 7013377 > > rbd 0 0 0 0 > 0 0 0 0 0 > > vms 0 0 0 0 > 0 0 0 0 0 > > volumes 0 0 0 0 > 0 0 0 0 0 > > total used 14190236 860 > > total avail 8550637828 > > total space 8564828064 > > > > > [root@OSKVM2 ~]# ceph auth list > > installed auth entries: > > > mds.OSKVM1 > > key: AQCK6XtXNBFdDBAAXmX73gBqK3lyakSxxP+XjA== > > caps: [mds] allow > > caps: [mon] allow profile mds > > caps: [osd] allow rwx > > osd.0 > > key: AQAB4HtX7q27KBAAEqcuJXwXAJyD6a1Qu/MXqA== > > caps: [mon] allow profile osd > > caps: [osd] allow * > > osd.1 > > key: AQC/4ntXFJGdFBAAADYH03iQTF4jWI1LnBZeJg== > > caps: [mon] allow profile osd > > caps: [osd] allow * > > osd.2 > > key: AQCa43tXr12fDhAAzbq6FO2+8m9qg1B12/99Og== > > caps: [mon] allow profile osd > > caps: [osd] allow * > > osd.3 > > key: AQA/5HtXDNfcLxAAJWawgxc1nd8CB+4uH/8fdQ== > > caps: [mon] allow profile osd > > caps: [osd] allow * > > client.admin > > key: AQBNknJXE/I2FRAA+caW02eje7GZ/uv1O6aUgA== > > caps: [mds] allow > > caps: [mon] allow * > > caps: [osd] allow * > > client.bootstrap-mds > > key: AQBOknJXjLloExAAGjMRfjp5okI1honz9Nx4wg== > > caps: [mon] allow profile bootstrap-mds > > client.bootstrap-osd > > key: AQBNknJXDUMFKBAAZ8/TfDkS0N7Q6CbaOG3DyQ== > > caps: [mon] allow profile bootstrap-osd > > client.bootstrap-rgw > > key: AQBOknJXQAUiABAA6IB4p4RyUmrsxXk+pv4u7g== > > caps: [mon] allow profile bootstrap-rgw > > client.cinder > > key: AQCIAHxX9ga8LxAAU+S3Vybdu+Cm2bP3lplGnA== > > caps: [mon] allow r > > caps: [osd] allow class-read object_prefix rbd_children, allow rwx > pool=volumes, allow rwx pool=vms, allow rx pool=images > > client.cinder-backup > > key: AQCXAHxXAVSNKhAAV1d/ZRMsrriDOt+7pYgJIg== > > caps: [mon] allow r > > caps: [osd] allow class-read object_prefix rbd_children, allow rwx > pool=backups > > client.glance > > key: AQCVAHxXupPdLBAA7hh1TJZnvSmFSDWbQiaiEQ== > > caps: [mon] allow r > > caps: [osd] allow class-read object_prefix rbd_children, allow rwx > pool=images > > > Regards > > Gaurav Goyal > > On Thu, Jul 7, 2016 at 2:54 AM, Kees Meijs <k...@nefos.nl> wrote: > >> Hi Gaurav, >> >> Unfortunately I'm not completely sure about your setup, but I guess it >> makes sense to configure Cinder and Glance to use RBD for a backend. It >> seems to me, you're trying to store VM images directly on an OSD >> filesystem. >> >> Please refer to http://docs.ceph.com/docs/master/rbd/rbd-openstack/ for >> details. >> >> Regards, >> Kees >> >> On 06-07-16 23:03, Gaurav Goyal wrote: >> > >> > I am installing ceph hammer and integrating it with openstack Liberty >> > for the first time. >> > >> > My local disk has only 500 GB but i need to create 600 GB VM. SO i >> > have created a soft link to ceph filesystem as >> > >> > lrwxrwxrwx 1 root root 34 Jul 6 13:02 instances -> >> > /var/lib/ceph/osd/ceph-0/instances [root@OSKVM1 nova]# pwd >> > /var/lib/nova [root@OSKVM1 nova]# >> > >> > now when i am trying to create an instance it is giving the following >> > error as checked from nova-compute.log >> > I need your help to fix this issue. >> > >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com