Glance should just require something like the following under [default] rbd_store_user=glance rbd_store_pool=images rbd_store_ceph_conf=/etc/ceph/ceph.conf rbd_store_chunk_size=8 default_store=rbd
Also make sure the keyring is in /etc/ceph and you may want to explicitly define the user and keyring in your ceph.conf [client.glance] keyring = /etc/ceph/ceph.client.glance.keyring I also note one strange change in the docs and perhaps it's a matter of where you're getting your packages from, but the doc now says you need python-rbd for Glance when it used to be python-ceph. I have python-ceph for my Juno -> Firefly integration and I don't even see an python-rbd package available in any repo so I make sure you have python-ceph installed. -Erik On Thu, Apr 2, 2015 at 11:54 AM, Quentin Hartman < qhart...@direwolfdigital.com> wrote: > As expected I had a typo in my config for cinder. Correcting that got > cinder working. Everything in glance looks correct according to the above > referenced page, but I'm not seeing any mention of rbd in the logs, and I > notice that the cephx authentication pieces that are present for cinder and > conspicuously absent from the sections on glance. Working on piecing that > together now... > > On Thu, Apr 2, 2015 at 5:34 AM, Iain Geddes <iain.ged...@cyaninc.com> > wrote: > >> Oh, apologies, I missed the versions ... >> >> # glance --version : 0.14.2 >> # cinder --version : 1.1.1 >> # ceph -v : ceph version 0.87.1 >> (283c2e7cfa2457799f534744d7d549f83ea1335e) >> >> From rpm I can confirm that Cinder and Glance are both of the February >> 2014 vintage: >> >> # rpm -qa |grep -e ceph -e glance -e cinder >> ceph-0.87.1-0.el7.x86_64 >> libcephfs1-0.87.1-0.el7.x86_64 >> ceph-common-0.87.1-0.el7.x86_64 >> python-ceph-0.87.1-0.el7.x86_64 >> openstack-cinder-2014.2.2-1.el7ost.noarch >> python-cinder-2014.2.2-1.el7ost.noarch >> python-cinderclient-1.1.1-1.el7ost.noarch >> python-glanceclient-0.14.2-2.el7ost.noarch >> python-glance-2014.2.2-1.el7ost.noarch >> python-glance-store-0.1.10-2.el7ost.noarch >> openstack-glance-2014.2.2-1.el7ost.noarch >> >> On Thu, Apr 2, 2015 at 4:24 AM, Iain Geddes <iain.ged...@cyaninc.com> >> wrote: >> >>> Thanks Karan/Quentin/Erik, >>> >>> I admit up front that this is all new to me as my background is optical >>> transport rather than server/storage admin! >>> >>> I'm reassured to know that it should work and this is why I'm completely >>> willing to believe that it's something that I'm doing wrong ... but >>> unfortunately I can't see it based on the RDO Havana/Ceph integration guide >>> or http://ceph.com/docs/master/rbd/rbd-openstack/. Essentially I have >>> extracted everything so that it can be copy/pasted so I am guaranteed >>> consistency - and this has the added advantage that it's easy to compare >>> what was done with what was documented. >>> >>> Just to keep everything clean, I've just restarted the Cinder and Glance >>> processes and do indeed see them establish with the same responses that you >>> showed: >>> >>> *Cinder* >>> >>> 2015-04-02 10:50:54.990 16447 INFO cinder.openstack.common.service [-] >>> Caught SIGTERM, stopping children >>> 2015-04-02 10:50:54.992 16447 INFO cinder.openstack.common.service [-] >>> Waiting on 1 children to exit >>> 2015-04-02 10:52:25.273 17366 INFO cinder.openstack.common.service [-] >>> Starting 1 workers >>> 2015-04-02 10:52:25.274 17366 INFO cinder.openstack.common.service [-] >>> Started child 17373 >>> 2015-04-02 10:52:25.275 17373 INFO cinder.service [-] Starting >>> cinder-volume node (version 2014.2.2) >>> 2015-04-02 10:52:25.276 17373 INFO cinder.volume.manager >>> [req-1b0774ff-1bd6-43bb-a271-e6d030aaa5e1 - - - - -] Starting volume driver >>> RBDDriver (1.1.0) >>> >>> *Glance* >>> 2015-04-02 10:58:37.141 18302 DEBUG glance.common.config [-] >>> glance_store.default_store = rbd log_opt_values >>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004 >>> 2015-04-02 10:58:37.141 18302 DEBUG glance.common.config [-] >>> glance_store.rbd_store_ceph_conf = /etc/ceph/ceph.conf log_opt_values >>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004 >>> 2015-04-02 10:58:37.142 18302 DEBUG glance.common.config [-] >>> glance_store.rbd_store_chunk_size = 8 log_opt_values >>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004 >>> 2015-04-02 10:58:37.142 18302 DEBUG glance.common.config [-] >>> glance_store.rbd_store_pool = images log_opt_values >>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004 >>> 2015-04-02 10:58:37.142 18302 DEBUG glance.common.config [-] >>> glance_store.rbd_store_user = glance log_opt_values >>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004 >>> 2015-04-02 10:58:37.143 18302 DEBUG glance.common.config [-] >>> glance_store.stores = ['rbd'] log_opt_values >>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004 >>> >>> >>> Debug of the api really doesn't reveal anything either as far as I can >>> see. Attempting an image-create from the CLI: >>> >>> glance image-create --name "cirros-0.3.3-x86_64" --file >>> cirros-0.3.3-x86_64-disk.raw --disk-format raw --container-format bare >>> --is-public True --progress >>> returns log entries that can be seen in the attached which appears to >>> show that the process has started ... but progress never moves beyond 4% >>> and I haven't seen any further log messages. openstack-status shows all the >>> processes to be up, and Glance images as saving. Given that the top one was >>> through the GUI yesterday I'm guessing it's not going to finish any time >>> soon! >>> >>> == Glance images == >>> >>> +--------------------------------------+---------------------+-------------+------------------+----------+--------+ >>> | ID | Name | Disk >>> Format | Container Format | Size | Status | >>> >>> +--------------------------------------+---------------------+-------------+------------------+----------+--------+ >>> | f77429b2-17fd-4ef6-97a8-f710862182c6 | Cirros Raw | raw >>> | bare | 41126400 | saving | >>> | 1b12e65a-01cd-4d05-91e8-9e9d86979229 | cirros-0.3.3-x86_64 | raw >>> | bare | 41126400 | saving | >>> | fd23c0f3-54b9-4698-b90b-8cdbd6e152c6 | cirros-0.3.3-x86_64 | raw >>> | bare | 41126400 | saving | >>> | db297a42-5242-4122-968e-33bf4ad3fe1f | cirros-0.3.3-x86_64 | raw >>> | bare | 41126400 | saving | >>> >>> +--------------------------------------+---------------------+-------------+------------------+----------+--------+ >>> >>> Was there a particular document that you referenced to perform your >>> install Karan? This should be the easy part ... but I've been saying that >>> about nearly everything for the past month or two!! >>> >>> Kind regards >>> >>> >>> Iain >>> >>> >>> >>> On Thu, Apr 2, 2015 at 3:28 AM, Karan Singh <karan.si...@csc.fi> wrote: >>> >>>> Fortunately Ceph Giant + OpenStack Juno works flawlessly for me. >>>> >>>> If you have configured cinder / glance correctly , then after >>>> restarting cinder and glance services , you should see something like this >>>> in cinder and glance logs. >>>> >>>> >>>> Cinder logs : >>>> >>>> volume.log:2015-04-02 13:20:43.943 2085 INFO cinder.volume.manager >>>> [req-526cb14e-42ef-4c49-b033-e9bf2096be8f - - - - -] Starting volume driver >>>> RBDDriver (1.1.0) >>>> >>>> >>>> Glance Logs: >>>> >>>> api.log:2015-04-02 13:20:50.448 1266 DEBUG glance.common.config [-] >>>> glance_store.default_store = rbd log_opt_values >>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004 >>>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-] >>>> glance_store.rbd_store_ceph_conf = /etc/ceph/ceph.conf log_opt_values >>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004 >>>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-] >>>> glance_store.rbd_store_chunk_size = 8 log_opt_values >>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004 >>>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-] >>>> glance_store.rbd_store_pool = images log_opt_values >>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004 >>>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-] >>>> glance_store.rbd_store_user = glance log_opt_values >>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004 >>>> api.log:2015-04-02 13:20:50.451 1266 DEBUG glance.common.config [-] >>>> glance_store.stores = ['rbd'] log_opt_values >>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004 >>>> >>>> >>>> If Cinder and Glance are able to initialize RBD driver , then >>>> everything should work like charm. >>>> >>>> >>>> **************************************************************** >>>> Karan Singh >>>> Systems Specialist , Storage Platforms >>>> CSC - IT Center for Science, >>>> Keilaranta 14, P. O. Box 405, FIN-02101 Espoo, Finland >>>> mobile: +358 503 812758 >>>> tel. +358 9 4572001 >>>> fax +358 9 4572302 >>>> http://www.csc.fi/ >>>> **************************************************************** >>>> >>>> On 02 Apr 2015, at 03:10, Erik McCormick <emccorm...@cirrusseven.com> >>>> wrote: >>>> >>>> Can you both set Cinder and / or Glance logging to debug and provide >>>> some logs? There was an issue with the first Juno release of Glance in some >>>> vendor packages, so make sure you're fully updated to 2014.2.2 >>>> On Apr 1, 2015 7:12 PM, "Quentin Hartman" <qhart...@direwolfdigital.com> >>>> wrote: >>>> >>>>> I am conincidentally going through the same process right now. The >>>>> best reference I've found is this: >>>>> http://ceph.com/docs/master/rbd/rbd-openstack/ >>>>> >>>>> When I did Firefly / icehouse, this (seemingly) same guide Just >>>>> Worked(tm), but now with Giant / Juno I'm running into similar trouble to >>>>> that which you describe. Everything _seems_ right, but creating volumes >>>>> via >>>>> openstack just sits and spins forever, never creating anything and (as far >>>>> as i've found so far) not logging anything interesting. Normal Rados >>>>> operations work fine. >>>>> >>>>> Feel free to hit me up off list if you want to confer and then we can >>>>> return here if we come up with anything to be shared with the group. >>>>> >>>>> QH >>>>> >>>>> On Wed, Apr 1, 2015 at 3:43 PM, Iain Geddes <iain.ged...@cyaninc.com> >>>>> wrote: >>>>> >>>>>> All, >>>>>> >>>>>> Apologies for my ignorance but I don't seem to be able to search an >>>>>> archive. >>>>>> >>>>>> I've spent a lot of time trying but am having difficulty in >>>>>> integrating Ceph (Giant) into Openstack (Juno). I don't appear to be >>>>>> recording any errors anywhere, but simply don't seem to be writing to the >>>>>> cluster if I try creating a new volume or importing an image. The cluster >>>>>> is good and I can create a static rbd mapping so I know the key >>>>>> components >>>>>> are in place. My problem is almost certainly finger trouble on my part >>>>>> but >>>>>> am completely lost and wondered if there was a well thumbed guide to >>>>>> integration? >>>>>> >>>>>> Thanks >>>>>> >>>>>> >>>>>> Iain >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> ceph-users mailing list >>>>>> ceph-users@lists.ceph.com >>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>>> >>>>>> >>>>> >>>>> _______________________________________________ >>>>> ceph-users mailing list >>>>> ceph-users@lists.ceph.com >>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>> >>>>> _______________________________________________ >>>> ceph-users mailing list >>>> ceph-users@lists.ceph.com >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> >>>> >>>> >>> >>> >>> -- >>> Iain Geddes >>> Customer Support Engineer[image: Cyan] <http://cyaninc.com/>1383 North >>> McDowell Blvd. >>> Petaluma, CA 94954M+353 89 432 6811eiain.ged...@cyaninc.com >>> www.cyaninc.com[image: Facebook] <http://www.facebook.com/CyanInc> [image: >>> LinkedIn] >>> <http://www.linkedin.com/company/cyan-inc?trk=hb_tab_compy_id_2171992> >>> [image: >>> Twitter] <http://twitter.com/CyanNews> >>> >> >> >> >> -- >> Iain Geddes >> Customer Support Engineer[image: Cyan] <http://cyaninc.com/>1383 North >> McDowell Blvd. >> Petaluma, CA 94954M+353 89 432 6811eiain.ged...@cyaninc.com >> www.cyaninc.com[image: Facebook] <http://www.facebook.com/CyanInc> [image: >> LinkedIn] >> <http://www.linkedin.com/company/cyan-inc?trk=hb_tab_compy_id_2171992> >> [image: >> Twitter] <http://twitter.com/CyanNews> >> > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com