Thanks Erik,

Maybe this is related as I have:

[DEFAULT]
   verbose = True
   notification_driver = noop
   default_store = rbd
   show_image_direct_url = true
   debug=True

[database]
   connection = mysql://glance:glancepw@ps-sw-ctrl1/glance

[keystone_authtoken]
   auth_uri = http://ps-sw-ctrl1:5000/v2.0
   identity_uri = http://ps-sw-ctrl1:35357
   admin_tenant_name = service
   admin_user = glance
   admin_password = glancepw

[paste_deploy]
   flavor = keystone

[glance_store]
   stores = rbd
   rbd_store_pool = images
   rbd_store_user = glance
   rbd_store_ceph_conf = /etc/ceph/ceph.conf
   rbd_store_chunk_size = 8

[filesystem_store_datadir]
   /var/lib/glance/images/ =

The instructions are that the entries should be found under glance_store
for Juno ....? The keyring is indeed in /etc/ceph but I hadn't specifically
called in out in the ceph.conf file.

Kind regards


Iain

On Thu, Apr 2, 2015 at 5:17 PM, Erik McCormick <emccorm...@cirrusseven.com>
wrote:

> Glance should just require something like the following under [default]
>
> rbd_store_user=glance
> rbd_store_pool=images
> rbd_store_ceph_conf=/etc/ceph/ceph.conf
> rbd_store_chunk_size=8
> default_store=rbd
>
> Also make sure the keyring is in /etc/ceph and you may want to explicitly
> define the user and keyring in your ceph.conf
>
> [client.glance]
> keyring = /etc/ceph/ceph.client.glance.keyring
>
> I also note one strange change in the docs and perhaps it's a matter of
> where you're getting your packages from, but the doc now says you need
> python-rbd for Glance when it used to be python-ceph. I have python-ceph
> for my Juno -> Firefly integration and I don't even see an python-rbd
> package available in any repo so I make sure you have python-ceph installed.
>
> -Erik
>
>
> On Thu, Apr 2, 2015 at 11:54 AM, Quentin Hartman <
> qhart...@direwolfdigital.com> wrote:
>
>> As expected I had a typo in my config for cinder. Correcting that got
>> cinder working. Everything in glance looks correct according to the above
>> referenced page, but I'm not seeing any mention of rbd in the logs, and I
>> notice that the cephx authentication pieces that are present for cinder and
>> conspicuously absent from the sections on glance. Working on piecing that
>> together now...
>>
>> On Thu, Apr 2, 2015 at 5:34 AM, Iain Geddes <iain.ged...@cyaninc.com>
>> wrote:
>>
>>> Oh, apologies, I missed the versions ...
>>>
>>> # glance --version   :   0.14.2
>>> # cinder --version   :   1.1.1
>>> # ceph -v    :   ceph version 0.87.1
>>> (283c2e7cfa2457799f534744d7d549f83ea1335e)
>>>
>>> From rpm I can confirm that Cinder and Glance are both of the February
>>> 2014 vintage:
>>>
>>> # rpm -qa |grep -e ceph -e glance -e cinder
>>> ceph-0.87.1-0.el7.x86_64
>>> libcephfs1-0.87.1-0.el7.x86_64
>>> ceph-common-0.87.1-0.el7.x86_64
>>> python-ceph-0.87.1-0.el7.x86_64
>>> openstack-cinder-2014.2.2-1.el7ost.noarch
>>> python-cinder-2014.2.2-1.el7ost.noarch
>>> python-cinderclient-1.1.1-1.el7ost.noarch
>>> python-glanceclient-0.14.2-2.el7ost.noarch
>>> python-glance-2014.2.2-1.el7ost.noarch
>>> python-glance-store-0.1.10-2.el7ost.noarch
>>> openstack-glance-2014.2.2-1.el7ost.noarch
>>>
>>> On Thu, Apr 2, 2015 at 4:24 AM, Iain Geddes <iain.ged...@cyaninc.com>
>>> wrote:
>>>
>>>> Thanks Karan/Quentin/Erik,
>>>>
>>>> I admit up front that this is all new to me as my background is optical
>>>> transport rather than server/storage admin!
>>>>
>>>> I'm reassured to know that it should work and this is why I'm
>>>> completely willing to believe that it's something that I'm doing wrong ...
>>>> but unfortunately I can't see it based on the RDO Havana/Ceph integration
>>>> guide or http://ceph.com/docs/master/rbd/rbd-openstack/. Essentially I
>>>> have extracted everything so that it can be copy/pasted so I am guaranteed
>>>> consistency - and this has the added advantage that it's easy to compare
>>>> what was done with what was documented.
>>>>
>>>> Just to keep everything clean, I've just restarted the Cinder and
>>>> Glance processes and do indeed see them establish with the same responses
>>>> that you showed:
>>>>
>>>> *Cinder*
>>>>
>>>> 2015-04-02 10:50:54.990 16447 INFO cinder.openstack.common.service [-]
>>>> Caught SIGTERM, stopping children
>>>> 2015-04-02 10:50:54.992 16447 INFO cinder.openstack.common.service [-]
>>>> Waiting on 1 children to exit
>>>> 2015-04-02 10:52:25.273 17366 INFO cinder.openstack.common.service [-]
>>>> Starting 1 workers
>>>> 2015-04-02 10:52:25.274 17366 INFO cinder.openstack.common.service [-]
>>>> Started child 17373
>>>> 2015-04-02 10:52:25.275 17373 INFO cinder.service [-] Starting
>>>> cinder-volume node (version 2014.2.2)
>>>> 2015-04-02 10:52:25.276 17373 INFO cinder.volume.manager
>>>> [req-1b0774ff-1bd6-43bb-a271-e6d030aaa5e1 - - - - -] Starting volume driver
>>>> RBDDriver (1.1.0)
>>>>
>>>> *Glance*
>>>> 2015-04-02 10:58:37.141 18302 DEBUG glance.common.config [-]
>>>> glance_store.default_store     = rbd log_opt_values
>>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>>> 2015-04-02 10:58:37.141 18302 DEBUG glance.common.config [-]
>>>> glance_store.rbd_store_ceph_conf = /etc/ceph/ceph.conf log_opt_values
>>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>>> 2015-04-02 10:58:37.142 18302 DEBUG glance.common.config [-]
>>>> glance_store.rbd_store_chunk_size = 8 log_opt_values
>>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>>> 2015-04-02 10:58:37.142 18302 DEBUG glance.common.config [-]
>>>> glance_store.rbd_store_pool    = images log_opt_values
>>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>>> 2015-04-02 10:58:37.142 18302 DEBUG glance.common.config [-]
>>>> glance_store.rbd_store_user    = glance log_opt_values
>>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>>> 2015-04-02 10:58:37.143 18302 DEBUG glance.common.config [-]
>>>> glance_store.stores            = ['rbd'] log_opt_values
>>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>>>
>>>>
>>>> Debug of the api really doesn't reveal anything either as far as I can
>>>> see. Attempting an image-create from the CLI:
>>>>
>>>> glance image-create --name "cirros-0.3.3-x86_64" --file
>>>> cirros-0.3.3-x86_64-disk.raw --disk-format raw --container-format bare
>>>> --is-public True --progress
>>>> returns log entries that can be seen in the attached which appears to
>>>> show that the process has started ... but progress never moves beyond 4%
>>>> and I haven't seen any further log messages. openstack-status shows all the
>>>> processes to be up, and Glance images as saving. Given that the top one was
>>>> through the GUI yesterday I'm guessing it's not going to finish any time
>>>> soon!
>>>>
>>>> == Glance images ==
>>>>
>>>> +--------------------------------------+---------------------+-------------+------------------+----------+--------+
>>>> | ID                                   | Name                | Disk
>>>> Format | Container Format | Size     | Status |
>>>>
>>>> +--------------------------------------+---------------------+-------------+------------------+----------+--------+
>>>> | f77429b2-17fd-4ef6-97a8-f710862182c6 | Cirros Raw          | raw
>>>>     | bare             | 41126400 | saving |
>>>> | 1b12e65a-01cd-4d05-91e8-9e9d86979229 | cirros-0.3.3-x86_64 | raw
>>>>     | bare             | 41126400 | saving |
>>>> | fd23c0f3-54b9-4698-b90b-8cdbd6e152c6 | cirros-0.3.3-x86_64 | raw
>>>>     | bare             | 41126400 | saving |
>>>> | db297a42-5242-4122-968e-33bf4ad3fe1f | cirros-0.3.3-x86_64 | raw
>>>>     | bare             | 41126400 | saving |
>>>>
>>>> +--------------------------------------+---------------------+-------------+------------------+----------+--------+
>>>>
>>>> Was there a particular document that you referenced to perform your
>>>> install Karan? This should be the easy part ... but I've been saying that
>>>> about nearly everything for the past month or two!!
>>>>
>>>> Kind regards
>>>>
>>>>
>>>> Iain
>>>>
>>>>
>>>>
>>>> On Thu, Apr 2, 2015 at 3:28 AM, Karan Singh <karan.si...@csc.fi> wrote:
>>>>
>>>>> Fortunately Ceph Giant + OpenStack Juno works flawlessly for me.
>>>>>
>>>>> If you have configured cinder / glance correctly , then after
>>>>> restarting  cinder and glance services , you should see something like 
>>>>> this
>>>>> in cinder and glance logs.
>>>>>
>>>>>
>>>>> Cinder logs :
>>>>>
>>>>> volume.log:2015-04-02 13:20:43.943 2085 INFO cinder.volume.manager
>>>>> [req-526cb14e-42ef-4c49-b033-e9bf2096be8f - - - - -] Starting volume 
>>>>> driver
>>>>> RBDDriver (1.1.0)
>>>>>
>>>>>
>>>>> Glance Logs:
>>>>>
>>>>> api.log:2015-04-02 13:20:50.448 1266 DEBUG glance.common.config [-]
>>>>> glance_store.default_store     = rbd log_opt_values
>>>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>>>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-]
>>>>> glance_store.rbd_store_ceph_conf = /etc/ceph/ceph.conf log_opt_values
>>>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>>>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-]
>>>>> glance_store.rbd_store_chunk_size = 8 log_opt_values
>>>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>>>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-]
>>>>> glance_store.rbd_store_pool    = images log_opt_values
>>>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>>>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-]
>>>>> glance_store.rbd_store_user    = glance log_opt_values
>>>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>>>> api.log:2015-04-02 13:20:50.451 1266 DEBUG glance.common.config [-]
>>>>> glance_store.stores            = ['rbd'] log_opt_values
>>>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>>>>
>>>>>
>>>>> If Cinder and Glance are able to initialize RBD driver , then
>>>>> everything should work like charm.
>>>>>
>>>>>
>>>>> ****************************************************************
>>>>> Karan Singh
>>>>> Systems Specialist , Storage Platforms
>>>>> CSC - IT Center for Science,
>>>>> Keilaranta 14, P. O. Box 405, FIN-02101 Espoo, Finland
>>>>> mobile: +358 503 812758
>>>>> tel. +358 9 4572001
>>>>> fax +358 9 4572302
>>>>> http://www.csc.fi/
>>>>> ****************************************************************
>>>>>
>>>>> On 02 Apr 2015, at 03:10, Erik McCormick <emccorm...@cirrusseven.com>
>>>>> wrote:
>>>>>
>>>>> Can you both set Cinder and / or Glance logging to debug and provide
>>>>> some logs? There was an issue with the first Juno release of Glance in 
>>>>> some
>>>>> vendor packages, so make sure you're fully updated to 2014.2.2
>>>>> On Apr 1, 2015 7:12 PM, "Quentin Hartman" <
>>>>> qhart...@direwolfdigital.com> wrote:
>>>>>
>>>>>> I am conincidentally going through the same process right now. The
>>>>>> best reference I've found is this:
>>>>>> http://ceph.com/docs/master/rbd/rbd-openstack/
>>>>>>
>>>>>> When I did Firefly / icehouse, this (seemingly) same guide Just
>>>>>> Worked(tm), but now with Giant / Juno I'm running into similar trouble  
>>>>>> to
>>>>>> that which you describe. Everything _seems_ right, but creating volumes 
>>>>>> via
>>>>>> openstack just sits and spins forever, never creating anything and (as 
>>>>>> far
>>>>>> as i've found so far) not logging anything interesting. Normal Rados
>>>>>> operations work fine.
>>>>>>
>>>>>> Feel free to hit me up off list if you want to confer and then we can
>>>>>> return here if we come up with anything to be shared with the group.
>>>>>>
>>>>>> QH
>>>>>>
>>>>>> On Wed, Apr 1, 2015 at 3:43 PM, Iain Geddes <iain.ged...@cyaninc.com>
>>>>>> wrote:
>>>>>>
>>>>>>> All,
>>>>>>>
>>>>>>> Apologies for my ignorance but I don't seem to be able to search an
>>>>>>> archive.
>>>>>>>
>>>>>>> I've spent a lot of time trying but am having difficulty in
>>>>>>> integrating Ceph (Giant) into Openstack (Juno). I don't appear to be
>>>>>>> recording any errors anywhere, but simply don't seem to be writing to 
>>>>>>> the
>>>>>>> cluster if I try creating a new volume or importing an image. The 
>>>>>>> cluster
>>>>>>> is good and I can create a static rbd mapping so I know the key 
>>>>>>> components
>>>>>>> are in place. My problem is almost certainly finger trouble on my part 
>>>>>>> but
>>>>>>> am completely lost and wondered if there was a well thumbed guide to
>>>>>>> integration?
>>>>>>>
>>>>>>> Thanks
>>>>>>>
>>>>>>>
>>>>>>> Iain
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> ceph-users mailing list
>>>>>>> ceph-users@lists.ceph.com
>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>>
>>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> ceph-users mailing list
>>>>>> ceph-users@lists.ceph.com
>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>
>>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> ceph-users@lists.ceph.com
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> Iain Geddes
>>>> Customer Support Engineer[image: Cyan] <http://cyaninc.com/>1383 North
>>>> McDowell Blvd.
>>>> Petaluma, CA 94954M+353 89 432 6811eiain.ged...@cyaninc.com
>>>> www.cyaninc.com[image: Facebook] <http://www.facebook.com/CyanInc> [image:
>>>> LinkedIn]
>>>> <http://www.linkedin.com/company/cyan-inc?trk=hb_tab_compy_id_2171992> 
>>>> [image:
>>>> Twitter] <http://twitter.com/CyanNews>
>>>>
>>>
>>>
>>>
>>> --
>>> Iain Geddes
>>> Customer Support Engineer[image: Cyan] <http://cyaninc.com/>1383 North
>>> McDowell Blvd.
>>> Petaluma, CA 94954M+353 89 432 6811eiain.ged...@cyaninc.com
>>> www.cyaninc.com[image: Facebook] <http://www.facebook.com/CyanInc> [image:
>>> LinkedIn]
>>> <http://www.linkedin.com/company/cyan-inc?trk=hb_tab_compy_id_2171992> 
>>> [image:
>>> Twitter] <http://twitter.com/CyanNews>
>>>
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>


-- 
Iain Geddes
Customer Support Engineer[image: Cyan] <http://cyaninc.com/>1383 North
McDowell Blvd.
Petaluma, CA 94954M+353 89 432
6811eiain.ged...@cyaninc.comwww.cyaninc.com[image:
Facebook] <http://www.facebook.com/CyanInc> [image: LinkedIn]
<http://www.linkedin.com/company/cyan-inc?trk=hb_tab_compy_id_2171992> [image:
Twitter] <http://twitter.com/CyanNews>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to