Oh, apologies, I missed the versions ...

# glance --version   :   0.14.2
# cinder --version   :   1.1.1
# ceph -v    :   ceph version 0.87.1
(283c2e7cfa2457799f534744d7d549f83ea1335e)

>From rpm I can confirm that Cinder and Glance are both of the February 2014
vintage:

# rpm -qa |grep -e ceph -e glance -e cinder
ceph-0.87.1-0.el7.x86_64
libcephfs1-0.87.1-0.el7.x86_64
ceph-common-0.87.1-0.el7.x86_64
python-ceph-0.87.1-0.el7.x86_64
openstack-cinder-2014.2.2-1.el7ost.noarch
python-cinder-2014.2.2-1.el7ost.noarch
python-cinderclient-1.1.1-1.el7ost.noarch
python-glanceclient-0.14.2-2.el7ost.noarch
python-glance-2014.2.2-1.el7ost.noarch
python-glance-store-0.1.10-2.el7ost.noarch
openstack-glance-2014.2.2-1.el7ost.noarch

On Thu, Apr 2, 2015 at 4:24 AM, Iain Geddes <iain.ged...@cyaninc.com> wrote:

> Thanks Karan/Quentin/Erik,
>
> I admit up front that this is all new to me as my background is optical
> transport rather than server/storage admin!
>
> I'm reassured to know that it should work and this is why I'm completely
> willing to believe that it's something that I'm doing wrong ... but
> unfortunately I can't see it based on the RDO Havana/Ceph integration guide
> or http://ceph.com/docs/master/rbd/rbd-openstack/. Essentially I have
> extracted everything so that it can be copy/pasted so I am guaranteed
> consistency - and this has the added advantage that it's easy to compare
> what was done with what was documented.
>
> Just to keep everything clean, I've just restarted the Cinder and Glance
> processes and do indeed see them establish with the same responses that you
> showed:
>
> *Cinder*
>
> 2015-04-02 10:50:54.990 16447 INFO cinder.openstack.common.service [-]
> Caught SIGTERM, stopping children
> 2015-04-02 10:50:54.992 16447 INFO cinder.openstack.common.service [-]
> Waiting on 1 children to exit
> 2015-04-02 10:52:25.273 17366 INFO cinder.openstack.common.service [-]
> Starting 1 workers
> 2015-04-02 10:52:25.274 17366 INFO cinder.openstack.common.service [-]
> Started child 17373
> 2015-04-02 10:52:25.275 17373 INFO cinder.service [-] Starting
> cinder-volume node (version 2014.2.2)
> 2015-04-02 10:52:25.276 17373 INFO cinder.volume.manager
> [req-1b0774ff-1bd6-43bb-a271-e6d030aaa5e1 - - - - -] Starting volume driver
> RBDDriver (1.1.0)
>
> *Glance*
> 2015-04-02 10:58:37.141 18302 DEBUG glance.common.config [-]
> glance_store.default_store     = rbd log_opt_values
> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
> 2015-04-02 10:58:37.141 18302 DEBUG glance.common.config [-]
> glance_store.rbd_store_ceph_conf = /etc/ceph/ceph.conf log_opt_values
> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
> 2015-04-02 10:58:37.142 18302 DEBUG glance.common.config [-]
> glance_store.rbd_store_chunk_size = 8 log_opt_values
> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
> 2015-04-02 10:58:37.142 18302 DEBUG glance.common.config [-]
> glance_store.rbd_store_pool    = images log_opt_values
> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
> 2015-04-02 10:58:37.142 18302 DEBUG glance.common.config [-]
> glance_store.rbd_store_user    = glance log_opt_values
> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
> 2015-04-02 10:58:37.143 18302 DEBUG glance.common.config [-]
> glance_store.stores            = ['rbd'] log_opt_values
> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>
>
> Debug of the api really doesn't reveal anything either as far as I can
> see. Attempting an image-create from the CLI:
>
> glance image-create --name "cirros-0.3.3-x86_64" --file
> cirros-0.3.3-x86_64-disk.raw --disk-format raw --container-format bare
> --is-public True --progress
> returns log entries that can be seen in the attached which appears to show
> that the process has started ... but progress never moves beyond 4% and I
> haven't seen any further log messages. openstack-status shows all the
> processes to be up, and Glance images as saving. Given that the top one was
> through the GUI yesterday I'm guessing it's not going to finish any time
> soon!
>
> == Glance images ==
>
> +--------------------------------------+---------------------+-------------+------------------+----------+--------+
> | ID                                   | Name                | Disk Format
> | Container Format | Size     | Status |
>
> +--------------------------------------+---------------------+-------------+------------------+----------+--------+
> | f77429b2-17fd-4ef6-97a8-f710862182c6 | Cirros Raw          | raw
> | bare             | 41126400 | saving |
> | 1b12e65a-01cd-4d05-91e8-9e9d86979229 | cirros-0.3.3-x86_64 | raw
> | bare             | 41126400 | saving |
> | fd23c0f3-54b9-4698-b90b-8cdbd6e152c6 | cirros-0.3.3-x86_64 | raw
> | bare             | 41126400 | saving |
> | db297a42-5242-4122-968e-33bf4ad3fe1f | cirros-0.3.3-x86_64 | raw
> | bare             | 41126400 | saving |
>
> +--------------------------------------+---------------------+-------------+------------------+----------+--------+
>
> Was there a particular document that you referenced to perform your
> install Karan? This should be the easy part ... but I've been saying that
> about nearly everything for the past month or two!!
>
> Kind regards
>
>
> Iain
>
>
>
> On Thu, Apr 2, 2015 at 3:28 AM, Karan Singh <karan.si...@csc.fi> wrote:
>
>> Fortunately Ceph Giant + OpenStack Juno works flawlessly for me.
>>
>> If you have configured cinder / glance correctly , then after restarting
>>  cinder and glance services , you should see something like this in cinder
>> and glance logs.
>>
>>
>> Cinder logs :
>>
>> volume.log:2015-04-02 13:20:43.943 2085 INFO cinder.volume.manager
>> [req-526cb14e-42ef-4c49-b033-e9bf2096be8f - - - - -] Starting volume driver
>> RBDDriver (1.1.0)
>>
>>
>> Glance Logs:
>>
>> api.log:2015-04-02 13:20:50.448 1266 DEBUG glance.common.config [-]
>> glance_store.default_store     = rbd log_opt_values
>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-]
>> glance_store.rbd_store_ceph_conf = /etc/ceph/ceph.conf log_opt_values
>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-]
>> glance_store.rbd_store_chunk_size = 8 log_opt_values
>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-]
>> glance_store.rbd_store_pool    = images log_opt_values
>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-]
>> glance_store.rbd_store_user    = glance log_opt_values
>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>> api.log:2015-04-02 13:20:50.451 1266 DEBUG glance.common.config [-]
>> glance_store.stores            = ['rbd'] log_opt_values
>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>
>>
>> If Cinder and Glance are able to initialize RBD driver , then everything
>> should work like charm.
>>
>>
>> ****************************************************************
>> Karan Singh
>> Systems Specialist , Storage Platforms
>> CSC - IT Center for Science,
>> Keilaranta 14, P. O. Box 405, FIN-02101 Espoo, Finland
>> mobile: +358 503 812758
>> tel. +358 9 4572001
>> fax +358 9 4572302
>> http://www.csc.fi/
>> ****************************************************************
>>
>> On 02 Apr 2015, at 03:10, Erik McCormick <emccorm...@cirrusseven.com>
>> wrote:
>>
>> Can you both set Cinder and / or Glance logging to debug and provide some
>> logs? There was an issue with the first Juno release of Glance in some
>> vendor packages, so make sure you're fully updated to 2014.2.2
>> On Apr 1, 2015 7:12 PM, "Quentin Hartman" <qhart...@direwolfdigital.com>
>> wrote:
>>
>>> I am conincidentally going through the same process right now. The best
>>> reference I've found is this:
>>> http://ceph.com/docs/master/rbd/rbd-openstack/
>>>
>>> When I did Firefly / icehouse, this (seemingly) same guide Just
>>> Worked(tm), but now with Giant / Juno I'm running into similar trouble  to
>>> that which you describe. Everything _seems_ right, but creating volumes via
>>> openstack just sits and spins forever, never creating anything and (as far
>>> as i've found so far) not logging anything interesting. Normal Rados
>>> operations work fine.
>>>
>>> Feel free to hit me up off list if you want to confer and then we can
>>> return here if we come up with anything to be shared with the group.
>>>
>>> QH
>>>
>>> On Wed, Apr 1, 2015 at 3:43 PM, Iain Geddes <iain.ged...@cyaninc.com>
>>> wrote:
>>>
>>>> All,
>>>>
>>>> Apologies for my ignorance but I don't seem to be able to search an
>>>> archive.
>>>>
>>>> I've spent a lot of time trying but am having difficulty in integrating
>>>> Ceph (Giant) into Openstack (Juno). I don't appear to be recording any
>>>> errors anywhere, but simply don't seem to be writing to the cluster if I
>>>> try creating a new volume or importing an image. The cluster is good and I
>>>> can create a static rbd mapping so I know the key components are in place.
>>>> My problem is almost certainly finger trouble on my part but am completely
>>>> lost and wondered if there was a well thumbed guide to integration?
>>>>
>>>> Thanks
>>>>
>>>>
>>>> Iain
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@lists.ceph.com
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>>
>
>
> --
> Iain Geddes
> Customer Support Engineer[image: Cyan] <http://cyaninc.com/>1383 North
> McDowell Blvd.
> Petaluma, CA 94954M+353 89 432 
> 6811eiain.ged...@cyaninc.comwww.cyaninc.com[image:
> Facebook] <http://www.facebook.com/CyanInc> [image: LinkedIn]
> <http://www.linkedin.com/company/cyan-inc?trk=hb_tab_compy_id_2171992> [image:
> Twitter] <http://twitter.com/CyanNews>
>



-- 
Iain Geddes
Customer Support Engineer[image: Cyan] <http://cyaninc.com/>1383 North
McDowell Blvd.
Petaluma, CA 94954M+353 89 432
6811eiain.ged...@cyaninc.comwww.cyaninc.com[image:
Facebook] <http://www.facebook.com/CyanInc> [image: LinkedIn]
<http://www.linkedin.com/company/cyan-inc?trk=hb_tab_compy_id_2171992> [image:
Twitter] <http://twitter.com/CyanNews>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to