The RDO glance-store package had a bug in it that miscalculated the chunk
size. I should hope that it's been patched by Redhat now since the fix was
committed upstream before the first Juno rleease, but perhaps not. The
symptom of the bug was horribly slow uploads to glance.

Run this and send back the output:

grep chunk_size
/usr/lib/python2.7/site-packages/glance_store/_drivers/rbd.py

-Erik

On Thu, Apr 2, 2015 at 7:34 AM, Iain Geddes <iain.ged...@cyaninc.com> wrote:

> Oh, apologies, I missed the versions ...
>
> # glance --version   :   0.14.2
> # cinder --version   :   1.1.1
> # ceph -v    :   ceph version 0.87.1
> (283c2e7cfa2457799f534744d7d549f83ea1335e)
>
> From rpm I can confirm that Cinder and Glance are both of the February
> 2014 vintage:
>
> # rpm -qa |grep -e ceph -e glance -e cinder
> ceph-0.87.1-0.el7.x86_64
> libcephfs1-0.87.1-0.el7.x86_64
> ceph-common-0.87.1-0.el7.x86_64
> python-ceph-0.87.1-0.el7.x86_64
> openstack-cinder-2014.2.2-1.el7ost.noarch
> python-cinder-2014.2.2-1.el7ost.noarch
> python-cinderclient-1.1.1-1.el7ost.noarch
> python-glanceclient-0.14.2-2.el7ost.noarch
> python-glance-2014.2.2-1.el7ost.noarch
> python-glance-store-0.1.10-2.el7ost.noarch
> openstack-glance-2014.2.2-1.el7ost.noarch
>
> On Thu, Apr 2, 2015 at 4:24 AM, Iain Geddes <iain.ged...@cyaninc.com>
> wrote:
>
>> Thanks Karan/Quentin/Erik,
>>
>> I admit up front that this is all new to me as my background is optical
>> transport rather than server/storage admin!
>>
>> I'm reassured to know that it should work and this is why I'm completely
>> willing to believe that it's something that I'm doing wrong ... but
>> unfortunately I can't see it based on the RDO Havana/Ceph integration guide
>> or http://ceph.com/docs/master/rbd/rbd-openstack/. Essentially I have
>> extracted everything so that it can be copy/pasted so I am guaranteed
>> consistency - and this has the added advantage that it's easy to compare
>> what was done with what was documented.
>>
>> Just to keep everything clean, I've just restarted the Cinder and Glance
>> processes and do indeed see them establish with the same responses that you
>> showed:
>>
>> *Cinder*
>>
>> 2015-04-02 10:50:54.990 16447 INFO cinder.openstack.common.service [-]
>> Caught SIGTERM, stopping children
>> 2015-04-02 10:50:54.992 16447 INFO cinder.openstack.common.service [-]
>> Waiting on 1 children to exit
>> 2015-04-02 10:52:25.273 17366 INFO cinder.openstack.common.service [-]
>> Starting 1 workers
>> 2015-04-02 10:52:25.274 17366 INFO cinder.openstack.common.service [-]
>> Started child 17373
>> 2015-04-02 10:52:25.275 17373 INFO cinder.service [-] Starting
>> cinder-volume node (version 2014.2.2)
>> 2015-04-02 10:52:25.276 17373 INFO cinder.volume.manager
>> [req-1b0774ff-1bd6-43bb-a271-e6d030aaa5e1 - - - - -] Starting volume driver
>> RBDDriver (1.1.0)
>>
>> *Glance*
>> 2015-04-02 10:58:37.141 18302 DEBUG glance.common.config [-]
>> glance_store.default_store     = rbd log_opt_values
>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>> 2015-04-02 10:58:37.141 18302 DEBUG glance.common.config [-]
>> glance_store.rbd_store_ceph_conf = /etc/ceph/ceph.conf log_opt_values
>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>> 2015-04-02 10:58:37.142 18302 DEBUG glance.common.config [-]
>> glance_store.rbd_store_chunk_size = 8 log_opt_values
>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>> 2015-04-02 10:58:37.142 18302 DEBUG glance.common.config [-]
>> glance_store.rbd_store_pool    = images log_opt_values
>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>> 2015-04-02 10:58:37.142 18302 DEBUG glance.common.config [-]
>> glance_store.rbd_store_user    = glance log_opt_values
>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>> 2015-04-02 10:58:37.143 18302 DEBUG glance.common.config [-]
>> glance_store.stores            = ['rbd'] log_opt_values
>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>
>>
>> Debug of the api really doesn't reveal anything either as far as I can
>> see. Attempting an image-create from the CLI:
>>
>> glance image-create --name "cirros-0.3.3-x86_64" --file
>> cirros-0.3.3-x86_64-disk.raw --disk-format raw --container-format bare
>> --is-public True --progress
>> returns log entries that can be seen in the attached which appears to
>> show that the process has started ... but progress never moves beyond 4%
>> and I haven't seen any further log messages. openstack-status shows all the
>> processes to be up, and Glance images as saving. Given that the top one was
>> through the GUI yesterday I'm guessing it's not going to finish any time
>> soon!
>>
>> == Glance images ==
>>
>> +--------------------------------------+---------------------+-------------+------------------+----------+--------+
>> | ID                                   | Name                | Disk
>> Format | Container Format | Size     | Status |
>>
>> +--------------------------------------+---------------------+-------------+------------------+----------+--------+
>> | f77429b2-17fd-4ef6-97a8-f710862182c6 | Cirros Raw          | raw
>>   | bare             | 41126400 | saving |
>> | 1b12e65a-01cd-4d05-91e8-9e9d86979229 | cirros-0.3.3-x86_64 | raw
>>   | bare             | 41126400 | saving |
>> | fd23c0f3-54b9-4698-b90b-8cdbd6e152c6 | cirros-0.3.3-x86_64 | raw
>>   | bare             | 41126400 | saving |
>> | db297a42-5242-4122-968e-33bf4ad3fe1f | cirros-0.3.3-x86_64 | raw
>>   | bare             | 41126400 | saving |
>>
>> +--------------------------------------+---------------------+-------------+------------------+----------+--------+
>>
>> Was there a particular document that you referenced to perform your
>> install Karan? This should be the easy part ... but I've been saying that
>> about nearly everything for the past month or two!!
>>
>> Kind regards
>>
>>
>> Iain
>>
>>
>>
>> On Thu, Apr 2, 2015 at 3:28 AM, Karan Singh <karan.si...@csc.fi> wrote:
>>
>>> Fortunately Ceph Giant + OpenStack Juno works flawlessly for me.
>>>
>>> If you have configured cinder / glance correctly , then after restarting
>>>  cinder and glance services , you should see something like this in cinder
>>> and glance logs.
>>>
>>>
>>> Cinder logs :
>>>
>>> volume.log:2015-04-02 13:20:43.943 2085 INFO cinder.volume.manager
>>> [req-526cb14e-42ef-4c49-b033-e9bf2096be8f - - - - -] Starting volume driver
>>> RBDDriver (1.1.0)
>>>
>>>
>>> Glance Logs:
>>>
>>> api.log:2015-04-02 13:20:50.448 1266 DEBUG glance.common.config [-]
>>> glance_store.default_store     = rbd log_opt_values
>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-]
>>> glance_store.rbd_store_ceph_conf = /etc/ceph/ceph.conf log_opt_values
>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-]
>>> glance_store.rbd_store_chunk_size = 8 log_opt_values
>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-]
>>> glance_store.rbd_store_pool    = images log_opt_values
>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>> api.log:2015-04-02 13:20:50.449 1266 DEBUG glance.common.config [-]
>>> glance_store.rbd_store_user    = glance log_opt_values
>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>> api.log:2015-04-02 13:20:50.451 1266 DEBUG glance.common.config [-]
>>> glance_store.stores            = ['rbd'] log_opt_values
>>> /usr/lib/python2.7/site-packages/oslo/config/cfg.py:2004
>>>
>>>
>>> If Cinder and Glance are able to initialize RBD driver , then everything
>>> should work like charm.
>>>
>>>
>>> ****************************************************************
>>> Karan Singh
>>> Systems Specialist , Storage Platforms
>>> CSC - IT Center for Science,
>>> Keilaranta 14, P. O. Box 405, FIN-02101 Espoo, Finland
>>> mobile: +358 503 812758
>>> tel. +358 9 4572001
>>> fax +358 9 4572302
>>> http://www.csc.fi/
>>> ****************************************************************
>>>
>>> On 02 Apr 2015, at 03:10, Erik McCormick <emccorm...@cirrusseven.com>
>>> wrote:
>>>
>>> Can you both set Cinder and / or Glance logging to debug and provide
>>> some logs? There was an issue with the first Juno release of Glance in some
>>> vendor packages, so make sure you're fully updated to 2014.2.2
>>> On Apr 1, 2015 7:12 PM, "Quentin Hartman" <qhart...@direwolfdigital.com>
>>> wrote:
>>>
>>>> I am conincidentally going through the same process right now. The best
>>>> reference I've found is this:
>>>> http://ceph.com/docs/master/rbd/rbd-openstack/
>>>>
>>>> When I did Firefly / icehouse, this (seemingly) same guide Just
>>>> Worked(tm), but now with Giant / Juno I'm running into similar trouble  to
>>>> that which you describe. Everything _seems_ right, but creating volumes via
>>>> openstack just sits and spins forever, never creating anything and (as far
>>>> as i've found so far) not logging anything interesting. Normal Rados
>>>> operations work fine.
>>>>
>>>> Feel free to hit me up off list if you want to confer and then we can
>>>> return here if we come up with anything to be shared with the group.
>>>>
>>>> QH
>>>>
>>>> On Wed, Apr 1, 2015 at 3:43 PM, Iain Geddes <iain.ged...@cyaninc.com>
>>>> wrote:
>>>>
>>>>> All,
>>>>>
>>>>> Apologies for my ignorance but I don't seem to be able to search an
>>>>> archive.
>>>>>
>>>>> I've spent a lot of time trying but am having difficulty in
>>>>> integrating Ceph (Giant) into Openstack (Juno). I don't appear to be
>>>>> recording any errors anywhere, but simply don't seem to be writing to the
>>>>> cluster if I try creating a new volume or importing an image. The cluster
>>>>> is good and I can create a static rbd mapping so I know the key components
>>>>> are in place. My problem is almost certainly finger trouble on my part but
>>>>> am completely lost and wondered if there was a well thumbed guide to
>>>>> integration?
>>>>>
>>>>> Thanks
>>>>>
>>>>>
>>>>> Iain
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> ceph-users@lists.ceph.com
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>
>>>>>
>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@lists.ceph.com
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>>
>>
>>
>> --
>> Iain Geddes
>> Customer Support Engineer[image: Cyan] <http://cyaninc.com/>1383 North
>> McDowell Blvd.
>> Petaluma, CA 94954M+353 89 432 6811eiain.ged...@cyaninc.com
>> www.cyaninc.com[image: Facebook] <http://www.facebook.com/CyanInc> [image:
>> LinkedIn]
>> <http://www.linkedin.com/company/cyan-inc?trk=hb_tab_compy_id_2171992> 
>> [image:
>> Twitter] <http://twitter.com/CyanNews>
>>
>
>
>
> --
> Iain Geddes
> Customer Support Engineer[image: Cyan] <http://cyaninc.com/>1383 North
> McDowell Blvd.
> Petaluma, CA 94954M+353 89 432 
> 6811eiain.ged...@cyaninc.comwww.cyaninc.com[image:
> Facebook] <http://www.facebook.com/CyanInc> [image: LinkedIn]
> <http://www.linkedin.com/company/cyan-inc?trk=hb_tab_compy_id_2171992> [image:
> Twitter] <http://twitter.com/CyanNews>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to