VM live migration is supported and should work Can you add engine and cinderlib logs?
On Tue, Jul 9, 2019 at 11:01 AM Dan Poltawski <[email protected]> wrote: > On Tue, 2019-07-09 at 08:00 +0100, Dan Poltawski wrote: > > I've now managed to succesfully create/mount/delete volumes! > > However, I'm seeing live migrations stay stuck. Is this supported? > > (gdb) py-list > 345 client.conf_set('rados_osd_op_timeout', > timeout) > 346 client.conf_set('rados_mon_op_timeout', > timeout) > 347 client.conf_set('client_mount_timeout', > timeout) > 348 > 349 client.connect() > >350 ioctx = client.open_ioctx(pool) > 351 return client, ioctx > 352 except self.rados.Error: > 353 msg = _("Error connecting to ceph cluster.") > 354 LOG.exception(msg) > 355 client.shutdown() > > > (gdb) py-bt > #15 Frame 0x3ea0e50, for file /usr/lib/python2.7/site- > packages/cinder/volume/drivers/rbd.py, line 350, in _do_conn > (pool='storage-ssd', remote=None, timeout=-1, name='ceph', > conf='/etc/ceph/ceph.conf', user='ovirt', client=<rados.Rados at remote > 0x7fb1f4f83a60>) > ioctx = client.open_ioctx(pool) > #20 Frame 0x3ea4620, for file /usr/lib/python2.7/site- > packages/retrying.py, line 217, in call > (self=<Retrying(_retry_on_exception=<function at remote > 0x7fb1f4f23488>, _wait_exponential_max=1073741823, > _wait_incrementing_start=0, stop=<function at remote 0x7fb1f4f23578>, > _stop_max_attempt_number=5, _wait_incrementing_increment=100, > _wait_random_max=1000, _retry_on_result=<instancemethod at remote > 0x7fb1f51da550>, _stop_max_delay=100, _wait_fixed=1000, > _wrap_exception=False, _wait_random_min=0, > _wait_exponential_multiplier=1, wait=<function at remote > 0x7fb1f4f23500>) at remote 0x7fb1f4f1ae90>, fn=<function at remote > 0x7fb1f4f23668>, args=(None, None, None), kwargs={}, > start_time=1562658179214, attempt_number=1) > attempt = Attempt(fn(*args, **kwargs), attempt_number, False) > #25 Frame 0x3e49d50, for file /usr/lib/python2.7/site- > packages/cinder/utils.py, line 818, in _wrapper (args=(None, None, > None), kwargs={}, r=<Retrying(_retry_on_exception=<function at remote > 0x7fb1f4f23488>, _wait_exponential_max=1073741823, > _wait_incrementing_start=0, stop=<function at remote 0x7fb1f4f23578>, > _stop_max_attempt_number=5, _wait_incrementing_increment=100, > _wait_random_max=1000, _retry_on_result=<instancemethod at remote > 0x7fb1f51da550>, _stop_max_delay=100, _wait_fixed=1000, > _wrap_exception=False, _wait_random_min=0, > _wait_exponential_multiplier=1, wait=<function at remote > 0x7fb1f4f23500>) at remote 0x7fb1f4f1ae90>) > return r.call(f, *args, **kwargs) > #29 Frame 0x7fb1f4f9a810, for file /usr/lib/python2.7/site- > packages/cinder/volume/drivers/rbd.py, line 358, in _connect_to_rados > (self=<RBDDriver(_target_names=[], rbd=<module at remote > 0x7fb20583e830>, _is_replication_enabled=False, _execute=<function at > remote 0x7fb2041242a8>, _active_config={'name': 'ceph', 'conf': > '/etc/ceph/ceph.conf', 'user': 'ovirt'}, _active_backend_id=None, > _initialized=False, db=<DBAPI(_backend=<module at remote > 0x7fb203f8d520>, qos_specs_get=<instancemethod at remote > 0x7fb1f677d460>, _lock=<Semaphore(counter=1, > _waiters=<collections.deque at remote 0x7fb1f5246d70>) at remote > 0x7fb1f5205bd0>, _wrap_db_kwargs={'max_retries': 20, > 'inc_retry_interval': True, 'retry_interval': 1, 'max_retry_interval': > 10}, _backend_mapping={'sqlalchemy': 'cinder.db.sqlalchemy.api'}, > _backend_name='sqlalchemy', use_db_reconnect=False, > get_by_id=<instancemethod at remote 0x7fb1f61d8050>, > volume_type_get=<instancemethod at remote 0x7fb1f61c0f50>) at remote > 0x7fb2003aab10>, target_mapping={'tgtadm': 'cinder.vol...(truncated) > return _do_conn(pool, remote, timeout) > #33 Frame 0x7fb1f4f5b220, for file /usr/lib/python2.7/site- > packages/cinder/volume/drivers/rbd.py, line 177, in __init__ > (self=<RADOSClient(driver=<RBDDriver(_target_names=[], rbd=<module at > remote 0x7fb20583e830>, _is_replication_enabled=False, > _execute=<function at remote 0x7fb2041242a8>, _active_config={'name': > 'ceph', 'conf': '/etc/ceph/ceph.conf', 'user': 'ovirt'}, > _active_backend_id=None, _initialized=False, db=<DBAPI(_backend=<module > at remote 0x7fb203f8d520>, qos_specs_get=<instancemethod at remote > 0x7fb1f677d460>, _lock=<Semaphore(counter=1, > _waiters=<collections.deque at remote 0x7fb1f5246d70>) at remote > 0x7fb1f5205bd0>, _wrap_db_kwargs={'max_retries': 20, > 'inc_retry_interval': True, 'retry_interval': 1, 'max_retry_interval': > 10}, _backend_mapping={'sqlalchemy': 'cinder.db.sqlalchemy.api'}, > _backend_name='sqlalchemy', use_db_reconnect=False, > get_by_id=<instancemethod at remote 0x7fb1f61d8050>, > volume_type_get=<instancemethod at remote 0x7fb1f61c0f50>) at remote > 0x7fb2003aab10>, target_mapping={'tgtadm': ...(truncated) > self.cluster, self.ioctx = driver._connect_to_rados(pool) > #44 Frame 0x7fb1f4f9a620, for file /usr/lib/python2.7/site- > packages/cinder/volume/drivers/rbd.py, line 298, in > check_for_setup_error (self=<RBDDriver(_target_names=[], rbd=<module at > remote 0x7fb20583e830>, _is_replication_enabled=False, > _execute=<function at remote 0x7fb2041242a8>, _active_config={'name': > 'ceph', 'conf': '/etc/ceph/ceph.conf', 'user': 'ovirt'}, > _active_backend_id=None, _initialized=False, db=<DBAPI(_backend=<module > at remote 0x7fb203f8d520>, qos_specs_get=<instancemethod at remote > 0x7fb1f677d460>, _lock=<Semaphore(counter=1, > _waiters=<collections.deque at remote 0x7fb1f5246d70>) at remote > 0x7fb1f5205bd0>, _wrap_db_kwargs={'max_retries': 20, > 'inc_retry_interval': True, 'retry_interval': 1, 'max_retry_interval': > 10}, _backend_mapping={'sqlalchemy': 'cinder.db.sqlalchemy.api'}, > _backend_name='sqlalchemy', use_db_reconnect=False, > get_by_id=<instancemethod at remote 0x7fb1f61d8050>, > volume_type_get=<instancemethod at remote 0x7fb1f61c0f50>) at remote > 0x7fb2003aab10>, target_mapping={'tgtadm': 'cinder...(truncated) > with RADOSClient(self): > #48 Frame 0x3e5bef0, for file /usr/lib/python2.7/site- > packages/cinderlib/cinderlib.py, line 88, in __init__ > (self=<Backend(driver=<RBDDriver(_target_names=[], rbd=<module at > remote 0x7fb20583e830>, _is_replication_enabled=False, > _execute=<function at remote 0x7fb2041242a8>, _active_config={'name': > 'ceph', 'conf': '/etc/ceph/ceph.conf', 'user': 'ovirt'}, > _active_backend_id=None, _initialized=False, db=<DBAPI(_backend=<module > at remote 0x7fb203f8d520>, qos_specs_get=<instancemethod at remote > 0x7fb1f677d460>, _lock=<Semaphore(counter=1, > _waiters=<collections.deque at remote 0x7fb1f5246d70>) at remote > 0x7fb1f5205bd0>, _wrap_db_kwargs={'max_retries': 20, > 'inc_retry_interval': True, 'retry_interval': 1, 'max_retry_interval': > 10}, _backend_mapping={'sqlalchemy': 'cinder.db.sqlalchemy.api'}, > _backend_name='sqlalchemy', use_db_reconnect=False, > get_by_id=<instancemethod at remote 0x7fb1f61d8050>, > volume_type_get=<instancemethod at remote 0x7fb1f61c0f50>) at remote > 0x7fb2003aab10>, target_mapping={'tgtadm': > 'cinder.volume.t...(truncated) > self.driver.check_for_setup_error() > #58 Frame 0x3d15cf0, for file ./cinderlib-client.py, line 210, in > load_backend (args=<Namespace(driver='{"volume_backend_name":"ceph- > storage-ssd-ovirt","volume---Type <return> to continue, or q <return> > to quit--- > _driver":"cinder.volume.drivers.rbd.RBDDriver","rbd_ceph_conf":"/etc/ce > ph/ceph.conf","rbd_user":"ovirt","rbd_keyring_conf":"/etc/ceph/ceph.cli > ent.ovirt.keyring","rbd_pool":"storage- > ssd","use_multipath_for_image_xfer":"true"}', > db_url='postgresql+psycopg2://ovirt_cinderlib:ViRsNB3Dnwy5wmL0lLuDEq@lo > calhost:5432/ovirt_cinderlib', correlation_id='7733e7cc', > command=<function at remote 0x7fb1f52725f0>, volume_id='9d858d39-bbd4- > 4cbe-9f2c-5bef25ed0525', > connector_info='{"ip":null,"host":"compute00.hq.tnp.infra","os_type":"l > inux2","platform":"x86_64","initiator":"iqn.1994- > 05.com.redhat:c6ab662d439c","multipath":true,"do_local_attach":false}') > at remote 0x7fb1f5215910>, persistence_config={'connection': > 'postgresql+psycopg2://ovirt_cinderlib:ViRsNB3Dnwy5wmL0lLuDEq@localhost > :5432/ovirt_cinderlib', 'storage': 'db'}) > Python Exception <type 'exceptions.IOError'> [Errno 2] No such file or > directory: './cinderlib-client.py': > Error occurred in Python command: [Errno 2] No such file or directory: > './cinderlib-client.py' > > > > > ________________________________ > > The Networking People (TNP) Limited. Registered office: Network House, > Caton Rd, Lancaster, LA1 3PE. Registered in England & Wales with company > number: 07667393 > > This email and any files transmitted with it are confidential and intended > solely for the use of the individual or entity to whom they are addressed. > If you have received this email in error please notify the system manager. > This message contains confidential information and is intended only for the > individual named. If you are not the named addressee you should not > disseminate, distribute or copy this e-mail. Please notify the sender > immediately by e-mail if you have received this e-mail by mistake and > delete this e-mail from your system. If you are not the intended recipient > you are notified that disclosing, copying, distributing or taking any > action in reliance on the contents of this information is strictly > prohibited. >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/ISM3AFUBRF275TIXTABFWRVFFUE3S6OV/

