Looks like LSM with libgfapi enabled can't work at the moment[1]

[1] - https://bugzilla.redhat.com/show_bug.cgi?id=1481688#c38

On Thu, Jul 18, 2019 at 12:05 PM Christoph Köhler <
[email protected]> wrote:

> The engine runs on version 4.3.4
>
> The hypervisors run on 4.2.8, (version vdsm: 4.20.46-1.el7)
> with libvirt 4.5.0-10.el7_6.7
>
> By vdsm.log :
>
> 2019-07-18 09:29:09,744+0200 INFO  (jsonrpc/2) [storage.StorageDomain]
> Creating domain run directory
> u'/var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640' (fileSD:577)
> 2019-07-18 09:29:09,744+0200 INFO  (jsonrpc/2) [storage.fileUtils]
> Creating directory:
> /var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640 mode: None
> (fileUtils:197)
>
> 2019-07-18 09:29:09,744+0200 INFO  (jsonrpc/2) [storage.StorageDomain]
> Creating symlink from
> /rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50
>
> to /var/run/vdsm/storage/4dabb6d6-4be5-458
> c-811d-6d5e87699640/d2964ff9-10f7-4b92-8327-d68f3cfd5b50 (fileSD:580)
>
> 2019-07-18 09:29:09,789+0200 INFO  (jsonrpc/2) [vdsm.api] FINISH
> prepareImage return={'info': {'path':
> u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
>
> 'type': 'network', 'hosts': [{'port':
> '0', 'transport': 'tcp', 'name': '192.168.11.20'}, {'port': '0',
> 'transport': 'tcp', 'name': 'glusrv01'}, {'port': '0', 'transport':
> 'tcp', 'name': 'glusrv02'}, {'port': '0', 'transport': 'tcp', 'name':
> 'glusrv03'}], 'protocol': 'gluster'}, 'path': u'/rhev/data-center/mn
> t/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
>
> 'imgVolumesInfo': [{'domainID': u'4dabb6d6-4be5-458c-811d-6d5e87699640',
> 'leaseOffset': 0, 'path': u'gluvol4/4dabb6d
> 6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
>
> 'volumeID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', 'leasePath':
> u'/rhev/data-center/mnt/glusterSD/192.168.11.20:
> _gluvol4/4dabb6d6-4be5-458c-811d-6d5e8769964
> 0/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98.lease',
>
> 'imageID': u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}, {'domainID':
> u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0, 'path':
> u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/
> images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a',
>
> 'volumeID': u'43dbb053-c5fe-45bf-9464-acf77546b96a', 'leasePath':
> u'/rhev/data-center/mnt/glusterSD/192.168.11.20:
> _gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-83
> 27-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a.lease', 'imageID':
> u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}]} from=::ffff:10.4.8.242,45784,
> flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f,
> task_id=b825b7c7-ddae-441f-b9c8-2bd50ec144b5 (api:52)
> 2019-07-18 09:29:09,790+0200 INFO  (jsonrpc/2) [vds] prepared volume
> path:
> gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98
>
> (clientIF:497)
> 2019-07-18 09:29:09,791+0200 INFO  (jsonrpc/2) [vdsm.api] START
> teardownImage(sdUUID=u'4dabb6d6-4be5-458c-811d-6d5e87699640',
> spUUID=u'00000001-0001-0001-0001-000000000311',
> imgUUID=u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50', volUUID=None)
> from=::ffff:10.4.8.242,45784, flow
> _id=c957e011-37e0-43aa-abe7-9bb633c38c5f,
> task_id=c13f7ee9-8d01-4b73-8200-3052b4953385 (api:46)
>
> 2019-07-18 09:29:09,792+0200 INFO  (jsonrpc/2) [storage.StorageDomain]
> Removing image rundir link
> u'/var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640/d2964ff9-10f7-4b92-8327-d68f3cfd5b50'
>
> (fileSD:600)
>
> 2019-07-18 09:29:09,792+0200 INFO  (jsonrpc/2) [vdsm.api] FINISH
> teardownImage return=None from=::ffff:10.4.8.242,45784,
> flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f,
> task_id=c13f7ee9-8d01-4b73-8200-3052b4953385 (api:52)
>
> 2019-07-18 09:29:09,792+0200 ERROR (jsonrpc/2) [virt.vm]
> (vmId='3b79d0c0-47e9-47c3-8511-980a8cfe147c') Unable to start
> replication for sda to {u'domainID':
> u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'volumeInfo': {'path':
> u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/imag
> es/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
>
> 'type': 'network', 'hosts': [{'port': '0', 'transport': 'tcp', 'name':
> '192.168.11.20'}, {'port': '0', 'transport': 'tcp', 'name': 'glusrv01'},
> {'port': '0', 'transport': 'tcp', 'name': 'glusr
> v02'}, {'port': '0', 'transport': 'tcp', 'name': 'glusrv03'}],
> 'protocol': 'gluster'}, 'format': 'cow', u'poolID':
> u'00000001-0001-0001-0001-000000000311', u'device': 'disk', 'protocol':
> 'gluster', 'propagateErrors': 'off', u'diskType': u'network', 'cache':
> 'none', u'vol
> umeID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', u'imageID':
> u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50', 'hosts': [{'port': '0',
> 'transport': 'tcp', 'name': '192.168.11.20'}], 'path':
>
> u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50
> /62656632-8984-4b7e-8be1-fd2547ca0f98', 'volumeChain': [{'domainID':
> u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0, 'path':
> u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
>
> 'volu
> meID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', 'leasePath':
> u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98.lease',
>
> 'imageID': u'd2964ff9-10f7-
> 4b92-8327-d68f3cfd5b50'}, {'domainID':
> u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0, 'path':
> u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a',
>
> 'volumeID': u'43dbb053-c5fe-45bf-94
> 64-acf77546b96a', 'leasePath':
> u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a.lease',
>
> 'imageID': u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}]} (vm
> :4710)
> Traceback (most recent call last):
>    File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4704,
> in diskReplicateStart
>      self._startDriveReplication(drive)
>    File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4843,
> in _startDriveReplication
>      self._dom.blockCopy(drive.name, destxml, flags=flags)
>    File "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line
> 98, in f
>      ret = attr(*args, **kwargs)
>    File
> "/usr/lib/python2.7/site-packages/vdsm/common/libvirtconnection.py",
> line 130, in wrapper
>      ret = f(*args, **kwargs)
>    File "/usr/lib/python2.7/site-packages/vdsm/common/function.py", line
> 92, in wrapper
>      return func(inst, *args, **kwargs)
>    File "/usr/lib64/python2.7/site-packages/libvirt.py", line 729, in
> blockCopy
>      if ret == -1: raise libvirtError ('virDomainBlockCopy() failed',
> dom=self)
> libvirtError: argument unsupported: non-file destination not supported yet
>
> 2019-07-18 09:29:09,796+0200 INFO  (jsonrpc/2) [api.virt] FINISH
> diskReplicateStart return={'status': {'message': 'Drive replication
> error', 'code': 55}} from=::ffff:10.4.8.242,45784,
> flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f,
> vmId=3b79d0c0-47e9-47c3-8511-980a8cfe147c
>   (api:52)
>
>
> On 18.07.19 10:42, Benny Zlotnik wrote:
> > It should work, what is the engine and vdsm versions?
> > Can you add vdsm logs as well?
> >
> > On Thu, Jul 18, 2019 at 11:16 AM Christoph Köhler
> > <[email protected] <mailto:[email protected]>>
> wrote:
> >
> >     Hello,
> >
> >     I try to migrate a disk of a running vm from gluster 3.12.15 to
> gluster
> >     3.12.15 but it fails. libGfApi set to true by engine-config.
> >
> >     ° taking a snapshot first, is working. Then at engine-log:
> >
> >     2019-07-18 09:29:13,932+02 ERROR
> >
>  [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
> >
> >     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> >     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed in
> >     'VmReplicateDiskStartVDS' method
> >
> >     2019-07-18 09:29:13,936+02 ERROR
> >
>  [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> >     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> >     [c957e011-37e0-43aa-abe7-9bb633c38c5f] EVENT_ID:
> >     VDS_BROKER_COMMAND_FAILURE(10,802), VDSM ovvirt07 command
> >     VmReplicateDiskStartVDS failed: Drive replication error
> >
> >     2019-07-18 09:29:13,936+02 INFO
> >
>  [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
> >
> >     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> >     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command
> >
>  'org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand'
> >
> >     return value 'StatusOnlyReturn [status=Status [code=55, message=Drive
> >     replication error]]'
> >
> >     2019-07-18 09:29:13,936+02 INFO
> >
>  [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
> >
> >     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> >     [c957e011-37e0-43aa-abe7-9bb633c38c5f] HostName = ovvirt07
> >
> >     2019-07-18 09:29:13,937+02 ERROR
> >
>  [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
> >
> >     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> >     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command
> >     'VmReplicateDiskStartVDSCommand(HostName = ovvirt07,
> >
>  VmReplicateDiskParameters:{hostId='3a7bf85c-e92d-4559-908e-5eed2f5608d4',
> >
> >     vmId='3b79d0c0-47e9-47c3-8511-980a8cfe147c',
> >     storagePoolId='00000001-0001-0001-0001-000000000311',
> >     srcStorageDomainId='e54d835a-d8a5-44ae-8e17-fcba1c54e46f',
> >     targetStorageDomainId='4dabb6d6-4be5-458c-811d-6d5e87699640',
> >     imageGroupId='d2964ff9-10f7-4b92-8327-d68f3cfd5b50',
> >     imageId='62656632-8984-4b7e-8be1-fd2547ca0f98'})' execution failed:
> >     VDSGenericException: VDSErrorException: Failed to
> >     VmReplicateDiskStartVDS, error = Drive replication error, code = 55
> >
> >     2019-07-18 09:29:13,937+02 INFO
> >
>  [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
> >
> >     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> >     [c957e011-37e0-43aa-abe7-9bb633c38c5f] FINISH,
> >     VmReplicateDiskStartVDSCommand, return: , log id: 5b2afb0b
> >
> >     2019-07-18 09:29:13,937+02 ERROR
> >     [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand]
> >     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> >     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed VmReplicateDiskStart
> >     (Disk
> >     'd2964ff9-10f7-4b92-8327-d68f3cfd5b50' , VM
> >     '3b79d0c0-47e9-47c3-8511-980a8cfe147c')
> >
> >     2019-07-18 09:29:13,938+02 ERROR
> >     [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
> >     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> >     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command 'LiveMigrateDisk' id:
> >     '8174c74c-8ab0-49fa-abfc-44d8b7c691e0' with children
> >     [03672b60-443b-47ba-834c-ac306d7129d0,
> >     562522fc-6691-47fe-93bf-ef2c45e85676] failed when attempting to
> perform
> >     the next operation, marking as 'ACTIVE'
> >
> >     2019-07-18 09:29:13,938+02 ERROR
> >     [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
> >     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> >     [c957e011-37e0-43aa-abe7-9bb633c38c5f] EngineException: Drive
> >     replication error (Failed with error replicaErr and code 55):
> >     org.ovirt.engine.core.common.errors.EngineException: EngineException:
> >     Drive replication error (Failed with error replicaErr and code 55)
> >               at
> >
>  
> org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand.replicateDiskStart(LiveMigrateDiskCommand.java:526)
> >
> >     [bll.jar:]
> >               at
> >
>  
> org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand.performNextOperation(LiveMigrateDiskCommand.java:233)
> >
> >     [bll.jar:]
> >               at
> >
>  
> org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback.childCommandsExecutionEnded(SerialChildCommandsExecutionCallback.java:32)
> >
> >     [bll.jar:]
> >               at
> >
>  
> org.ovirt.engine.core.bll.ChildCommandsCallbackBase.doPolling(ChildCommandsCallbackBase.java:77)
> >
> >     [bll.jar:]
> >               at
> >
>  
> org.ovirt.engine.core.bll.tasks.CommandCallbacksPoller.invokeCallbackMethodsImpl(CommandCallbacksPoller.java:175)
> >
> >     [bll.jar:]
> >               at
> >
>  
> org.ovirt.engine.core.bll.tasks.CommandCallbacksPoller.invokeCallbackMethods(CommandCallbacksPoller.java:109)
> >
> >     [bll.jar:]
> >               at
> >
>  java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> >     [rt.jar:1.8.0_212]
> >               at
> >     java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
> >     [rt.jar:1.8.0_212]
> >               at
> >
>  
> org.glassfish.enterprise.concurrent.internal.ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask.access$201(ManagedScheduledThreadPoolExecutor.java:383)
> >
> >     [javax.enterprise.concurrent-1.0.jar:]
> >               at
> >
>  
> org.glassfish.enterprise.concurrent.internal.ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask.run(ManagedScheduledThreadPoolExecutor.java:534)
> >
> >     [javax.enterprise.concurrent-1.0.jar:]
> >               at
> >
>  
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> >
> >     [rt.jar:1.8.0_212]
> >               at
> >
>  
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> >
> >     [rt.jar:1.8.0_212]
> >               at java.lang.Thread.run(Thread.java:748) [rt.jar:1.8.0_212]
> >               at
> >
>  
> org.glassfish.enterprise.concurrent.ManagedThreadFactoryImpl$ManagedThread.run(ManagedThreadFactoryImpl.java:250)
> >
> >     [javax.enterprise.concurrent-1.0.jar:]
> >               at
> >
>  
> org.jboss.as.ee.concurrent.service.ElytronManagedThreadFactory$ElytronManagedThread.run(ElytronManagedThreadFactory.java:78)
> >
> >     2019-07-18 09:29:13,938+02 INFO
> >     [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
> >     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> >     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command 'LiveMigrateDisk' id:
> >     '8174c74c-8ab0-49fa-abfc-44d8b7c691e0' child commands
> >     '[03672b60-443b-47ba-834c-ac306d7129d0,
> >     562522fc-6691-47fe-93bf-ef2c45e85676]' executions were completed,
> >     status
> >     'FAILED'
> >
> >     2019-07-18 09:29:15,019+02 ERROR
> >     [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand]
> >     (EE-ManagedThreadFactory-engineScheduled-Thread-1)
> >     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Ending command
> >     'org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand' with
> >     failure.
> >
> >     2019-07-18 09:29:15,019+02 ERROR
> >     [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand]
> >     (EE-ManagedThreadFactory-engineScheduled-Thread-1)
> >     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed during live storage
> >     migration of disk 'd2964ff9-10f7-4b92-8327-d68f3cfd5b50' of vm
> >     '3b79d0c0-47e9-47c3-8511-980a8cfe147c', attempting to end replication
> >     before deleting the target disk
> >
> >     //
> >
> >     LiveStorageMigration on gluster - should that work at all? Has
> someone
> >     tried it?
> >
> >     Greetings!
> >     Christoph Köhler
> >     _______________________________________________
> >     Users mailing list -- [email protected] <mailto:[email protected]>
> >     To unsubscribe send an email to [email protected]
> >     <mailto:[email protected]>
> >     Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> >     oVirt Code of Conduct:
> >     https://www.ovirt.org/community/about/community-guidelines/
> >     List Archives:
> >
> https://lists.ovirt.org/archives/list/[email protected]/message/WSF45K47AJNSAJWYSAMSRKW34T5MWRWZ/
> >
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/LP64ZQVEM7LLQFXVCCUUJSM7UAVGJ3NY/

Reply via email to