Looks like LSM with libgfapi enabled can't work at the moment[1]
[1] - https://bugzilla.redhat.com/show_bug.cgi?id=1481688#c38 On Thu, Jul 18, 2019 at 12:05 PM Christoph Köhler < [email protected]> wrote: > The engine runs on version 4.3.4 > > The hypervisors run on 4.2.8, (version vdsm: 4.20.46-1.el7) > with libvirt 4.5.0-10.el7_6.7 > > By vdsm.log : > > 2019-07-18 09:29:09,744+0200 INFO (jsonrpc/2) [storage.StorageDomain] > Creating domain run directory > u'/var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640' (fileSD:577) > 2019-07-18 09:29:09,744+0200 INFO (jsonrpc/2) [storage.fileUtils] > Creating directory: > /var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640 mode: None > (fileUtils:197) > > 2019-07-18 09:29:09,744+0200 INFO (jsonrpc/2) [storage.StorageDomain] > Creating symlink from > /rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50 > > to /var/run/vdsm/storage/4dabb6d6-4be5-458 > c-811d-6d5e87699640/d2964ff9-10f7-4b92-8327-d68f3cfd5b50 (fileSD:580) > > 2019-07-18 09:29:09,789+0200 INFO (jsonrpc/2) [vdsm.api] FINISH > prepareImage return={'info': {'path': > u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98', > > 'type': 'network', 'hosts': [{'port': > '0', 'transport': 'tcp', 'name': '192.168.11.20'}, {'port': '0', > 'transport': 'tcp', 'name': 'glusrv01'}, {'port': '0', 'transport': > 'tcp', 'name': 'glusrv02'}, {'port': '0', 'transport': 'tcp', 'name': > 'glusrv03'}], 'protocol': 'gluster'}, 'path': u'/rhev/data-center/mn > t/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98', > > 'imgVolumesInfo': [{'domainID': u'4dabb6d6-4be5-458c-811d-6d5e87699640', > 'leaseOffset': 0, 'path': u'gluvol4/4dabb6d > 6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98', > > 'volumeID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', 'leasePath': > u'/rhev/data-center/mnt/glusterSD/192.168.11.20: > _gluvol4/4dabb6d6-4be5-458c-811d-6d5e8769964 > 0/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98.lease', > > 'imageID': u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}, {'domainID': > u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0, 'path': > u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/ > images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a', > > 'volumeID': u'43dbb053-c5fe-45bf-9464-acf77546b96a', 'leasePath': > u'/rhev/data-center/mnt/glusterSD/192.168.11.20: > _gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-83 > 27-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a.lease', 'imageID': > u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}]} from=::ffff:10.4.8.242,45784, > flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f, > task_id=b825b7c7-ddae-441f-b9c8-2bd50ec144b5 (api:52) > 2019-07-18 09:29:09,790+0200 INFO (jsonrpc/2) [vds] prepared volume > path: > gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98 > > (clientIF:497) > 2019-07-18 09:29:09,791+0200 INFO (jsonrpc/2) [vdsm.api] START > teardownImage(sdUUID=u'4dabb6d6-4be5-458c-811d-6d5e87699640', > spUUID=u'00000001-0001-0001-0001-000000000311', > imgUUID=u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50', volUUID=None) > from=::ffff:10.4.8.242,45784, flow > _id=c957e011-37e0-43aa-abe7-9bb633c38c5f, > task_id=c13f7ee9-8d01-4b73-8200-3052b4953385 (api:46) > > 2019-07-18 09:29:09,792+0200 INFO (jsonrpc/2) [storage.StorageDomain] > Removing image rundir link > u'/var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640/d2964ff9-10f7-4b92-8327-d68f3cfd5b50' > > (fileSD:600) > > 2019-07-18 09:29:09,792+0200 INFO (jsonrpc/2) [vdsm.api] FINISH > teardownImage return=None from=::ffff:10.4.8.242,45784, > flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f, > task_id=c13f7ee9-8d01-4b73-8200-3052b4953385 (api:52) > > 2019-07-18 09:29:09,792+0200 ERROR (jsonrpc/2) [virt.vm] > (vmId='3b79d0c0-47e9-47c3-8511-980a8cfe147c') Unable to start > replication for sda to {u'domainID': > u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'volumeInfo': {'path': > u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/imag > es/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98', > > 'type': 'network', 'hosts': [{'port': '0', 'transport': 'tcp', 'name': > '192.168.11.20'}, {'port': '0', 'transport': 'tcp', 'name': 'glusrv01'}, > {'port': '0', 'transport': 'tcp', 'name': 'glusr > v02'}, {'port': '0', 'transport': 'tcp', 'name': 'glusrv03'}], > 'protocol': 'gluster'}, 'format': 'cow', u'poolID': > u'00000001-0001-0001-0001-000000000311', u'device': 'disk', 'protocol': > 'gluster', 'propagateErrors': 'off', u'diskType': u'network', 'cache': > 'none', u'vol > umeID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', u'imageID': > u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50', 'hosts': [{'port': '0', > 'transport': 'tcp', 'name': '192.168.11.20'}], 'path': > > u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50 > /62656632-8984-4b7e-8be1-fd2547ca0f98', 'volumeChain': [{'domainID': > u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0, 'path': > u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98', > > 'volu > meID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', 'leasePath': > u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98.lease', > > 'imageID': u'd2964ff9-10f7- > 4b92-8327-d68f3cfd5b50'}, {'domainID': > u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0, 'path': > u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a', > > 'volumeID': u'43dbb053-c5fe-45bf-94 > 64-acf77546b96a', 'leasePath': > u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a.lease', > > 'imageID': u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}]} (vm > :4710) > Traceback (most recent call last): > File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4704, > in diskReplicateStart > self._startDriveReplication(drive) > File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4843, > in _startDriveReplication > self._dom.blockCopy(drive.name, destxml, flags=flags) > File "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line > 98, in f > ret = attr(*args, **kwargs) > File > "/usr/lib/python2.7/site-packages/vdsm/common/libvirtconnection.py", > line 130, in wrapper > ret = f(*args, **kwargs) > File "/usr/lib/python2.7/site-packages/vdsm/common/function.py", line > 92, in wrapper > return func(inst, *args, **kwargs) > File "/usr/lib64/python2.7/site-packages/libvirt.py", line 729, in > blockCopy > if ret == -1: raise libvirtError ('virDomainBlockCopy() failed', > dom=self) > libvirtError: argument unsupported: non-file destination not supported yet > > 2019-07-18 09:29:09,796+0200 INFO (jsonrpc/2) [api.virt] FINISH > diskReplicateStart return={'status': {'message': 'Drive replication > error', 'code': 55}} from=::ffff:10.4.8.242,45784, > flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f, > vmId=3b79d0c0-47e9-47c3-8511-980a8cfe147c > (api:52) > > > On 18.07.19 10:42, Benny Zlotnik wrote: > > It should work, what is the engine and vdsm versions? > > Can you add vdsm logs as well? > > > > On Thu, Jul 18, 2019 at 11:16 AM Christoph Köhler > > <[email protected] <mailto:[email protected]>> > wrote: > > > > Hello, > > > > I try to migrate a disk of a running vm from gluster 3.12.15 to > gluster > > 3.12.15 but it fails. libGfApi set to true by engine-config. > > > > ° taking a snapshot first, is working. Then at engine-log: > > > > 2019-07-18 09:29:13,932+02 ERROR > > > [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand] > > > > (EE-ManagedThreadFactory-engineScheduled-Thread-84) > > [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed in > > 'VmReplicateDiskStartVDS' method > > > > 2019-07-18 09:29:13,936+02 ERROR > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > > (EE-ManagedThreadFactory-engineScheduled-Thread-84) > > [c957e011-37e0-43aa-abe7-9bb633c38c5f] EVENT_ID: > > VDS_BROKER_COMMAND_FAILURE(10,802), VDSM ovvirt07 command > > VmReplicateDiskStartVDS failed: Drive replication error > > > > 2019-07-18 09:29:13,936+02 INFO > > > [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand] > > > > (EE-ManagedThreadFactory-engineScheduled-Thread-84) > > [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command > > > 'org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand' > > > > return value 'StatusOnlyReturn [status=Status [code=55, message=Drive > > replication error]]' > > > > 2019-07-18 09:29:13,936+02 INFO > > > [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand] > > > > (EE-ManagedThreadFactory-engineScheduled-Thread-84) > > [c957e011-37e0-43aa-abe7-9bb633c38c5f] HostName = ovvirt07 > > > > 2019-07-18 09:29:13,937+02 ERROR > > > [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand] > > > > (EE-ManagedThreadFactory-engineScheduled-Thread-84) > > [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command > > 'VmReplicateDiskStartVDSCommand(HostName = ovvirt07, > > > VmReplicateDiskParameters:{hostId='3a7bf85c-e92d-4559-908e-5eed2f5608d4', > > > > vmId='3b79d0c0-47e9-47c3-8511-980a8cfe147c', > > storagePoolId='00000001-0001-0001-0001-000000000311', > > srcStorageDomainId='e54d835a-d8a5-44ae-8e17-fcba1c54e46f', > > targetStorageDomainId='4dabb6d6-4be5-458c-811d-6d5e87699640', > > imageGroupId='d2964ff9-10f7-4b92-8327-d68f3cfd5b50', > > imageId='62656632-8984-4b7e-8be1-fd2547ca0f98'})' execution failed: > > VDSGenericException: VDSErrorException: Failed to > > VmReplicateDiskStartVDS, error = Drive replication error, code = 55 > > > > 2019-07-18 09:29:13,937+02 INFO > > > [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand] > > > > (EE-ManagedThreadFactory-engineScheduled-Thread-84) > > [c957e011-37e0-43aa-abe7-9bb633c38c5f] FINISH, > > VmReplicateDiskStartVDSCommand, return: , log id: 5b2afb0b > > > > 2019-07-18 09:29:13,937+02 ERROR > > [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand] > > (EE-ManagedThreadFactory-engineScheduled-Thread-84) > > [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed VmReplicateDiskStart > > (Disk > > 'd2964ff9-10f7-4b92-8327-d68f3cfd5b50' , VM > > '3b79d0c0-47e9-47c3-8511-980a8cfe147c') > > > > 2019-07-18 09:29:13,938+02 ERROR > > [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback] > > (EE-ManagedThreadFactory-engineScheduled-Thread-84) > > [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command 'LiveMigrateDisk' id: > > '8174c74c-8ab0-49fa-abfc-44d8b7c691e0' with children > > [03672b60-443b-47ba-834c-ac306d7129d0, > > 562522fc-6691-47fe-93bf-ef2c45e85676] failed when attempting to > perform > > the next operation, marking as 'ACTIVE' > > > > 2019-07-18 09:29:13,938+02 ERROR > > [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback] > > (EE-ManagedThreadFactory-engineScheduled-Thread-84) > > [c957e011-37e0-43aa-abe7-9bb633c38c5f] EngineException: Drive > > replication error (Failed with error replicaErr and code 55): > > org.ovirt.engine.core.common.errors.EngineException: EngineException: > > Drive replication error (Failed with error replicaErr and code 55) > > at > > > > org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand.replicateDiskStart(LiveMigrateDiskCommand.java:526) > > > > [bll.jar:] > > at > > > > org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand.performNextOperation(LiveMigrateDiskCommand.java:233) > > > > [bll.jar:] > > at > > > > org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback.childCommandsExecutionEnded(SerialChildCommandsExecutionCallback.java:32) > > > > [bll.jar:] > > at > > > > org.ovirt.engine.core.bll.ChildCommandsCallbackBase.doPolling(ChildCommandsCallbackBase.java:77) > > > > [bll.jar:] > > at > > > > org.ovirt.engine.core.bll.tasks.CommandCallbacksPoller.invokeCallbackMethodsImpl(CommandCallbacksPoller.java:175) > > > > [bll.jar:] > > at > > > > org.ovirt.engine.core.bll.tasks.CommandCallbacksPoller.invokeCallbackMethods(CommandCallbacksPoller.java:109) > > > > [bll.jar:] > > at > > > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > > [rt.jar:1.8.0_212] > > at > > java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308) > > [rt.jar:1.8.0_212] > > at > > > > org.glassfish.enterprise.concurrent.internal.ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask.access$201(ManagedScheduledThreadPoolExecutor.java:383) > > > > [javax.enterprise.concurrent-1.0.jar:] > > at > > > > org.glassfish.enterprise.concurrent.internal.ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask.run(ManagedScheduledThreadPoolExecutor.java:534) > > > > [javax.enterprise.concurrent-1.0.jar:] > > at > > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > > > > [rt.jar:1.8.0_212] > > at > > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > > > > [rt.jar:1.8.0_212] > > at java.lang.Thread.run(Thread.java:748) [rt.jar:1.8.0_212] > > at > > > > org.glassfish.enterprise.concurrent.ManagedThreadFactoryImpl$ManagedThread.run(ManagedThreadFactoryImpl.java:250) > > > > [javax.enterprise.concurrent-1.0.jar:] > > at > > > > org.jboss.as.ee.concurrent.service.ElytronManagedThreadFactory$ElytronManagedThread.run(ElytronManagedThreadFactory.java:78) > > > > 2019-07-18 09:29:13,938+02 INFO > > [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback] > > (EE-ManagedThreadFactory-engineScheduled-Thread-84) > > [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command 'LiveMigrateDisk' id: > > '8174c74c-8ab0-49fa-abfc-44d8b7c691e0' child commands > > '[03672b60-443b-47ba-834c-ac306d7129d0, > > 562522fc-6691-47fe-93bf-ef2c45e85676]' executions were completed, > > status > > 'FAILED' > > > > 2019-07-18 09:29:15,019+02 ERROR > > [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand] > > (EE-ManagedThreadFactory-engineScheduled-Thread-1) > > [c957e011-37e0-43aa-abe7-9bb633c38c5f] Ending command > > 'org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand' with > > failure. > > > > 2019-07-18 09:29:15,019+02 ERROR > > [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand] > > (EE-ManagedThreadFactory-engineScheduled-Thread-1) > > [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed during live storage > > migration of disk 'd2964ff9-10f7-4b92-8327-d68f3cfd5b50' of vm > > '3b79d0c0-47e9-47c3-8511-980a8cfe147c', attempting to end replication > > before deleting the target disk > > > > // > > > > LiveStorageMigration on gluster - should that work at all? Has > someone > > tried it? > > > > Greetings! > > Christoph Köhler > > _______________________________________________ > > Users mailing list -- [email protected] <mailto:[email protected]> > > To unsubscribe send an email to [email protected] > > <mailto:[email protected]> > > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > > oVirt Code of Conduct: > > https://www.ovirt.org/community/about/community-guidelines/ > > List Archives: > > > https://lists.ovirt.org/archives/list/[email protected]/message/WSF45K47AJNSAJWYSAMSRKW34T5MWRWZ/ > > >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/LP64ZQVEM7LLQFXVCCUUJSM7UAVGJ3NY/

