The engine runs on version 4.3.4

The hypervisors run on 4.2.8, (version vdsm: 4.20.46-1.el7)
with libvirt 4.5.0-10.el7_6.7

By vdsm.log :

2019-07-18 09:29:09,744+0200 INFO (jsonrpc/2) [storage.StorageDomain] Creating domain run directory u'/var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640' (fileSD:577) 2019-07-18 09:29:09,744+0200 INFO (jsonrpc/2) [storage.fileUtils] Creating directory: /var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640 mode: None (fileUtils:197)

2019-07-18 09:29:09,744+0200 INFO (jsonrpc/2) [storage.StorageDomain] Creating symlink from /rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50 to /var/run/vdsm/storage/4dabb6d6-4be5-458
c-811d-6d5e87699640/d2964ff9-10f7-4b92-8327-d68f3cfd5b50 (fileSD:580)

2019-07-18 09:29:09,789+0200 INFO (jsonrpc/2) [vdsm.api] FINISH prepareImage return={'info': {'path': u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98', 'type': 'network', 'hosts': [{'port': '0', 'transport': 'tcp', 'name': '192.168.11.20'}, {'port': '0', 'transport': 'tcp', 'name': 'glusrv01'}, {'port': '0', 'transport': 'tcp', 'name': 'glusrv02'}, {'port': '0', 'transport': 'tcp', 'name': 'glusrv03'}], 'protocol': 'gluster'}, 'path': u'/rhev/data-center/mn t/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98', 'imgVolumesInfo': [{'domainID': u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0, 'path': u'gluvol4/4dabb6d 6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98', 'volumeID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', 'leasePath': u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e8769964 0/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98.lease', 'imageID': u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}, {'domainID': u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0, 'path': u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/ images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a', 'volumeID': u'43dbb053-c5fe-45bf-9464-acf77546b96a', 'leasePath': u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-83 27-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a.lease', 'imageID': u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}]} from=::ffff:10.4.8.242,45784, flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f, task_id=b825b7c7-ddae-441f-b9c8-2bd50ec144b5 (api:52) 2019-07-18 09:29:09,790+0200 INFO (jsonrpc/2) [vds] prepared volume path: gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98 (clientIF:497) 2019-07-18 09:29:09,791+0200 INFO (jsonrpc/2) [vdsm.api] START teardownImage(sdUUID=u'4dabb6d6-4be5-458c-811d-6d5e87699640', spUUID=u'00000001-0001-0001-0001-000000000311', imgUUID=u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50', volUUID=None) from=::ffff:10.4.8.242,45784, flow _id=c957e011-37e0-43aa-abe7-9bb633c38c5f, task_id=c13f7ee9-8d01-4b73-8200-3052b4953385 (api:46)

2019-07-18 09:29:09,792+0200 INFO (jsonrpc/2) [storage.StorageDomain] Removing image rundir link u'/var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640/d2964ff9-10f7-4b92-8327-d68f3cfd5b50' (fileSD:600)

2019-07-18 09:29:09,792+0200 INFO (jsonrpc/2) [vdsm.api] FINISH teardownImage return=None from=::ffff:10.4.8.242,45784, flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f, task_id=c13f7ee9-8d01-4b73-8200-3052b4953385 (api:52)

2019-07-18 09:29:09,792+0200 ERROR (jsonrpc/2) [virt.vm] (vmId='3b79d0c0-47e9-47c3-8511-980a8cfe147c') Unable to start replication for sda to {u'domainID': u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'volumeInfo': {'path': u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/imag es/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98', 'type': 'network', 'hosts': [{'port': '0', 'transport': 'tcp', 'name': '192.168.11.20'}, {'port': '0', 'transport': 'tcp', 'name': 'glusrv01'}, {'port': '0', 'transport': 'tcp', 'name': 'glusr v02'}, {'port': '0', 'transport': 'tcp', 'name': 'glusrv03'}], 'protocol': 'gluster'}, 'format': 'cow', u'poolID': u'00000001-0001-0001-0001-000000000311', u'device': 'disk', 'protocol': 'gluster', 'propagateErrors': 'off', u'diskType': u'network', 'cache': 'none', u'vol umeID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', u'imageID': u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50', 'hosts': [{'port': '0', 'transport': 'tcp', 'name': '192.168.11.20'}], 'path': u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50 /62656632-8984-4b7e-8be1-fd2547ca0f98', 'volumeChain': [{'domainID': u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0, 'path': u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98', 'volu meID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', 'leasePath': u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98.lease', 'imageID': u'd2964ff9-10f7- 4b92-8327-d68f3cfd5b50'}, {'domainID': u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0, 'path': u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a', 'volumeID': u'43dbb053-c5fe-45bf-94 64-acf77546b96a', 'leasePath': u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a.lease', 'imageID': u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}]} (vm
:4710)
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4704, in diskReplicateStart
    self._startDriveReplication(drive)
File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4843, in _startDriveReplication
    self._dom.blockCopy(drive.name, destxml, flags=flags)
File "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line 98, in f
    ret = attr(*args, **kwargs)
File "/usr/lib/python2.7/site-packages/vdsm/common/libvirtconnection.py", line 130, in wrapper
    ret = f(*args, **kwargs)
File "/usr/lib/python2.7/site-packages/vdsm/common/function.py", line 92, in wrapper
    return func(inst, *args, **kwargs)
File "/usr/lib64/python2.7/site-packages/libvirt.py", line 729, in blockCopy if ret == -1: raise libvirtError ('virDomainBlockCopy() failed', dom=self)
libvirtError: argument unsupported: non-file destination not supported yet

2019-07-18 09:29:09,796+0200 INFO (jsonrpc/2) [api.virt] FINISH diskReplicateStart return={'status': {'message': 'Drive replication error', 'code': 55}} from=::ffff:10.4.8.242,45784, flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f, vmId=3b79d0c0-47e9-47c3-8511-980a8cfe147c
 (api:52)


On 18.07.19 10:42, Benny Zlotnik wrote:
It should work, what is the engine and vdsm versions?
Can you add vdsm logs as well?

On Thu, Jul 18, 2019 at 11:16 AM Christoph Köhler <[email protected] <mailto:[email protected]>> wrote:

    Hello,

    I try to migrate a disk of a running vm from gluster 3.12.15 to gluster
    3.12.15 but it fails. libGfApi set to true by engine-config.

    ° taking a snapshot first, is working. Then at engine-log:

    2019-07-18 09:29:13,932+02 ERROR
    [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]

    (EE-ManagedThreadFactory-engineScheduled-Thread-84)
    [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed in
    'VmReplicateDiskStartVDS' method

    2019-07-18 09:29:13,936+02 ERROR
    [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
    (EE-ManagedThreadFactory-engineScheduled-Thread-84)
    [c957e011-37e0-43aa-abe7-9bb633c38c5f] EVENT_ID:
    VDS_BROKER_COMMAND_FAILURE(10,802), VDSM ovvirt07 command
    VmReplicateDiskStartVDS failed: Drive replication error

    2019-07-18 09:29:13,936+02 INFO
    [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]

    (EE-ManagedThreadFactory-engineScheduled-Thread-84)
    [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command
    'org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand'

    return value 'StatusOnlyReturn [status=Status [code=55, message=Drive
    replication error]]'

    2019-07-18 09:29:13,936+02 INFO
    [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]

    (EE-ManagedThreadFactory-engineScheduled-Thread-84)
    [c957e011-37e0-43aa-abe7-9bb633c38c5f] HostName = ovvirt07

    2019-07-18 09:29:13,937+02 ERROR
    [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]

    (EE-ManagedThreadFactory-engineScheduled-Thread-84)
    [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command
    'VmReplicateDiskStartVDSCommand(HostName = ovvirt07,
    VmReplicateDiskParameters:{hostId='3a7bf85c-e92d-4559-908e-5eed2f5608d4',

    vmId='3b79d0c0-47e9-47c3-8511-980a8cfe147c',
    storagePoolId='00000001-0001-0001-0001-000000000311',
    srcStorageDomainId='e54d835a-d8a5-44ae-8e17-fcba1c54e46f',
    targetStorageDomainId='4dabb6d6-4be5-458c-811d-6d5e87699640',
    imageGroupId='d2964ff9-10f7-4b92-8327-d68f3cfd5b50',
    imageId='62656632-8984-4b7e-8be1-fd2547ca0f98'})' execution failed:
    VDSGenericException: VDSErrorException: Failed to
    VmReplicateDiskStartVDS, error = Drive replication error, code = 55

    2019-07-18 09:29:13,937+02 INFO
    [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]

    (EE-ManagedThreadFactory-engineScheduled-Thread-84)
    [c957e011-37e0-43aa-abe7-9bb633c38c5f] FINISH,
    VmReplicateDiskStartVDSCommand, return: , log id: 5b2afb0b

    2019-07-18 09:29:13,937+02 ERROR
    [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand]
    (EE-ManagedThreadFactory-engineScheduled-Thread-84)
    [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed VmReplicateDiskStart
    (Disk
    'd2964ff9-10f7-4b92-8327-d68f3cfd5b50' , VM
    '3b79d0c0-47e9-47c3-8511-980a8cfe147c')

    2019-07-18 09:29:13,938+02 ERROR
    [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
    (EE-ManagedThreadFactory-engineScheduled-Thread-84)
    [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command 'LiveMigrateDisk' id:
    '8174c74c-8ab0-49fa-abfc-44d8b7c691e0' with children
    [03672b60-443b-47ba-834c-ac306d7129d0,
    562522fc-6691-47fe-93bf-ef2c45e85676] failed when attempting to perform
    the next operation, marking as 'ACTIVE'

    2019-07-18 09:29:13,938+02 ERROR
    [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
    (EE-ManagedThreadFactory-engineScheduled-Thread-84)
    [c957e011-37e0-43aa-abe7-9bb633c38c5f] EngineException: Drive
    replication error (Failed with error replicaErr and code 55):
    org.ovirt.engine.core.common.errors.EngineException: EngineException:
    Drive replication error (Failed with error replicaErr and code 55)
              at
    
org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand.replicateDiskStart(LiveMigrateDiskCommand.java:526)

    [bll.jar:]
              at
    
org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand.performNextOperation(LiveMigrateDiskCommand.java:233)

    [bll.jar:]
              at
    
org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback.childCommandsExecutionEnded(SerialChildCommandsExecutionCallback.java:32)

    [bll.jar:]
              at
    
org.ovirt.engine.core.bll.ChildCommandsCallbackBase.doPolling(ChildCommandsCallbackBase.java:77)

    [bll.jar:]
              at
    
org.ovirt.engine.core.bll.tasks.CommandCallbacksPoller.invokeCallbackMethodsImpl(CommandCallbacksPoller.java:175)

    [bll.jar:]
              at
    
org.ovirt.engine.core.bll.tasks.CommandCallbacksPoller.invokeCallbackMethods(CommandCallbacksPoller.java:109)

    [bll.jar:]
              at
    java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    [rt.jar:1.8.0_212]
              at
    java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    [rt.jar:1.8.0_212]
              at
    
org.glassfish.enterprise.concurrent.internal.ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask.access$201(ManagedScheduledThreadPoolExecutor.java:383)

    [javax.enterprise.concurrent-1.0.jar:]
              at
    
org.glassfish.enterprise.concurrent.internal.ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask.run(ManagedScheduledThreadPoolExecutor.java:534)

    [javax.enterprise.concurrent-1.0.jar:]
              at
    
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)

    [rt.jar:1.8.0_212]
              at
    
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)

    [rt.jar:1.8.0_212]
              at java.lang.Thread.run(Thread.java:748) [rt.jar:1.8.0_212]
              at
    
org.glassfish.enterprise.concurrent.ManagedThreadFactoryImpl$ManagedThread.run(ManagedThreadFactoryImpl.java:250)

    [javax.enterprise.concurrent-1.0.jar:]
              at
    
org.jboss.as.ee.concurrent.service.ElytronManagedThreadFactory$ElytronManagedThread.run(ElytronManagedThreadFactory.java:78)

    2019-07-18 09:29:13,938+02 INFO
    [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
    (EE-ManagedThreadFactory-engineScheduled-Thread-84)
    [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command 'LiveMigrateDisk' id:
    '8174c74c-8ab0-49fa-abfc-44d8b7c691e0' child commands
    '[03672b60-443b-47ba-834c-ac306d7129d0,
    562522fc-6691-47fe-93bf-ef2c45e85676]' executions were completed,
    status
    'FAILED'

    2019-07-18 09:29:15,019+02 ERROR
    [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand]
    (EE-ManagedThreadFactory-engineScheduled-Thread-1)
    [c957e011-37e0-43aa-abe7-9bb633c38c5f] Ending command
    'org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand' with
    failure.

    2019-07-18 09:29:15,019+02 ERROR
    [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand]
    (EE-ManagedThreadFactory-engineScheduled-Thread-1)
    [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed during live storage
    migration of disk 'd2964ff9-10f7-4b92-8327-d68f3cfd5b50' of vm
    '3b79d0c0-47e9-47c3-8511-980a8cfe147c', attempting to end replication
    before deleting the target disk

    //

    LiveStorageMigration on gluster - should that work at all? Has someone
    tried it?

    Greetings!
    Christoph Köhler
    _______________________________________________
    Users mailing list -- [email protected] <mailto:[email protected]>
    To unsubscribe send an email to [email protected]
    <mailto:[email protected]>
    Privacy Statement: https://www.ovirt.org/site/privacy-policy/
    oVirt Code of Conduct:
    https://www.ovirt.org/community/about/community-guidelines/
    List Archives:
    
https://lists.ovirt.org/archives/list/[email protected]/message/WSF45K47AJNSAJWYSAMSRKW34T5MWRWZ/

_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/L7PY54OG5SXSQ6S3MMFPMVNKW3LZICQU/

Reply via email to