On Wed, Jul 10, 2019, 14:57 Neil <[email protected]> wrote: > To provide a slight update on this. > > I put one of my hosts into maintenance and it then migrated the two VM's > off of it, I then upgraded the host to 4.3. > > I have 12 VM's running on the remaining host, if I put it into maintenance > will it try migrate all 12 VM's at once or will it stagger them until they > are all migrated? > If you have a good migration network (at least 10Gbps) the it should be fine. You could also just manually migrate one by one.
> > Thank you. > > Regards. > > Neil Wilson. > > > > > > > On Wed, Jul 10, 2019 at 9:44 AM Neil <[email protected]> wrote: > >> Hi Michal, >> >> Thanks for assisting. >> >> I've just done as requested however nothing is logged in the engine.log >> at the time I click Migrate, below is the log and I hit the Migrate button >> about 4 times between 09:35 and 09:36 and nothing was logged about this... >> >> 2019-07-10 09:35:57,967+02 INFO >> [org.ovirt.engine.core.sso.utils.AuthenticationUtils] (default task-14) [] >> User trouble@internal successfully logged in with scopes: >> ovirt-app-admin ovirt-app-api ovirt-app-portal >> ovirt-ext=auth:sequence-priority=~ ovirt-ext=revoke:revoke-all >> ovirt-ext=token-info:authz-search ovirt-ext=token-info:public-authz-search >> ovirt-ext=token-info:validate ovirt-ext=token:password-access >> 2019-07-10 09:35:58,012+02 INFO >> [org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand] (default task-14) >> [2997034] Running command: CreateUserSessionCommand internal: false. >> 2019-07-10 09:35:58,021+02 INFO >> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >> (default task-14) [2997034] EVENT_ID: USER_VDC_LOGIN(30), User >> trouble@internal-authz connecting from '160.128.20.85' using session >> 'bv55G0wZznETUiQwjgjfUNje7wOsG4UDCuFunSslVeAFQkhdY2zzTY7du36ynTF5nW5U7JiPyr7gl9QDHfWuig==' >> logged in. >> 2019-07-10 09:36:58,304+02 INFO >> [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] >> (EE-ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] Thread pool >> 'default' is using 0 threads out of 1, 5 threads waiting for tasks. >> 2019-07-10 09:36:58,305+02 INFO >> [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] >> (EE-ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] Thread pool >> 'engine' is using 0 threads out of 500, 16 threads waiting for tasks and 0 >> tasks in queue. >> 2019-07-10 09:36:58,305+02 INFO >> [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] >> (EE-ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] Thread pool >> 'engineScheduled' is using 0 threads out of 100, 100 threads waiting for >> tasks. >> 2019-07-10 09:36:58,305+02 INFO >> [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] >> (EE-ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] Thread pool >> 'engineThreadMonitoring' is using 1 threads out of 1, 0 threads waiting for >> tasks. >> 2019-07-10 09:36:58,305+02 INFO >> [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] >> (EE-ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] Thread pool >> 'hostUpdatesChecker' is using 0 threads out of 5, 2 threads waiting for >> tasks. >> >> The same is observed in the vdsm.log too, below is the log during the >> attempted migration.... >> >> 2019-07-10 09:39:57,034+0200 INFO (jsonrpc/7) [jsonrpc.JsonRpcServer] >> RPC call Host.getStats succeeded in 0.01 seconds (__init__:573) >> 2019-07-10 09:39:57,994+0200 INFO (jsonrpc/2) [api.host] START >> getStats() from=::ffff:10.0.1.1,57934 (api:46) >> 2019-07-10 09:39:57,994+0200 INFO (jsonrpc/2) [vdsm.api] START >> repoStats(domains=()) from=::ffff:10.0.1.1,57934, >> task_id=e2529cfc-4293-42b4-91fa-7f5558e279dd (api:46) >> 2019-07-10 09:39:57,994+0200 INFO (jsonrpc/2) [vdsm.api] FINISH >> repoStats return={u'8a607f8a-542a-473c-bb18-25c05fe2a3d4': {'code': 0, >> 'actual': True, 'version': 4, 'acquired': True, 'delay': '0.000194846', >> 'lastCheck': '2.4', 'valid': True}, >> u'37b1a5d7-4e29-4763-9337-63c51dbc5fc8': {'code': 0, 'actual': True, >> 'version': 0, 'acquired': True, 'delay': '0.000277154', 'lastCheck': '6.0', >> 'valid': True}, u'2558679a-2214-466b-8f05-06fdda9146e5': {'code': 0, >> 'actual': True, 'version': 4, 'acquired': True, 'delay': '0.000421988', >> 'lastCheck': '2.4', 'valid': True}, >> u'640a5875-3d82-43c0-860f-7bb3e4a7e6f0': {'code': 0, 'actual': True, >> 'version': 4, 'acquired': True, 'delay': '0.000228443', 'lastCheck': '2.4', >> 'valid': True}} from=::ffff:10.0.1.1,57934, >> task_id=e2529cfc-4293-42b4-91fa-7f5558e279dd (api:52) >> 2019-07-10 09:39:57,995+0200 INFO (jsonrpc/2) [vdsm.api] START >> multipath_health() from=::ffff:10.0.1.1,57934, >> task_id=fd7ad703-5096-4f09-99fa-54672cb4aad9 (api:46) >> 2019-07-10 09:39:57,995+0200 INFO (jsonrpc/2) [vdsm.api] FINISH >> multipath_health return={} from=::ffff:10.0.1.1,57934, >> task_id=fd7ad703-5096-4f09-99fa-54672cb4aad9 (api:52) >> 2019-07-10 09:39:58,002+0200 INFO (jsonrpc/2) [api.host] FINISH getStats >> return={'status': {'message': 'Done', 'code': 0}, 'info': {'cpuStatistics': >> {'42': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.13', 'cpuIdle': >> '99.87'}, '43': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', >> 'cpuIdle': '100.00'}, '24': {'cpuUser': '0.73', 'nodeIndex': 0, 'cpuSys': >> '0.07', 'cpuIdle': '99.20'}, '25': {'cpuUser': '0.07', 'nodeIndex': 1, >> 'cpuSys': '0.00', 'cpuIdle': '99.93'}, '26': {'cpuUser': '5.59', >> 'nodeIndex': 0, 'cpuSys': '1.20', 'cpuIdle': '93.21'}, '27': {'cpuUser': >> '0.87', 'nodeIndex': 1, 'cpuSys': '0.60', 'cpuIdle': '98.53'}, '20': >> {'cpuUser': '0.53', 'nodeIndex': 0, 'cpuSys': '0.13', 'cpuIdle': '99.34'}, >> '21': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.07', 'cpuIdle': >> '99.93'}, '22': {'cpuUser': '0.40', 'nodeIndex': 0, 'cpuSys': '0.20', >> 'cpuIdle': '99.40'}, '23': {'cpuUser': '0.07', 'nodeIndex': 1, 'cpuSys': >> '0.07', 'cpuIdle': '99.86'}, '46': {'cpuUser': '0.13', 'nodeIndex': 0, >> 'cpuSys': '0.00', 'cpuIdle': '99.87'}, '47': {'cpuUser': '0.00', >> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '44': {'cpuUser': >> '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '45': >> {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, >> '28': {'cpuUser': '0.60', 'nodeIndex': 0, 'cpuSys': '0.07', 'cpuIdle': >> '99.33'}, '29': {'cpuUser': '1.07', 'nodeIndex': 1, 'cpuSys': '0.20', >> 'cpuIdle': '98.73'}, '40': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': >> '0.00', 'cpuIdle': '100.00'}, '41': {'cpuUser': '0.00', 'nodeIndex': 1, >> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '1': {'cpuUser': '1.07', >> 'nodeIndex': 1, 'cpuSys': '1.13', 'cpuIdle': '97.80'}, '0': {'cpuUser': >> '0.60', 'nodeIndex': 0, 'cpuSys': '0.20', 'cpuIdle': '99.20'}, '3': >> {'cpuUser': '0.20', 'nodeIndex': 1, 'cpuSys': '0.07', 'cpuIdle': '99.73'}, >> '2': {'cpuUser': '3.00', 'nodeIndex': 0, 'cpuSys': '0.53', 'cpuIdle': >> '96.47'}, '5': {'cpuUser': '0.20', 'nodeIndex': 1, 'cpuSys': '0.13', >> 'cpuIdle': '99.67'}, '4': {'cpuUser': '0.47', 'nodeIndex': 0, 'cpuSys': >> '0.20', 'cpuIdle': '99.33'}, '7': {'cpuUser': '0.40', 'nodeIndex': 1, >> 'cpuSys': '0.20', 'cpuIdle': '99.40'}, '6': {'cpuUser': '0.67', >> 'nodeIndex': 0, 'cpuSys': '0.20', 'cpuIdle': '99.13'}, '9': {'cpuUser': >> '0.47', 'nodeIndex': 1, 'cpuSys': '0.40', 'cpuIdle': '99.13'}, '8': >> {'cpuUser': '0.13', 'nodeIndex': 0, 'cpuSys': '0.07', 'cpuIdle': '99.80'}, >> '39': {'cpuUser': '0.33', 'nodeIndex': 1, 'cpuSys': '0.13', 'cpuIdle': >> '99.54'}, '38': {'cpuUser': '0.07', 'nodeIndex': 0, 'cpuSys': '0.00', >> 'cpuIdle': '99.93'}, '11': {'cpuUser': '0.67', 'nodeIndex': 1, 'cpuSys': >> '0.27', 'cpuIdle': '99.06'}, '10': {'cpuUser': '0.13', 'nodeIndex': 0, >> 'cpuSys': '0.13', 'cpuIdle': '99.74'}, '13': {'cpuUser': '0.07', >> 'nodeIndex': 1, 'cpuSys': '0.07', 'cpuIdle': '99.86'}, '12': {'cpuUser': >> '0.07', 'nodeIndex': 0, 'cpuSys': '0.27', 'cpuIdle': '99.66'}, '15': >> {'cpuUser': '0.27', 'nodeIndex': 1, 'cpuSys': '0.13', 'cpuIdle': '99.60'}, >> '14': {'cpuUser': '0.27', 'nodeIndex': 0, 'cpuSys': '0.07', 'cpuIdle': >> '99.66'}, '17': {'cpuUser': '0.07', 'nodeIndex': 1, 'cpuSys': '0.27', >> 'cpuIdle': '99.66'}, '16': {'cpuUser': '0.53', 'nodeIndex': 0, 'cpuSys': >> '0.07', 'cpuIdle': '99.40'}, '19': {'cpuUser': '0.00', 'nodeIndex': 1, >> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '18': {'cpuUser': '1.00', >> 'nodeIndex': 0, 'cpuSys': '0.27', 'cpuIdle': '98.73'}, '31': {'cpuUser': >> '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '30': >> {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, >> '37': {'cpuUser': '0.07', 'nodeIndex': 1, 'cpuSys': '0.07', 'cpuIdle': >> '99.86'}, '36': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', >> 'cpuIdle': '100.00'}, '35': {'cpuUser': '0.20', 'nodeIndex': 1, 'cpuSys': >> '0.33', 'cpuIdle': '99.47'}, '34': {'cpuUser': '0.00', 'nodeIndex': 0, >> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '33': {'cpuUser': '0.07', >> 'nodeIndex': 1, 'cpuSys': '0.13', 'cpuIdle': '99.80'}, '32': {'cpuUser': >> '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}}, >> 'numaNodeMemFree': {'1': {'memPercent': 5, 'memFree': '94165'}, '0': >> {'memPercent': 22, 'memFree': '77122'}}, 'memShared': 0, 'haScore': 3400, >> 'thpState': 'always', 'ksmMergeAcrossNodes': True, 'vmCount': 2, 'memUsed': >> '11', 'storageDomains': {u'8a607f8a-542a-473c-bb18-25c05fe2a3d4': {'code': >> 0, 'actual': True, 'version': 4, 'acquired': True, 'delay': '0.000194846', >> 'lastCheck': '2.4', 'valid': True}, >> u'37b1a5d7-4e29-4763-9337-63c51dbc5fc8': {'code': 0, 'actual': True, >> 'version': 0, 'acquired': True, 'delay': '0.000277154', 'lastCheck': '6.0', >> 'valid': True}, u'2558679a-2214-466b-8f05-06fdda9146e5': {'code': 0, >> 'actual': True, 'version': 4, 'acquired': True, 'delay': '0.000421988', >> 'lastCheck': '2.4', 'valid': True}, >> u'640a5875-3d82-43c0-860f-7bb3e4a7e6f0': {'code': 0, 'actual': True, >> 'version': 4, 'acquired': True, 'delay': '0.000228443', 'lastCheck': '2.4', >> 'valid': True}}, 'incomingVmMigrations': 0, 'network': {'em4': {'txErrors': >> '0', 'state': 'up', 'sampleTime': 1562744396.40508, 'name': 'em4', 'tx': >> '2160', 'txDropped': '0', 'rx': '261751836', 'rxErrors': '0', 'speed': >> '1000', 'rxDropped': '1'}, 'ovirtmgmt': {'txErrors': '0', 'state': 'up', >> 'sampleTime': 1562744396.40508, 'name': 'ovirtmgmt', 'tx': '193005142', >> 'txDropped': '0', 'rx': '4300879104', 'rxErrors': '0', 'speed': '1000', >> 'rxDropped': '478'}, 'restores': {'txErrors': '0', 'state': 'up', >> 'sampleTime': 1562744396.40508, 'name': 'restores', 'tx': '1362', >> 'txDropped': '0', 'rx': '226442665', 'rxErrors': '0', 'speed': '1000', >> 'rxDropped': '478'}, 'em2': {'txErrors': '0', 'state': 'down', >> 'sampleTime': 1562744396.40508, 'name': 'em2', 'tx': '0', 'txDropped': '0', >> 'rx': '0', 'rxErrors': '0', 'speed': '1000', 'rxDropped': '0'}, 'vnet0': >> {'txErrors': '0', 'state': 'up', 'sampleTime': 1562744396.40508, 'name': >> 'vnet0', 'tx': '2032610435', 'txDropped': '686', 'rx': '4287479548', >> 'rxErrors': '0', 'speed': '1000', 'rxDropped': '0'}, ';vdsmdummy;': >> {'txErrors': '0', 'state': 'down', 'sampleTime': 1562744396.40508, 'name': >> ';vdsmdummy;', 'tx': '0', 'txDropped': '0', 'rx': '0', 'rxErrors': '0', >> 'speed': '1000', 'rxDropped': '0'}, 'em1': {'txErrors': '0', 'state': 'up', >> 'sampleTime': 1562744396.40508, 'name': 'em1', 'tx': '4548433238', >> 'txDropped': '0', 'rx': '6476729588', 'rxErrors': '0', 'speed': '1000', >> 'rxDropped': '1'}, 'em3': {'txErrors': '0', 'state': 'down', 'sampleTime': >> 1562744396.40508, 'name': 'em3', 'tx': '0', 'txDropped': '0', 'rx': '0', >> 'rxErrors': '0', 'speed': '1000', 'rxDropped': '0'}, 'lo': {'txErrors': >> '0', 'state': 'up', 'sampleTime': 1562744396.40508, 'name': 'lo', 'tx': >> '397962377', 'txDropped': '0', 'rx': '397962377', 'rxErrors': '0', 'speed': >> '1000', 'rxDropped': '0'}, 'vnet1': {'txErrors': '0', 'state': 'up', >> 'sampleTime': 1562744396.40508, 'name': 'vnet1', 'tx': '526185708', >> 'txDropped': '0', 'rx': '118512222', 'rxErrors': '0', 'speed': '1000', >> 'rxDropped': '0'}}, 'txDropped': '686', 'anonHugePages': '18532', >> 'ksmPages': 100, 'elapsedTime': '85176.64', 'cpuLoad': '0.06', 'cpuSys': >> '0.17', 'diskStats': {'/var/log': {'free': '6850'}, '/var/run/vdsm/': >> {'free': '96410'}, '/tmp': {'free': '1825'}}, 'cpuUserVdsmd': '1.07', >> 'netConfigDirty': 'False', 'memCommitted': 24706, 'ksmState': False, >> 'vmMigrating': 0, 'ksmCpu': 0, 'memAvailable': 166010, 'bootTime': >> '1562659184', 'haStats': {'active': True, 'configured': True, 'score': >> 3400, 'localMaintenance': False, 'globalMaintenance': False}, 'momStatus': >> 'active', 'multipathHealth': {}, 'rxDropped': '958', >> 'outgoingVmMigrations': 0, 'swapTotal': 4095, 'swapFree': 4095, >> 'hugepages': defaultdict(<type 'dict'>, {1048576: {'resv_hugepages': 0, >> 'free_hugepages': 0, 'nr_overcommit_hugepages': 0, 'surplus_hugepages': 0, >> 'vm.free_hugepages': 0, 'nr_hugepages': 0, 'nr_hugepages_mempolicy': 0}, >> 2048: {'resv_hugepages': 0, 'free_hugepages': 0, 'nr_overcommit_hugepages': >> 0, 'surplus_hugepages': 0, 'vm.free_hugepages': 0, 'nr_hugepages': 0, >> 'nr_hugepages_mempolicy': 0}}), 'dateTime': '2019-07-10T07:39:57 GMT', >> 'cpuUser': '0.44', 'memFree': 172451, 'cpuIdle': '99.39', 'vmActive': 2, >> 'v2vJobs': {}, 'cpuSysVdsmd': '0.60'}} from=::ffff:10.0.1.1,57934 (api:52) >> 2019-07-10 09:39:58,004+0200 INFO (jsonrpc/2) [jsonrpc.JsonRpcServer] >> RPC call Host.getStats succeeded in 0.01 seconds (__init__:573) >> >> Please let me know if you need further info. >> >> Thank you. >> >> Regards. >> >> Neil Wilson >> >> >> On Tue, Jul 9, 2019 at 5:52 PM Michal Skrivanek < >> [email protected]> wrote: >> >>> Can you share the engine.log please? And highlight the exact time when >>> you attempt that migrate action >>> >>> Thanks, >>> michal >>> >>> > On 9 Jul 2019, at 16:42, Neil <[email protected]> wrote: >>> > >>> > --000000000000166784058d409302 >>> > Content-Type: text/plain; charset="UTF-8" >>> > Content-Transfer-Encoding: quoted-printable >>> > >>> > I remember seeing the bug earlier but because it was closed thought it >>> was >>> > unrelated, this appears to be it.... >>> > >>> > https://bugzilla.redhat.com/show_bug.cgi?id=3D1670701 >>> > >>> > Perhaps I'm not understanding your question about the VM guest agent, >>> but I >>> > don't have any guest agent currently installed on the VM, not sure if >>> the >>> > output of my qemu-kvm process maybe answers this question?.... >>> > >>> > /usr/libexec/qemu-kvm -name >>> guest=3DHeadoffice.cbl-ho.local,debug-threads= >>> > =3Don >>> > -S -object >>> > >>> secret,id=3DmasterKey0,format=3Draw,file=3D/var/lib/libvirt/qemu/domain-1-H= >>> > eadoffice.cbl-ho.lo/master-key.aes >>> > -machine >>> pc-i440fx-rhel7.3.0,accel=3Dkvm,usb=3Doff,dump-guest-core=3Doff -c= >>> > pu >>> > >>> Broadwell,vme=3Don,f16c=3Don,rdrand=3Don,hypervisor=3Don,arat=3Don,xsaveopt= >>> > =3Don,abm=3Don,rtm=3Don,hle=3Don >>> > -m 8192 -realtime mlock=3Doff -smp >>> 8,maxcpus=3D64,sockets=3D16,cores=3D4,th= >>> > reads=3D1 >>> > -numa node,nodeid=3D0,cpus=3D0-7,mem=3D8192 -uuid >>> > 9a6561b8-5702-43dc-9e92-1dc5dfed4eef -smbios >>> > type=3D1,manufacturer=3DoVirt,product=3DoVirt >>> > >>> Node,version=3D7-3.1611.el7.centos,serial=3D4C4C4544-0034-5810-8033-C2C04F4= >>> > E4B32,uuid=3D9a6561b8-5702-43dc-9e92-1dc5dfed4eef >>> > -no-user-config -nodefaults -chardev >>> > socket,id=3Dcharmonitor,fd=3D31,server,nowait -mon >>> > chardev=3Dcharmonitor,id=3Dmonitor,mode=3Dcontrol -rtc >>> > base=3D2019-07-09T10:26:53,driftfix=3Dslew -global >>> > kvm-pit.lost_tick_policy=3Ddelay -no-hpet -no-shutdown -boot >>> strict=3Don >>> > -device piix3-usb-uhci,id=3Dusb,bus=3Dpci.0,addr=3D0x1.0x2 -device >>> > virtio-scsi-pci,id=3Dscsi0,bus=3Dpci.0,addr=3D0x4 -device >>> > >>> virtio-serial-pci,id=3Dvirtio-serial0,max_ports=3D16,bus=3Dpci.0,addr=3D0x5= >>> > -drive >>> > if=3Dnone,id=3Ddrive-ide0-1-0,readonly=3Don -device >>> > ide-cd,bus=3Dide.1,unit=3D0,drive=3Ddrive-ide0-1-0,id=3Dide0-1-0 -drive >>> > >>> file=3D/rhev/data-center/59831b91-00a5-01e4-0294-000000000018/8a607f8a-542a= >>> > >>> -473c-bb18-25c05fe2a3d4/images/56e8240c-a172-4f52-b0c1-2bddc4f34f93/9f24546= >>> > >>> 7-d31d-4f5a-8037-7c5012a4aa84,format=3Dqcow2,if=3Dnone,id=3Ddrive-virtio-di= >>> > >>> sk0,serial=3D56e8240c-a172-4f52-b0c1-2bddc4f34f93,werror=3Dstop,rerror=3Dst= >>> > op,cache=3Dnone,aio=3Dnative >>> > -device >>> > >>> virtio-blk-pci,scsi=3Doff,bus=3Dpci.0,addr=3D0x7,drive=3Ddrive-virtio-disk0= >>> > ,id=3Dvirtio-disk0,bootindex=3D1,write-cache=3Don >>> > -netdev tap,fd=3D33,id=3Dhostnet0,vhost=3Don,vhostfd=3D34 -device >>> > >>> virtio-net-pci,netdev=3Dhostnet0,id=3Dnet0,mac=3D00:1a:4a:16:01:5b,bus=3Dpc= >>> > i.0,addr=3D0x3 >>> > -chardev socket,id=3Dcharchannel0,fd=3D35,server,nowait -device >>> > >>> virtserialport,bus=3Dvirtio-serial0.0,nr=3D1,chardev=3Dcharchannel0,id=3Dch= >>> > annel0,name=3Dcom.redhat.rhevm.vdsm >>> > -chardev socket,id=3Dcharchannel1,fd=3D36,server,nowait -device >>> > >>> virtserialport,bus=3Dvirtio-serial0.0,nr=3D2,chardev=3Dcharchannel1,id=3Dch= >>> > annel1,name=3Dorg.qemu.guest_agent.0 >>> > -chardev spicevmc,id=3Dcharchannel2,name=3Dvdagent -device >>> > >>> virtserialport,bus=3Dvirtio-serial0.0,nr=3D3,chardev=3Dcharchannel2,id=3Dch= >>> > annel2,name=3Dcom.redhat.spice.0 >>> > -spice >>> > >>> tls-port=3D5900,addr=3D10.0.1.11,x509-dir=3D/etc/pki/vdsm/libvirt-spice,tls= >>> > >>> -channel=3Ddefault,tls-channel=3Dmain,tls-channel=3Ddisplay,tls-channel=3Di= >>> > >>> nputs,tls-channel=3Dcursor,tls-channel=3Dplayback,tls-channel=3Drecord,tls-= >>> > channel=3Dsmartcard,tls-channel=3Dusbredir,seamless-migration=3Don >>> > -device >>> > >>> qxl-vga,id=3Dvideo0,ram_size=3D67108864,vram_size=3D8388608,vram64_size_mb= >>> > =3D0,vgamem_mb=3D16,max_outputs=3D1,bus=3Dpci.0,addr=3D0x2 >>> > -incoming defer -device >>> virtio-balloon-pci,id=3Dballoon0,bus=3Dpci.0,addr= >>> > =3D0x6 >>> > -object rng-random,id=3Dobjrng0,filename=3D/dev/urandom -device >>> > virtio-rng-pci,rng=3Dobjrng0,id=3Drng0,bus=3Dpci.0,addr=3D0x8 -sandbox >>> > >>> on,obsolete=3Ddeny,elevateprivileges=3Ddeny,spawn=3Ddeny,resourcecontrol=3D= >>> > deny >>> > -msg timestamp=3Don >>> > >>> > Please shout if you need further info. >>> > >>> > Thanks. >>> > >>> > >>> > >>> > >>> > >>> > >>> > On Tue, Jul 9, 2019 at 4:17 PM Strahil Nikolov <[email protected]> >>> > wrote: >>> > >>> >> Shouldn't cause that problem. >>> >> >>> >> You have to find the bug in bugzilla and report a regression (if it's >>> not >>> >> closed) , or open a new one and report the regression. >>> >> As far as I remember , only the dashboard was affected due to new >>> feature= >>> > s >>> >> about vdo disk savings. >>> >> >>> >> About the VM - this should be another issue. What agent are you using >>> in >>> >> the VMs (ovirt or qemu) ? >>> >> >>> >> Best Regards, >>> >> Strahil Nikolov >>> >> >>> >> =D0=92 =D0=B2=D1=82=D0=BE=D1=80=D0=BD=D0=B8=D0=BA, 9 >>> =D1=8E=D0=BB=D0=B8 2= >>> > 019 =D0=B3., 10:09:05 =D1=87. >>> =D0=93=D1=80=D0=B8=D0=BD=D1=83=D0=B8=D1=87-4,= >>> > Neil < >>> >> [email protected]> =D0=BD=D0=B0=D0=BF=D0=B8=D1=81=D0=B0: >>> >> >>> >> >>> >> Hi Strahil, >>> >> >>> >> Thanks for the quick reply. >>> >> I put the cluster into global maintenance, then installed the 4.3 >>> repo, >>> >> then "yum update ovirt\*setup\*" then "engine-upgrade-check", >>> >> "engine-setup", then "yum update", once completed, I rebooted the >>> >> hosted-engine VM, and took the cluster out of global maintenance. >>> >> >>> >> Thinking back to the upgrade from 4.1 to 4.2 I don't recall doing a >>> "yum >>> >> update" after doing the engine-setup, not sure if this would cause it >>> >> perhaps? >>> >> >>> >> Thank you. >>> >> Regards. >>> >> Neil Wilson. >>> >> >>> >> On Tue, Jul 9, 2019 at 3:47 PM Strahil Nikolov <[email protected] >>> > >>> >> wrote: >>> >> >>> >> Hi Neil, >>> >> >>> >> for "Could not fetch data needed for VM migrate operation" - there >>> was a >>> >> bug and it was fixed. >>> >> Are you sure you have fully updated ? >>> >> What procedure did you use ? >>> >> >>> >> Best Regards, >>> >> Strahil Nikolov >>> >> >>> >> =D0=92 =D0=B2=D1=82=D0=BE=D1=80=D0=BD=D0=B8=D0=BA, 9 >>> =D1=8E=D0=BB=D0=B8 2= >>> > 019 =D0=B3., 7:26:21 =D1=87. >>> =D0=93=D1=80=D0=B8=D0=BD=D1=83=D0=B8=D1=87-4, = >>> > Neil <[email protected]> >>> >> =D0=BD=D0=B0=D0=BF=D0=B8=D1=81=D0=B0: >>> >> >>> >> >>> >> Hi guys. >>> >> >>> >> I have two problems since upgrading from 4.2.x to 4.3.4 >>> >> >>> >> First issue is I can no longer manually migrate VM's between hosts, I >>> get >>> >> an error in the ovirt GUI that says "Could not fetch data needed for >>> VM >>> >> migrate operation" and nothing gets logged either in my engine.log or >>> my >>> >> vdsm.log >>> >> >>> >> Then the other issue is my Dashboard says the following "Error! Could >>> not >>> >> fetch dashboard data. Please ensure that data warehouse is properly >>> >> installed and configured." >>> >> >>> >> If I look at my ovirt-engine-dwhd.log I see the following if I try >>> restar= >>> > t >>> >> the dwh service... >>> >> >>> >> 2019-07-09 11:48:04|ETL Service Started >>> >> ovirtEngineDbDriverClass|org.postgresql.Driver >>> >> >>> >> >>> ovirtEngineHistoryDbJdbcConnection|jdbc:postgresql://localhost:5432/ovirt= >>> > _engine_history?sslfactory=3Dorg.postgresql.ssl.NonValidatingFactory >>> >> hoursToKeepDaily|0 >>> >> hoursToKeepHourly|720 >>> >> ovirtEngineDbPassword|********************** >>> >> runDeleteTime|3 >>> >> >>> >> >>> ovirtEngineDbJdbcConnection|jdbc:postgresql://localhost:5432/engine?sslfa= >>> > ctory=3Dorg.postgresql.ssl.NonValidatingFactory >>> >> runInterleave|60 >>> >> limitRows|limit 1000 >>> >> ovirtEngineHistoryDbUser|ovirt_engine_history >>> >> ovirtEngineDbUser|engine >>> >> deleteIncrement|10 >>> >> timeBetweenErrorEvents|300000 >>> >> hoursToKeepSamples|24 >>> >> deleteMultiplier|1000 >>> >> lastErrorSent|2011-07-03 12:46:47.000000 >>> >> etlVersion|4.3.0 >>> >> dwhAggregationDebug|false >>> >> dwhUuid|dca0ebd3-c58f-4389-a1f8-6aecc20b1316 >>> >> ovirtEngineHistoryDbDriverClass|org.postgresql.Driver >>> >> ovirtEngineHistoryDbPassword|********************** >>> >> 2019-07-09 11:48:10|ETL Service Stopped >>> >> 2019-07-09 11:49:59|ETL Service Started >>> >> ovirtEngineDbDriverClass|org.postgresql.Driver >>> >> >>> >> >>> ovirtEngineHistoryDbJdbcConnection|jdbc:postgresql://localhost:5432/ovirt= >>> > _engine_history?sslfactory=3Dorg.postgresql.ssl.NonValidatingFactory >>> >> hoursToKeepDaily|0 >>> >> hoursToKeepHourly|720 >>> >> ovirtEngineDbPassword|********************** >>> >> runDeleteTime|3 >>> >> >>> >> >>> ovirtEngineDbJdbcConnection|jdbc:postgresql://localhost:5432/engine?sslfa= >>> > ctory=3Dorg.postgresql.ssl.NonValidatingFactory >>> >> runInterleave|60 >>> >> limitRows|limit 1000 >>> >> ovirtEngineHistoryDbUser|ovirt_engine_history >>> >> ovirtEngineDbUser|engine >>> >> deleteIncrement|10 >>> >> timeBetweenErrorEvents|300000 >>> >> hoursToKeepSamples|24 >>> >> deleteMultiplier|1000 >>> >> lastErrorSent|2011-07-03 12:46:47.000000 >>> >> etlVersion|4.3.0 >>> >> dwhAggregationDebug|false >>> >> dwhUuid|dca0ebd3-c58f-4389-a1f8-6aecc20b1316 >>> >> ovirtEngineHistoryDbDriverClass|org.postgresql.Driver >>> >> ovirtEngineHistoryDbPassword|********************** >>> >> 2019-07-09 11:52:56|ETL Service Stopped >>> >> 2019-07-09 11:52:57|ETL Service Started >>> >> ovirtEngineDbDriverClass|org.postgresql.Driver >>> >> >>> >> >>> ovirtEngineHistoryDbJdbcConnection|jdbc:postgresql://localhost:5432/ovirt= >>> > _engine_history?sslfactory=3Dorg.postgresql.ssl.NonValidatingFactory >>> >> hoursToKeepDaily|0 >>> >> hoursToKeepHourly|720 >>> >> ovirtEngineDbPassword|********************** >>> >> runDeleteTime|3 >>> >> >>> >> >>> ovirtEngineDbJdbcConnection|jdbc:postgresql://localhost:5432/engine?sslfa= >>> > ctory=3Dorg.postgresql.ssl.NonValidatingFactory >>> >> runInterleave|60 >>> >> limitRows|limit 1000 >>> >> ovirtEngineHistoryDbUser|ovirt_engine_history >>> >> ovirtEngineDbUser|engine >>> >> deleteIncrement|10 >>> >> timeBetweenErrorEvents|300000 >>> >> hoursToKeepSamples|24 >>> >> deleteMultiplier|1000 >>> >> lastErrorSent|2011-07-03 12:46:47.000000 >>> >> etlVersion|4.3.0 >>> >> dwhAggregationDebug|false >>> >> dwhUuid|dca0ebd3-c58f-4389-a1f8-6aecc20b1316 >>> >> ovirtEngineHistoryDbDriverClass|org.postgresql.Driver >>> >> ovirtEngineHistoryDbPassword|********************** >>> >> 2019-07-09 12:16:01|ETL Service Stopped >>> >> 2019-07-09 12:16:45|ETL Service Started >>> >> ovirtEngineDbDriverClass|org.postgresql.Driver >>> >> >>> >> >>> ovirtEngineHistoryDbJdbcConnection|jdbc:postgresql://localhost:5432/ovirt= >>> > _engine_history?sslfactory=3Dorg.postgresql.ssl.NonValidatingFactory >>> >> hoursToKeepDaily|0 >>> >> hoursToKeepHourly|720 >>> >> ovirtEngineDbPassword|********************** >>> >> runDeleteTime|3 >>> >> >>> >> >>> ovirtEngineDbJdbcConnection|jdbc:postgresql://localhost:5432/engine?sslfa= >>> > ctory=3Dorg.postgresql.ssl.NonValidatingFactory >>> >> runInterleave|60 >>> >> limitRows|limit 1000 >>> >> ovirtEngineHistoryDbUser|ovirt_engine_history >>> >> ovirtEngineDbUser|engine >>> >> deleteIncrement|10 >>> >> timeBetweenErrorEvents|300000 >>> >> hoursToKeepSamples|24 >>> >> deleteMultiplier|1000 >>> >> lastErrorSent|2011-07-03 12:46:47.000000 >>> >> etlVersion|4.3.0 >>> >> dwhAggregationDebug|false >>> >> dwhUuid|dca0ebd3-c58f-4389-a1f8-6aecc20b1316 >>> >> ovirtEngineHistoryDbDriverClass|org.postgresql.Driver >>> >> ovirtEngineHistoryDbPassword|********************** >>> >> >>> >> >>> >> >>> >> >>> >> >>> >> I have a hosted engine, and I have two hosts and my storage is FC >>> based. >>> >> The hosts are still running on 4.2 because I'm unable to migrate VM's >>> off= >>> >> _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/BPVRXXGDGZ6MYMCFUXLHBGVAO7BJJ75S/ >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/NUYBACV5Q3GGVW5NUYRVAR2MA233TFB4/

