On Tue, Dec 6, 2016 at 4:02 PM, Marcelo Leandro <[email protected]> wrote: > Hello. > > Logs after rotate.
Seem like the same files, only with more newer stuff. We need older ones, from the time frame when the migration failed. Thanks. > > Thanks > > > 2016-12-06 10:50 GMT-03:00 Yedidyah Bar David <[email protected]>: >> >> On Tue, Dec 6, 2016 at 2:14 PM, Marcelo Leandro <[email protected]> >> wrote: >> > Hello, >> > >> > I Tried this solution , but to some vms not was resolved. >> > Logs: >> > >> > src logs: >> > >> > Thread-12::DEBUG::2016-12-06 >> > 08:50:58,112::check::327::storage.check::(_check_completed) FINISH check >> > >> > u'/rhev/data-center/mnt/192.168.144.6:_home_iso/b5fa054f-0d3d-458b-a891-13fd9383ee7d/dom_md/metadata' >> > rc=0 err=bytearray(b'0+1 records in\n0+1 records out\n328 bytes (328 B) >> > copied, 0.000474535 s, 691 kB/s\n') elapsed=0.08 >> > Thread-2374888::WARNING::2016-12-06 >> > 08:50:58,815::migration::671::virt.vm::(monitor_migration) >> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::The migration took 520 >> > seconds >> > which is exceeding the configured maximum time for migrations of 512 >> > seconds. The migration will be aborted. >> > Thread-2374888::DEBUG::2016-12-06 >> > 08:50:58,816::migration::715::virt.vm::(stop) >> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration monitor >> > thread >> > Thread-2374888::DEBUG::2016-12-06 >> > 08:50:58,816::migration::570::virt.vm::(stop) >> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration downtime >> > thread >> > Thread-2374888::DEBUG::2016-12-06 >> > 08:50:58,817::migration::629::virt.vm::(run) >> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopped migration monitor >> > thread >> > Thread-2374886::DEBUG::2016-12-06 >> > 08:50:59,098::migration::715::virt.vm::(stop) >> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration monitor >> > thread >> > Thread-2374886::ERROR::2016-12-06 >> > 08:50:59,098::migration::252::virt.vm::(_recover) >> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::operation aborted: >> > migration >> > job: canceled by client >> > Thread-2374886::DEBUG::2016-12-06 >> > 08:50:59,098::stompreactor::408::jsonrpc.AsyncoreClient::(send) Sending >> > response >> > Thread-2374886::DEBUG::2016-12-06 >> > 08:50:59,321::__init__::208::jsonrpc.Notification::(emit) Sending event >> > {"params": {"notify_time": 6040272640, >> > "9b5ab7b4-1045-4858-b24c-1f5a9f6172c3": {"status": "Migration Source"}}, >> > "jsonrpc": "2.0", "method": >> > "|virt|VM_status|9b5ab7b4-1045-4858-b24c-1f5a9f6172c3"} >> > Thread-2374886::ERROR::2016-12-06 >> > 08:50:59,322::migration::381::virt.vm::(run) >> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Failed to migrate >> > Traceback (most recent call last): >> > File "/usr/share/vdsm/virt/migration.py", line 363, in run >> > self._startUnderlyingMigration(time.time()) >> > File "/usr/share/vdsm/virt/migration.py", line 438, in >> > _startUnderlyingMigration >> > self._perform_with_downtime_thread(duri, muri) >> > File "/usr/share/vdsm/virt/migration.py", line 489, in >> > _perform_with_downtime_thread >> > self._perform_migration(duri, muri) >> > File "/usr/share/vdsm/virt/migration.py", line 476, in >> > _perform_migration >> > self._vm._dom.migrateToURI3(duri, params, flags) >> > File "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line >> > 69, >> > in f >> > ret = attr(*args, **kwargs) >> > File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py", >> > line >> > 123, in wrapper >> > ret = f(*args, **kwargs) >> > File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 916, in >> > wrapper >> > return func(inst, *args, **kwargs) >> > File "/usr/lib64/python2.7/site-packages/libvirt.py", line 1836, in >> > migrateToURI3 >> > if ret == -1: raise libvirtError ('virDomainMigrateToURI3() failed', >> > dom=self) >> > libvirtError: operation aborted: migration job: canceled by client >> > Thread-12::DEBUG::2016-12-06 >> > 08:50:59,875::check::296::storage.check::(_start_process) START check >> > '/dev/c2dc0101-748e-4a7b-9913-47993eaa52bd/metadata' >> > cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd', >> > 'if=/dev/c2dc0101-748e-4a7b-9913-47993eaa52bd/metadata', 'of=/dev/null', >> > 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00 >> > mailbox.SPMMonitor::DEBUG::2016-12-06 >> > 08:50:59,914::storage_mailbox::733::Storage.Misc.excCmd::(_checkForMail) >> > /usr/bin/taskset --cpu-list 0-31 dd >> > >> > if=/rhev/data-center/77e24b20-9d21-4952-a089-3c5c592b4e6d/mastersd/dom_md/inbox >> > iflag=direct,fullblock count=1 bs=1024000 (cwd None) >> >> This snippet is not enough, and the attached logs are too new. Can you >> check/share >> more of the relevant log? Thanks. >> >> > >> > dst log: >> > >> > libvirtEventLoop::DEBUG::2016-12-06 >> > 08:50:59,080::task::995::Storage.TaskManager.Task::(_decref) >> > Task=`7446f040-c5f4-497c-b4a5-8934921a7b89`::ref 1 aborting False >> > libvirtEventLoop::DEBUG::2016-12-06 >> > 08:50:59,080::fileUtils::190::Storage.fileUtils::(cleanupdir) Removing >> > directory: /var/run/vdsm/storage/6e5cce71-3438-4045-9d54-6071 >> > 23e0557e/413a560d-4919-4870-88d6-f7fedbb77523 >> > libvirtEventLoop::DEBUG::2016-12-06 >> > 08:50:59,080::lvm::288::Storage.Misc.excCmd::(cmd) /usr/bin/taskset >> > --cpu-list 0-31 /usr/bin/sudo -n /usr/sbin/lvm lvs --config ' de >> > vices { preferred_names = ["^/dev/mapper/"] ignore_suspended_devices=1 >> > write_cache_state=0 disable_after_error_count=3 filter = [ >> > '\''a|/dev/mapper/36005076300810a4db80 >> > 0000000000002|'\'', '\''r|.*|'\'' ] } global { locking_type=1 >> > prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 } backup { >> > retain_min = 50 retain_days = 0 >> > } ' --noheadings --units b --nosuffix --separator '|' >> > --ignoreskippedcluster -o >> > uuid,name,vg_name,attr,size,seg_start_pe,devices,tags >> > 6e5cce71-3438-4045-9d54-607123e05 >> > 57e (cwd None) >> > jsonrpc.Executor/6::DEBUG::2016-12-06 >> > 08:50:59,090::__init__::530::jsonrpc.JsonRpcServer::(_handle_request) >> > Calling 'VM.destroy' in bridge with {u'vmID': u'9b5ab7b4-104 >> > 5-4858-b24c-1f5a9f6172c3'} >> > jsonrpc.Executor/6::DEBUG::2016-12-06 >> > 08:50:59,090::API::314::vds::(destroy) >> > About to destroy VM 9b5ab7b4-1045-4858-b24c-1f5a9f6172c3 >> > jsonrpc.Executor/6::DEBUG::2016-12-06 >> > 08:50:59,091::vm::4171::virt.vm::(destroy) >> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::destroy Called >> > Thread-51550::ERROR::2016-12-06 >> > 08:50:59,091::vm::759::virt.vm::(_startUnderlyingVm) >> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Failed to start a migration >> > destinatio >> > n vm >> > Traceback (most recent call last): >> > File "/usr/share/vdsm/virt/vm.py", line 725, in _startUnderlyingVm >> > self._completeIncomingMigration() >> > File "/usr/share/vdsm/virt/vm.py", line 3071, in >> > _completeIncomingMigration >> > self._incomingMigrationFinished.isSet(), usedTimeout) >> > File "/usr/share/vdsm/virt/vm.py", line 3154, in >> > _attachLibvirtDomainAfterMigration >> > raise MigrationError(e.get_error_message()) >> > MigrationError: Domain not found: no domain with matching uuid >> > '9b5ab7b4-1045-4858-b24c-1f5a9f6172c3' >> > Thread-51550::INFO::2016-12-06 >> > 08:50:59,093::vm::1308::virt.vm::(setDownStatus) >> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Changed state to Down: VM >> > failed to migrate (code=8) >> > >> > >> > >> > Logs Attached. >> > >> > any ideas? >> > >> > Very thanks. >> > >> > >> > 2016-12-06 4:16 GMT-03:00 Yedidyah Bar David <[email protected]>: >> >> >> >> On Mon, Dec 5, 2016 at 10:25 PM, Marcelo Leandro >> >> <[email protected]> >> >> wrote: >> >> > Hello >> >> > I am with problem, I can not migrate vm. >> >> > Can someone help me? >> >> > >> >> > Message in log: >> >> > >> >> > Src vdsm log: >> >> > >> >> > Thread-89514 :: WARNING :: 2016-12-05 17: 01: 49,542 :: migration :: >> >> > 683 >> >> > :: >> >> > virt.vm ::: monitor_migration vmId = >> >> > `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` >> >> > :: Migration Stalling: remaining >> >> > (56MiB)> lowmark (2MiB). Refer to RHBZ # 919201. >> >> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,543 :: migration :: >> >> > 689 >> >> > :: >> >> > virt.vm ::: monitor_migration vmId = >> >> > `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` >> >> > :: new Iteration detected: 15 >> >> > Thread-89514 :: WARNING :: 2016-12-05 17: 01: 49,543 :: migration :: >> >> > 704 >> >> > :: >> >> > virt.vm ::: monitor_migration vmId = >> >> > `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` >> >> > :: Migration Is stuck: Has not pro >> >> > Gressed in 240.071660042 seconds. Aborting. >> >> >> >> This is usually a result of a too-busy VM, changing its memory faster >> >> than >> >> the migration process can copy the changes to the destination. >> >> >> >> You can try changing the cluster migration policy to "suspend workload >> >> if needed". >> >> >> >> For more details/background, see also: >> >> >> >> >> >> >> >> https://www.ovirt.org/develop/release-management/features/migration-enhancements/ >> >> >> >> Best, >> >> >> >> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,544 :: migration :: >> >> > 715 >> >> > :: >> >> > virt.vm ::: stop} vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` :: >> >> > stopping >> >> > Migration monitor thread >> >> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,545 :: migration :: >> >> > 570 >> >> > :: >> >> > virt.vm ::: stop) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` :: >> >> > stopping >> >> > Migration downtime thread >> >> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,545 :: migration :: >> >> > 629 >> >> > :: >> >> > virt.vm ::: (run) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` :: >> >> > stopped >> >> > Migration monitor thread >> >> > Thread-89513 :: DEBUG :: 2016-12-05 17: 01: 49,766 :: migration :: >> >> > 715 >> >> > :: >> >> > virt.vm ::: stop} vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` :: >> >> > stopping >> >> > Migration monitor thread >> >> > Thread-89513 :: ERROR :: 2016-12-05 17: 01: 49,767 :: migration :: >> >> > 252 >> >> > :: >> >> > virt.vm :: (_ recover) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` >> >> > :: >> >> > operation Aborted: migration job: cancel >> >> > D by cliente >> >> > >> >> > Dst vdsm.log: >> >> > >> >> > Periodic / 13 :: WARNING :: 2016-12-05 17: 01: 49,791 :: sampling :: >> >> > 483 >> >> > :: >> >> > virt.sampling.StatsCache: :( put) dropped stale old sample: sampled >> >> > 4303678.080000 stored 4303693.070000 >> >> > Periodic / 13 :: DEBUG :: 2016-12-05 17: 01: 49,791 :: executor :: >> >> > 221 >> >> > :: >> >> > Executor :: (_ run) Worker was discarded >> >> > Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17: 01: 49,792 :: __ init >> >> > __ >> >> > :: >> >> > 530 :: jsonrpc.JsonRpcServer :: (_ handle_request) Calling >> >> > 'VM.destroy' >> >> > in >> >> > bridge with {u'vmID ' : U'f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e '} >> >> > Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17: 01: 49,793 :: API :: >> >> > 314 >> >> > :: >> >> > vds :::( destroy) About to destroy VM >> >> > f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e >> >> > Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17:01:49,793 :: vm :: >> >> > 4171 >> >> > :: >> >> > virt.vm :::( destroy) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`: >> >> > : >> >> > Destroy Called >> >> > Thread-483 :: ERROR :: 2016-12-05 17: 01: 49,793 :: vm :: 759 :: >> >> > virt.vm >> >> > :: >> >> > (_ startUnderlyingVm) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` >> >> > :: >> >> > Failed To start a migration destination vm >> >> > Traceback (most recent call last): >> >> > File "/usr/share/vdsm/virt/vm.py", line 725, in _startUnderlyingVm >> >> > Self._completeIncomingMigration () >> >> > File "/usr/share/vdsm/virt/vm.py", line 3071, in >> >> > _completeIncomingMigration >> >> > Self._incomingMigrationFinished.isSet (), usedTimeout) >> >> > File "/usr/share/vdsm/virt/vm.py", line 3154, in >> >> > _attachLibvirtDomainAfterMigration >> >> > Raise MigrationError (e.get_error_message ()) >> >> > MigrationError: Domain not found: no domain with matching uuid >> >> > 'f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e' >> >> > >> >> > The logs attached. >> >> > Thanks. >> >> > >> >> > _______________________________________________ >> >> > Users mailing list >> >> > [email protected] >> >> > http://lists.ovirt.org/mailman/listinfo/users >> >> > >> >> >> >> >> >> >> >> -- >> >> Didi >> > >> > >> >> >> >> -- >> Didi > > -- Didi _______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

