Using the auto-generated snapshot is generally a bad idea as it's inconsistent, you should remove it before moving further
On Fri, May 11, 2018 at 7:25 PM, Juan Pablo <[email protected]> wrote: > I rebooted it with no luck, them I used the auto-gen snapshot , same luck. > attaching the logs in gdrive > > thanks in advance > > 2018-05-11 12:50 GMT-03:00 Benny Zlotnik <[email protected]>: > >> I see here a failed attempt: >> 2018-05-09 16:00:20,129-03 ERROR [org.ovirt.engine.core.dal.dbb >> roker.auditloghandling.AuditLogDirector] >> (EE-ManagedThreadFactory-engineScheduled-Thread-67) >> [bd8eeb1d-f49a-4f91-a521-e0f31b4a7cbd] EVENT_ID: >> USER_MOVED_DISK_FINISHED_FAILURE(2,011), User admin@internal-authz have >> failed to move disk mail02-int_Disk1 to domain 2penLA. >> >> Then another: >> 2018-05-09 16:15:06,998-03 ERROR [org.ovirt.engine.core.dal.dbb >> roker.auditloghandling.AuditLogDirector] >> (EE-ManagedThreadFactory-engineScheduled-Thread-34) >> [] EVENT_ID: USER_MOVED_DISK_FINISHED_FAILURE(2,011), User >> admin@internal-authz have failed to move disk mail02-int_Disk1 to domain >> 2penLA. >> >> Here I see a successful attempt: >> 2018-05-09 21:58:42,628-03 INFO [org.ovirt.engine.core.dal.dbb >> roker.auditloghandling.AuditLogDirector] (default task-50) >> [940b051c-8c63-4711-baf9-f3520bb2b825] EVENT_ID: USER_MOVED_DISK(2,008), >> User admin@internal-authz moving disk mail02-int_Disk1 to domain 2penLA. >> >> >> Then, in the last attempt I see the attempt was successful but live merge >> failed: >> 2018-05-11 03:37:59,509-03 ERROR >> [org.ovirt.engine.core.bll.MergeStatusCommand] >> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) >> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Failed to live merge, still in >> volume chain: [5d9d2958-96bc-49fa-9100-2f33a3ba737f, >> 52532d05-970e-4643-9774-96c31796062c] >> 2018-05-11 03:38:01,495-03 INFO [org.ovirt.engine.core.bll.Ser >> ialChildCommandsExecutionCallback] >> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'LiveMigrateDisk' (id: >> '115fc375-6018-4d59-b9f2-51ee05ca49f8') waiting on child command id: >> '26bc52a4-4509-4577-b342-44a679bc628f' type:'RemoveSnapshot' to complete >> 2018-05-11 03:38:01,501-03 ERROR [org.ovirt.engine.core.bll.sna >> pshots.RemoveSnapshotSingleDiskLiveCommand] >> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command id: >> '4936d196-a891-4484-9cf5-fceaafbf3364 failed child command status for >> step 'MERGE_STATUS' >> 2018-05-11 03:38:01,501-03 INFO [org.ovirt.engine.core.bll.sna >> pshots.RemoveSnapshotSingleDiskLiveCommandCallback] >> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command >> 'RemoveSnapshotSingleDiskLive' id: '4936d196-a891-4484-9cf5-fceaafbf3364' >> child commands '[8da5f261-7edd-4930-8d9d-d34f232d84b3, >> 1c320f4b-7296-43c4-a3e6-8a868e23fc35, a0e9e70c-cd65-4dfb-bd00-076c4e99556a]' >> executions were completed, status 'FAILED' >> 2018-05-11 03:38:02,513-03 ERROR [org.ovirt.engine.core.bll.sna >> pshots.RemoveSnapshotSingleDiskLiveCommand] >> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Merging of snapshot >> '319e8bbb-9efe-4de4-a9a6-862e3deb891f' images >> '52532d05-970e-4643-9774-96c31796062c'..'5d9d2958-96bc-49fa-9100-2f33a3ba737f' >> failed. Images have been marked illegal and can no longer be previewed or >> reverted to. Please retry Live Merge on the snapshot to complete the >> operation. >> 2018-05-11 03:38:02,519-03 ERROR [org.ovirt.engine.core.bll.sna >> pshots.RemoveSnapshotSingleDiskLiveCommand] >> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand' >> with failure. >> 2018-05-11 03:38:03,530-03 INFO [org.ovirt.engine.core.bll.Con >> currentChildCommandsExecutionCallback] >> (EE-ManagedThreadFactory-engineScheduled-Thread-37) >> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'RemoveSnapshot' id: >> '26bc52a4-4509-4577-b342-44a679bc628f' child commands >> '[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were completed, >> status 'FAILED' >> 2018-05-11 03:38:04,548-03 ERROR >> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with failure. >> 2018-05-11 03:38:04,557-03 INFO >> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object >> 'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39-16b579 >> 8213ac=LIVE_STORAGE_MIGRATION]', sharedLocks=''}' >> >> I do not see the merge attempt in the vdsm.log, so please send vdsm logs >> for node02.phy.eze.ampgn.com.ar from that time. >> >> Also, did you use the auto-generated snapshot to start the vm? >> >> >> On Fri, May 11, 2018 at 6:11 PM, Juan Pablo <[email protected]> >> wrote: >> >>> after the xfs_repair, it says: sorry I could not find valid secondary >>> superblock >>> >>> 2018-05-11 12:09 GMT-03:00 Juan Pablo <[email protected]>: >>> >>>> hi, >>>> Alias: >>>> mail02-int_Disk1 >>>> Description: >>>> ID: >>>> 65ec515e-0aae-4fe6-a561-387929c7fb4d >>>> Alignment: >>>> Unknown >>>> Disk Profile: >>>> Wipe After Delete: >>>> No >>>> >>>> that one >>>> >>>> 2018-05-11 11:12 GMT-03:00 Benny Zlotnik <[email protected]>: >>>> >>>>> I looked at the logs and I see some disks have moved successfully and >>>>> some failed. Which disk is causing the problems? >>>>> >>>>> On Fri, May 11, 2018 at 5:02 PM, Juan Pablo <[email protected] >>>>> > wrote: >>>>> >>>>>> Hi, just sent you via drive the files. attaching some extra info, >>>>>> thanks thanks and thanks : >>>>>> >>>>>> from inside the migrated vm I had the following attached dmesg output >>>>>> before rebooting >>>>>> >>>>>> regards and thanks again for the help, >>>>>> >>>>>> 2018-05-11 10:45 GMT-03:00 Benny Zlotnik <[email protected]>: >>>>>> >>>>>>> Dropbox or google drive I guess. Also, can you attach engine.log? >>>>>>> >>>>>>> On Fri, May 11, 2018 at 4:43 PM, Juan Pablo < >>>>>>> [email protected]> wrote: >>>>>>> >>>>>>>> >>>>>>>> vdsm is too big for gmail ...any other way I can share it with you? >>>>>>>> >>>>>>>> >>>>>>>> ---------- Forwrded message ---------- >>>>>>>> From: Juan Pablo <[email protected]> >>>>>>>> Date: 2018-05-11 10:40 GMT-03:00 >>>>>>>> Subject: Re: [ovirt-users] strange issue: vm lost info on disk >>>>>>>> To: Benny Zlotnik <[email protected]> >>>>>>>> Cc: users <[email protected]> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> Benny, thanks for your reply! it was a Live migration. sorry, it >>>>>>>> was from nfs to iscsi, not otherwise. I have reboot the vm for rescue >>>>>>>> and >>>>>>>> it does not detect any partitions with fdisk, Im running a xfs_repair >>>>>>>> with >>>>>>>> -n and found some corrupted primary superblock., its still running... ( >>>>>>>> so... there's info in the disk maybe?) >>>>>>>> >>>>>>>> attaching logs, let me know if those are the ones. >>>>>>>> thanks again! >>>>>>>> >>>>>>>> 2018-05-11 9:45 GMT-03:00 Benny Zlotnik <[email protected]>: >>>>>>>> >>>>>>>>> Can you provide the logs? engine and vdsm. >>>>>>>>> Did you perform a live migration (the VM is running) or cold? >>>>>>>>> >>>>>>>>> On Fri, May 11, 2018 at 2:49 PM, Juan Pablo < >>>>>>>>> [email protected]> wrote: >>>>>>>>> >>>>>>>>>> Hi! , Im strugled about an ongoing problem: >>>>>>>>>> after migrating a vm's disk from an iscsi domain to a nfs and >>>>>>>>>> ovirt reporting the migration was successful, I see there's no data >>>>>>>>>> 'inside' the vm's disk. we never had this issues with ovirt so Im >>>>>>>>>> stranged >>>>>>>>>> about the root cause and if theres a chance of recovering the >>>>>>>>>> information. >>>>>>>>>> >>>>>>>>>> can you please help me out troubleshooting this one? I would >>>>>>>>>> really appreciate it =) >>>>>>>>>> running ovirt 4.2.1 here! >>>>>>>>>> >>>>>>>>>> thanks in advance, >>>>>>>>>> JP >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> Users mailing list -- [email protected] >>>>>>>>>> To unsubscribe send an email to [email protected] >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected]

