2 clues: -the original size of the disk was 750G and was extended a month ago to 1100G. The System rebooted fine several times, and took the new size with no problems.
-I run fdisk from a centos 7 rescue cd and '/dev/vda' reported 750G. then, I took a snapshot of the disk to play with recovery tools and now fdisk reports 1100G... ¬¬ so my guess is on the extend and later migration to a different storage domain caused the issue. Im currently running testdisk to see if theres any partition to recover. regards, 2018-05-13 12:31 GMT-03:00 Juan Pablo <[email protected]>: > I removed the auto-snapshot and still no lucky. no bootable disk found. =( > ideas? > > > 2018-05-13 12:26 GMT-03:00 Juan Pablo <[email protected]>: > >> benny, thanks for your reply: >> ok, so the steps are : removing the snapshot on the first place. then >> what do you suggest? >> >> >> 2018-05-12 15:23 GMT-03:00 Nir Soffer <[email protected]>: >> >>> On Sat, 12 May 2018, 11:32 Benny Zlotnik, <[email protected]> wrote: >>> >>>> Using the auto-generated snapshot is generally a bad idea as it's >>>> inconsistent, >>>> >>> >>> What do you mean by inconsistant? >>> >>> >>> you should remove it before moving further >>>> >>>> On Fri, May 11, 2018 at 7:25 PM, Juan Pablo <[email protected]> >>>> wrote: >>>> >>>>> I rebooted it with no luck, them I used the auto-gen snapshot , same >>>>> luck. >>>>> attaching the logs in gdrive >>>>> >>>>> thanks in advance >>>>> >>>>> 2018-05-11 12:50 GMT-03:00 Benny Zlotnik <[email protected]>: >>>>> >>>>>> I see here a failed attempt: >>>>>> 2018-05-09 16:00:20,129-03 ERROR [org.ovirt.engine.core.dal.dbb >>>>>> roker.auditloghandling.AuditLogDirector] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-67) >>>>>> [bd8eeb1d-f49a-4f91-a521-e0f31b4a7cbd] EVENT_ID: >>>>>> USER_MOVED_DISK_FINISHED_FAILURE(2,011), User admin@internal-authz >>>>>> have failed to move disk mail02-int_Disk1 to domain 2penLA. >>>>>> >>>>>> Then another: >>>>>> 2018-05-09 16:15:06,998-03 ERROR [org.ovirt.engine.core.dal.dbb >>>>>> roker.auditloghandling.AuditLogDirector] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-34) [] EVENT_ID: >>>>>> USER_MOVED_DISK_FINISHED_FAILURE(2,011), User admin@internal-authz >>>>>> have failed to move disk mail02-int_Disk1 to domain 2penLA. >>>>>> >>>>>> Here I see a successful attempt: >>>>>> 2018-05-09 21:58:42,628-03 INFO [org.ovirt.engine.core.dal.dbb >>>>>> roker.auditloghandling.AuditLogDirector] (default task-50) >>>>>> [940b051c-8c63-4711-baf9-f3520bb2b825] EVENT_ID: >>>>>> USER_MOVED_DISK(2,008), User admin@internal-authz moving disk >>>>>> mail02-int_Disk1 to domain 2penLA. >>>>>> >>>>>> >>>>>> Then, in the last attempt I see the attempt was successful but live >>>>>> merge failed: >>>>>> 2018-05-11 03:37:59,509-03 ERROR >>>>>> [org.ovirt.engine.core.bll.MergeStatusCommand] >>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Failed to live merge, still >>>>>> in volume chain: [5d9d2958-96bc-49fa-9100-2f33a3ba737f, >>>>>> 52532d05-970e-4643-9774-96c31796062c] >>>>>> 2018-05-11 03:38:01,495-03 INFO [org.ovirt.engine.core.bll.Ser >>>>>> ialChildCommandsExecutionCallback] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'LiveMigrateDisk' >>>>>> (id: '115fc375-6018-4d59-b9f2-51ee05ca49f8') waiting on child >>>>>> command id: '26bc52a4-4509-4577-b342-44a679bc628f' >>>>>> type:'RemoveSnapshot' to complete >>>>>> 2018-05-11 03:38:01,501-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command id: >>>>>> '4936d196-a891-4484-9cf5-fceaafbf3364 failed child command status >>>>>> for step 'MERGE_STATUS' >>>>>> 2018-05-11 03:38:01,501-03 INFO [org.ovirt.engine.core.bll.sna >>>>>> pshots.RemoveSnapshotSingleDiskLiveCommandCallback] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command >>>>>> 'RemoveSnapshotSingleDiskLive' id: '4936d196-a891-4484-9cf5-fceaafbf3364' >>>>>> child commands '[8da5f261-7edd-4930-8d9d-d34f232d84b3, >>>>>> 1c320f4b-7296-43c4-a3e6-8a868e23fc35, >>>>>> a0e9e70c-cd65-4dfb-bd00-076c4e99556a]' >>>>>> executions were completed, status 'FAILED' >>>>>> 2018-05-11 03:38:02,513-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Merging of snapshot >>>>>> '319e8bbb-9efe-4de4-a9a6-862e3deb891f' images >>>>>> '52532d05-970e-4643-9774-96c31796062c'..'5d9d2958-96bc-49fa-9100-2f33a3ba737f' >>>>>> failed. Images have been marked illegal and can no longer be previewed or >>>>>> reverted to. Please retry Live Merge on the snapshot to complete the >>>>>> operation. >>>>>> 2018-05-11 03:38:02,519-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand' >>>>>> with failure. >>>>>> 2018-05-11 03:38:03,530-03 INFO [org.ovirt.engine.core.bll.Con >>>>>> currentChildCommandsExecutionCallback] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-37) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'RemoveSnapshot' id: >>>>>> '26bc52a4-4509-4577-b342-44a679bc628f' child commands >>>>>> '[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were completed, >>>>>> status 'FAILED' >>>>>> 2018-05-11 03:38:04,548-03 ERROR >>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with >>>>>> failure. >>>>>> 2018-05-11 03:38:04,557-03 INFO >>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object >>>>>> 'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39-16b579 >>>>>> 8213ac=LIVE_STORAGE_MIGRATION]', sharedLocks=''}' >>>>>> >>>>>> I do not see the merge attempt in the vdsm.log, so please send vdsm >>>>>> logs for node02.phy.eze.ampgn.com.ar from that time. >>>>>> >>>>>> Also, did you use the auto-generated snapshot to start the vm? >>>>>> >>>>>> >>>>>> On Fri, May 11, 2018 at 6:11 PM, Juan Pablo < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> after the xfs_repair, it says: sorry I could not find valid >>>>>>> secondary superblock >>>>>>> >>>>>>> 2018-05-11 12:09 GMT-03:00 Juan Pablo <[email protected]>: >>>>>>> >>>>>>>> hi, >>>>>>>> Alias: >>>>>>>> mail02-int_Disk1 >>>>>>>> Description: >>>>>>>> ID: >>>>>>>> 65ec515e-0aae-4fe6-a561-387929c7fb4d >>>>>>>> Alignment: >>>>>>>> Unknown >>>>>>>> Disk Profile: >>>>>>>> Wipe After Delete: >>>>>>>> No >>>>>>>> >>>>>>>> that one >>>>>>>> >>>>>>>> 2018-05-11 11:12 GMT-03:00 Benny Zlotnik <[email protected]>: >>>>>>>> >>>>>>>>> I looked at the logs and I see some disks have moved successfully >>>>>>>>> and some failed. Which disk is causing the problems? >>>>>>>>> >>>>>>>>> On Fri, May 11, 2018 at 5:02 PM, Juan Pablo < >>>>>>>>> [email protected]> wrote: >>>>>>>>> >>>>>>>>>> Hi, just sent you via drive the files. attaching some extra info, >>>>>>>>>> thanks thanks and thanks : >>>>>>>>>> >>>>>>>>>> from inside the migrated vm I had the following attached dmesg >>>>>>>>>> output before rebooting >>>>>>>>>> >>>>>>>>>> regards and thanks again for the help, >>>>>>>>>> >>>>>>>>>> 2018-05-11 10:45 GMT-03:00 Benny Zlotnik <[email protected]>: >>>>>>>>>> >>>>>>>>>>> Dropbox or google drive I guess. Also, can you attach engine.log? >>>>>>>>>>> >>>>>>>>>>> On Fri, May 11, 2018 at 4:43 PM, Juan Pablo < >>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> vdsm is too big for gmail ...any other way I can share it with >>>>>>>>>>>> you? >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> ---------- Forwrded message ---------- >>>>>>>>>>>> From: Juan Pablo <[email protected]> >>>>>>>>>>>> Date: 2018-05-11 10:40 GMT-03:00 >>>>>>>>>>>> Subject: Re: [ovirt-users] strange issue: vm lost info on disk >>>>>>>>>>>> To: Benny Zlotnik <[email protected]> >>>>>>>>>>>> Cc: users <[email protected]> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Benny, thanks for your reply! it was a Live migration. sorry, >>>>>>>>>>>> it was from nfs to iscsi, not otherwise. I have reboot the vm for >>>>>>>>>>>> rescue >>>>>>>>>>>> and it does not detect any partitions with fdisk, Im running a >>>>>>>>>>>> xfs_repair >>>>>>>>>>>> with -n and found some corrupted primary superblock., its still >>>>>>>>>>>> running... >>>>>>>>>>>> ( so... there's info in the disk maybe?) >>>>>>>>>>>> >>>>>>>>>>>> attaching logs, let me know if those are the ones. >>>>>>>>>>>> thanks again! >>>>>>>>>>>> >>>>>>>>>>>> 2018-05-11 9:45 GMT-03:00 Benny Zlotnik <[email protected]>: >>>>>>>>>>>> >>>>>>>>>>>>> Can you provide the logs? engine and vdsm. >>>>>>>>>>>>> Did you perform a live migration (the VM is running) or cold? >>>>>>>>>>>>> >>>>>>>>>>>>> On Fri, May 11, 2018 at 2:49 PM, Juan Pablo < >>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hi! , Im strugled about an ongoing problem: >>>>>>>>>>>>>> after migrating a vm's disk from an iscsi domain to a nfs >>>>>>>>>>>>>> and ovirt reporting the migration was successful, I see there's >>>>>>>>>>>>>> no data >>>>>>>>>>>>>> 'inside' the vm's disk. we never had this issues with ovirt so >>>>>>>>>>>>>> Im stranged >>>>>>>>>>>>>> about the root cause and if theres a chance of recovering the >>>>>>>>>>>>>> information. >>>>>>>>>>>>>> >>>>>>>>>>>>>> can you please help me out troubleshooting this one? I would >>>>>>>>>>>>>> really appreciate it =) >>>>>>>>>>>>>> running ovirt 4.2.1 here! >>>>>>>>>>>>>> >>>>>>>>>>>>>> thanks in advance, >>>>>>>>>>>>>> JP >>>>>>>>>>>>>> >>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>> Users mailing list -- [email protected] >>>>>>>>>>>>>> To unsubscribe send an email to [email protected] >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> _______________________________________________ >>>> Users mailing list -- [email protected] >>>> To unsubscribe send an email to [email protected] >>>> >>> >> >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected]

