OK, I got a core dump, what do I do with it now? It is 1.2G in size.
On Wed, May 19, 2010 at 10:54 AM, John Andrunas <j...@andrunas.net> wrote: > Hmmm... no coredump even though I configured it. > > Here is the trace though I will see what I can do about the coredump > > r...@cluster:/export/home/admin# zfs mount vol2/vm2 > > panic[cpu3]/thread=ffffff001f45ec60: BAD TRAP: type=e (#pf Page fault) > rp=ffffff001f45e950 addr=30 occurred in module "zfs" due to a NULL > pointer deree > > zpool-vol2: #pf Page fault > Bad kernel fault at addr=0x30 > pid=1469, pc=0xfffffffff795d054, sp=0xffffff001f45ea48, eflags=0x10296 > cr0: 8005003b<pg,wp,ne,et,ts,mp,pe> cr4: 6f8<xmme,fxsr,pge,mce,pae,pse,de> > cr2: 30cr3: 5000000cr8: c > > rdi: 0 rsi: ffffff05208b2388 rdx: ffffff001f45e888 > rcx: 0 r8: 3000900ff r9: 198f5ff6 > rax: 0 rbx: 200 rbp: ffffff001f45ea50 > r10: c0130803 r11: ffffff001f45ec60 r12: ffffff05208b2388 > r13: ffffff0521fc4000 r14: ffffff050c0167e0 r15: ffffff050c0167e8 > fsb: 0 gsb: ffffff04eb9b8080 ds: 4b > es: 4b fs: 0 gs: 1c3 > trp: e err: 2 rip: fffffffff795d054 > cs: 30 rfl: 10296 rsp: ffffff001f45ea48 > ss: 38 > > ffffff001f45e830 unix:die+dd () > ffffff001f45e940 unix:trap+177b () > ffffff001f45e950 unix:cmntrap+e6 () > ffffff001f45ea50 zfs:ddt_phys_decref+c () > ffffff001f45ea80 zfs:zio_ddt_free+55 () > ffffff001f45eab0 zfs:zio_execute+8d () > ffffff001f45eb50 genunix:taskq_thread+248 () > ffffff001f45eb60 unix:thread_start+8 () > > syncing file systems... done > skipping system dump - no dump device configured > rebooting... > > > On Wed, May 19, 2010 at 8:55 AM, Michael Schuster > <michael.schus...@oracle.com> wrote: >> On 19.05.10 17:53, John Andrunas wrote: >>> >>> Not to my knowledge, how would I go about getting one? (CC'ing discuss) >> >> man savecore and dumpadm. >> >> Michael >>> >>> >>> On Wed, May 19, 2010 at 8:46 AM, Mark J Musante<mark.musa...@oracle.com> >>> wrote: >>>> >>>> Do you have a coredump? Or a stack trace of the panic? >>>> >>>> On Wed, 19 May 2010, John Andrunas wrote: >>>> >>>>> Running ZFS on a Nexenta box, I had a mirror get broken and apparently >>>>> the metadata is corrupt now. If I try and mount vol2 it works but if >>>>> I try and mount -a or mount vol2/vm2 is instantly kernel panics and >>>>> reboots. Is it possible to recover from this? I don't care if I lose >>>>> the file listed below, but the other data in the volume would be >>>>> really nice to get back. I have scrubbed the volume to no avail. Any >>>>> other thoughts. >>>>> >>>>> >>>>> zpool status -xv vol2 >>>>> pool: vol2 >>>>> state: ONLINE >>>>> status: One or more devices has experienced an error resulting in data >>>>> corruption. Applications may be affected. >>>>> action: Restore the file in question if possible. Otherwise restore the >>>>> entire pool from backup. >>>>> see: http://www.sun.com/msg/ZFS-8000-8A >>>>> scrub: none requested >>>>> config: >>>>> >>>>> NAME STATE READ WRITE CKSUM >>>>> vol2 ONLINE 0 0 0 >>>>> mirror-0 ONLINE 0 0 0 >>>>> c3t3d0 ONLINE 0 0 0 >>>>> c3t2d0 ONLINE 0 0 0 >>>>> >>>>> errors: Permanent errors have been detected in the following files: >>>>> >>>>> vol2/v...@snap-daily-1-2010-05-06-0000:/as5/as5-flat.vmdk >>>>> >>>>> -- >>>>> John >>>>> _______________________________________________ >>>>> zfs-discuss mailing list >>>>> zfs-discuss@opensolaris.org >>>>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss >>>>> >>>> >>>> >>>> Regards, >>>> markm >>>> >>> >>> >>> >> >> >> -- >> michael.schus...@oracle.com http://blogs.sun.com/recursion >> Recursion, n.: see 'Recursion' >> > > > > -- > John > -- John _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss