Hmmm... no coredump even though I configured it.

Here is the trace though  I will see what I can do about the coredump

r...@cluster:/export/home/admin# zfs mount vol2/vm2

panic[cpu3]/thread=ffffff001f45ec60: BAD TRAP: type=e (#pf Page fault)
rp=ffffff001f45e950 addr=30 occurred in module "zfs" due to a NULL
pointer deree

zpool-vol2: #pf Page fault
Bad kernel fault at addr=0x30
pid=1469, pc=0xfffffffff795d054, sp=0xffffff001f45ea48, eflags=0x10296
cr0: 8005003b<pg,wp,ne,et,ts,mp,pe> cr4: 6f8<xmme,fxsr,pge,mce,pae,pse,de>
cr2: 30cr3: 5000000cr8: c

        rdi:                0 rsi: ffffff05208b2388 rdx: ffffff001f45e888
        rcx:                0  r8:        3000900ff  r9:         198f5ff6
        rax:                0 rbx:              200 rbp: ffffff001f45ea50
        r10:         c0130803 r11: ffffff001f45ec60 r12: ffffff05208b2388
        r13: ffffff0521fc4000 r14: ffffff050c0167e0 r15: ffffff050c0167e8
        fsb:                0 gsb: ffffff04eb9b8080  ds:               4b
         es:               4b  fs:                0  gs:              1c3
        trp:                e err:                2 rip: fffffffff795d054
         cs:               30 rfl:            10296 rsp: ffffff001f45ea48
         ss:               38

ffffff001f45e830 unix:die+dd ()
ffffff001f45e940 unix:trap+177b ()
ffffff001f45e950 unix:cmntrap+e6 ()
ffffff001f45ea50 zfs:ddt_phys_decref+c ()
ffffff001f45ea80 zfs:zio_ddt_free+55 ()
ffffff001f45eab0 zfs:zio_execute+8d ()
ffffff001f45eb50 genunix:taskq_thread+248 ()
ffffff001f45eb60 unix:thread_start+8 ()

syncing file systems... done
skipping system dump - no dump device configured
rebooting...


On Wed, May 19, 2010 at 8:55 AM, Michael Schuster
<michael.schus...@oracle.com> wrote:
> On 19.05.10 17:53, John Andrunas wrote:
>>
>> Not to my knowledge, how would I go about getting one?  (CC'ing discuss)
>
> man savecore and dumpadm.
>
> Michael
>>
>>
>> On Wed, May 19, 2010 at 8:46 AM, Mark J Musante<mark.musa...@oracle.com>
>>  wrote:
>>>
>>> Do you have a coredump?  Or a stack trace of the panic?
>>>
>>> On Wed, 19 May 2010, John Andrunas wrote:
>>>
>>>> Running ZFS on a Nexenta box, I had a mirror get broken and apparently
>>>> the metadata is corrupt now.  If I try and mount vol2 it works but if
>>>> I try and mount -a or mount vol2/vm2 is instantly kernel panics and
>>>> reboots.  Is it possible to recover from this?  I don't care if I lose
>>>> the file listed below, but the other data in the volume would be
>>>> really nice to get back.  I have scrubbed the volume to no avail.  Any
>>>> other thoughts.
>>>>
>>>>
>>>> zpool status -xv vol2
>>>>  pool: vol2
>>>> state: ONLINE
>>>> status: One or more devices has experienced an error resulting in data
>>>>       corruption.  Applications may be affected.
>>>> action: Restore the file in question if possible.  Otherwise restore the
>>>>       entire pool from backup.
>>>>  see: http://www.sun.com/msg/ZFS-8000-8A
>>>> scrub: none requested
>>>> config:
>>>>
>>>>       NAME        STATE     READ WRITE CKSUM
>>>>       vol2        ONLINE       0     0     0
>>>>         mirror-0  ONLINE       0     0     0
>>>>           c3t3d0  ONLINE       0     0     0
>>>>           c3t2d0  ONLINE       0     0     0
>>>>
>>>> errors: Permanent errors have been detected in the following files:
>>>>
>>>>       vol2/v...@snap-daily-1-2010-05-06-0000:/as5/as5-flat.vmdk
>>>>
>>>> --
>>>> John
>>>> _______________________________________________
>>>> zfs-discuss mailing list
>>>> zfs-discuss@opensolaris.org
>>>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>>>>
>>>
>>>
>>> Regards,
>>> markm
>>>
>>
>>
>>
>
>
> --
> michael.schus...@oracle.com     http://blogs.sun.com/recursion
> Recursion, n.: see 'Recursion'
>



-- 
John
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to