Seems like core.vold.* are not being created until I try to boot from zfsBE, 
just creating zfsBE gets onlu core.cpio created.



[10:29:48] @adas: /var/crash > mdb core.cpio.5545
Loading modules: [ libc.so.1 libavl.so.1 ld.so.1 ]
> ::status
debugging core file of cpio (32-bit) from adas
file: /usr/bin/cpio
initial argv: /usr/bin/cpio -pPcdum /.alt.tmp.b-Prb.mnt
threading model: multi-threaded
status: process terminated by SIGBUS (Bus Error)
> $C
ffbfe5b0 libc.so.1`_malloc_unlocked+0x164(30, 0, 39c28, ff, 2e2f2e2f, 0)
ffbfe610 libc.so.1`malloc+0x4c(30, 1, e8070, 0, ff33e3c0, ff3485b8)
ffbfe670 libsec.so.1`cacl_get+0x138(ffbfe7c4, 2, 0, 35bc0, 0, 35f98)
ffbfe768 libsec.so.1`acl_get+0x14(37fe2, 2, 35bc0, 354c0, 1000, 1)
ffbfe7d0 0x183b4(1, 35800, 359e8, 346b0, 34874, 34870)
ffbfec30 main+0x28c(34708, 1, 35bc0, 166fc, 35800, 34400)
ffbfec90 _start+0x108(0, 0, 0, 0, 0, 0)
> $r
%g0 = 0x00000000                 %l0 = 0x00000000
%g1 = 0xff25638c libc.so.1`malloc+0x44 %l1 = 0x00039c28
%g2 = 0x00037fe0                 %l2 = 0x2e2f2e2f
%g3 = 0x00008000                 %l3 = 0x000003c8
%g4 = 0x00000000                 %l4 = 0x2e2f2e2f
%g5 = 0x00000000                 %l5 = 0x00000000
%g6 = 0x00000000                 %l6 = 0xffffdc00
%g7 = 0xff382a00                 %l7 = 0xff347344 libc.so.1`Lfree
%o0 = 0x00000000                 %i0 = 0x00000030
%o1 = 0x00000000                 %i1 = 0x00000000
%o2 = 0x000e70c4                 %i2 = 0x00039c28
%o3 = 0x00000000                 %i3 = 0x000000ff
%o4 = 0xff33e3c0                 %i4 = 0x2e2f2e2f
%o5 = 0xff347344 libc.so.1`Lfree %i5 = 0x00000000
%o6 = 0xffbfe5b0                 %i6 = 0xffbfe610
%o7 = 0xff2564a4 libc.so.1`_malloc_unlocked+0xf4 %i7 = 0xff256394
libc.so.1`malloc+0x4c

  %psr = 0xfe001002 impl=0xf ver=0xe icc=nzvc
                    ec=0 ef=4096 pil=0 s=0 ps=0 et=0 cwp=0x2
    %y = 0x00000000
   %pc = 0xff256514 libc.so.1`_malloc_unlocked+0x164
  %npc = 0xff2564d8 libc.so.1`_malloc_unlocked+0x128
   %sp = 0xffbfe5b0
   %fp = 0xffbfe610

  %wim = 0x00000000
  %tbr = 0x00000000
>







On Thu, 6 Nov 2008, Enda O'Connor wrote:

> Hi
> try and get the stack trace from the core
> ie mdb core.vold.24978
> ::status
> $C
> $r
>
> also run the same 3 mdb commands on the cpio core dump.
>
> also if you could extract some data from the truss log, ie a few hundred 
> lines before the first SIGBUS
>
>
> Enda
>
> On 11/06/08 01:25, Krzys wrote:
>> THis is so bizare, I am unable to pass this problem. I though I had not 
>> enough space on my hard drive (new one) so I replaced it with 72gb drive, 
>> but still getting that bus error. Originally when I restarted my server it 
>> did not want to boot, do I had to power it off and then back on and it then 
>> booted up. But constantly I am getting this "Bus Error - core dumped"
>> 
>> anyway in my /var/crash I see hundreds of core.void files and 3 core.cpio 
>> files. I would imagine core.cpio are the ones that are direct result of 
>> what I am probably eperiencing.
>> 
>> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24854
>> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24867
>> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24880
>> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24893
>> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24906
>> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24919
>> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24932
>> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24950
>> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24978
>> drwxr-xr-x   3 root     root       81408 Nov  5 20:06 .
>> -rw-------   1 root     root     31351099 Nov  5 20:06 core.cpio.6208
>> 
>> 
>> 
>> On Wed, 5 Nov 2008, Enda O'Connor wrote:
>> 
>>> Hi
>>> Looks ok, some mounts left over from pervious fail.
>>> In regards to swap and dump on zpool you can set them
>>> zfs set volsize=1G rootpool/dump
>>> zfs set volsize=1G rootpool/swap
>>> 
>>> for instance, of course above are only an example of how to do it.
>>> or make the zvol doe rootpool/dump etc before lucreate, in which case it 
>>> will take the swap and dump size you have preset.
>>> 
>>> But I think we need to see the coredump/truss at this point to get an idea 
>>> of where things went wrong.
>>> Enda
>>> 
>>> On 11/05/08 15:38, Krzys wrote:
>>>> I did upgrade my U5 to U6 from DVD, went trough the upgrade process.
>>>> my file system is setup as follow:
>>>> [10:11:54] [EMAIL PROTECTED]: /root > df -h | egrep -v 
>>>> "platform|sharefs|objfs|mnttab|proc|ctfs|devices|fd|nsr"
>>>> Filesystem             size   used  avail capacity  Mounted on
>>>> /dev/dsk/c1t0d0s0       16G   7.2G   8.4G    47%    /
>>>> swap                   8.3G   1.5M   8.3G     1%    /etc/svc/volatile
>>>> /dev/dsk/c1t0d0s6       16G   8.7G   6.9G    56%    /usr
>>>> /dev/dsk/c1t0d0s1       16G   2.5G    13G    17%    /var
>>>> swap                   8.5G   229M   8.3G     3%    /tmp
>>>> swap                   8.3G    40K   8.3G     1%    /var/run
>>>> /dev/dsk/c1t0d0s7       78G   1.2G    76G     2%    /export/home
>>>> rootpool                33G    19K    21G     1%    /rootpool
>>>> rootpool/ROOT           33G    18K    21G     1%    /rootpool/ROOT
>>>> rootpool/ROOT/zfsBE     33G    31M    21G     1%    /.alt.tmp.b-UUb.mnt
>>>> /export/home            78G   1.2G    76G     2% 
>>>> /.alt.tmp.b-UUb.mnt/export/home
>>>> /rootpool               21G    19K    21G     1% 
>>>> /.alt.tmp.b-UUb.mnt/rootpool
>>>> /rootpool/ROOT          21G    18K    21G     1% 
>>>> /.alt.tmp.b-UUb.mnt/rootpool/ROOT
>>>> swap                   8.3G     0K   8.3G     0% 
>>>> /.alt.tmp.b-UUb.mnt/var/run
>>>> swap                   8.3G     0K   8.3G     0% 
>>>> /.alt.tmp.b-UUb.mnt/tmp
>>>> [10:12:00] [EMAIL PROTECTED]: /root >
>>>> 
>>>> 
>>>> so I have /, /usr, /var and /export/home on that primary disk. Original 
>>>> disk is 140gb, this new one is only 36gb, but disk utilization on that 
>>>> primary disk is much less utilized so easily should fit on it.
>>>> 
>>>> / 7.2GB
>>>> /usr 8.7GB
>>>> /var 2.5GB
>>>> /export/home 1.2GB
>>>> total space 19.6GB
>>>> I did notice that lucreate did alocate 8GB to SWAP and 4GB to DUMP
>>>> total space needed 31.6GB
>>>> seems like total available disk space on my disk should be 33.92GB
>>>> so its quite close as both numbers do approach. So to make sure I will 
>>>> change disk for 72gb and will try again. I do not beleive that I need to 
>>>> match my main disk size as 146gb as I am not using that much disk space 
>>>> on it. But let me try this and it might be why I am getting this 
>>>> problem...
>>>> 
>>>> 
>>>> 
>>>> On Wed, 5 Nov 2008, Enda O'Connor wrote:
>>>> 
>>>>> Hi Krzys
>>>>> Also some info on the actual system
>>>>> ie what was it upgraded to u6 from and how.
>>>>> and an idea of how the filesystems are laid out, ie is usr seperate from 
>>>>> / and so on ( maybe a df -k ). Don't appear to have any zones installed, 
>>>>> just to confirm.
>>>>> Enda
>>>>> 
>>>>> On 11/05/08 14:07, Enda O'Connor wrote:
>>>>>> Hi
>>>>>> did you get a core dump?
>>>>>> would be nice to see the core file to get an idea of what dumped core,
>>>>>> might configure coreadm if not already done
>>>>>> run coreadm first, if the output looks like
>>>>>> 
>>>>>> # coreadm
>>>>>>      global core file pattern: /var/crash/core.%f.%p
>>>>>>      global core file content: default
>>>>>>        init core file pattern: core
>>>>>>        init core file content: default
>>>>>>             global core dumps: enabled
>>>>>>        per-process core dumps: enabled
>>>>>>       global setid core dumps: enabled
>>>>>>  per-process setid core dumps: disabled
>>>>>>      global core dump logging: enabled
>>>>>> 
>>>>>> then all should be good, and cores should appear in /var/crash
>>>>>> 
>>>>>> otherwise the following should configure coreadm:
>>>>>> coreadm -g /var/crash/core.%f.%p
>>>>>> coreadm -G all
>>>>>> coreadm -e global
>>>>>> coreadm -e per-process
>>>>>> 
>>>>>> 
>>>>>> coreadm -u to load the new settings without rebooting.
>>>>>> 
>>>>>> also might need to set the size of the core dump via
>>>>>> ulimit -c unlimited
>>>>>> check ulimit -a first.
>>>>>> 
>>>>>> then rerun test and check /var/crash for core dump.
>>>>>> 
>>>>>> If that fails a truss via say truss -fae -o /tmp/truss.out lucreate -c 
>>>>>> ufsBE -n zfsBE -p rootpool
>>>>>> 
>>>>>> might give an indication, look for SIGBUS in the truss log
>>>>>> 
>>>>>> NOTE, that you might want to reset the coreadm and ulimit for coredumps 
>>>>>> after this, in order to not risk filling the system with coredumps in 
>>>>>> the case of some utility coredumping in a loop say.
>>>>>> 
>>>>>> 
>>>>>> Enda
>
>
> -- 
> Enda O'Connor x19781  Software Product Engineering
> Patch System Test : Ireland : x19781/353-1-8199718
>
>
> !DSPAM:122,4912d10015286266247132!
>
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to