Hi Wierd, almost like some kind of memory corruption. Could I see the upgrade logs, that got you to u6 ie /var/sadm/system/logs/upgrade_log for the u6 env. What kind of upgrade did you do, liveupgrade, text based etc?
Enda On 11/06/08 15:41, Krzys wrote: > Seems like core.vold.* are not being created until I try to boot from zfsBE, > just creating zfsBE gets onlu core.cpio created. > > > > [10:29:48] @adas: /var/crash > mdb core.cpio.5545 > Loading modules: [ libc.so.1 libavl.so.1 ld.so.1 ] >> ::status > debugging core file of cpio (32-bit) from adas > file: /usr/bin/cpio > initial argv: /usr/bin/cpio -pPcdum /.alt.tmp.b-Prb.mnt > threading model: multi-threaded > status: process terminated by SIGBUS (Bus Error) >> $C > ffbfe5b0 libc.so.1`_malloc_unlocked+0x164(30, 0, 39c28, ff, 2e2f2e2f, 0) > ffbfe610 libc.so.1`malloc+0x4c(30, 1, e8070, 0, ff33e3c0, ff3485b8) > ffbfe670 libsec.so.1`cacl_get+0x138(ffbfe7c4, 2, 0, 35bc0, 0, 35f98) > ffbfe768 libsec.so.1`acl_get+0x14(37fe2, 2, 35bc0, 354c0, 1000, 1) > ffbfe7d0 0x183b4(1, 35800, 359e8, 346b0, 34874, 34870) > ffbfec30 main+0x28c(34708, 1, 35bc0, 166fc, 35800, 34400) > ffbfec90 _start+0x108(0, 0, 0, 0, 0, 0) >> $r > %g0 = 0x00000000 %l0 = 0x00000000 > %g1 = 0xff25638c libc.so.1`malloc+0x44 %l1 = 0x00039c28 > %g2 = 0x00037fe0 %l2 = 0x2e2f2e2f > %g3 = 0x00008000 %l3 = 0x000003c8 > %g4 = 0x00000000 %l4 = 0x2e2f2e2f > %g5 = 0x00000000 %l5 = 0x00000000 > %g6 = 0x00000000 %l6 = 0xffffdc00 > %g7 = 0xff382a00 %l7 = 0xff347344 libc.so.1`Lfree > %o0 = 0x00000000 %i0 = 0x00000030 > %o1 = 0x00000000 %i1 = 0x00000000 > %o2 = 0x000e70c4 %i2 = 0x00039c28 > %o3 = 0x00000000 %i3 = 0x000000ff > %o4 = 0xff33e3c0 %i4 = 0x2e2f2e2f > %o5 = 0xff347344 libc.so.1`Lfree %i5 = 0x00000000 > %o6 = 0xffbfe5b0 %i6 = 0xffbfe610 > %o7 = 0xff2564a4 libc.so.1`_malloc_unlocked+0xf4 %i7 = 0xff256394 > libc.so.1`malloc+0x4c > > %psr = 0xfe001002 impl=0xf ver=0xe icc=nzvc > ec=0 ef=4096 pil=0 s=0 ps=0 et=0 cwp=0x2 > %y = 0x00000000 > %pc = 0xff256514 libc.so.1`_malloc_unlocked+0x164 > %npc = 0xff2564d8 libc.so.1`_malloc_unlocked+0x128 > %sp = 0xffbfe5b0 > %fp = 0xffbfe610 > > %wim = 0x00000000 > %tbr = 0x00000000 > > > > > > > > On Thu, 6 Nov 2008, Enda O'Connor wrote: > >> Hi >> try and get the stack trace from the core >> ie mdb core.vold.24978 >> ::status >> $C >> $r >> >> also run the same 3 mdb commands on the cpio core dump. >> >> also if you could extract some data from the truss log, ie a few hundred >> lines before the first SIGBUS >> >> >> Enda >> >> On 11/06/08 01:25, Krzys wrote: >>> THis is so bizare, I am unable to pass this problem. I though I had not >>> enough space on my hard drive (new one) so I replaced it with 72gb drive, >>> but still getting that bus error. Originally when I restarted my server it >>> did not want to boot, do I had to power it off and then back on and it then >>> booted up. But constantly I am getting this "Bus Error - core dumped" >>> >>> anyway in my /var/crash I see hundreds of core.void files and 3 core.cpio >>> files. I would imagine core.cpio are the ones that are direct result of >>> what I am probably eperiencing. >>> >>> -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24854 >>> -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24867 >>> -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24880 >>> -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24893 >>> -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24906 >>> -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24919 >>> -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24932 >>> -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24950 >>> -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24978 >>> drwxr-xr-x 3 root root 81408 Nov 5 20:06 . >>> -rw------- 1 root root 31351099 Nov 5 20:06 core.cpio.6208 >>> >>> >>> >>> On Wed, 5 Nov 2008, Enda O'Connor wrote: >>> >>>> Hi >>>> Looks ok, some mounts left over from pervious fail. >>>> In regards to swap and dump on zpool you can set them >>>> zfs set volsize=1G rootpool/dump >>>> zfs set volsize=1G rootpool/swap >>>> >>>> for instance, of course above are only an example of how to do it. >>>> or make the zvol doe rootpool/dump etc before lucreate, in which case it >>>> will take the swap and dump size you have preset. >>>> >>>> But I think we need to see the coredump/truss at this point to get an idea >>>> of where things went wrong. >>>> Enda >>>> >>>> On 11/05/08 15:38, Krzys wrote: >>>>> I did upgrade my U5 to U6 from DVD, went trough the upgrade process. >>>>> my file system is setup as follow: >>>>> [10:11:54] [EMAIL PROTECTED]: /root > df -h | egrep -v >>>>> "platform|sharefs|objfs|mnttab|proc|ctfs|devices|fd|nsr" >>>>> Filesystem size used avail capacity Mounted on >>>>> /dev/dsk/c1t0d0s0 16G 7.2G 8.4G 47% / >>>>> swap 8.3G 1.5M 8.3G 1% /etc/svc/volatile >>>>> /dev/dsk/c1t0d0s6 16G 8.7G 6.9G 56% /usr >>>>> /dev/dsk/c1t0d0s1 16G 2.5G 13G 17% /var >>>>> swap 8.5G 229M 8.3G 3% /tmp >>>>> swap 8.3G 40K 8.3G 1% /var/run >>>>> /dev/dsk/c1t0d0s7 78G 1.2G 76G 2% /export/home >>>>> rootpool 33G 19K 21G 1% /rootpool >>>>> rootpool/ROOT 33G 18K 21G 1% /rootpool/ROOT >>>>> rootpool/ROOT/zfsBE 33G 31M 21G 1% /.alt.tmp.b-UUb.mnt >>>>> /export/home 78G 1.2G 76G 2% >>>>> /.alt.tmp.b-UUb.mnt/export/home >>>>> /rootpool 21G 19K 21G 1% >>>>> /.alt.tmp.b-UUb.mnt/rootpool >>>>> /rootpool/ROOT 21G 18K 21G 1% >>>>> /.alt.tmp.b-UUb.mnt/rootpool/ROOT >>>>> swap 8.3G 0K 8.3G 0% >>>>> /.alt.tmp.b-UUb.mnt/var/run >>>>> swap 8.3G 0K 8.3G 0% >>>>> /.alt.tmp.b-UUb.mnt/tmp >>>>> [10:12:00] [EMAIL PROTECTED]: /root > >>>>> >>>>> >>>>> so I have /, /usr, /var and /export/home on that primary disk. Original >>>>> disk is 140gb, this new one is only 36gb, but disk utilization on that >>>>> primary disk is much less utilized so easily should fit on it. >>>>> >>>>> / 7.2GB >>>>> /usr 8.7GB >>>>> /var 2.5GB >>>>> /export/home 1.2GB >>>>> total space 19.6GB >>>>> I did notice that lucreate did alocate 8GB to SWAP and 4GB to DUMP >>>>> total space needed 31.6GB >>>>> seems like total available disk space on my disk should be 33.92GB >>>>> so its quite close as both numbers do approach. So to make sure I will >>>>> change disk for 72gb and will try again. I do not beleive that I need to >>>>> match my main disk size as 146gb as I am not using that much disk space >>>>> on it. But let me try this and it might be why I am getting this >>>>> problem... >>>>> >>>>> >>>>> >>>>> On Wed, 5 Nov 2008, Enda O'Connor wrote: >>>>> >>>>>> Hi Krzys >>>>>> Also some info on the actual system >>>>>> ie what was it upgraded to u6 from and how. >>>>>> and an idea of how the filesystems are laid out, ie is usr seperate from >>>>>> / and so on ( maybe a df -k ). Don't appear to have any zones installed, >>>>>> just to confirm. >>>>>> Enda >>>>>> >>>>>> On 11/05/08 14:07, Enda O'Connor wrote: >>>>>>> Hi >>>>>>> did you get a core dump? >>>>>>> would be nice to see the core file to get an idea of what dumped core, >>>>>>> might configure coreadm if not already done >>>>>>> run coreadm first, if the output looks like >>>>>>> >>>>>>> # coreadm >>>>>>> global core file pattern: /var/crash/core.%f.%p >>>>>>> global core file content: default >>>>>>> init core file pattern: core >>>>>>> init core file content: default >>>>>>> global core dumps: enabled >>>>>>> per-process core dumps: enabled >>>>>>> global setid core dumps: enabled >>>>>>> per-process setid core dumps: disabled >>>>>>> global core dump logging: enabled >>>>>>> >>>>>>> then all should be good, and cores should appear in /var/crash >>>>>>> >>>>>>> otherwise the following should configure coreadm: >>>>>>> coreadm -g /var/crash/core.%f.%p >>>>>>> coreadm -G all >>>>>>> coreadm -e global >>>>>>> coreadm -e per-process >>>>>>> >>>>>>> >>>>>>> coreadm -u to load the new settings without rebooting. >>>>>>> >>>>>>> also might need to set the size of the core dump via >>>>>>> ulimit -c unlimited >>>>>>> check ulimit -a first. >>>>>>> >>>>>>> then rerun test and check /var/crash for core dump. >>>>>>> >>>>>>> If that fails a truss via say truss -fae -o /tmp/truss.out lucreate -c >>>>>>> ufsBE -n zfsBE -p rootpool >>>>>>> >>>>>>> might give an indication, look for SIGBUS in the truss log >>>>>>> >>>>>>> NOTE, that you might want to reset the coreadm and ulimit for coredumps >>>>>>> after this, in order to not risk filling the system with coredumps in >>>>>>> the case of some utility coredumping in a loop say. >>>>>>> >>>>>>> >>>>>>> Enda >> >> -- >> Enda O'Connor x19781 Software Product Engineering >> Patch System Test : Ireland : x19781/353-1-8199718 >> >> >> !DSPAM:122,4912d10015286266247132! >> > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- Enda O'Connor x19781 Software Product Engineering Patch System Test : Ireland : x19781/353-1-8199718 _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss