what makes me wonder is why I am not even able to see anything under boot -L ? and it is just not seeing this disk as a boot device? so strange.
On Wed, 5 Nov 2008, Krzys wrote: > THis is so bizare, I am unable to pass this problem. I though I had not enough > space on my hard drive (new one) so I replaced it with 72gb drive, but still > getting that bus error. Originally when I restarted my server it did not want > to > boot, do I had to power it off and then back on and it then booted up. But > constantly I am getting this "Bus Error - core dumped" > > anyway in my /var/crash I see hundreds of core.void files and 3 core.cpio > files. > I would imagine core.cpio are the ones that are direct result of what I am > probably eperiencing. > > -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24854 > -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24867 > -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24880 > -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24893 > -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24906 > -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24919 > -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24932 > -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24950 > -rw------- 1 root root 4126301 Nov 5 19:22 core.vold.24978 > drwxr-xr-x 3 root root 81408 Nov 5 20:06 . > -rw------- 1 root root 31351099 Nov 5 20:06 core.cpio.6208 > > > > On Wed, 5 Nov 2008, Enda O'Connor wrote: > >> Hi >> Looks ok, some mounts left over from pervious fail. >> In regards to swap and dump on zpool you can set them >> zfs set volsize=1G rootpool/dump >> zfs set volsize=1G rootpool/swap >> >> for instance, of course above are only an example of how to do it. >> or make the zvol doe rootpool/dump etc before lucreate, in which case it will >> take the swap and dump size you have preset. >> >> But I think we need to see the coredump/truss at this point to get an idea of >> where things went wrong. >> Enda >> >> On 11/05/08 15:38, Krzys wrote: >>> I did upgrade my U5 to U6 from DVD, went trough the upgrade process. >>> my file system is setup as follow: >>> [10:11:54] [EMAIL PROTECTED]: /root > df -h | egrep -v >>> "platform|sharefs|objfs|mnttab|proc|ctfs|devices|fd|nsr" >>> Filesystem size used avail capacity Mounted on >>> /dev/dsk/c1t0d0s0 16G 7.2G 8.4G 47% / >>> swap 8.3G 1.5M 8.3G 1% /etc/svc/volatile >>> /dev/dsk/c1t0d0s6 16G 8.7G 6.9G 56% /usr >>> /dev/dsk/c1t0d0s1 16G 2.5G 13G 17% /var >>> swap 8.5G 229M 8.3G 3% /tmp >>> swap 8.3G 40K 8.3G 1% /var/run >>> /dev/dsk/c1t0d0s7 78G 1.2G 76G 2% /export/home >>> rootpool 33G 19K 21G 1% /rootpool >>> rootpool/ROOT 33G 18K 21G 1% /rootpool/ROOT >>> rootpool/ROOT/zfsBE 33G 31M 21G 1% /.alt.tmp.b-UUb.mnt >>> /export/home 78G 1.2G 76G 2% >>> /.alt.tmp.b-UUb.mnt/export/home >>> /rootpool 21G 19K 21G 1% >>> /.alt.tmp.b-UUb.mnt/rootpool >>> /rootpool/ROOT 21G 18K 21G 1% >>> /.alt.tmp.b-UUb.mnt/rootpool/ROOT >>> swap 8.3G 0K 8.3G 0% >>> /.alt.tmp.b-UUb.mnt/var/run >>> swap 8.3G 0K 8.3G 0% /.alt.tmp.b-UUb.mnt/tmp >>> [10:12:00] [EMAIL PROTECTED]: /root > >>> >>> >>> so I have /, /usr, /var and /export/home on that primary disk. Original >>> disk is 140gb, this new one is only 36gb, but disk utilization on that >>> primary disk is much less utilized so easily should fit on it. >>> >>> / 7.2GB >>> /usr 8.7GB >>> /var 2.5GB >>> /export/home 1.2GB >>> total space 19.6GB >>> I did notice that lucreate did alocate 8GB to SWAP and 4GB to DUMP >>> total space needed 31.6GB >>> seems like total available disk space on my disk should be 33.92GB >>> so its quite close as both numbers do approach. So to make sure I will >>> change disk for 72gb and will try again. I do not beleive that I need to >>> match my main disk size as 146gb as I am not using that much disk space on >>> it. But let me try this and it might be why I am getting this problem... >>> >>> >>> >>> On Wed, 5 Nov 2008, Enda O'Connor wrote: >>> >>>> Hi Krzys >>>> Also some info on the actual system >>>> ie what was it upgraded to u6 from and how. >>>> and an idea of how the filesystems are laid out, ie is usr seperate from / >>>> and so on ( maybe a df -k ). Don't appear to have any zones installed, >>>> just to confirm. >>>> Enda >>>> >>>> On 11/05/08 14:07, Enda O'Connor wrote: >>>>> Hi >>>>> did you get a core dump? >>>>> would be nice to see the core file to get an idea of what dumped core, >>>>> might configure coreadm if not already done >>>>> run coreadm first, if the output looks like >>>>> >>>>> # coreadm >>>>> global core file pattern: /var/crash/core.%f.%p >>>>> global core file content: default >>>>> init core file pattern: core >>>>> init core file content: default >>>>> global core dumps: enabled >>>>> per-process core dumps: enabled >>>>> global setid core dumps: enabled >>>>> per-process setid core dumps: disabled >>>>> global core dump logging: enabled >>>>> >>>>> then all should be good, and cores should appear in /var/crash >>>>> >>>>> otherwise the following should configure coreadm: >>>>> coreadm -g /var/crash/core.%f.%p >>>>> coreadm -G all >>>>> coreadm -e global >>>>> coreadm -e per-process >>>>> >>>>> >>>>> coreadm -u to load the new settings without rebooting. >>>>> >>>>> also might need to set the size of the core dump via >>>>> ulimit -c unlimited >>>>> check ulimit -a first. >>>>> >>>>> then rerun test and check /var/crash for core dump. >>>>> >>>>> If that fails a truss via say truss -fae -o /tmp/truss.out lucreate -c >>>>> ufsBE -n zfsBE -p rootpool >>>>> >>>>> might give an indication, look for SIGBUS in the truss log >>>>> >>>>> NOTE, that you might want to reset the coreadm and ulimit for coredumps >>>>> after this, in order to not risk filling the system with coredumps in the >>>>> case of some utility coredumping in a loop say. >>>>> >>>>> >>>>> Enda > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > > !DSPAM:122,4912484314258371292! > _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss