what makes me wonder is why I am not even able to see anything under boot -L ? 
and it is just not seeing this disk as a boot device? so strange.

On Wed, 5 Nov 2008, Krzys wrote:

> THis is so bizare, I am unable to pass this problem. I though I had not enough
> space on my hard drive (new one) so I replaced it with 72gb drive, but still
> getting that bus error. Originally when I restarted my server it did not want 
> to
> boot, do I had to power it off and then back on and it then booted up. But
> constantly I am getting this "Bus Error - core dumped"
>
> anyway in my /var/crash I see hundreds of core.void files and 3 core.cpio 
> files.
> I would imagine core.cpio are the ones that are direct result of what I am
> probably eperiencing.
>
> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24854
> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24867
> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24880
> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24893
> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24906
> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24919
> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24932
> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24950
> -rw-------   1 root     root     4126301 Nov  5 19:22 core.vold.24978
> drwxr-xr-x   3 root     root       81408 Nov  5 20:06 .
> -rw-------   1 root     root     31351099 Nov  5 20:06 core.cpio.6208
>
>
>
> On Wed, 5 Nov 2008, Enda O'Connor wrote:
>
>> Hi
>> Looks ok, some mounts left over from pervious fail.
>> In regards to swap and dump on zpool you can set them
>> zfs set volsize=1G rootpool/dump
>> zfs set volsize=1G rootpool/swap
>>
>> for instance, of course above are only an example of how to do it.
>> or make the zvol doe rootpool/dump etc before lucreate, in which case it will
>> take the swap and dump size you have preset.
>>
>> But I think we need to see the coredump/truss at this point to get an idea of
>> where things went wrong.
>> Enda
>>
>> On 11/05/08 15:38, Krzys wrote:
>>> I did upgrade my U5 to U6 from DVD, went trough the upgrade process.
>>> my file system is setup as follow:
>>> [10:11:54] [EMAIL PROTECTED]: /root > df -h | egrep -v
>>> "platform|sharefs|objfs|mnttab|proc|ctfs|devices|fd|nsr"
>>> Filesystem             size   used  avail capacity  Mounted on
>>> /dev/dsk/c1t0d0s0       16G   7.2G   8.4G    47%    /
>>> swap                   8.3G   1.5M   8.3G     1%    /etc/svc/volatile
>>> /dev/dsk/c1t0d0s6       16G   8.7G   6.9G    56%    /usr
>>> /dev/dsk/c1t0d0s1       16G   2.5G    13G    17%    /var
>>> swap                   8.5G   229M   8.3G     3%    /tmp
>>> swap                   8.3G    40K   8.3G     1%    /var/run
>>> /dev/dsk/c1t0d0s7       78G   1.2G    76G     2%    /export/home
>>> rootpool                33G    19K    21G     1%    /rootpool
>>> rootpool/ROOT           33G    18K    21G     1%    /rootpool/ROOT
>>> rootpool/ROOT/zfsBE     33G    31M    21G     1%    /.alt.tmp.b-UUb.mnt
>>> /export/home            78G   1.2G    76G     2%
>>> /.alt.tmp.b-UUb.mnt/export/home
>>> /rootpool               21G    19K    21G     1%
>>> /.alt.tmp.b-UUb.mnt/rootpool
>>> /rootpool/ROOT          21G    18K    21G     1%
>>> /.alt.tmp.b-UUb.mnt/rootpool/ROOT
>>> swap                   8.3G     0K   8.3G     0%
>>> /.alt.tmp.b-UUb.mnt/var/run
>>> swap                   8.3G     0K   8.3G     0%    /.alt.tmp.b-UUb.mnt/tmp
>>> [10:12:00] [EMAIL PROTECTED]: /root >
>>>
>>>
>>> so I have /, /usr, /var and /export/home on that primary disk. Original
>>> disk is 140gb, this new one is only 36gb, but disk utilization on that
>>> primary disk is much less utilized so easily should fit on it.
>>>
>>> / 7.2GB
>>> /usr 8.7GB
>>> /var 2.5GB
>>> /export/home 1.2GB
>>> total space 19.6GB
>>> I did notice that lucreate did alocate 8GB to SWAP and 4GB to DUMP
>>> total space needed 31.6GB
>>> seems like total available disk space on my disk should be 33.92GB
>>> so its quite close as both numbers do approach. So to make sure I will
>>> change disk for 72gb and will try again. I do not beleive that I need to
>>> match my main disk size as 146gb as I am not using that much disk space on
>>> it. But let me try this and it might be why I am getting this problem...
>>>
>>>
>>>
>>> On Wed, 5 Nov 2008, Enda O'Connor wrote:
>>>
>>>> Hi Krzys
>>>> Also some info on the actual system
>>>> ie what was it upgraded to u6 from and how.
>>>> and an idea of how the filesystems are laid out, ie is usr seperate from /
>>>> and so on ( maybe a df -k ). Don't appear to have any zones installed,
>>>> just to confirm.
>>>> Enda
>>>>
>>>> On 11/05/08 14:07, Enda O'Connor wrote:
>>>>> Hi
>>>>> did you get a core dump?
>>>>> would be nice to see the core file to get an idea of what dumped core,
>>>>> might configure coreadm if not already done
>>>>> run coreadm first, if the output looks like
>>>>>
>>>>> # coreadm
>>>>>      global core file pattern: /var/crash/core.%f.%p
>>>>>      global core file content: default
>>>>>        init core file pattern: core
>>>>>        init core file content: default
>>>>>             global core dumps: enabled
>>>>>        per-process core dumps: enabled
>>>>>       global setid core dumps: enabled
>>>>>  per-process setid core dumps: disabled
>>>>>      global core dump logging: enabled
>>>>>
>>>>> then all should be good, and cores should appear in /var/crash
>>>>>
>>>>> otherwise the following should configure coreadm:
>>>>> coreadm -g /var/crash/core.%f.%p
>>>>> coreadm -G all
>>>>> coreadm -e global
>>>>> coreadm -e per-process
>>>>>
>>>>>
>>>>> coreadm -u to load the new settings without rebooting.
>>>>>
>>>>> also might need to set the size of the core dump via
>>>>> ulimit -c unlimited
>>>>> check ulimit -a first.
>>>>>
>>>>> then rerun test and check /var/crash for core dump.
>>>>>
>>>>> If that fails a truss via say truss -fae -o /tmp/truss.out lucreate -c
>>>>> ufsBE -n zfsBE -p rootpool
>>>>>
>>>>> might give an indication, look for SIGBUS in the truss log
>>>>>
>>>>> NOTE, that you might want to reset the coreadm and ulimit for coredumps
>>>>> after this, in order to not risk filling the system with coredumps in the
>>>>> case of some utility coredumping in a loop say.
>>>>>
>>>>>
>>>>> Enda
> _______________________________________________
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
>
> !DSPAM:122,4912484314258371292!
>
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to