Hi Looks ok, some mounts left over from pervious fail. In regards to swap and dump on zpool you can set them zfs set volsize=1G rootpool/dump zfs set volsize=1G rootpool/swap
for instance, of course above are only an example of how to do it. or make the zvol doe rootpool/dump etc before lucreate, in which case it will take the swap and dump size you have preset. But I think we need to see the coredump/truss at this point to get an idea of where things went wrong. Enda On 11/05/08 15:38, Krzys wrote: > I did upgrade my U5 to U6 from DVD, went trough the upgrade process. > my file system is setup as follow: > [10:11:54] [EMAIL PROTECTED]: /root > df -h | egrep -v > "platform|sharefs|objfs|mnttab|proc|ctfs|devices|fd|nsr" > Filesystem size used avail capacity Mounted on > /dev/dsk/c1t0d0s0 16G 7.2G 8.4G 47% / > swap 8.3G 1.5M 8.3G 1% /etc/svc/volatile > /dev/dsk/c1t0d0s6 16G 8.7G 6.9G 56% /usr > /dev/dsk/c1t0d0s1 16G 2.5G 13G 17% /var > swap 8.5G 229M 8.3G 3% /tmp > swap 8.3G 40K 8.3G 1% /var/run > /dev/dsk/c1t0d0s7 78G 1.2G 76G 2% /export/home > rootpool 33G 19K 21G 1% /rootpool > rootpool/ROOT 33G 18K 21G 1% /rootpool/ROOT > rootpool/ROOT/zfsBE 33G 31M 21G 1% /.alt.tmp.b-UUb.mnt > /export/home 78G 1.2G 76G 2% > /.alt.tmp.b-UUb.mnt/export/home > /rootpool 21G 19K 21G 1% > /.alt.tmp.b-UUb.mnt/rootpool > /rootpool/ROOT 21G 18K 21G 1% > /.alt.tmp.b-UUb.mnt/rootpool/ROOT > swap 8.3G 0K 8.3G 0% > /.alt.tmp.b-UUb.mnt/var/run > swap 8.3G 0K 8.3G 0% /.alt.tmp.b-UUb.mnt/tmp > [10:12:00] [EMAIL PROTECTED]: /root > > > > so I have /, /usr, /var and /export/home on that primary disk. Original > disk is 140gb, this new one is only 36gb, but disk utilization on that > primary disk is much less utilized so easily should fit on it. > > / 7.2GB > /usr 8.7GB > /var 2.5GB > /export/home 1.2GB > total space 19.6GB > I did notice that lucreate did alocate 8GB to SWAP and 4GB to DUMP > total space needed 31.6GB > seems like total available disk space on my disk should be 33.92GB > so its quite close as both numbers do approach. So to make sure I will > change disk for 72gb and will try again. I do not beleive that I need to > match my main disk size as 146gb as I am not using that much disk space > on it. But let me try this and it might be why I am getting this problem... > > > > On Wed, 5 Nov 2008, Enda O'Connor wrote: > >> Hi Krzys >> Also some info on the actual system >> ie what was it upgraded to u6 from and how. >> and an idea of how the filesystems are laid out, ie is usr seperate >> from / and so on ( maybe a df -k ). Don't appear to have any zones >> installed, just to confirm. >> Enda >> >> On 11/05/08 14:07, Enda O'Connor wrote: >>> Hi >>> did you get a core dump? >>> would be nice to see the core file to get an idea of what dumped core, >>> might configure coreadm if not already done >>> run coreadm first, if the output looks like >>> >>> # coreadm >>> global core file pattern: /var/crash/core.%f.%p >>> global core file content: default >>> init core file pattern: core >>> init core file content: default >>> global core dumps: enabled >>> per-process core dumps: enabled >>> global setid core dumps: enabled >>> per-process setid core dumps: disabled >>> global core dump logging: enabled >>> >>> then all should be good, and cores should appear in /var/crash >>> >>> otherwise the following should configure coreadm: >>> coreadm -g /var/crash/core.%f.%p >>> coreadm -G all >>> coreadm -e global >>> coreadm -e per-process >>> >>> >>> coreadm -u to load the new settings without rebooting. >>> >>> also might need to set the size of the core dump via >>> ulimit -c unlimited >>> check ulimit -a first. >>> >>> then rerun test and check /var/crash for core dump. >>> >>> If that fails a truss via say truss -fae -o /tmp/truss.out lucreate >>> -c ufsBE -n zfsBE -p rootpool >>> >>> might give an indication, look for SIGBUS in the truss log >>> >>> NOTE, that you might want to reset the coreadm and ulimit for >>> coredumps after this, in order to not risk filling the system with >>> coredumps in the case of some utility coredumping in a loop say. >>> >>> >>> Enda >>> On 11/05/08 13:46, Krzys wrote: >>>> >>>> On Wed, 5 Nov 2008, Enda O'Connor wrote: >>>> >>>>> On 11/05/08 13:02, Krzys wrote: >>>>>> I am not sure what I did wrong but I did follow up all the steps >>>>>> to get my system moved from ufs to zfs and not I am unable to boot >>>>>> it... can anyone suggest what I could do to fix it? >>>>>> >>>>>> here are all my steps: >>>>>> >>>>>> [00:26:38] @adas: /root > zpool create rootpool c1t1d0s0 >>>>>> [00:26:57] @adas: /root > lucreate -c ufsBE -n zfsBE -p rootpool >>>>>> Analyzing system configuration. >>>>>> Comparing source boot environment <ufsBE> file systems with the file >>>>>> system(s) you specified for the new boot environment. Determining >>>>>> which >>>>>> file systems should be in the new boot environment. >>>>>> Updating boot environment description database on all BEs. >>>>>> Updating system configuration files. >>>>>> The device </dev/dsk/c1t1d0s0> is not a root device for any boot >>>>>> environment; cannot get BE ID. >>>>>> Creating configuration for boot environment <zfsBE>. >>>>>> Source boot environment is <ufsBE>. >>>>>> Creating boot environment <zfsBE>. >>>>>> Creating file systems on boot environment <zfsBE>. >>>>>> Creating <zfs> file system for </> in zone <global> on >>>>>> <rootpool/ROOT/zfsBE>. >>>>>> Populating file systems on boot environment <zfsBE>. >>>>>> Checking selection integrity. >>>>>> Integrity check OK. >>>>>> Populating contents of mount point </>. >>>>>> Copying. >>>>>> Bus Error - core dumped >>>>> hmm above might be relevant I'd guess. >>>>> >>>>> What release are you on , ie is this Solaris 10, or is this Nevada >>>>> build? >>>>> >>>>> Enda >>>>>> Creating shared file system mount points. >>>>>> Creating compare databases for boot environment <zfsBE>. >>>>>> Creating compare database for file system </var>. >>>>>> Creating compare database for file system </usr>. >>>>>> Creating compare database for file system </rootpool/ROOT>. >>>>>> Creating compare database for file system </>. >>>>>> Updating compare databases on boot environment <zfsBE>. >>>>>> Making boot environment <zfsBE> bootable. >>>> >>>> Anyway I did restart the whole process again, and I got again that >>>> Bus Error >>>> >>>> [07:59:01] [EMAIL PROTECTED]: /root > zpool create rootpool c1t1d0s0 >>>> [07:59:22] [EMAIL PROTECTED]: /root > zfs set compression=on rootpool/ROOT >>>> cannot open 'rootpool/ROOT': dataset does not exist >>>> [07:59:27] [EMAIL PROTECTED]: /root > zfs set compression=on rootpool >>>> [07:59:31] [EMAIL PROTECTED]: /root > lucreate -c ufsBE -n zfsBE -p >>>> rootpool >>>> Analyzing system configuration. >>>> Comparing source boot environment <ufsBE> file systems with the file >>>> system(s) you specified for the new boot environment. Determining which >>>> file systems should be in the new boot environment. >>>> Updating boot environment description database on all BEs. >>>> Updating system configuration files. >>>> The device </dev/dsk/c1t1d0s0> is not a root device for any boot >>>> environment; cannot get BE ID. >>>> Creating configuration for boot environment <zfsBE>. >>>> Source boot environment is <ufsBE>. >>>> Creating boot environment <zfsBE>. >>>> Creating file systems on boot environment <zfsBE>. >>>> Creating <zfs> file system for </> in zone <global> on >>>> <rootpool/ROOT/zfsBE>. >>>> Populating file systems on boot environment <zfsBE>. >>>> Checking selection integrity. >>>> Integrity check OK. >>>> Populating contents of mount point </>. >>>> Copying. >>>> Bus Error - core dumped >>>> Creating shared file system mount points. >>>> Creating compare databases for boot environment <zfsBE>. >>>> Creating compare database for file system </var>. >>>> Creating compare database for file system </usr>. >>>> >>>> >>>> >>>> _______________________________________________ >>>> zfs-discuss mailing list >>>> zfs-discuss@opensolaris.org >>>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss >>> >>> >> >> >> -- >> Enda O'Connor x19781 Software Product Engineering >> Patch System Test : Ireland : x19781/353-1-8199718 >> >> >> !DSPAM:122,4911ac7c27292151120594! >> -- Enda O'Connor x19781 Software Product Engineering Patch System Test : Ireland : x19781/353-1-8199718 _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss