Hi Amy,

This is a known problem with ZFS and live upgrade.  I believe the docs for 
s10u6 discourage the config you show here.  A patch should be ready some 
time next month with a fix for this.


On Fri, 16 Jan 2009, amy.r...@tufts.edu wrote:

> I've installed an s10u6 machine with no UFS partitions at all.  I've created a
> dataset for zones and one for a zone named "default."  I then do an lucreate
> and luactivate and a subsequent boot off the new BE.  All of that appears to
> go just fine (though I've found that I MUST call the zone dataset zoneds for
> some reason, or it will rename it ot that for me).  When I try to delete the
> old BE, it fails with the following message:
>
> # ludelete s10-RC
> ERROR: cannot mount '/zoneds': directory is not empty
> ERROR: cannot mount mount point </.alt.tmp.b-VK.mnt/zoneds> device 
> <rpool/ROOT/s10-RC/zoneds>
> ERROR: failed to mount file system <rpool/ROOT/s10-RC/zoneds> on 
> </.alt.tmp.b-VK.mnt/zoneds>
> ERROR: unmounting partially mounted boot environment file systems
> ERROR: cannot mount boot environment by icf file </etc/lu/ICF.1>
> ERROR: Cannot mount BE <s10-RC>.
> Unable to delete boot environment.
>
> It's obvious that luactivate is not correctly resetting the mount point of
> /zoneds and /zoneds/default (the zone named default) in the old BE so that
> it's under /.alt like the rest of the ROOT dataset:
>
> # zfs list |grep s10-RC
> rpool/ROOT/s10-RC                  14.6M  57.3G  1.29G  /.alt.tmp.b-VK.mnt/
> rpool/ROOT/s10-RC/var              2.69M  57.3G  21.1M  
> /.alt.tmp.b-VK.mnt//var
> rpool/ROOT/s10-RC/zoneds           5.56M  57.3G    19K  /zoneds
> rpool/ROOT/s10-RC/zoneds/default   5.55M  57.3G  29.9M  /zoneds/default
>
> Obviously I can reset the mount points by hand with "zfs set mountpoint," but
> this seems like something that luactivate and the subsequent boot should
> handle.  Is this a bug, or am I missing a step/have something misconfigured?
>
> Also, once I run ludelete on a BE, it seems like it should also clean up the
> old ZFS filesystems for the BE s10-RC (the old BE) instead of me having to do
> an explicit zfs destroy.
>
> The very weird thing is that, if I run lucreate again (new BE is named bar)
> and boot off of the new BE, it does the right thing with the old BE (foo):
>
> rpool/ROOT/bar                     1.52G  57.2G  1.29G  /
> rpool/ROOT/b...@foo                 89.1M      -  1.29G  -
> rpool/ROOT/b...@bar                 84.1M      -  1.29G  -
> rpool/ROOT/bar/var                 24.7M  57.2G  21.2M  /var
> rpool/ROOT/bar/v...@foo             2.64M      -  21.0M  -
> rpool/ROOT/bar/v...@bar              923K      -  21.2M  -
> rpool/ROOT/bar/zoneds              32.7M  57.2G    20K  /zoneds
> rpool/ROOT/bar/zon...@foo            18K      -    19K  -
> rpool/ROOT/bar/zon...@bar            19K      -    20K  -
> rpool/ROOT/bar/zoneds/default      32.6M  57.2G  29.9M  /zoneds/default
> rpool/ROOT/bar/zoneds/defa...@foo  2.61M      -  27.0M  -
> rpool/ROOT/bar/zoneds/defa...@bar   162K      -  29.9M  -
> rpool/ROOT/foo                     2.93M  57.2G  1.29G  /.alt.foo
> rpool/ROOT/foo/var                  818K  57.2G  21.2M  /.alt.foo/var
> rpool/ROOT/foo/zoneds               270K  57.2G    20K  /.alt.foo/zoneds
> rpool/ROOT/foo/zoneds/default       253K  57.2G  29.9M  
> /.alt.foo/zoneds/default
>
> And then DOES clean up the zfs filesystem when I run ludelete.  Does anyone
> know where there's a discrepancy?  The same lucreate command (-n <BE> -p
> rpool) command was used both times.
>
>
>
>
>
> _______________________________________________
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>


Regards,
markm
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to