Hi Amy, This is a known problem with ZFS and live upgrade. I believe the docs for s10u6 discourage the config you show here. A patch should be ready some time next month with a fix for this.
On Fri, 16 Jan 2009, amy.r...@tufts.edu wrote: > I've installed an s10u6 machine with no UFS partitions at all. I've created a > dataset for zones and one for a zone named "default." I then do an lucreate > and luactivate and a subsequent boot off the new BE. All of that appears to > go just fine (though I've found that I MUST call the zone dataset zoneds for > some reason, or it will rename it ot that for me). When I try to delete the > old BE, it fails with the following message: > > # ludelete s10-RC > ERROR: cannot mount '/zoneds': directory is not empty > ERROR: cannot mount mount point </.alt.tmp.b-VK.mnt/zoneds> device > <rpool/ROOT/s10-RC/zoneds> > ERROR: failed to mount file system <rpool/ROOT/s10-RC/zoneds> on > </.alt.tmp.b-VK.mnt/zoneds> > ERROR: unmounting partially mounted boot environment file systems > ERROR: cannot mount boot environment by icf file </etc/lu/ICF.1> > ERROR: Cannot mount BE <s10-RC>. > Unable to delete boot environment. > > It's obvious that luactivate is not correctly resetting the mount point of > /zoneds and /zoneds/default (the zone named default) in the old BE so that > it's under /.alt like the rest of the ROOT dataset: > > # zfs list |grep s10-RC > rpool/ROOT/s10-RC 14.6M 57.3G 1.29G /.alt.tmp.b-VK.mnt/ > rpool/ROOT/s10-RC/var 2.69M 57.3G 21.1M > /.alt.tmp.b-VK.mnt//var > rpool/ROOT/s10-RC/zoneds 5.56M 57.3G 19K /zoneds > rpool/ROOT/s10-RC/zoneds/default 5.55M 57.3G 29.9M /zoneds/default > > Obviously I can reset the mount points by hand with "zfs set mountpoint," but > this seems like something that luactivate and the subsequent boot should > handle. Is this a bug, or am I missing a step/have something misconfigured? > > Also, once I run ludelete on a BE, it seems like it should also clean up the > old ZFS filesystems for the BE s10-RC (the old BE) instead of me having to do > an explicit zfs destroy. > > The very weird thing is that, if I run lucreate again (new BE is named bar) > and boot off of the new BE, it does the right thing with the old BE (foo): > > rpool/ROOT/bar 1.52G 57.2G 1.29G / > rpool/ROOT/b...@foo 89.1M - 1.29G - > rpool/ROOT/b...@bar 84.1M - 1.29G - > rpool/ROOT/bar/var 24.7M 57.2G 21.2M /var > rpool/ROOT/bar/v...@foo 2.64M - 21.0M - > rpool/ROOT/bar/v...@bar 923K - 21.2M - > rpool/ROOT/bar/zoneds 32.7M 57.2G 20K /zoneds > rpool/ROOT/bar/zon...@foo 18K - 19K - > rpool/ROOT/bar/zon...@bar 19K - 20K - > rpool/ROOT/bar/zoneds/default 32.6M 57.2G 29.9M /zoneds/default > rpool/ROOT/bar/zoneds/defa...@foo 2.61M - 27.0M - > rpool/ROOT/bar/zoneds/defa...@bar 162K - 29.9M - > rpool/ROOT/foo 2.93M 57.2G 1.29G /.alt.foo > rpool/ROOT/foo/var 818K 57.2G 21.2M /.alt.foo/var > rpool/ROOT/foo/zoneds 270K 57.2G 20K /.alt.foo/zoneds > rpool/ROOT/foo/zoneds/default 253K 57.2G 29.9M > /.alt.foo/zoneds/default > > And then DOES clean up the zfs filesystem when I run ludelete. Does anyone > know where there's a discrepancy? The same lucreate command (-n <BE> -p > rpool) command was used both times. > > > > > > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > Regards, markm _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss