Hi,
 
I know, this would be more a message to the LU-list, but I'm not
subscribed there (is there such a list?), but here, and the problem is
at least zfs related:
 
I have a Solaris 10 installation from last year, it's a jumpstarted u6
with zfs rootpool and some (zfs-rooted) zones. Last week I did a
luupgrade which worked like a charm.
Now I inteded to install some more patches which I overlooked last week
and decided to do that with live upgrade again.
 
But though I'm able to create a new BE (with some ERRORs, see below), a
luupgrade fails and leaves an orphaned /a/var/run and a broken
/a/zones/myzonename.
 
The errors lucreate delivers:
----8<----
Creating snapshot for <rootpool/ROOT/s10u8-01/zones> on
rootpool/ROOT/s10u8-01/zo...@s10u8-2010010
<mailto:rootpool/ROOT/s10u8-01/zo...@s10u8-20100106> 6.
Creating clone for rootpool/ROOT/s10u8-01/zo...@s10u8-2010010
<mailto:rootpool/ROOT/s10u8-01/zo...@s10u8-20100106> 6 on
<rootpool/ROOT/s10u8-20100106/zones>.
Setting canmount=noauto for </zones> in zone <global> on
<rootpool/ROOT/s10u8-20100106/zones>.
Creating snapshot for <rootpool/ROOT/s10u8-01/zones/myzonename> on
<rootpool/ROOT/s10u8-01/zones/myzonen...@s10u8-20100106>.
Creating clone for rootpool/ROOT/s10u8-01/zones/myzonem...@s10u8-2010010
<mailto:rootpool/ROOT/s10u8-01/zones/myzonem...@s10u8-20100106> 6 on
<rootpool/ROOT/s10u8-20100106/zones/myzonename-s10u8-20100106>.
cannot mount
'rootpool/ROOT/s10u8-20100106/zones/myzonename-s10u8-20100106': legacy
mountpoint
use mount(1M) to mount this filesystem
ERROR: Failed to mount dataset
<rootpool/ROOT/s10u8-20100106/zones/myzonename-s10u8-20100106>
legacy is not an absolute path.
Population of boot environment <s10u8-20100106> successful.
Creation of boot environment <s10u8-20100106> successful.
---8<---
 
and here is, what luupgrade does:
 
---8<---
[0]r...@global[~]>>luupgrade -t -n s10u8-20100106 -s /root/patches
 
Validating the contents of the media </root/patches>.
The media contains 106 software patches that can be added.
All 106 patches will be added because you did not specify any specific
patches to add.
Mounting the BE <s10u8-20100106>.
ERROR: unable to mount zones:
zoneadm: zone 'myzonename': zone root /zones/myzonename/root already in
use by zone myzonename
zoneadm: zone 'myzonename': call to zoneadmd failed
ERROR: unable to mount zone <myzonename> in </a>
ERROR: unmounting partially mounted boot environment file systems
ERROR: cannot mount boot environment by icf file
</tmp/.luupgrade.beicf.22173>
cat: cannot open /tmp/.luupgrade.tmp.22173
ERROR: Unable to mount ABE disk slices: < >.
ERROR: Unable to mount the BE <s10u8-20100106>.
[0]r...@global[~]>>df -hl
Filesystem             size   used  avail capacity  Mounted on
rootpool/ROOT/s10u8-01
                       134G   8.4G    91G     9%    /
/devices                 0K     0K     0K     0%    /devices
ctfs                     0K     0K     0K     0%    /system/contract
proc                     0K     0K     0K     0%    /proc
mnttab                   0K     0K     0K     0%    /etc/mnttab
swap                    10G   416K    10G     1%    /etc/svc/volatile
objfs                    0K     0K     0K     0%    /system/object
sharefs                  0K     0K     0K     0%    /etc/dfs/sharetab
/platform/SUNW,SPARC-Enterprise-T5220/lib/libc_psr/libc_psr_hwcap2.so.1
                        99G   8.4G    91G     9%
/platform/sun4v/lib/libc_psr.so.1
/platform/SUNW,SPARC-Enterprise-T5220/lib/sparcv9/libc_psr/libc_psr_hwca
p2.so.1
                        99G   8.4G    91G     9%
/platform/sun4v/lib/sparcv9/libc_psr.so.1
fd                       0K     0K     0K     0%    /dev/fd
swap                    10G    56K    10G     1%    /tmp
swap                    10G    80K    10G     1%    /var/run
rootpool/ROOT/s10u8-01/zones
                       134G    27K    91G     1%    /zones
rootpool/ROOT/s10u8-01/zones/myzonename
                       134G   1.3G    91G     2%    /zones/myzonename
rootpool/export        134G    20K    91G     1%    /export
rootpool/export/home   134G    18K    91G     1%    /export/home
rootpool               134G    96K    91G     1%    /rootpool
rootpool/ROOT/s10u8-20100106
                       134G    13G    91G    13%    /a
swap                    10G     0K    10G     0%    /a/var/run
df: cannot statvfs /a/zones/myzonename: No such file or directory
---8<---
 
anyone here, who can help out with what I did wrong?

 

Greetings

Jan Dreyer

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to