Thanks for the reply!
My /var issue was a problem with my process.. I did:
install UFS root, install extra stuff, create UFS flar.
lucreate ZFS root <- this is where my /var dissappeared
flar create zfs flar
In my straight zfsroot install profiles I have a seperate /var - I didn't even
think to check post lucreate. Easy enough to split /var back out into a dataset
before I do the flar create. I'm trying to get my zfsroot flar and my ufs flar
to be as similar as possible for the enterprise environment.
/tmp is still a mystery.. it's tmpfs on the source system and the vfstab entry
didn't survive the zfsroot flash install. That's super simple to fix in a
finish script
Source system:
10:57:53 uigvbp234(5)> df -lk /tmp /
Filesystem kbytes used avail capacity Mounted on
swap 6066328 2248 6064080 1% /tmp
rpool/ROOT/s10-zfs 34836480 5905781 19547778 24% /
zfsroot flash installed system:
10:57:39 uigvbp234v(8)> df -lk /tmp /
Filesystem kbytes used avail capacity Mounted on
rpool/ROOT/s10-zfs 70189056 6153504 49241246 12% /
rpool/ROOT/s10-zfs 70189056 6153504 49241246 12% /
Thanks again for all of the information, and for helping make zfs flash
installs a reality!
--Kris
Yesterday at 15:13, Lori Alt <lori....@sun.com> wrote:
On 08/19/09 14:57, Kris Kasner wrote:
Hi.
First - Thanks very much for releasing this as a patch and not making
us wait until U8. It's very much appreciated. Things we put on hold
can start moving again which is a good thing.
Is there further documentation on this yet?
I just asked Cindy Swearingen, the tech writer for ZFS, about this and
sadly, it appears that there isn't any documentation for this available
outside of Sun yet. The documentation for using flash archives to set
up systems with zfs roots won't be available until the release of U8. I
will give you the very short summary and Cindy is going to look into
putting some more information out on the zfs-boot page of the
opensolaris website (which doesn't really make sense since this isnt'
supported on OpenSolaris, but whatever).
I haven't been able to find it, and it looks like most of the
interesting stuff is in 'pfinstall' so I can't just go look and see
how it works.. I have a couple of questions on how zfsroot interacts
with flash install - I'm happy to go figure it out if you can point me
at the docs.. Failing that, hopefully someone can help me answer the
below questions.
on our zfs root systems, we like to keep /var as a separate dataset on
servers so we can monitor it's usage. We did that before with the line:
bootenv installbe bename zfsroot dataset /var
Bottom line: whatever the configuration is of the system that you use as
the basis for generating the flash archive, that's the configuration
that will exist on systems installed using the flar. If the original
system has a separate /var, the systems created from it will have a
separate /var.
additional datasets I create in a finish script.
Can a zfsroot flash installed system have a separate dataset for /var?
It seems a little counter intuitive to me to say
partitioning explicit
when everything is going into a zfs pool with a single dataset - are
there other profile tokens that make this.
Yes, I know, but that's the way flash archive support works (even
before zfs was added to it).
One last thing - My flash installs are working, but I'm ending up with
/tmp as a non tmpfs filesystem (no vfstab entry for it..) I can easily
fix this in a postinstall/finish script, but if there is a reason it's
not happening that can be fixed in my JS profile, that would be better.
I don't know why you wouldn't have a separate /tmp. If you're sure that
the system from which the flar is generated has a /tmp file system, but
the system that results from the install of the flash archive does not
have a separate /tmp, there might be a bug. Get back to me on it
later if you confirm this.
So here's the quick summary of how to do flash archive installs of
systems with zfs root file systems:
1. Set up the system that you want to replicate or re-create. I.e.
install a system with a zfs root file system and add whatever
customizations you want to it. By default, ALL datasets in the root
pool (except for swap and dump) are included in the flash archive, not
just those that are part of the Solaris name space.
2. Use flarcreate to create the flash archive on the system. There is
a new -D <dataset> option that can be used (repeatedly, if you wish) on
the command to exclude datasets from the archive.
So an example command might be:
# flarcreate -n ourU8 -D rpool/somedata /some-remote-directory/ourU8.flar
3. Set up the system you want to install for jumpstart install (however
you do that). Use a profile something like this:
install_type flash_install
archive_location nfs schubert:/export/home/lalt/U8.flar
partitioning explicit
pool rpool auto auto auto mirror c0t1d0s0 c0t0d0s0
I think that covers the basics.
Lori
Thanks again.
Kris Kasner
Qualcomm Inc.
Jul 9 at 16:41, Lori Alt <lori....@sun.com> wrote:
On 07/09/09 17:25, Mark Michael wrote:
Thanks for the info. Hope that the pfinstall changes to support zfs
root flash jumpstarts can be extended to support luupgrade -f at
some point soon.
BTW, where can I find an example profile? do I just substitute in the
install_type flash_install
archive_location ...
for
install_type initial_install
??
Here's a sample:
install_type flash_install
archive_location nfs schubert:/export/home/lalt/mirror.flar
partitioning explicit
pool rpool auto auto auto mirror c0t1d0s0 c0t0d0s0
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
--
Thomas Kris Kasner
Qualcomm Inc.
5775 Morehouse Drive
San Diego, CA 92121
(858)658-4932
Don't burn the day...
-- Dave Matthews "Pig"
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss