On Oct 3, 2009, at 7:46 AM, Ray Clark wrote:
Richard, with respect to:
"This has been answered several times in this thread already.
set checksum=sha256 filesystem
copy your files -- all newly written data will have the sha256
checksums."
I understand that. I understood it before the thread started. I
did not ask this. It is a fact that there is no feature to convert
checksums as a part of resilver or some such.
There is no such feature. There is a long-awaited RFE for block
pointer rewriting (checksums are stored in the block pointer)
which would add the underlying capabilities for this.
I started what utility to use, but quickly zeroed in on zfs send/
receive as being the native and presumably best method, but had
questions as to how to get the properly set when it was
automatically created. etc.
Say for example I have a pool called zwimming with some stuff in it
and compression=on. To create a copy of the data using send/recv
in the same pool but with compression=sha256 do:
zfs snapshot zwimm...@now
zfs send zwimm...@now | zfs receive zwimming/new
You will now have a new file system called "zwimming/new" with
the same data as zwimming, but with compression=sha256.
If you then want to get back to the original directory structure
you can the mountpoint properties, as desired. There are dozens
of other ways to accomplish the copy.
Note that my focus in recent portions of the thread has changed to
the underlying zpool.
Simply changing the checksum=sha356 and copying my data is analogous
to hanging my data from a hierarchy of 0.256" welded steel chain,
with the top of the hierarchy hanging it all from an 0.001" steel
thread. Well, that is not quite fair because there are
probabilities involved. Someone is going to pick a link randomly
and go after it with a fingernail clipper. If they pick a thick
one, I have very little to worry about to say the least. If they
pick one of the few dozen? hundred? thousand? (I don't know how
many) that contain the structure and services of the underlying
zpool, then the nailclipper will not be stopped by the 0.001"
thread. I do have 8,000,000,000 links in the chain, and only a a
very small fraction are 0.001" thick, and that is strongly in my
favor, but I would expect the heads to also spend a disproportionate
amount of time over the intent log. It is hard to know how it comes
out. I just don't want and 0.001" steel threads protecting my data
from the
gremlins. I moved to ZFS to avoid gambles. If I wanted gambles I
would use linux raid and lvm2. They work well enough if there are
no errors.
I think you are missing the concept of pools. Pools contain datasets.
One form of dataset is a file system. Pools do not contain data per se,
datasets contain data. Reviewing the checksums used with this
heirarchy in mind:
Pool
Label [SHA-256]
Uberblock [SHA-256]
Metadata [fletcher4]
Gang block [SHA-256]
ZIL log [fletcher2]
Dataset (file system or volume)
Metadata [fletcher4]
Data [fletcher2 (default, today), fletcher4, or SHA-256]
Send stream [fletcher4]
With this in mind, I don't understand your steel analogy.
wrt the ZIL, it is rarely used for normal file system access. ZIL
blocks
are allocated from the pool as needed and freed no more than 30
seconds later, unless there is a sudden halt. If the system is halted
then the ZIL is used to roll forward transactions. The heads do not
"spend a disproportionate amount of time over the intent log."
-- richard
I should have enumerated the knowns and unknowns in my list last
night, then I would not have annoyed you with my apparent deafness.
(Hopefully I am not still being deaf). I will clarify below, as I
should have last night:
Given that I only have 1.6TB of data in a 4TB pool, what can I do to
change those blocks to sha256 or Fletcher4:
(1) Without destroying and recreating the zpool under U4
I know how to fix the user data (just change checksum property on
the pool using zfs specifying the pool vs. a zfs file system, then
copy the data).
I don't know (am ignorant of) blocks comprising the underlying
zpool, and how to fix them without recreating the pool. It makes
sense to me that at least some would be rewritten in the course of
using the system, but (1) I have had no confirmation or denial that
this is the case, (2) I don't know if this is all of them or some of
them, (3) I don't know if the checksum= parameter would effect these
(relling's Oct 2 at 3:26 post implies that it does not by lack of
reference to checksum property). So I don't know yet how much
exposure will remain. I would think that if the user specified a
stronger checksum for their data that the system would abandon its
use of weaker ones in the underlying structure, but Richard's list
seems to imply otherwise.
(2) With destroying and recreating the zpool under U4 (Which I don't
really have the resources to pull off)
Due to some of the non-technical factors in the situation, I cannot
actually execute an experimental valid zpool command, but "zpool
create -o garbage" gives me a usage that does not include any -o or -
O. So it appears that under U4 I cannot do this. I wish there were
someone who could confirm that I can or cannot do this before I
arrange for and propose that we dive into this massive undertaking.
Also, from Richard's Oct 2 3:26 note, I infer that this will not
change the checksum used by the underlying zpool anyway, so this
might be fruitless. But I am infering... Richard gave a quick list,
his attitude was not that of providing all level of precise detail
so I really don't know. Many of the answers I have received have
turned out to recommend features that are not available in U4 but in
later versions, even unreleased versions. I have no way of sorting
this out without the information being qualified with a version.
(3) With upgrading to U7 (Perhaps in a few months)
Not clear what this will support on zpool, or if it would be
effective (similar to U4 above)
(4) With upgrading to U8
Not sure when it will come out, what it will support, or if it will
be effective (similar to U7, U4 above).
So I can enable robust protection on my user data, but perhaps not
the infrastructure needed to get at that user data, and perhaps not
the intent log.
The answer may be that I cannot be helped. That is not the desired
answer, but if that is the case, so be it. Let's lay out the facts
and the best way to move on from here, for me and everybody else.
Why leave us thrashing in the dark? Am I a Mac user?
I personally will still believe ZFS is the way to go in the short
term because it is still a vastly better gamble, and in the long
term because this too will pass as file systems are rebuilt. I
would question why anything but the best would be used for the
underlying zpool, any why there is absolutely zero presentation of
the tradeoffs between the three algorithms in the admin guide, but
that is another story.
I know that someone out there, and probably people reading this
thread know the answers to these questions. I hate to stop without
that simple knowledge being communicated. I do greatly appreciate
the attempts to work with me, I don't understand how I could be
clearer.
--
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss