I think ZFS should look for more opportunities to write to disk rather than
leaving it to the last second (5seconds) as it appears it does. e.g.
if a file has record size worth of data outstanding it should be queued within
ZFS to be written out. If the record is updated again before a txg, the
I am talking about having a write queue, which points to ready to write, full
stripes.
Ready to write full stripes would be
*The last byte of the full stripe has been updated.
*The file has been closed for writing. (Exception to the above rule)
I believe there is now a scheduler for ZFS, to han
Sorry, Full Stripe on a RaidZ is the recordsize ie if the record size is 128k
on a RaidZ and its made up of 5 disks, then 128k is spread across 4 disks with
the calc parity on the 5 disk, which means the writes are 32k to each disk.
For a RaidZ, when data is written to a disk, are individual 32k
>
> > For a RaidZ, when data is written to a disk, are
> individual 32k join together to the same disk and
> written out as a single I/O to the disk?
>
> I/Os can be coalesced, but there is no restriction as
> to what can be coalesced.
> In other words, subsequent writes can also be
> coalesced i
In /etc/hosts for the format is
IP FQDN Alias...
Which would means "1.1.1.1 aserver.google.com aserver aserver-le0"
I have seen a lot of sysadmins do the following:
"1.1.1.1 aserver aserver.google.com"
which means the host file (or NIS) does not match DNS
As the first entry is FQDN it is then "nam
pantzer5 wrote:
>
> > These days I am a fan for forward check access
> lists, because any one who
> > owns a DNS server can say that for IPAddressX
> returns aserver.google.com.
> > They can not set the forward lookup outside of
> their domain but they can
> > setup a reverse lookup. The other adv
I vote for zfs needing a backup and restore command against a snapshot.
backup command should output on stderr at least
Full_Filename SizeBytes Modification_Date_1970secSigned
so backup software can build indexes and stdout contains the data.
The advantage of zfs providing the command is that as
A system with 100TB of data its 80% full and the a user ask their local system
admin to restore a directory with large files, as it was 30days ago with all
Windows/CIFS ACLS and NFSv4/ACLS etc.
If we used zfs send, we need to go back to a zfs send some 30days ago, and find
80TB of disk space t
You could try copying the file to /tmp (ie swap/ram) and do a continues loop of
checksums e.g.
while [ ! -f ibdlpi.so.1.x ] ; do sleep 1; cp libdlpi.so.1 libdlpi.so.1.x ;
A="`sha512sum -b libdlpi.so.1.x`" ; [ "$A" == "what it should be
libdlpi.so.1.x" ] && rm libdlpi.so.1.x ; done ; date
Ass
you could also use psradm to take a CPU off-line.
At boot I would ??assume?? the system boots the same way every time unless
something changes, so you could be hiting the came CPU core every time or the
same bit of RAM until booted fully.
Or even run SunVTS "Validation Test Suite" which I beliv
In the "Thoughts on ZFS Pool Backup Strategies thread" it was stated that zfs
send, sends uncompress data and uses the ARC.
If "zfs send" sends uncompress data which has already been compress this is not
very efficient, and it would be *nice* to see it send the original compress
data. (or an op
NFSv4 has a concept of a root of the overall exported filesystem
(Pseudofilesystem).
FileHandle 0 in terms of Linux it is setting fsid=0 when exporting.
Which would explain why someone said Linux (NFSv4) automounts an exported
filesystem under another exported filesystem
ie mount servername:/
Why do we still need "/etc/zfs/zpool.cache" file???
(I could understand it was useful when zfs import was slow)
zpool import is now multi-threaded
(http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6844191), hence a
lot faster, each disk contains the hostname
(http://bugs.opensolaris
I assume the swap, dumpadm, grub is because the pool has a different name now,
but is it still a problem if you take it to a *different system* boot off a CD
change it back to rpool. (which is most likley unsupported, ie no help to get
it working)
Over 10 years ago (way before flash archive exi
> You mean /usr/sbin/sys-unconfig?
No, it does not reset a system back far enough.
You still left with the orginal path_to_inst and the device tree.
e.g. take a disk to a different system and the first disk might end up
being sd10 and c15t0d0s0 instead of sd0 and c0 without cleaning up the system
If power failure happens you will lose anything in cache. So you could lose the
entire file on power failure if the system is not busy (ie ZFS does delay
writes, unless you do a fsync before closing the file). I would still like to
see a file system option "sync on close" or even "wait for txg
Because ZFS is transaction, (effectively preserves order), the rename trick
will work.
If you find the ".filename" delete create a new ".filename" and when finish
writing rename it to "filename". If "filename" exists you no all writes were
completed. If you have a batch system which looks for th
What about removing attach/deattach and replace it with
zpool add [-fn] 'pool' submirror 'device/mirrorname' 'new_device'
e.g.
NAMESTATE READ WRITE CKSUM
rpoolONLINE 0 0 0
mirror-01 ONLINE 0 0 0
c4d0s0 ONLINE
Any small updates to a file cause the file to be fragmented.
The best Mail Box to use under Dovecot for ZFS is MailDir, each email is store
as a individual file. Fair bit of statis info is kept in a index file, but also
the filename it self is also used. The only problem with it is backups take
Before Veritas VM had support for this, you needed to use a different server to
import a disk group. You could use a different server for ZFS, which will also
take the backup load off the Server?
Cheers
--
This message posted from opensolaris.org
___
Read this http://wiki.dovecot.org/MailLocation/SharedDisk
If you were running Dovecot on the Thumper, mmap has issues under ZFS, old
versions of ZFS (not sure if it is fixed in Sol10), so switch this off
mmap_disable = yes as per the URL above for over NFS.
Ensure NFS is tuned to 32K read and
One thing which may help is the zfs import was single threaded, ie it open
every disk one disk (maybe slice) at a time and processed it, as of 128b it is
multi-threaded, ie it opens N disks/slices at once and process N disks/slices
at once. When N is the number of threads it decides to use.
htt
According to DoveCot Wiki dbox files are re-written by a secondary process. ie
delete do not happen immediately, they happen latter as a background process
and the whole message file is re-written. You can set a size limit on message
files.
Some time ago I email Tim, on a few ideas to make it m
In my previous post I was refering more to mdbox (Multi-dbox) rather than dbox,
however I beleive the meta data is store with the mail msg in version 1.x and
2.x meta is not updated within the msg which would be better for ZFS.
What I am saying is msg per file which is not updated is better for
Here is the output
-bash-4.0# uname -a
SunOS 5.11 snv_130 i86pc i386 i86pc
-bash-4.0# zfs get -r -o all compression mainfs01 | egrep -v \@
NAMEPROPERTY VALUE RECEIVED SOURCE
mainfs01 compression gzip-3- local
mainfs01/home compression g
May be look at rsync and rsync lib (http://librsync.sourceforge.net/) code to
see if a ZFS API could be design to help rsync/librsync in the future as well
as diff.
It might be a good idea for POSIX to have a single checksum and a
multi-checksum interface.
One problem could be block sizes, if
I would have thought that if I write 1k then ZFS txg times out in 30secs, then
the 1k will be written to disk in a 1k record block, and then if I write 4k
then 30secs latter txg happen another 4k record size block will be written, and
then if I write 130k a 128k and 2k record block will be writt
Create a new empty pool on the solaris system, let it format the disks etc
ie used the disk names cXtXd0 This should put the EFI label on the disks and
then setup the partitions for you. Just encase here is an example.
Go back to the Linux box, and see if you can use tools to see the same
partit
Do we still need the zpool.cache still. I believe early versions of
zpool used the cache to remember what zpools to import at boot.
I understand newer versions of zfs still use the cache but also check to
see if the pool contains the correct host name of the server, and will
only import if the
The zpool.cache file makes clustering complex. {Assume the man page is
still correct}
From the zpool man page:
cachefile=path | "none"
Controls the location of where the pool configuration is cached.
Discovering all pools on system startup requires a cached copy of the
configuration data that
The zpool.cache file makes clustering complex. {Assume the man page is
still correct}
From the zpool man page:
cachefile=path | "none"
Controls the location of where the pool configuration is cached.
Discovering all pools on system startup requires a cached copy of the
configuration data tha
With regards too http://blogs.sun.com/roch/entry/the_new_zfs_write_throttle
I would have thought that if you had enough data to be written, it is
worth just writing it, and not waiting X seconds or trying to adjust
things so it only takes 5 seconds
For example different Disk Bus have differe
There are some Disk Arrays that only support Mirroring or Raid5.
Raid5 stores the parity on a different disk, where StripeZ would store parity
in the standard record size block. ie as long as you can read the block, it has
a chance of recovering it. The Array H/W is responsible for ensure the
Guys,
What is the best way to ask for a feature enhancement to ZFS.
To allow ZFS to be usefull for DR disk replication, we need to be able
set an option against the pool or file system or both, called close
sync. ie When a programme closes a file any outstanding writes are flush
to disk, bef
>Date: Thu, 26 Jul 2007 20:39:09 PDT
>From: "Anton B. Rang" <[EMAIL PROTECTED]>
>That said, I?m not sure exactly what this buys you for disk replication.
>What?s special about files >which have been closed? Is the point that
>applications might close a file and then notify some other >process of
>Date: Thu, 26 Jul 2007 20:39:09 PDT
>From: "Anton B. Rang"
>That said, I?m not sure exactly what this buys you for disk replication.
>What?s special about files >which have been closed? Is the point that
>applications might close a file and then notify some other >process of the
>file?s availa
Close Sync on file systems option (ie when the app calls close the file
is flushed, including mmap, no data loss of closed files on system
crash) Atomic/Locked operations across all pools e.g. snapshot all or
selected pools at the same time.
Allowance for offline files, eg. first part of a file
ZFS should allow 31+NULL chars for a comment against each disk.
This would work well with the host name string (I assume is max_hostname
255+NULL)
If a disk fails it should report c6t4908029d0 failed "comment from
disk", it should also remember the comment until reboot
This would be useful for
38 matches
Mail list logo