Re: [zfs-discuss] send/recv over ssh

2010-05-20 Thread Gaëtan Lehmann


Le 21 mai 10 à 02:44, Freddie Cash a écrit :

On Thu, May 20, 2010 at 4:40 PM, Brent Jones   
wrote:
On Thu, May 20, 2010 at 3:42 PM, Brandon High   
wrote:
> On Thu, May 20, 2010 at 1:23 PM, Thomas Burgess  
 wrote:
>> I know i'm probably doing something REALLY stupid.but for  
some reason i
>> can't get send/recv to work over ssh.  I just built a new media  
server and

>
> Unless you need to have the send to be encrypted, ssh is going to  
slow

> you down a lot.
>
> I've used mbuffer when doing sends on the same network, it's  
worked well.

>

pre-start an Mbuffer session on the receiving end somehow.
SSH is always running on the receiving end, so no issues there.

And you can always patch OpenSSH with HPN, thus enabling the NONE  
cipher, which disable encryption for the data transfer  
(authentication is always encrypted).  And twiddle the internal  
buffers that OpenSSH uses to improve transfer rates, especially on  
100 Mbps or faster links.



Without patching anything, using arcfour instead of the default cipher  
increase the transfer rate dramatically.

It is bounded by the Gb/s network here, not by ssh.

Gaëtan


--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] compressed fs taking up more space than uncompressed equivalent

2009-10-23 Thread Gaëtan Lehmann


Le 23 oct. 09 à 08:46, Stathis Kamperis a écrit :


2009/10/23 michael schuster :

Stathis Kamperis wrote:


Salute.

I have a filesystem where I store various source repositories (cvs +
git). I have compression enabled on and zfs get compressratio  
reports

1.46x. When I copy all the stuff to another filesystem without
compression, the data take up _less_ space (3.5GB vs 2.5GB). How's
that possible ?


just a few thoughts:
- how do you measure how much space your data consumes?

With zfs list, under the 'USED' column. du(1) gives the same results
as well (the individual fs sizes aren't enterily identical with those
that zfs list reports , but the difference still exists).

tank/sources   3.73G   620G  3.73G  /export/sources
 <--- compressed
tank/test  2.32G   620G  2.32G  /tank/test
 <--- uncompressed



USED includes the size of the children and the size of the snapshot. I  
see below that you don't have snapshots on that pull, but in general,  
I found more useful to use


  zfs list -o space,compress,ratio

to look at how the space is used.


- how do you copy?

With cp(1). Should I be using zfs send | zfs receive ?


zfs send/receive or rsync -aH may do a better job by preserving hard  
links.


Regards,

Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Zfs improvements to compression in Solaris 10?

2009-10-30 Thread Gaëtan Lehmann


Le 4 août 09 à 20:25, Prabahar Jeyaram a écrit :


On Tue, Aug 04, 2009 at 01:01:40PM -0500, Bob Friesenhahn wrote:

On Tue, 4 Aug 2009, Prabahar Jeyaram wrote:


You seem to be hitting :

http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6586537

The fix is available in OpenSolaris build 115 and later not for  
Solaris 10 yet.


It is interesting that this is a simple thread priority issue.  The
system has a ton of available CPU but the higher priority compression
thread seems to cause scheduling lockout.  The Perfmeter tool shows  
that
compression is a very short-term spike in CPU. Of course since  
Perfmeter

and other apps stop running, it might be missing some sample data.

I could put the X11 server into the real-time scheduling class but  
hate to
think about what would happen as soon as Firefox visits a web  
site. :-)


Compression is only used for the intermittently-used backup pool so  
it
would be a shame to reduce overall system performance for the rest  
of the

time.

Do you know if this fix is planned to be integrated into a future  
Solaris

10 update?



Yes. It is planned for S10U9.



In the mean time, is there a patch available for Solaris 10?
I can't find it on sunsolve.

Thanks,

Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs zend is very slow

2009-12-16 Thread Gaëtan Lehmann


Le 17 déc. 09 à 03:19, Brent Jones a écrit :


Something must've changed in either SSH, or the ZFS receive bits to
cause this, but sadly since I upgrade my pool, I cannot roll back
these hosts  :(



I'm not sure that's the best way, but to look at how ssh is slowing  
down the transfer, I'm usually doing something like that:


  zfs send -R data/vo...@2009-12-14_01h44m47 | pv | ssh hex "cat > / 
dev/null"


I usually get approximately 60 MB/s.

Changing the cipher can enhance the transfer rate noticeably: "-c  
arcfour" gives an extra 10 MB/s here.
If you found that ssh is what is slowing down the transfer, maybe  
another cipher than the default one can help.


Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] directory size on compressed file system on Solaris 10

2009-12-21 Thread Gaëtan Lehmann


Hi,

On opensolaris, I use du with the -b option to get the uncompressed  
size of a directory):


  r...@opensolaris:~# du -sh /usr/local/
  399M/usr/local/
  r...@opensolaris:~# du -sbh /usr/local/
  915M/usr/local/
  r...@opensolaris:~# zfs list -o space,refer,ratio,compress data/local
  NAMEAVAIL   USED  USEDSNAP  USEDDS  USEDREFRESERV   
USEDCHILD  REFER  RATIO  COMPRESS
  data/local   228G   643M  249M394M  0   
0   394M  2.51xon

  r...@opensolaris:~# which du
  /usr/gnu/bin/du

but on Solaris 10, there is no such option.

So what is the best way to get the uncompressed size of a directory on  
Solaris 10?


Regards,

Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] directory size on compressed file system on Solaris 10

2009-12-21 Thread Gaëtan Lehmann


Le 21 déc. 09 à 19:28, Matthew Ahrens a écrit :


Gaëtan Lehmann wrote:

Hi,
On opensolaris, I use du with the -b option to get the uncompressed  
size of a directory):

 r...@opensolaris:~# du -sh /usr/local/
 399M/usr/local/
 r...@opensolaris:~# du -sbh /usr/local/
 915M/usr/local/
 r...@opensolaris:~# zfs list -o space,refer,ratio,compress data/ 
local
 NAMEAVAIL   USED  USEDSNAP  USEDDS  USEDREFRESERV   
USEDCHILD  REFER  RATIO  COMPRESS
 data/local   228G   643M  249M394M  0   
0   394M  2.51xon

 r...@opensolaris:~# which du
 /usr/gnu/bin/du
but on Solaris 10, there is no such option.
So what is the best way to get the uncompressed size of a directory  
on Solaris 10?


Install GNU du on solaris 10?


That's an option of course, but I'd prefer something that I can use  
installing any extra program.


 Although the answer will be just as (in)accurate as GNU du on  
solaris 10.  Note that it reports the compression ratio as 915/399 =  
2.29x, actual is 2.51x.  This could be due to sparse files, or  
metadata like directories, whose "apparent size" (st_size) is not  
what GNU du expects.


At least it gives a not so bad estimation :-)

And the compression ratio includes the data in the snapshots, so it  
may be inaccurate in that case also.

The actual compression ratio, given on the last snapshot, is 2.41x.

Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] directory size on compressed file system on Solaris 10

2009-12-21 Thread Gaëtan Lehmann


Le 21 déc. 09 à 20:23, Joerg Schilling a écrit :


Matthew Ahrens  wrote:


Gaëtan Lehmann wrote:


Hi,

On opensolaris, I use du with the -b option to get the  
uncompressed size

of a directory):

 r...@opensolaris:~# du -sh /usr/local/
 399M/usr/local/
 r...@opensolaris:~# du -sbh /usr/local/
 915M/usr/local/
 r...@opensolaris:~# zfs list -o space,refer,ratio,compress data/ 
local
 NAMEAVAIL   USED  USEDSNAP  USEDDS  USEDREFRESERV   
USEDCHILD

REFER  RATIO  COMPRESS
 data/local   228G   643M  249M394M   
0  0

394M  2.51xon
 r...@opensolaris:~# which du
 /usr/gnu/bin/du

but on Solaris 10, there is no such option.

So what is the best way to get the uncompressed size of a  
directory on

Solaris 10?


Install GNU du on solaris 10?  Although the answer will be just as
(in)accurate as GNU du on solaris 10.  Note that it reports the  
compression


How about:

find . -type f -ls | awk '{ sum += $7} END {print sum}'



sounds good

  r...@opensolaris:~# find /usr/local/ -ls | awk '{sum += $7} END  
{print sum/1024**2}'

  914.039

but maybe a little longer to write

  r...@opensolaris:~# echo "du -sbh ." | wc -c
  10
  r...@opensolaris:~# echo "find . -ls | awk '{sum += $7} END {print  
sum/1024**2}'" | wc -c

  53

Thanks!

Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Backing up a ZFS pool

2010-01-17 Thread Gaëtan Lehmann


Le 17 janv. 10 à 11:38, Edward Ned Harvey a écrit :


Personally, I use "zfs send | zfs receive" to an external disk.

Initially a

full image, and later incrementals.


Do these incrementals go into the same filesystem that received the
original zfs stream?


Yes.  In fact, I think that's the only way possible.  The end result  
is ... On my external disk, I have a ZFS filesystem, with  
snapshots.  Each snapshot corresponds to each incremental send| 
receive.


Personally, I like to start with a fresh "full" image once a month,  
and then do daily incrementals for the rest of the month.


There is one drawback:  If I have >500G filesystem to backup, and I  
have 1Tb target media ...  Once per month, I have to "zpool destroy"  
the target media before I can write a new full backup onto it.  This  
leaves a gap where the backup has been destroyed and the new image  
has yet to be written.


To solve this problem, I have more than one external disk, and  
occasionally rotate them.  So there's still another offline backup  
available, if something were to happen to my system during the  
moment when the backup was being destroyed once per month.



ZFS can check the pool and make sure that there is no error.
Running 'zpool scrub' on the two pools from time to time - let's say  
every month - should give you a similar level of protection without  
the need for a full backup.


Even when backing up with rsync+zfs snapshot, a full incremental every  
month may not be required. A rsync run with the --checksum option  
every month may be good enough. It forces the read of the full data on  
both sides, but at least it avoids the network transfer if the pools  
are on different hosts, and it avoids increasing the space used by the  
snapshots.


Gaëtan


--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Backing up a ZFS pool

2010-01-18 Thread Gaëtan Lehmann


Le 18 janv. 10 à 09:24, Edward Ned Harvey a écrit :


Personally, I like to start with a fresh "full" image once a month,

and then do daily incrementals for the rest of the month.

This doesn't buy you anything. ZFS isn't like traditional backups.


If you never send another full, then eventually the delta from the  
original
to the present will become large.  Not a problem, you're correct, as  
long as

your destination media is sufficiently large.

Unless I am mistaken, I believe, the following is not possible:

On the source, create snapshot "1"
Send snapshot "1" to destination
On the source, create snapshot "2"
Send incremental, from "1" to "2" to the destination.
On the source, destroy snapshot "1"
On the destination, destroy snapshot "1"

I think, since snapshot "2" was derived from "1" you can't destroy "1"
unless you've already destroyed "2"



This is definitely possible with zfs. Just try!

Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zones and other filesystems

2010-01-21 Thread Gaëtan Lehmann


Le 21 janv. 10 à 12:33, Thomas Burgess a écrit :


I'm pretty new to opensolaris.  I come from FreeBSD.

Naturally, after using FreeBSD forr awhile i've been big on the use  
of FreeBSD jails so i just had to try zones.  I've figured out how  
to get zones running but now i'm stuck and need help.  Is there  
anything like "nullfs" in opensolaris...


or maybe there is a more "solaris" way of doing what i need to do.

Basically, what i'd like to do is give a specific zone access to 2  
zfs filesystems which are available to the global zone.

my new zones are in:

/export/home/zone1
/export/home/zone2




the path of the root of your zone is not important for that feature.


What i'd like to do is give them access to:

/tank/nas/Video
/tank/nas/JeffB




with zonecfg, you can add a configuration like this one to your zone:

add fs
set dir=/some/path/Video
set special=/tank/nas/Video
set type=lofs
end
add fs
set dir=/some/path/JeffB
set special=/tank/nas/JeffB
set type=lofs
end

Your filesystems will appear in /some/path/Video and /some/path/JeffB  
in your zone, and still be accessible in the global zone.


http://docs.sun.com/app/docs/doc/817-1592/z.conf.start-29?a=view

This option don't let you manage the filesystems from the zone though.
You must use "add dataset" in that case.

http://docs.sun.com/app/docs/doc/819-5461/gbbst?a=view

Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zones and other filesystems

2010-01-21 Thread Gaëtan Lehmann


Le 21 janv. 10 à 14:14, Thomas Burgess a écrit :


now i'm stuck again.sorry to clog the tubes with my nubishness.

i can't seem to create users inside the zonei'm sure it's due to  
zfs privelages somewhere but i'm not exactly sure how to fix iti  
dont' mind if i need to manage the zfs filesystem outside of the  
zone, i'm just not sure WHERE i'm supposed to do it



when i try to create a home dir i get this:

mkdir: Failed to make directory "wonslung"; Operation not applicable


when i try to do it via adduser i get this:

UX: useradd: ERROR: Unable to create the home directory: Operation  
not applicable.



and when i try to enter the zone home dir from the global zone i get  
this, even as root:


bash: cd: home: Not owner


have i seriously screwed up or did i again miss something vital.



Maybe it's because of the automounter.
If you don't need that feature, try to disable it in your zone with

  svcadm disable autofs

Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Wrice cache for compressed file system

2009-07-06 Thread Gaëtan Lehmann


Hi,

I'd like to compress quite well compressable (~4x) data on a file  
server using ZFS compression, and still get good transfer speed. The  
users are transferring several GB of data (typically, 8-10 GB). The  
host is a X4150 with 16 GB of RAM.
Looking at ZFS layer described at http://www.opensolaris.org/os/community/zfs/source/ 
, it seems that the ZIL can play this role, as it is able to store the  
data uncompressed on a disk before passing the data to ZIO for the  
compression and the transfer to the final drives.


Fortunately, I got two idle drives in the server that I can use as log  
devices for my data pool. Unfortunately, it has no effect on write  
performances, so I guess I was wrong in thinking that the ZIL can  
cache the data before compression.


So is it possible to use the ZIL cache the data before compression?  
Did I miss an option somewhere?


Thanks!

Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Wrice cache for compressed file system

2009-07-07 Thread Gaëtan Lehmann


Hi Darren,

Le 7 juil. 09 à 13:41, Darren J Moffat a écrit :


Gaëtan Lehmann wrote:
I'd like to compress quite well compressable (~4x) data on a file  
server using ZFS compression, and still get good transfer speed.  
The users are transferring several GB of data (typically, 8-10 GB).  
The host is a X4150 with 16 GB of RAM.


What protocol is being used for file transfer ?

FTP, sftp, NFSv3, NFSv4, CIFS/SMB, iSCSI, FCoE, other ?


There will be two kinds of transfer protocol, once in production -  
CIFS and one specific to the application.

But for a quick test, the test was made with scp.



It is really important to know that.

Also really important is what the data access pattern is ?  Is it  
read mostly or write mostly ?  Are the files that are written  
accessed for read soon ?  Is it sequential or random reads ?


It is read mostly. Files are usually rarely accessed, but are likely  
to be accessed several times when they begin to be accessed. This is  
sequential read.
The data set is transfered on the server, and the files are not  
modified after that (or very rarely). The whole data set is transfered  
at once.




What is "good" by your definition ?


The data set is transfered in a little less than 3 minutes without  
compression - that's good!
With compression, the data set is transfered in 15 minutes. I was  
hopping that a disk write cache can keep the transfer speed close to  
the 3 minutes obtained without compression.




What kind of networking do you have ? 1G or 10G ? what about the  
clients ?


1G on both sides.



What OS are the clients running ?


windows, mac and linux.
The test was made with a linux client.



Looking at ZFS layer described at http://www.opensolaris.org/os/community/zfs/source/ 
, it seems that the ZIL can play this role, as it is able to store  
the data uncompressed on a disk before passing the data to ZIO for  
the compression and the transfer to the final drives.


The ZIL is not a cache it is used to provide the synchronous  
requirements for calls such as fsync() and the requirements of NFS.


Fortunately, I got two idle drives in the server that I can use as  
log devices for my data pool.


To be useful as a log device it must be faster than the pool  
devices. You may see some benefit using a 15k RPM drive if your main  
pool devices are 7200rpm.


> Unfortunately, it has no effect on write
performances, so I guess I was wrong in thinking that the ZIL can  
cache the data before compression.


However it very much depends on your workload (which you didn't  
describe).  You may need to use a fast (and expensive) SSD or you  
may not have a workload that is even helped by the ZIL.


So is it possible to use the ZIL cache the data before compression?  
Did I miss an option somewhere?


The fact that the ZIL doesn't compress data is not a feature it is a  
very low level implementation detail.




My workload is very simple: the user copy approximately 10GB of data  
on the server, and then only read it from time to time.
During the copy on the server, the transfer is cpu bounded, and there  
is a lot of cpu time available when there is no copy to the server.  
Using a disk to store the uncompressed data, as I guess it is done by  
the memory cache, may have helped.


I thought the ZIL may have played this role.

I understand that this is very specific to a compressed filesystem,  
but this caching behavior would greatly enhance this kind of workload.  
For now, I guess I have to tune the compression rate and/or increase  
the amount of RAM so that ZFS can cache more data.


Regards,

Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Wrice cache for compressed file system

2009-07-07 Thread Gaëtan Lehmann


Le 7 juil. 09 à 15:21, Darren J Moffat a écrit :


Gaëtan Lehmann wrote:
There will be two kinds of transfer protocol, once in production -  
CIFS and one specific to the application.

But for a quick test, the test was made with scp.


CIFS and scp are very different protocols with very different  
performance characteristics.


Also really important is what the data access pattern is ?  Is it  
read mostly or write mostly ?  Are the files that are written  
accessed for read soon ?  Is it sequential or random reads ?
It is read mostly. Files are usually rarely accessed, but are  
likely to be accessed several times when they begin to be accessed.  
This is sequential read.
The data set is transfered on the server, and the files are not  
modified after that (or very rarely). The whole data set is  
transfered at once.


What is "good" by your definition ?
The data set is transfered in a little less than 3 minutes without  
compression - that's good!
With compression, the data set is transfered in 15 minutes. I was  
hopping that a disk write cache can keep the transfer speed close  
to the 3 minutes obtained without compression.


That is a big difference and without seeing data my initial thought  
is that the system appears to be CPU bound, particularly if the only  
difference really was a compressed ZFS dataset versus an  
uncompressed one.


Is "transfer" in this case a read from the server or a write to it ?


It is a write to the server. The server is cpu bound, because of the  
compression.




My workload is very simple: the user copy approximately 10GB of  
data on the server, and then only read it from time to time.
During the copy on the server, the transfer is cpu bounded, and  
there is a lot of cpu time available when there is no copy to the  
server.


Are you talking here about the server's CPU or the clients ?


I'm talking about the server's cpu. Client is not cpu bounded by scp.



>Using a
disk to store the uncompressed data, as I guess it is done by the  
memory cache, may have helped.

I thought the ZIL may have played this role.


I think you are maybe confusing the ZIL and the L2ARC.


I think L2ARC is for reading data - I'd like an L2ARC for writing data.



What compression algorithm are you using ?  The default "on" value  
of lzjb or are you doing something like gzip-9 ?



gzip-6. There is no speed problem with lzjb, but also not the same  
compression ratio :-)


Gaëtan


--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Wrice cache for compressed file system

2009-07-07 Thread Gaëtan Lehmann


Le 7 juil. 09 à 15:54, Darren J Moffat a écrit :

What compression algorithm are you using ?  The default "on" value  
of lzjb or are you doing something like gzip-9 ?
gzip-6. There is no speed problem with lzjb, but also not the same  
compression ratio :-)


What build of OpenSolaris are you running ?

The fix for "6812655 need larger kmem caches for newer workloads"  
might help but you need to be running build 114 or higher which  
means you need to be using the pkg.opensolaris.org/dev repository  
not the /release one.


It could also be "6586537 async zio taskqs can block out userland  
commands" which is fixed in 115 and was the "real" fix.


This is guess work though since I haven't seen perf data from your  
particular system.


I think it's 2008.11.
I won't be able to access that host until next week. I'll update the  
OS and post the result on the list.


Thanks,

Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Why is Solaris 10 ZFS performance so terrible?

2009-07-13 Thread Gaëtan Lehmann
r2m41.81s
sys 9m46.76s

Feel free to clean up with 'zfs destroy rpool/zfscachetest'.

I am interested to hear about systems which do not suffer from this  
bug.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Why is Solaris 10 ZFS performance so terrible?

2009-07-14 Thread Gaëtan Lehmann


Le 14 juil. 09 à 18:09, Bob Friesenhahn a écrit :


On Tue, 14 Jul 2009, Jorgen Lundman wrote:

I have no idea. I downloaded the script from Bob without  
modifications and ran it specifying only the name of our pool.  
Should I have changed something to run the test?


If your system has quite a lot of memory, the number of files should  
be increased to at least match the amount of memory.


We have two kinds of x4500/x4540, those with Sol 10 10/08, and 2  
running svn117 for ZFS quotas. Worth trying on both?


It is useful to test as much as possible in order to fully  
understand the situation.


Since results often get posted without system details, the script is  
updated to dump some system info and the pool configuration.   
Refresh from


http://www.simplesystems.org/users/bfriesen/zfs-discuss/zfs-cache-test.ksh




Here is the result on another host with faster drives (sas 1 rpm)  
and solaris 10u7.


System Configuration: Sun Microsystems SUN FIRE X4150
System architecture: i386
System release level: 5.10 Generic_139556-08
CPU ISA list: amd64 pentium_pro+mmx pentium_pro pentium+mmx pentium  
i486 i386 i86


Pool configuration:
  pool : rpool
 état : ONLINE
 purger : aucun requis
configuration :

NAME  STATE READ WRITE CKSUM
rpool ONLINE   0 0 0
  mirror  ONLINE   0 0 0
c1t0d0s0  ONLINE   0 0 0
c1t1d0s0  ONLINE   0 0 0

erreurs : aucune erreur de données connue

zfs unmount rpool/zfscachetest
zfs mount rpool/zfscachetest

Doing initial (unmount/mount) 'cpio -C 131072 -o > /dev/null'
48000256 blocs

real4m56.84s
user0m1.72s
sys 0m28.48s

Doing second 'cpio -C 131072 -o > /dev/null'
48000256 blocs

real13m48.19s
user0m2.07s
sys 0m44.45s

Feel free to clean up with 'zfs destroy rpool/zfscachetest'.


--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How to "mirror" an entire zfs pool to another pool

2009-07-28 Thread Gaëtan Lehmann


Le 28 juil. 09 à 15:54, Darren J Moffat a écrit :


> How do I monitor the progress of the transfer?  Once

Unfortunately there is no easy way to do that just now.  When the  
'zfs recv' finishes is it is done.


I've just found pv (pipe viewer) today (http://www.ivarch.com/programs/pv.shtml 
) which is packaged in /contrib (http://pkg.opensolaris.org/contrib/p5i/0/pv.p5i 
).


You can do

  zfs send -R oldp...@sendit | pv -s 3T | zfs recv -vFd newpool

and you'll see a message like that:

  8GO 0:00:05 [5,71GO/s] [=> ]   
7% ETA 0:00:58


A nice and simple way to get a progress report!

Gaëtan


--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] feature proposal

2009-07-31 Thread Gaëtan Lehmann


Le 31 juil. 09 à 10:24, dick hoogendijk a écrit :


On Wed, 29 Jul 2009 17:34:53 -0700
Roman V Shaposhnik  wrote:


On the read-write front: wouldn't it be cool to be able to snapshot
things by:
   $ mkdir .zfs/snapshot/


I've followed this thread but I fail to see the advantages of this. I
guess I miss something here. Can you explain to me why the above would
be better (nice to have) then "zfs create whate...@now?"



Because it can be done on any host mounting this file system through a  
network protocol like NFS or CIFS.

A nice feature for a NAS.

Gaëtan


--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs receive -o

2009-08-13 Thread Gaëtan Lehmann


Hi,

One thing I miss in zfs is the ability to override an attribute value  
in zfs receive - something like the -o option in zfs create. This  
option would be particularly useful with zfs send -R to make a backup  
and be sure that the destination won't be mounted


  zfs send -R f...@snap | ssh myhost zfs receive -d -o canmount=off  
backup/host


or to have compressed backups, because the compression attribute may  
be sent with the file system and thus is not necessarily inherited


  zfs send -R f...@snap | ssh myhost zfs receive -d -o canmount=off -o  
compression=gzip backup/host


It is possible to use -u for the mount problem, but it is not  
absolutely safe until the property is actually changed. An unexpected  
reboot can put the backup server in maintenance mode because the file  
system may have the same mountpoint as another one already there.


Do I miss something? Would it be possible to have such an option?

Regards,

Gaëtan


--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snv_110 -> snv_121 produces checksum errors on Raid-Z pool

2009-09-02 Thread Gaëtan Lehmann


Le 2 sept. 09 à 15:27, Frank Middleton a écrit :


On 09/02/09 05:40 AM, Henrik Johansson wrote:


For those of us which have already upgraded and written data to our
raidz pools, are there any risks of inconsistency, wrong checksums in
the pool? Is there a bug id?


This may not be a new problem insofar as it may also affect mirrors.
As part of the ancient "mirrored drives should not have checksum
errors thread", I used Richard Elling's amazing zcksummon script
http://www.richardelling.com/Home/scripts-and-programs-1/zcksummon
to help diagnose this (thanks, Richard, for all your help).

The bottom line is that hardware glitches (as found on cheap PCs
without ECC on buses and memory) can put ZFS into a mode where it
detects bogus checksum errors. If you set copies=2, it seems to
always be able to repair them, but they are never actually repaired.
Every time you scrub, it finds a checksum error on the affected  
file(s)

and it pretends to repair it (or may fail if you have copies=1 set).

Note: I have not tried this on raidz, only mirrors, where it is
highly reproducible. It would be really interesting to see if
raidz gets results similar to the mirror case when running zcksummon.
Note I have NEVER had this problem on SPARC, only on certain
bargain-basement PCs (used as X-Terminals) which as it turns out
have mobos notorious for not detecting bus parity errors.

If this is the same problem, you can certainly mitigate it by
setting copies=2 and actually copying the files (e.g., by
promoting a snapshot, which I believe will do this - can someone
confirm?). My guess is that snv121 has done something to make
the problem more likely to occur, but the problem itself is
quite old (predates snv100). Could you share with us some details
of your hardware, especially how much memory and if it has ECC
orbus parity?



I see the same problem on a workstation with ECC RAM and disks in  
mirror.

The host is a Dell T5500 with 2 cpus and 24 GB of RAM.

Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] nuke lots of snapshots

2009-09-03 Thread Gaëtan Lehmann


Le 3 sept. 09 à 19:57, Jacob Ritorto a écrit :

Sorry if this is a faq, but I just got a time sensitive dictim from  
the higherups to disable and remove all remnants of rolling  
snapshots on our DR filer.  Is there a way for me to nuke all  
snapshots with a single command, or to I have to manually destroy  
all 600+ snapshots with zfs destroy?



  zfs list -r -t snapshot -o name -H pool | xargs -tl zfs destroy

should destroy all the snapshots in a pool

Gaëtan


--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snv_110 -> snv_121 produces checksum errors on Raid-Z pool

2009-09-03 Thread Gaëtan Lehmann


Le 3 sept. 09 à 21:59, Simon Breden a écrit :

So what's the consensus on checksum errors appearing within mirror  
vdevs?
Is it caused the same bug announced by Adam, or is something else  
causing it?

If so, what's the bug id?



Sorry, I forgot to report the end of my experiments. They shown that  
the checksum error I've seen are unlikely related to snv 121.
The checksum errors I've seen is more likely caused by the iommu bug  
on intel platforms which has lead to several crash some time ago, and  
a lack of scrub in the last days. After the errors reported during the  
scrub on snv 121, I run a scrub on snv 118 and find the same amount of  
error, all on rpool/dump. I dropped that zvol, rerun the scrub again  
still on snv 118 without any error. After a reboot on snv 121 and a  
new scrub, no checksum error are reported.


Regards,

Gaëtan

--
Gaëtan Lehmann
Biologie du Développement et de la Reproduction
INRA de Jouy-en-Josas (France)
tel: +33 1 34 65 29 66fax: 01 34 65 29 09
http://voxel.jouy.inra.fr  http://www.itk.org
http://www.mandriva.org  http://www.bepo.fr



PGP.sig
Description: Ceci est une signature électronique PGP
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss