[zfs-discuss] one time passwords - apache infrastructure incident report 8/28/2009

2009-09-04 Thread russell aspinwall
Hi, Just be reading about apache.org incident report for 8/28/2009 ( https://blogs.apache.org/infra/entry/apache_org_downtime_report ) The use of Solaris and ZFS on the European server was interesting including the recovery. However, what I found more interesting was the use of one time passwo

Re: [zfs-discuss] Read about ZFS backup - Still confused

2009-09-04 Thread Thomas Burgess
Let me explain what i have and you decide if it's what you're looking for. I run a home NAS based on ZFS (due to hardware issues i am using FreeBSD 7.2 as my os but all the data is on ZFS) This system has multiple uses. I have about 10 users and 4 HTPC's connected via gigabit. I have ZFS filesyst

Re: [zfs-discuss] How to find poor performing disks

2009-09-04 Thread Roch
Scott Lawson writes: > Also you may wish to look at the output of 'iostat -xnce 1' as well. > > You can post those to the list if you have a specific problem. > > You want to be looking for error counts increasing and specifically 'asvc_t' > for the service times on the disks. I higher num

Re: [zfs-discuss] ARC limits not obeyed in OSol 2009.06

2009-09-04 Thread Roch
Do you have the zfs primarycache property on this release ? if so, you could set it to 'metadata' or none. primarycache=all | none | metadata Controls what is cached in the primary cache (ARC). If this property is set to "all", then both user data and metadat

Re: [zfs-discuss] Petabytes on a budget - blog

2009-09-04 Thread Marc Bevand
Bill Moore sun.com> writes: > > Moving on, modern high-capacity SATA drives are in the 100-120MB/s > range. Let's call it 125MB/s for easier math. A 5-port port multiplier > (PM) has 5 links to the drives, and 1 uplink. SATA-II speed is 3Gb/s, > which after all the framing overhead, can get yo

Re: [zfs-discuss] Petabytes on a budget - blog

2009-09-04 Thread Marc Bevand
Marc Bevand gmail.com> writes: > > So in conclusion, my SBNSWAG (scientific but not so wild-ass guess) > is that the max I/O throughput when reading from all the disks on > 1 of their storage pod is about 1000MB/s. Correction: the SiI3132 are on x1 (not x2) links, so my guess as to the aggregate

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Roch
"100% random writes produce around 200 IOPS with a 4-6 second pause around every 10 seconds. " This indicates that the bandwidth you're able to transfer through the protocol is about 50% greater than the bandwidth the pool can offer to ZFS. Since, this is is not sustainable, you s

Re: [zfs-discuss] Change the volblocksize of a ZFS volume

2009-09-04 Thread Roch
stuart anderson writes: > > > > Question : > > > > > > > > Is there a way to change the volume blocksize > > say > > > via 'zfs snapshot send/receive'? > > > > > > > > As I see things, this isn't possible as the > > target > > > volume (including property values) gets > > overwritten >

Re: [zfs-discuss] check a zfs rcvd file

2009-09-04 Thread dick hoogendijk
Lori Alt wrote: The -n option does some verification. It verifies that the record headers distributed throughout the stream are syntactically valid. Since each record header contains a length field which allows the next header to be found, one bad header will cause the processing of the stre

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Scott Meilicke
Roch Bourbonnais Wrote: ""100% random writes produce around 200 IOPS with a 4-6 second pause around every 10 seconds. " This indicates that the bandwidth you're able to transfer through the protocol is about 50% greater than the bandwidth the pool can offer to ZFS. Since, this is is not sustainabl

Re: [zfs-discuss] check a zfs rcvd file

2009-09-04 Thread Lori Alt
On 09/04/09 09:41, dick hoogendijk wrote: Lori Alt wrote: The -n option does some verification. It verifies that the record headers distributed throughout the stream are syntactically valid. Since each record header contains a length field which allows the next header to be found, one bad he

[zfs-discuss] question about my hardware choice

2009-09-04 Thread Eugen Leitl
Hi zfs congnoscenti, a few quick question about my hardware choice (a bit late, since the box is up already): A 3U supermicro chassis with 16x SATA/SAS hotplug Supermicro X8DDAi (2x Xeon QC 1.26 GHz S1366, 24 GByte RAM, IPMI) 2x LSI SAS3081E-R 16x WD2002FYPS Right now I'm running Solaris 10 5/9

Re: [zfs-discuss] Archiving and Restoring Snapshots

2009-09-04 Thread Richard Elling
On Sep 3, 2009, at 10:32 PM, Tim Cook wrote: On Fri, Sep 4, 2009 at 12:17 AM, Ross wrote: Hi Richard, Actually, reading your reply has made me realise I was overlooking something when I talked about tar, star, etc... How do you backup a ZFS volume? That's something traditional tools can'

Re: [zfs-discuss] check a zfs rcvd file

2009-09-04 Thread dick hoogendijk
Lori Alt wrote: The -u option to zfs recv (which was just added to support flash archive installs, but it's useful for other reasons too) suppresses all mounts of the received file systems. So you can mount them yourself afterward in whatever order is appropriate, or do a 'zfs mount -a'. You

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Neil Perrin
On 09/04/09 09:54, Scott Meilicke wrote: Roch Bourbonnais Wrote: ""100% random writes produce around 200 IOPS with a 4-6 second pause around every 10 seconds. " This indicates that the bandwidth you're able to transfer through the protocol is about 50% greater than the bandwidth the po

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Scott Meilicke
I am still not buying it :) I need to research this to satisfy myself. I can understand that the writes come from memory to disk during a txg write for async, and that is the behavior I see in testing. But for sync, data must be committed, and a SSD/ZIL makes that faster because you are writing

Re: [zfs-discuss] Understanding when (and how) ZFS will use spare disks

2009-09-04 Thread Scott Meilicke
This sounds like the same behavior as opensolaris 2009.06. I had several disks recently go UNAVAIL, and the spares did not take over. But as soon as I physically removed a disk, the spare started replacing the removed disk. It seems UNAVAIL is not the same as the disk not being there. I wish the

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Eric Sproul
Scott Meilicke wrote: > So what happens during the txg commit? > > For example, if the ZIL is a separate device, SSD for this example, does it > not work like: > > 1. A sync operation commits the data to the SSD > 2. A txg commit happens, and the data from the SSD are written to the > spinning

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Scott Meilicke
So what happens during the txg commit? For example, if the ZIL is a separate device, SSD for this example, does it not work like: 1. A sync operation commits the data to the SSD 2. A txg commit happens, and the data from the SSD are written to the spinning disk So this is two writes, correct?

[zfs-discuss] Understanding when (and how) ZFS will use spare disks

2009-09-04 Thread Chris Siebenmann
We have a number of shared spares configured in our ZFS pools, and we're seeing weird issues where spares don't get used under some circumstances. We're running Solaris 10 U6 using pools made up of mirrored vdevs, and what I've seen is: * if ZFS detects enough checksum errors on an active disk,

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Scott Meilicke
Doh! I knew that, but then forgot... So, for the case of no separate device for the ZIL, the ZIL lives on the disk pool. In which case, the data are written to the pool twice during a sync: 1. To the ZIL (on disk) 2. From RAM to disk during tgx If this is correct (and my history in this thread

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Bob Friesenhahn
On Fri, 4 Sep 2009, Scott Meilicke wrote: So what happens during the txg commit? For example, if the ZIL is a separate device, SSD for this example, does it not work like: 1. A sync operation commits the data to the SSD 2. A txg commit happens, and the data from the SSD are written to the spi

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Scott Meilicke
So, I just re-read the thread, and you can forget my last post. I had thought the argument was that the data were not being written to disk twice (assuming no separate device for the ZIL), but it was just explaining to me that the data are not read from the ZIL to disk, but rather from memory to

Re: [zfs-discuss] Petabytes on a budget - blog

2009-09-04 Thread Tim Cook
On Fri, Sep 4, 2009 at 5:36 AM, Marc Bevand wrote: > Marc Bevand gmail.com> writes: > > > > So in conclusion, my SBNSWAG (scientific but not so wild-ass guess) > > is that the max I/O throughput when reading from all the disks on > > 1 of their storage pod is about 1000MB/s. > > Correction: the

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Kyle McDonald
Scott Meilicke wrote: I am still not buying it :) I need to research this to satisfy myself. I can understand that the writes come from memory to disk during a txg write for async, and that is the behavior I see in testing. But for sync, data must be committed, and a SSD/ZIL makes that faster

[zfs-discuss] zfs compression algorithm : jpeg ??

2009-09-04 Thread Len Zaifman
We have groups generating terabytes a day of image data from lab instruments and saving them to an X4500. We have tried lzbj : compressratio = 1.13 in 11 hours , 1.3 TB -> 1.1 TB gzip -9 : compress ratio = 1.68 in > 37 hours, 1.3 TB -> .75 TB The filesystem performance was

[zfs-discuss] PMP support in Opensolaris

2009-09-04 Thread Brandon High
On Wed, Sep 2, 2009 at 4:56 PM, David Magda wrote: > Said support was committed only two to three weeks ago: > >> PSARC/2009/394 SATA Framework Port Multiplier Support >> 6422924 sata framework has to support port multipliers >> 6691950 ahci driver needs to support SIL3726/4726 SATA port multiplier

Re: [zfs-discuss] PMP support in Opensolaris

2009-09-04 Thread Nigel Smith
Hi Brandon To answer your question, all you need to do is look up those bug numbers: http://bugs.opensolaris.org/view_bug.do?bug_id=6422924 http://bugs.opensolaris.org/view_bug.do?bug_id=6691950 ..and you see the fix should be in release snv_122. Your in luck, as the OpenSolaris dev repository

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Ross Walker
On Sep 4, 2009, at 2:22 PM, Scott Meilicke > wrote: So, I just re-read the thread, and you can forget my last post. I had thought the argument was that the data were not being written to disk twice (assuming no separate device for the ZIL), but it was just explaining to me that the data are

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Scott Meilicke
Yes, I was getting confused. Thanks to you (and everyone else) for clarifying. Sync or async, I see the txg flushing to disk starve read IO. Scott -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http

Re: [zfs-discuss] zfs compression algorithm : jpeg ??

2009-09-04 Thread Richard Elling
On Sep 4, 2009, at 12:23 PM, Len Zaifman wrote: We have groups generating terabytes a day of image data from lab instruments and saving them to an X4500. Wouldn't it be easier to compress at the application, or between the application and the archiving file system? We have tried lzbj : com

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Ross Walker
On Sep 4, 2009, at 4:33 PM, Scott Meilicke > wrote: Yes, I was getting confused. Thanks to you (and everyone else) for clarifying. Sync or async, I see the txg flushing to disk starve read IO. Well try the kernel setting and see how it helps. Honestly though if you can say it's all sync wr

Re: [zfs-discuss] zfs compression algorithm : jpeg ??

2009-09-04 Thread Louis-Frédéric Feuillette
On Fri, 2009-09-04 at 13:41 -0700, Richard Elling wrote: > On Sep 4, 2009, at 12:23 PM, Len Zaifman wrote: > > > We have groups generating terabytes a day of image data from lab > > instruments and saving them to an X4500. > > Wouldn't it be easier to compress at the application, or between th

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Scott Meilicke
I only see the blocking while load testing, not during regular usage, so I am not so worried. I will try the kernel settings to see if that helps if/when I see the issue in production. For what it is worth, here is the pattern I see when load testing NFS (iometer, 60% random, 65% read, 8k chun

Re: [zfs-discuss] zfs compression algorithm : jpeg ??

2009-09-04 Thread Nicolas Williams
On Fri, Sep 04, 2009 at 01:41:15PM -0700, Richard Elling wrote: > On Sep 4, 2009, at 12:23 PM, Len Zaifman wrote: > >We have groups generating terabytes a day of image data from lab > >instruments and saving them to an X4500. > > Wouldn't it be easier to compress at the application, or between

Re: [zfs-discuss] check a zfs rcvd file

2009-09-04 Thread Lori Alt
On 09/04/09 10:17, dick hoogendijk wrote: Lori Alt wrote: The -u option to zfs recv (which was just added to support flash archive installs, but it's useful for other reasons too) suppresses all mounts of the received file systems. So you can mount them yourself afterward in whatever order is

Re: [zfs-discuss] zfs compression algorithm : jpeg ??

2009-09-04 Thread Bob Friesenhahn
On Fri, 4 Sep 2009, Louis-Frédéric Feuillette wrote: JPEG2000 uses arithmetic encoding to do the final compression step. Arithmetic encoding has a higher compression rate (in general) than gzip-9, lzbj or others. There is an opensource implementation of jpeg2000 called jasper[1]. Jasper is the

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Bob Friesenhahn
On Fri, 4 Sep 2009, Scott Meilicke wrote: I only see the blocking while load testing, not during regular usage, so I am not so worried. I will try the kernel settings to see if that helps if/when I see the issue in production. The flipside of the "pulsing" is that the deferred writes dimish

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Ross Walker
On Sep 4, 2009, at 6:33 PM, Bob Friesenhahn > wrote: On Fri, 4 Sep 2009, Scott Meilicke wrote: I only see the blocking while load testing, not during regular usage, so I am not so worried. I will try the kernel settings to see if that helps if/when I see the issue in production. The flips

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Ross Walker
On Sep 4, 2009, at 5:25 PM, Scott Meilicke > wrote: I only see the blocking while load testing, not during regular usage, so I am not so worried. I will try the kernel settings to see if that helps if/when I see the issue in production. For what it is worth, here is the pattern I see when l

Re: [zfs-discuss] PMP support in Opensolaris

2009-09-04 Thread Brandon High
On Fri, Sep 4, 2009 at 1:12 PM, Nigel Smith wrote: > Let us know if you can get the port multipliers working.. > > But remember, there is a problem with ZFS raidz in that release, so be > careful: I saw that, so I think I'll be waiting until snv_124 to update. The system that I'm thinking of usin

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Bob Friesenhahn
On Fri, 4 Sep 2009, Ross Walker wrote: I guess one can find a silver lining in any grey cloud, but for myself I'd just rather see a more linear approach to writes. Anyway I have never seen any reads happen during these write flushes. I have yet to see a read happen during the write flush eit

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Ross Walker
On Sep 4, 2009, at 8:59 PM, Bob Friesenhahn > wrote: On Fri, 4 Sep 2009, Ross Walker wrote: I guess one can find a silver lining in any grey cloud, but for myself I'd just rather see a more linear approach to writes. Anyway I have never seen any reads happen during these write flushes. I

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Bob Friesenhahn
On Fri, 4 Sep 2009, Ross Walker wrote: I have yet to see a read happen during the write flush either. That impacts my application since it needs to read in order to proceed, and it does a similar amount of writes as it does reads. The ARC makes it hard to tell if they are satisfied from cac

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread David Magda
On Sep 4, 2009, at 21:44, Ross Walker wrote: Though I have only heard good comments from my ESX admins since moving the VMs off iSCSI and on to ZFS over NFS, so it can't be that bad. What's your pool configuration? Striped mirrors? RAID-Z with SSDs? Other? __

Re: [zfs-discuss] Pulsing write performance

2009-09-04 Thread Ross Walker
On Sep 4, 2009, at 10:02 PM, David Magda wrote: On Sep 4, 2009, at 21:44, Ross Walker wrote: Though I have only heard good comments from my ESX admins since moving the VMs off iSCSI and on to ZFS over NFS, so it can't be that bad. What's your pool configuration? Striped mirrors? RAID-Z w

Re: [zfs-discuss] Motherboard for home zfs/solaris file server

2009-09-04 Thread Tim Cook
On Thu, Sep 3, 2009 at 4:57 AM, Karel Gardas wrote: > Hello, > your "(open)solaris for Ecc support (which seems to have been dropped from > 200906)" is misunderstanding. OS 2009.06 also supports ECC as 2005 did. Just > install it and use my updated ecccheck.pl script to get informed about > errors

[zfs-discuss] incremental send/recv larger than sum of snapshots?

2009-09-04 Thread Peter Baumgartner
I've been sending daily incrementals off-site for a while now, but recently they failed so I had to send an incremental covering a number of snapshots. I expected the incremental to be approximately the sum of the snapshots, but it seems to be considerably larger and still going. The source machine

Re: [zfs-discuss] Petabytes on a budget - blog

2009-09-04 Thread Marc Bevand
Tim Cook cook.ms> writes: > > Whats the point of arguing what the back-end can do anyways?  This is bulk data storage.  Their MAX input is ~100MB/sec.  The backend can more than satisfy that.  Who cares at that point whether it can push 500MB/s or 5000MB/s?  It's not a database processing tran

Re: [zfs-discuss] Petabytes on a budget - blog

2009-09-04 Thread Tim Cook
On Sat, Sep 5, 2009 at 12:30 AM, Marc Bevand wrote: > Tim Cook cook.ms> writes: > > > > Whats the point of arguing what the back-end can do anyways? This is > bulk > data storage. Their MAX input is ~100MB/sec. The backend can more than > satisfy that. Who cares at that point whether it can