Hi Richard, hi Daniel, hi Roy thanks for Your response..
Sorry for my delay, I spent a longer time ill in bed and didn't had the time to go on in testing the system: Am 19.06.2011 um 01:39 schrieb Richard Elling: > You're better off disabling dedup for this workload. If the dedup ratio was > more like > 10, 342, or some number > 2 dedup can be worthwhile. IMHO, dedup ratios < 2 > are > not good candidates for dedup. > > You can look at existing, non-deduped data to get an estimate of the > potential dedup > savings using: > zdb -S poolname > because the source isn't on a ZFS volume I can't check the possible dedup saving on that before copying it to the ZFS volume... Do I interpret the numbers correct? Dedup ratio of 1.18x = 18% saving in space? combined with the compress ratio of 1.48x I get a combined ratio of 1.67x ≈ 67% saving in space? I think that a good value. The data consist of a lot of pictures (jpeg, tiff, raw), Photoshop, InDesign files and about 1 TB of eMails. zdb -DD shows on the already dedupped and comrpessed data the following result on System 2: bucket allocated referenced ______ ______________________________ ______________________________ refcnt blocks LSIZE PSIZE DSIZE blocks LSIZE PSIZE DSIZE ------ ------ ----- ----- ----- ------ ----- ----- ----- 1 174M 21.8T 15.4T 15.4T 174M 21.8T 15.4T 15.4T 2 22.7M 2.84T 1.97T 1.98T 48.6M 6.08T 4.22T 4.23T 4 1.59M 204G 132G 132G 7.40M 947G 612G 614G 8 228K 28.5G 18.5G 18.6G 2.29M 293G 190G 190G 16 45.0K 5.63G 3.67G 3.68G 927K 116G 75.6G 75.9G 32 5.58K 714M 359M 362M 233K 29.1G 14.4G 14.5G 64 2.61K 334M 152M 153M 220K 27.4G 11.9G 12.0G 128 924 116M 30.0M 30.5M 154K 19.3G 5.19G 5.28G 256 496 62M 28.4M 28.6M 174K 21.7G 9.90G 9.98G 512 268 33.5M 14.3M 14.4M 168K 21.0G 9.22G 9.29G 1K 44 5.50M 94K 124K 52.2K 6.52G 116M 152M 2K 3 384K 1.50K 3.59K 7.19K 920M 3.59M 8.59M 4K 2 256K 1K 2.39K 10.9K 1.36G 5.43M 13.0M 16K 2 256K 1K 2.39K 35.2K 4.40G 17.6M 42.0M 256K 1 128K 512 1.20K 321K 40.2G 161M 384M Total 199M 24.9T 17.5T 17.6T 235M 29.4T 20.5T 20.6T dedup = 1.17, compress = 1.43, copies = 1.00, dedup * compress / copies = 1.67 zdb -D shows on System 2: DDT-sha256-zap-duplicate: 25742321 entries, size 445 on disk, 174 in core DDT-sha256-zap-unique: 182847350 entries, size 449 on disk, 186 in core dedup = 1.17, compress = 1.43, copies = 1.00, dedup * compress / copies = 1.67 I can't find a real explanation of this results in this table in any manual. Do I interpret the table correct? 199 Million Blocks used, using 24.9TB of space on the volume, referencing 29.4TB of "real data"? But for further testing I disabled the dedup, just leaving the gzip-6 compression. The values on System 1 will be the same, because it is the same data, only the compression rate is only 10%. >> Dedup is on "always" (not only at the start), also compression is activated: >> System 1 = compression on (lzib?) >> System 2 = compression on (gzip-6) >> compression rates: >> System 1 = 1.10x >> System 2 = 1.48x >> >> compression and dedup were some of the primary reasons to choose zfs in this >> situation. > > Compression is a win for a large number of cases. Dedup, not so much. > I was quite impressed about the compressratio with gzip-6 on System 2. Because it is the long therm backup-system in a data center (planned) the space on this system should be used at the best possible quotes. We don't want to be forced to add drives to the system every few weeks. The system was planned to hold the actual data and the new data for the next 1-2 years without the need to add more drives to it. But dedup must be a factor, because with dedup=on the same system I only get about 10-15Mb/s on writing data to the iSCSI volume, with dedup=off I get about 45-50MB/s. Because of the gzip-compression this looks like a good value. >> ok, I understand, more RAM will be no fault.. ;) >> >> On System 1 there is no such massive change in RAM usage while copying files >> to and from the volume. >> But the Performance is only about 20MB/s via GigaBit (iSCSI). > > This is likely not a dedup issue. More likely, Nagle is biting you or there > is a > serialization that is not immediately obvious. I have also seen questionable > network configurations cause strange slowdowns for iSCSI. > -- richard ok, but why didn't that occur at the beginning? We made a lot of tests, with some TB of data, performance was satisfying... This "nagle" problem shouldn't depend on the amount of storage used on the system.. As I understand the nagle problem, it could be a problem on the Client as on the Server side of the iSCSI-chain. So I need to have a look on the OS X-settings and on the OpenSolaris settings, to eliminate the nagle-problem. On the OS X-Server-System the value to disable the "nagle problem" is already set (long time before using iSCSI) /etc/sysctl.conf -> net.inet.tcp.delayed_ack=0 What do You mean with "serialization"? For the System 1 in our network we already planned to put it on a own network. So use one EtherNet interface on the OS X server only for iSCSI data and put it in a physically own network (with an own switch etc..). The slowdowns appear on different servers, I tried a OS X 10.5.8 (Server) Xserve Intel Dual Xeon 2Ghz OS X 10.6.7 (Server) Mac Pro Server Single Quad-Core Xeon 2.8Ghz OS X 10.6.7 (Server) Mac Mini Core2Duo 2.66Ghz The Xserve and Mac Pro use a LACP connection (Link aggregation) with 2 GigaBit connections, the Mac Mini a simple GigaBit connection. The transferrates in reading are comparable about about 70-80Mb/s. But on all systems the uploadrate is really low.. >> On Wed, Jun 15, 2011 at 07:19:05PM +0200, Roy Sigurd Karlsbakk wrote: >>> >>> Dedup is known to require a LOT of memory and/or L2ARC, and 24GB isn't >>> really much with 34TBs of data. >> >> The fact that your second system lacks the l2arc cache device is absolutely >> your prime suspect. I added a SSD for L2Arc and didn't get any change in speed.... But the problem must be somewhere else. Today I made a scrub on System 1: root@ACStorage:~# zpool status -v pool: ACBackPool_1 state: ONLINE scan: scrub in progress since Sat Jul 23 15:48:32 2011 35,9G scanned out of 45,4T at 5,36M/s, (scan is slow, no estimated time) 0 repaired, 0,08% done config: NAME STATE READ WRITE CKSUM ACBackPool_1 ONLINE 0 0 0 raidz2-0 ONLINE 0 0 0 c1t5000CCA369C76401d0 ONLINE 0 0 0 c1t5000CCA369C9E6A4d0 ONLINE 0 0 0 c1t5000CCA369CAD435d0 ONLINE 0 0 0 c1t5000CCA369CAF8CCd0 ONLINE 0 0 0 c1t5000CCA369CBA08Dd0 ONLINE 0 0 0 c1t5000CCA369CBA666d0 ONLINE 0 0 0 c1t5000CCA369CBAC4Ed0 ONLINE 0 0 0 c1t5000CCA369CBB08Ad0 ONLINE 0 0 0 c1t5000CCA369CBB102d0 ONLINE 0 0 0 c1t5000CCA369CBB30Fd0 ONLINE 0 0 0 raidz2-1 ONLINE 0 0 0 c1t5000CCA369C9EC38d0 ONLINE 0 0 0 c1t5000CCA369C9FA35d0 ONLINE 0 0 0 c1t5000CCA369C9FA83d0 ONLINE 0 0 0 c1t5000CCA369CA0188d0 ONLINE 0 0 0 c1t5000CCA369CA12FBd0 ONLINE 0 0 0 c1t5000CCA369CA3071d0 ONLINE 0 0 0 c1t5000CCA369CAB044d0 ONLINE 0 0 0 c1t5000CCA369CB928Ad0 ONLINE 0 0 0 c1t5000CCA369CBA1D5d0 ONLINE 0 0 0 c1t5000CCA369CBAAC1d0 ONLINE 0 0 0 raidz2-2 ONLINE 0 0 0 c1t5000CCA369C9F9B3d0 ONLINE 0 0 0 c1t5000CCA369CA09A6d0 ONLINE 0 0 0 c1t5000CCA369CA12AFd0 ONLINE 0 0 0 c1t5000CCA369CA1384d0 ONLINE 0 0 0 c1t5000CCA369CAAEC0d0 ONLINE 0 0 0 c1t5000CCA369CAD93Ad0 ONLINE 0 0 0 c1t5000CCA369CAD950d0 ONLINE 0 0 0 c1t5000CCA369CADA7Dd0 ONLINE 0 0 0 c1t5000CCA369CADA89d0 ONLINE 0 0 0 c1t5000CCA369CADA93d0 ONLINE 0 0 0 cache c1t5E83A97FEFD3F27Bd0 ONLINE 0 0 0 spares c1t5000CCA369C9ED1Bd0 AVAIL c1t5000CCA369CA09B3d0 AVAIL c1t5000CCA369CADA1Fd0 AVAIL c1t5000CCA369CADA88d0 AVAIL c1t5000CCA369CBA0E0d0 AVAIL c1t5000CCA369CBB15Dd0 AVAIL errors: No known data errors awful slow.... I tried some hints I found here (http://www.mail-archive.com/zfs-discuss@opensolaris.org/msg45685.html) echo "metaslab_min_alloc_size/Z 1000" | mdb -kw echo "zfs_scrub_delay/W0" | mdb -kw but that didn't change anything... a dd-test shows: root@ACStorage:~# dd if=/dev/zero of=/ACBackPool_1/ds.test bs=1024k count=5000 5000+0 records in 5000+0 records out 5242880000 bytes (5,2 GB) copied, 3,20443 s, 1,6 GB/s 1.6GB/s seems to be a real good value.. Or does the SSD manipulate this result? Why is a dd as fast and the scrub as slow? Any hints are appreciated. Many thank in advance Best Regards Sven C. Merckens -- Sven C. Merckens Michael-Ende-Straße 16 52499 Baesweiler Tel.: +49 2401 896074 Fax: +49 2401 801115 mercken...@mac.com _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss