> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Edward Ned Harvey
> 
> --- Performance loss:

I ran one more test, that is rather enlightening.  I repeated test #2 (tweak
arc_meta_limit, use the default primarycache=all) but this time I wrote 100%
duplicate data instead of unique.  Dedup=sha256 (no verify).  Ideally, you
would expect this to write very very fast... Because it's all duplicate
data, and it's all async, the system should just buffer a bunch of tiny
metadata changes, aggregate them, and occasionally write a single serial
block when it flushes the TXG.  It should be much faster to write dedup.

The results are:  With dedup, it writes several times slower.  Just the same
as test #2, minus the amount of time it takes to write the actual data.  For
example, here's one datapoint, which is representative of the whole test:
    time to write unique data without dedup:  7.090 sec
    time to write unique data with dedup:     47.379 sec

    time to write duplic data without dedup:  7.016 sec
    time to write duplic data with dedup:     39.852 sec

This clearly breaks it down:
    7 sec to write the actual data
    40 sec overhead caused by dedup
    <1 sec is about how fast it should have been writing duplicated data

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to