Greetings Gentlemen,

I'm currently testing a new setup for a ZFS based storage system with 
dedup enabled. The system is setup on OI 148, which seems quite stable 
w/ dedup enabled (compared to the OpenSolaris snv_136 build I used 
before).

One issue I ran into, however, is quite baffling: 
With iozone set to 32 threads, ZFS's ARC seems to consume all available 
memory, making the system completely unresponsive (no SSH or console 
access).
 
This happens (repeatably) during one of the READ tests, usually in 
iozone's random read test. The WRITE tests work perfectly fine (and 
blazingly fast).

The system has 12G RAM. It does not matter whether a reasonably fast 
L2ARC is added to the pool.

The only thing that seems to help is setting 'primarycache=metadata' for
the ZFS in question, which does not seem like a desirable configuration.

I have also tried setting the zfs_arc_max property to 0x80000000 (2G) 
in /etc/system. This helps a little (it won't lock up quite as quickly),
but at some point, it will still eat up all memory it can get.

IOzone command line: 
/usr/benchmarks/iozone/iozone -R -e -s 1g -t 32 -M -r 16k -F [...]

Primarycache property:
NAME                        PROPERTY      VALUE         SOURCE
somepool                    primarycache  all           local
rpool/swap                  primarycache  metadata      local

./arcstat.pl (last few lines before the system locks up):
    Time  read  miss  miss% dmis  dm%  pmis  pm%  mmis  mm%  arcsz     c
15:56:34  121K   247      0  243    0     4    0   178    1   151M  402M
15:56:35     0     0      0    0    0     0    0     0    0   137M  402M
15:56:36   14K    25      0   25    0     0    0    24    0   265M  402M
15:56:37  381K   575      0  575    0     0    0   574    1     4G  402M
15:56:38  362K   706      0  706    0     0    0   706    1     9G  402M
15:56:39  153K   364      0  306    0    58   18   362    2    11G  402M

zpool list 3 (last output before the system locks up):
somepool  2.72T  9.53M  2.72T     0%  34466.62x  ONLINE  -
rpool      37G  8.37G  28.6G    22%  1.00x  ONLINE  -

Is this a known issue with the ARC? Is there anything else I can do 
about it?

Also, does anyone have an idea what's actually in the ARC when the 
system locks up? 
In 'zpool list' output, you can see that only 9.53M on the zpool is in 
use, with a dedup rate of 34466.62x. So what's there to cache?

I realize that this test does not reflect a situation that is likely to
occur in a production environment, but I'd still be happy about any 
helpful comments.

-- Clemens
-- 
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to