> 
> What does vmstat look like ?
> Also zpool iostat 1.
> 

               capacity     operations    bandwidth
pool         used  avail   read  write   read  write
----------  -----  -----  -----  -----  -----  -----
tank         291M  9.65G      0     11   110K   694K
tank         301M  9.64G      0     32      0  87.9K
tank         301M  9.64G      0      0      0      0
tank         301M  9.64G     31      0  3.96M      0
tank         301M  9.64G      0     88      0  4.91M
tank         311M  9.63G     16     77  2.05M  2.64M
tank         311M  9.63G     31      0  3.88M      0
tank         311M  9.63G      0      0      0      0
tank         311M  9.63G     31     62  3.96M  3.88M
tank         321M  9.62G     15    101  1.90M  3.08M
tank         321M  9.62G      0      0      0      0
tank         321M  9.62G     31      0  3.96M      0
tank         321M  9.62G      0     88      0  4.47M

kthr      memory            page            disk          faults      cpu
 r b w   swap  free  re  mf pi po fr de sr dd s1 -- --   in   sy   cs us sy id
 0 0 0 8395576 67320  0  69 224 0  0  0  0 104 0  0  0  578 3463 2210 16 17 67
 13 0 0 8395456 67192 1 109 16  0  0  0  0 70  0  0  0  466 1176 1055  7 73 20
 0 0 0 8395416 67112  0  21 16  0  0  0  0  2  0  0  0  327  809  452  2  2 96
 0 0 0 8395416 67112  0   3  0  0  0  0  0  0  0  0  0  370 1947  818  6  4 90
 0 0 0 8395416 67112  0   2  0  0  0  0  0  0  0  0  0  306 1358  672  8  3 89
 0 0 0 8395416 67112  0   4  0  0  0  0  0  0  0  0  0  338  822  409  1  1 98
 1 0 0 8395416 67112  0  10  0  0  0  0  0  0  0  0  0  320 3152 1415 20  8 72
 0 0 0 8396568 68200  0  16  0  0  0  0  0 12  0  0  0  381 1273  633  5  5 90
 0 0 0 8396568 68200  0   6  8  0  0  0  0  1  0  0  0  320 1613  620  4  3 93
 0 0 0 8396568 68192  0   0  0  0  0  0  0  0  0  0  0  352 1198  595  5  2 93
 0 0 0 8396568 68192  0   1  0  0  0  0  0  0  0  0  0  292  843  413  2  2 96
 0 0 0 8396568 68192  0   0  0  0  0  0  0  0  0  0  0  343  818  405  1  1 98
 0 0 0 8396568 68192  0   0  0  0  0  0  0  0  0  0  0  308  803  412  1  1 98
 0 0 0 8396568 68192  0   0  0  0  0  0  0  0  0  0  0  345 1236  471  2  3 95
 0 0 0 8396568 68192  0   0  0  0  0  0  0  0  0  0  0  296 1570  709  6  2 92
 0 0 0 8396568 68192 13 142  0  0  0  0  0  0  0  0  0  380 3134 1182 14  6 80
 0 0 0 8396568 68192  0   4  8  0  0  0  0  1  0  0  0  301 1034  536  5  4 91
 0 0 0 8396568 68184  0   0  0  0  0  0  0  0  0  0  0  343  811  417  1  2 97
 0 0 0 8396568 68184  0   0  0  0  0  0  0  0  0  0  0  310 1220  452  1  2 97
 kthr      memory            page            disk          faults      cpu
 r b w   swap  free  re  mf pi po fr de sr dd s1 -- --   in   sy   cs us sy id
 0 0 0 8396568 68176  0   0  0  0  0  0  0  1  0  0  0  373 1715  651  4  2 94
 0 0 0 8396568 68176  0   0  0  0  0  0  0  0  0  0  0  336 1739  647  3  2 95
 0 0 0 8396160 67272 51 334 565 0  0  0  0 60  0  0  0  558 4029 1651 10 14 76
 0 0 0 8396776 68184  3  99  0  0  0  0  0  0  0  0  0  357 1204  577  4  3 93
 0 0 0 8396776 68184  0   8  8  0  0  0  0  1  0  0  0  356 3497 1353 16  7 77
 0 0 0 8396776 68176  0   0  0  0  0  0  0  0  0  0  0  311 1128  477  2  1 97
 0 0 0 8396776 68176  0   6  0  0  0  0  0  0  0  0  0  357 1259  518  3  2 95
 0 0 0 8396776 68176  0   1  0  0  0  0  0  0  0  0  0  312 1166  495  2  1 97
 0 0 0 8396776 68176  0  50 71  0  0  0  0  9  0  0  0  366 1207  540 25  3 72

> Do you have any disk based swap ?
> 
Yes, there is an 8GB swap partition on the system and 2GB of RAM.

> One best practice we probably will be coming out with
> is to
> configure at least physmem of swap with ZFS (at least
> as of
> this release).
> 
> The partly hung system could be this :
> 
>       http://bugs.opensolaris.org/bugdatabase/view_bug.do?b
> ug_id=6429205
> 6429205 each zpool needs to monitor it's  throughput
> t and throttle heavy writers
> 
> The fix state is "in-progress".
> 
I will look at this.

> What throughput do you get for the full untar
> (untared size / elapse time) ?
# tar xf thunderbird-1.5.0.4-source.tar  2.77s user 35.36s system 33% cpu 
1:54.19 

260M/114 =~ 2.28 MB/s on this IDE disk
 
 
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to