Hi all,

Carsten Aulbert wrote:
> More later.

OK, I'm completely puzzled right now (and sorry for this lengthy email).
 My first (and currently only idea) was that the size of the files is
related to this effect, but that does not seem to be the case:

(1) A 185 GB zfs file system was transferred yesterday with a speed of
about 60 MB/s to two different servers. The histogram of files looks like:

2822 files were investigated, total size is: 185.82 Gbyte

Summary of file sizes [bytes]:
zero:                  2
1 -> 2                 0
2 -> 4                 1
4 -> 8                 3
8 -> 16               26
16 -> 32               8
32 -> 64               6
64 -> 128             29
128 -> 256            11
256 -> 512            13
512 -> 1024           17
1024 -> 2k            33
2k -> 4k              45
4k -> 8k            9044  ************
8k -> 16k             60
16k -> 32k            41
32k -> 64k            19
64k -> 128k           22
128k -> 256k          12
256k -> 512k           5
512k -> 1024k       1218  **
1024k -> 2M        16004  *********************
2M -> 4M           46202
************************************************************
4M -> 8M               0
8M -> 16M              0
16M -> 32M             0
32M -> 64M             0
64M -> 128M            0
128M -> 256M           0
256M -> 512M           0
512M -> 1024M          0
1024M -> 2G            0
2G -> 4G               0
4G -> 8G               0
8G -> 16G              1

(2) Currently a much larger file system is being transferred, the same
script (even the same incarnation, i.e. process) is now running close to
22 hours:

28549 files were investigated, total size is: 646.67 Gbyte

Summary of file sizes [bytes]:
zero:               4954  **************************
1 -> 2                 0
2 -> 4                 0
4 -> 8                 1
8 -> 16                1
16 -> 32               0
32 -> 64               0
64 -> 128              1
128 -> 256             0
256 -> 512             9
512 -> 1024           71
1024 -> 2k             1
2k -> 4k            1095  ******
4k -> 8k            8449  *********************************************
8k -> 16k           2217  ************
16k -> 32k           503  ***
32k -> 64k             1
64k -> 128k            1
128k -> 256k           1
256k -> 512k           0
512k -> 1024k          0
1024k -> 2M            0
2M -> 4M               0
4M -> 8M              16
8M -> 16M              0
16M -> 32M             0
32M -> 64M         11218
************************************************************
64M -> 128M            0
128M -> 256M           0
256M -> 512M           0
512M -> 1024M          0
1024M -> 2G            0
2G -> 4G               5
4G -> 8G               1
8G -> 16G              3
16G -> 32G             1


When watching zpool iostat I get this (30 second average, NOT the first
output):

               capacity     operations    bandwidth
pool         used  avail   read  write   read  write
----------  -----  -----  -----  -----  -----  -----
atlashome   3.54T  17.3T    137      0  4.28M      0
  raidz2     833G  6.00T      1      0  30.8K      0
    c0t0d0      -      -      1      0  2.38K      0
    c1t0d0      -      -      1      0  2.18K      0
    c4t0d0      -      -      0      0  1.91K      0
    c6t0d0      -      -      0      0  1.76K      0
    c7t0d0      -      -      0      0  1.77K      0
    c0t1d0      -      -      0      0  1.79K      0
    c1t1d0      -      -      0      0  1.86K      0
    c4t1d0      -      -      0      0  1.97K      0
    c5t1d0      -      -      0      0  2.04K      0
    c6t1d0      -      -      1      0  2.25K      0
    c7t1d0      -      -      1      0  2.31K      0
    c0t2d0      -      -      1      0  2.21K      0
    c1t2d0      -      -      0      0  1.99K      0
    c4t2d0      -      -      0      0  1.99K      0
    c5t2d0      -      -      1      0  2.38K      0
  raidz2    1.29T  5.52T     67      0  2.09M      0
    c6t2d0      -      -     58      0   143K      0
    c7t2d0      -      -     58      0   141K      0
    c0t3d0      -      -     53      0   131K      0
    c1t3d0      -      -     53      0   130K      0
    c4t3d0      -      -     58      0   143K      0
    c5t3d0      -      -     58      0   145K      0
    c6t3d0      -      -     59      0   147K      0
    c7t3d0      -      -     59      0   146K      0
    c0t4d0      -      -     59      0   145K      0
    c1t4d0      -      -     58      0   145K      0
    c4t4d0      -      -     58      0   145K      0
    c6t4d0      -      -     58      0   143K      0
    c7t4d0      -      -     58      0   143K      0
    c0t5d0      -      -     58      0   145K      0
    c1t5d0      -      -     58      0   144K      0
  raidz2    1.43T  5.82T     69      0  2.16M      0
    c4t5d0      -      -     62      0   141K      0
    c5t5d0      -      -     60      0   138K      0
    c6t5d0      -      -     59      0   135K      0
    c7t5d0      -      -     60      0   138K      0
    c0t6d0      -      -     62      0   142K      0
    c1t6d0      -      -     61      0   138K      0
    c4t6d0      -      -     59      0   135K      0
    c5t6d0      -      -     60      0   138K      0
    c6t6d0      -      -     62      0   142K      0
    c7t6d0      -      -     61      0   138K      0
    c0t7d0      -      -     58      0   134K      0
    c1t7d0      -      -     60      0   137K      0
    c4t7d0      -      -     62      0   142K      0
    c5t7d0      -      -     61      0   139K      0
    c6t7d0      -      -     58      0   134K      0
    c7t7d0      -      -     60      0   138K      0
----------  -----  -----  -----  -----  -----  -----

Odd things:

(1) The zpool is not equally striped across the raidz2-pools
(2) The disks should be able to perform much much faster than they
currently output data at, I believe it;s 2008 and not 1995.
(3) The four cores of the X4500 are dying of boredom, i.e. idle >95% all
the time.

Has anyone a good idea, where the bottleneck could be? I'm running out
of ideas.

Cheers

Carsten
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to