Hi Lukasz,
The output of the first sript gives
bash-3.00# ./test.sh
dtrace: script './test.sh' matched 4 probes
CPU     ID                    FUNCTION:NAME
  0  42681                        :tick-10s

  0  42681                        :tick-10s

  0  42681                        :tick-10s

  0  42681                        :tick-10s

  0  42681                        :tick-10s

  0  42681                        :tick-10s

  0  42681                        :tick-10s



and it goes on.

The second script gives:

checking pool map size [B]: filer
mdb: failed to dereference symbol: unknown symbol name
423917216903435

Regards,
--Walter

On 11/7/07, Łukasz K <[EMAIL PROTECTED]> wrote:
>
> Hi,
>
>   I think your problem is filesystem fragmentation.
> When available space is less than 40% ZFS might have problems with
> finding free blocks. Use this script to check it:
>
> #!/usr/sbin/dtrace -s
>
> fbt::space_map_alloc:entry
> {
>    self->s = arg1;
> }
>
> fbt::space_map_alloc:return
> /arg1 != -1/
> {
>   self->s = 0;
> }
>
> fbt::space_map_alloc:return
> /self->s && (arg1 == -1)/
> {
>   @s = quantize(self->s);
>   self->s = 0;
> }
>
> tick-10s
> {
>   printa(@s);
> }
>
> Run script for few minutes.
>
>
> You might also have problems with space map size.
> This script will show you size of space map on disk:
> #!/bin/sh
>
> echo '::spa' | mdb -k | grep ACTIVE \
>   | while read pool_ptr state pool_name
> do
>   echo "checking pool map size [B]: $pool_name"
>
>   echo "${pool_ptr}::walk metaslab|::print -d struct metaslab
> ms_smo.smo_objsize" \
>     | mdb -k \
>     | nawk '{sub("^0t","",$3);sum+=$3}END{print sum}'
> done
>
> In memory space map takes 5 times more.
> All space map is loaded into memory all the time, but for example
> during snapshot remove all space map might be loaded, so check
> if you have enough RAM available on machine.
> Check ::kmastat in mdb.
> Space map uses kmem_alloc_40  ( on thumpers this is a real problem )
>
> Workaround:
> 1. first you can change pool recordsize
>   zfs set recordsize=64K POOL
>
> Maybe you wil have to use 32K or even 16K
>
> 2. You will have to disable ZIL, becuase ZIL always takes 128kB
> blocks.
>
> 3. Try to disable cache, tune vdev cache. Check:
> http://www.solarisinternals.com/wiki/index.php/ZFS_Evil_Tuning_Guide
>
> Lukas Karwacki
>
> Dnia 7-11-2007 o godz. 1:49 Walter Faleiro napisał(a):
> > Hi,
> > We have a zfs file system configured using a Sunfire 280R with a 10T
> > Raidweb array
> >
> > bash-3.00# zpool list
> > NAME                    SIZE    USED   AVAIL    CAP  HEALTH     ALTROOT
> > filer                           9.44T   6.97T   2.47T    73%  ONLINE
> -
> >
> >
> > bash-3.00# zpool status
> >   pool: backup
> >  state: ONLINE
> >  scrub: none requested
> > config:
> >
> >         NAME        STATE     READ WRITE CKSUM
> >         filer            ONLINE       0     0     0
> >           c1t2d1    ONLINE       0     0     0
> >           c1t2d2    ONLINE       0     0     0
> >           c1t2d3    ONLINE       0     0     0
> >           c1t2d4    ONLINE       0     0     0
> >           c1t2d5    ONLINE       0     0     0
> >
> >
> > the file system is shared via nfs. Off late we have seen that the file
> > system access slows down considerably. Running commands like find, du
> > on the zfs system did slow it down, but the intermittent slowdowns
> > cannot be explained.
> >
> > Is there a way to trace the I/O on the zfs so that we can list out
> > heavy read/writes to the file system to be responsible for the
> > slowness.
> >
> > Thanks,
> > --Walter
> > _______________________________________________
> > zfs-discuss mailing list
> > zfs-discuss@opensolaris.org
> > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
> ----------------------------------------------------
> Wojna z terrorem wkracza w decydującą fazę:
> Robert Redford, Meryl Streep i Tom Cruise w filmie
> UKRYTA STRATEGIA - w kinach od 9 listopada!
>
> http://klik.wp.pl/?adr=http%3A%2F%2Fcorto.www.wp.pl%2Fas%2Fstrategia.html&sid=90
>
>
>
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to