Re: Post 9.1 stable file system problems

2013-01-05 Thread Konstantin Belousov
On Tue, Jan 01, 2013 at 05:58:06PM +0200, Konstantin Belousov wrote:
> On Tue, Jan 01, 2013 at 02:39:44PM +0100, Dominic Fandrey wrote:
> > On 01/01/2013 07:51, Konstantin Belousov wrote:
> > > On Tue, Jan 01, 2013 at 02:05:11AM +0100, Dominic Fandrey wrote:
> > >> On 01/01/2013 01:49, Dominic Fandrey wrote:
> > >>> On 01/01/2013 01:29, Chris Rees wrote:
> >  On 1 Jan 2013 00:01, "Dominic Fandrey"  wrote:
> > >
> > > I have a Tinderbox that I just updated to the current RELENG_9.
> > > Following the update build times for packages have increased by a
> > > factor between 5 and 20. I.e. I have packages that used to build in
> > > 5 minutes and now take an hour.
> > >
> > > I'm suspecting the file system ever since I saw that the majority of 
> > > CPU
> > > load was caused by ls when I looked at top (more than 2 minutes of CPU
> > > time were counted that moment). The majority of the time most of the 
> > > CPU
> > > load is caused by bsdtar, pkg_add, qmake-qt4, etc. Without exception
> > > tools that access a lot of files.
> > >
> > > The file system on which packages are built is nullfs mounted from
> > > an async mounted UFS. I turned async off, to no avail.
> > >
> > > /usr/src/UPDATING says that there were nullfs optimisations. So I
> > > think this is where the problem originates. I might hack the 
> > > tinderbox to
> > > use 'ln -s' or set it up for NFS to verify this.
> > 
> >  Is your kernel newer than the Jail?  The converse causes problems.
> > >>>
> > >>> I ran makeJail for all jails after updating.
> Did you rebuild your modules together with the new kernel ?
> 
> > >>>
> > >>> I also seem to have similar problems when building in the host-system.
> > >>> The unzip for openjdk-7 has just passed the 11 minutes CPU time mark.
> > >>> On my notebook it takes less than 10 seconds.
> > >>
> > >> Just set WRKOBJDIRPREFIX to a tmpfs on the Tinderbox host system
> > >> and the extract takes less than a second. Originally WRKOBJDIRPREFIX
> > >> also pointed to a nullfs mount.
> > >>
> > >> Afterwards I pointed WRKOBJDIRPREFIX to a UFS file system (without
> > >> nullfs involvement). The entire make extract took 20s.
> > >>
> > >> So still faster by at least factor 30 than running it on a nullfs mount
> > >> (I eventually SIGINTed so I don't know how long it would've run).
> > > 
> > > Start providing some useful debugging information ?
> > 
> > That one might be interesting. It's all system time:
> > 
> > # time -lh make extract
> > ===>  License GPLv2 accepted by the user
> > ===>  Found saved configuration for openjdk-7.9.05_1
> > ===>  Extracting for openjdk-7.9.05_2
> > => SHA256 Checksum OK for openjdk-7u6-fcs-src-b24-09_aug_2012.zip.
> > => SHA256 Checksum OK for apache-ant-1.8.4-bin.zip.
> > ===>   openjdk-7.9.05_2 depends on file: /usr/local/bin/unzip - found
> > ^Ctime: command terminated abnormally
> > 4m29.30s real   3.03s user  4m22.55s sys
> >   5008  maximum resident set size
> >135  average shared memory size
> >   2932  average unshared data size
> >127  average unshared stack size
> >   7772  page reclaims
> >  0  page faults
> >  0  swaps
> > 19  block input operations
> >101  block output operations
> >  0  messages sent
> >  0  messages received
> > 41  signals received
> >   1597  voluntary context switches
> >  16590  involuntary context switches
> 
> Ok, from your mount -v output, are the three nullfs mounts the only
> nullfs mount ever used ?
> 
> Is it only unzip which demostrates the silly behaviour ? Or does it
> happen with any program ? E.g., does ls(1) or sha1 on the nullfs mount
> also slow ?
> 
> Could you try some low-tech profiling on the slow program. For instance,
> you could run ktrace/kdump -R to see which syscalls are slow.
> 
> Most darkly part of your report for me, is that I also use nullfs-backed
> jails both on HEAD and stable/9, with bigger scale, and I do not have
> an issue. I just did
> pooma32% time unzip -q 
> /usr/local/arch/freebsd/distfiles/openjdk-7u6-fcs-src-b24-09_aug_2012.zip
> unzip -q   3.25s user 23.77s system 78% cpu 34.482 total
> over nullfs mount of
> /usr/home on /usr/sfw/local8/opt/pooma32/usr/home (nullfs, local).
> 
> Please try the following patch, which changes nullfs behaviour to be
> non-cached by default. You could turn on the caching with the 'mount -t
> nullfs -o cache from to' mounting command. I am interested if use/non-use
> of -o cache makes a difference for you.

Ping. Any update ?


pgptVju7RyzVW.pgp
Description: PGP signature


load average 0.60 at 100% idle

2013-01-05 Thread Johan Broman
Hi!

Bare with me here as I'm totally new to FreeBSD (like 2 weeks).

>From what I understand, there is a change in FreeBSD 9.1 ACPI code
that makes detection of devices happen earlier? This can result in
eventtimer irqs being shared with another device and can therefor
throw off tools reporting load avg values (they report constantly
around 0.60 in load). I guess there were good reasons for making the
ACPI change, so is this a ACPI problem or should the code around the
tools, such as top, change? This is a 9.1 "regression" as the problem
is not there in FreeBSD 9.0.

Here's some more info on the bug and possible workarounds:

http://www.freebsd.org/cgi/query-pr.cgi?pr=173541&cat=

and

http://lists.freebsd.org/pipermail/freebsd-stable/2012-April/067382.html

Any input on this would be greatly appreciated.

Thanks
Johan
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: load average 0.60 at 100% idle

2013-01-05 Thread Adrian Chadd
Cc'ing the committer / ACPI owner!

jkim - any comments?



Adrian

On 5 January 2013 12:38, Johan Broman  wrote:
> Hi!
>
> Bare with me here as I'm totally new to FreeBSD (like 2 weeks).
>
> From what I understand, there is a change in FreeBSD 9.1 ACPI code
> that makes detection of devices happen earlier? This can result in
> eventtimer irqs being shared with another device and can therefor
> throw off tools reporting load avg values (they report constantly
> around 0.60 in load). I guess there were good reasons for making the
> ACPI change, so is this a ACPI problem or should the code around the
> tools, such as top, change? This is a 9.1 "regression" as the problem
> is not there in FreeBSD 9.0.
>
> Here's some more info on the bug and possible workarounds:
>
> http://www.freebsd.org/cgi/query-pr.cgi?pr=173541&cat=
>
> and
>
> http://lists.freebsd.org/pipermail/freebsd-stable/2012-April/067382.html
>
> Any input on this would be greatly appreciated.
>
> Thanks
> Johan
> ___
> freebsd-stable@freebsd.org mailing list
> http://lists.freebsd.org/mailman/listinfo/freebsd-stable
> To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


GELI performance on Atom D525

2013-01-05 Thread Wiktor Niesiobedzki
Hi,

Can anyone with access to Atom D525 processor can do a performance check for me:
# kldload geom_eli
# kldload geom_zero
# sysctl kern.geom.zero.clear=0
# geli onetime  -s 4096 -l 256 -e aes-cbc gzero
# dd if=/dev/gzero.eli of=/dev/null bs=1m count=4096
# geli kill gzero
# geli onetime  -s 4096 -l 256 -e aes-xts gzero
# dd if=/dev/gzero.eli of=/dev/null bs=1m count=4096

I'm interested in dd outputs ofcourse. I would like to compare this to
VIA Esther processor 1000MHz, which has padlock AES accelerator.

My results are:

AES-CBC:
4294967296 bytes transferred in 31.805894 secs (135036836 bytes/sec)

AES-XTS:
4294967296 bytes transferred in 765.232668 secs (5612629 bytes/sec)


Anyone can provide the results for Atom D525?

Cheers,

Wiktor Niesiobedzki
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: GELI performance on Atom D525

2013-01-05 Thread Maxim Khitrov
On Sat, Jan 5, 2013 at 4:58 PM, Wiktor Niesiobedzki  wrote:
> Hi,
>
> Can anyone with access to Atom D525 processor can do a performance check for 
> me:
> # kldload geom_eli
> # kldload geom_zero
> # sysctl kern.geom.zero.clear=0
> # geli onetime  -s 4096 -l 256 -e aes-cbc gzero
> # dd if=/dev/gzero.eli of=/dev/null bs=1m count=4096
> # geli kill gzero
> # geli onetime  -s 4096 -l 256 -e aes-xts gzero
> # dd if=/dev/gzero.eli of=/dev/null bs=1m count=4096
>
> I'm interested in dd outputs ofcourse. I would like to compare this to
> VIA Esther processor 1000MHz, which has padlock AES accelerator.
>
> My results are:
>
> AES-CBC:
> 4294967296 bytes transferred in 31.805894 secs (135036836 bytes/sec)
>
> AES-XTS:
> 4294967296 bytes transferred in 765.232668 secs (5612629 bytes/sec)
>
>
> Anyone can provide the results for Atom D525?

FreeBSD 9.0-RELEASE-p1 amd64:

CPU: Intel(R) Atom(TM) CPU D525   @ 1.80GHz (1800.04-MHz K8-class CPU)
  Origin = "GenuineIntel"  Id = 0x106ca  Family = 6  Model = 1c  Stepping = 10
  
Features=0xbfebfbff
  Features2=0x40e31d
  AMD Features=0x20100800
  AMD Features2=0x1
  TSC: P-state invariant, performance statistics

# sysctl kern.geom.zero.clear=0
kern.geom.zero.clear: 1 -> 0
# geli onetime  -s 4096 -l 256 -e aes-cbc gzero
# dd if=/dev/gzero.eli of=/dev/null bs=1m count=4096
4096+0 records in
4096+0 records out
4294967296 bytes transferred in 192.024101 secs (22366814 bytes/sec)
# geli kill gzero
# geli onetime  -s 4096 -l 256 -e aes-xts gzero
# dd if=/dev/gzero.eli of=/dev/null bs=1m count=4096
4096+0 records in
4096+0 records out
4294967296 bytes transferred in 232.956895 secs (18436747 bytes/sec)

- Max
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: GELI performance on Atom D525

2013-01-05 Thread Claude Buisson

On 01/05/2013 22:58, Wiktor Niesiobedzki wrote:

Hi,

Can anyone with access to Atom D525 processor can do a performance check for me:
# kldload geom_eli
# kldload geom_zero
# sysctl kern.geom.zero.clear=0
# geli onetime  -s 4096 -l 256 -e aes-cbc gzero
# dd if=/dev/gzero.eli of=/dev/null bs=1m count=4096
# geli kill gzero
# geli onetime  -s 4096 -l 256 -e aes-xts gzero
# dd if=/dev/gzero.eli of=/dev/null bs=1m count=4096

I'm interested in dd outputs ofcourse. I would like to compare this to
VIA Esther processor 1000MHz, which has padlock AES accelerator.

My results are:

AES-CBC:
4294967296 bytes transferred in 31.805894 secs (135036836 bytes/sec)

AES-XTS:
4294967296 bytes transferred in 765.232668 secs (5612629 bytes/sec)


Anyone can provide the results for Atom D525?



FreeBSD 9.1-STABLE (svn 243788)

CPU: Intel(R) Atom(TM) CPU D525   @ 1.80GHz (1795.75-MHz K8-class CPU)
Origin = "GenuineIntel"  Id = 0x106ca  Family = 0x6  Model = 0x1c  Stepping = 10

Features=0xbfebfbff
  Features2=0x40e31d
  AMD Features=0x20100800
  AMD Features2=0x1
  TSC: P-state invariant, performance statistics
real memory  = 4294967296 (4096 MB)
avail memory = 4083802112 (3894 MB)
Event timer "LAPIC" quality 400
ACPI APIC Table: <032111 APIC1732>
FreeBSD/SMP: Multiprocessor System Detected: 4 CPUs
FreeBSD/SMP: 1 package(s) x 2 core(s) x 2 HTT threads
 cpu0 (BSP): APIC ID:  0
 cpu1 (AP/HT): APIC ID:  1
 cpu2 (AP): APIC ID:  2
 cpu3 (AP/HT): APIC ID:  3

# kldload geom_eli
# kldload geom_zero
# sysctl kern.geom.zero.clear=0
kern.geom.zero.clear: 1 -> 0
# geli onetime  -s 4096 -l 256 -e aes-cbc gzero
# dd if=/dev/gzero.eli of=/dev/null bs=1m count=4096
4096+0 records in
4096+0 records out
4294967296 bytes transferred in 245.828912 secs (17471368 bytes/sec)
# geli kill gzero
# geli onetime  -s 4096 -l 256 -e aes-xts gzero
# dd if=/dev/gzero.eli of=/dev/null bs=1m count=4096
4096+0 records in
4096+0 records out
4294967296 bytes transferred in 291.027184 secs (14757959 bytes/sec)


Cheers,

Wiktor Niesiobedzki


Claude Buisson
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"