My understanding is that it is not supposed to be a loaded system. We
want to know what the load is.
gernot@tintenfass:~# intrstat 30
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
e1000g#0 | 1 0,0 0 0,0
ehci#0 | 0 0,0 4 0,0
ehci#1 | 3 0,0 0 0,0
hci1394#0 | 0 0,0 2 0,0
i8042#1 | 0 0,0 4 0,0
i915#1 | 0 0,0 2 0,0
pci-ide#0 | 15 0,1 0 0,0
uhci#0 | 0 0,0 2 0,0
uhci#1 | 0 0,0 0 0,0
uhci#2 | 3 0,0 0 0,0
uhci#3 | 0 0,0 2 0,0
uhci#4 | 0 0,0 4 0,0
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
e1000g#0 | 1 0,0 0 0,0
ehci#0 | 0 0,0 3 0,0
ehci#1 | 3 0,0 0 0,0
hci1394#0 | 0 0,0 1 0,0
i8042#1 | 0 0,0 6 0,0
i915#1 | 0 0,0 1 0,0
pci-ide#0 | 3 0,0 0 0,0
uhci#0 | 0 0,0 1 0,0
uhci#1 | 0 0,0 0 0,0
uhci#2 | 3 0,0 0 0,0
uhci#3 | 0 0,0 1 0,0
uhci#4 | 0 0,0 3 0,0
gernot@tintenfass:~# vmstat 5 10
kthr memory page disk faults
cpu
r b w swap free re mf pi po fr de sr cd s0 s1 s2 in sy cs us
sy id
0 0 0 4243840 1145720 1 6 0 0 0 0 2 0 1 1 1 9767 121 37073 0
54 46
0 0 0 4157824 1059796 4 11 0 0 0 0 0 0 0 0 0 9752 119 37132 0
54 46
0 0 0 4157736 1059752 0 0 0 0 0 0 0 0 0 0 0 9769 113 37194 0
54 46
0 0 0 4157744 1059788 0 0 0 0 0 0 0 0 0 0 0 9682 104 36941 0
54 46
0 0 0 4157744 1059788 0 0 0 0 0 0 0 0 0 0 0 9769 105 37208 0
54 46
0 0 0 4157728 1059772 0 1 0 0 0 0 0 0 0 0 0 9741 159 37104 0
54 46
0 0 0 4157728 1059772 0 0 0 0 0 0 0 0 0 0 0 9695 127 36931 0
54 46
0 0 0 4157744 1059788 0 0 0 0 0 0 0 0 0 0 0 9762 105 37188 0
54 46
0 0 0 4157744 1059788 0 0 0 0 0 0 0 0 0 0 0 9723 102 37058 0
54 46
0 0 0 4157744 1059788 0 0 0 0 0 0 0 0 0 0 0 9774 105 37263 0
54 46
Mike
On Thu, 2011-10-20 at 11:02 -0700, Rennie Allen wrote:
Sched is the scheduler itself. How long did you let this run? If only
for a couple of seconds, then that number is high, but not ridiculous for
a loaded system, so I think that this output rules out a high context
switch rate.
Try this command to see if some process is making an excessive number of
syscalls:
dtrace -n 'syscall:::entry { @[execname]=count()}'
If not, then I'd try looking at interrupts...
On 10/20/11 10:52 AM, "Gernot Wolf"<gw.i...@chello.at> wrote:
Yeah, I've been able to run this diagnostics on another OI box (at my
office, so much for OI not being used in production ;)), and noticed
that there were several values that were quite different. I just don't
have any idea on the meaning of this figures...
Anyway, here are the results of the dtrace command (I executed the
command twice, hence two result sets):
gernot@tintenfass:~# dtrace -n 'sched:::off-cpu { @[execname]=count()}'
dtrace: description 'sched:::off-cpu ' matched 3 probes
^C
ipmgmtd 1
gconfd-2 2
gnome-settings-d 2
idmapd 2
inetd 2
miniserv.pl 2
netcfgd 2
nscd 2
ospm-applet 2
ssh-agent 2
sshd 2
svc.startd 2
intrd 3
afpd 4
mdnsd 4
gnome-power-mana 5
clock-applet 7
sendmail 7
xscreensaver 7
fmd 9
fsflush 11
ntpd 11
updatemanagernot 13
isapython2.6 14
devfsadm 20
gnome-terminal 20
dtrace 23
mixer_applet2 25
smbd 39
nwam-manager 60
svc.configd 79
Xorg 100
sched 394078
gernot@tintenfass:~# dtrace -n 'sched:::off-cpu { @[execname]=count()}'
dtrace: description 'sched:::off-cpu ' matched 3 probes
^C
automountd 1
ipmgmtd 1
idmapd 2
in.routed 2
init 2
miniserv.pl 2
netcfgd 2
ssh-agent 2
sshd 2
svc.startd 2
fmd 3
hald 3
inetd 3
intrd 3
hald-addon-acpi 4
nscd 4
gnome-power-mana 5
sendmail 5
mdnsd 6
devfsadm 8
xscreensaver 9
fsflush 10
ntpd 14
updatemanagernot 16
mixer_applet2 21
isapython2.6 22
dtrace 24
gnome-terminal 24
smbd 39
nwam-manager 58
zpool-rpool 65
svc.configd 79
Xorg 82
sched 369939
So, quite obviously there is one executable standing out here, "sched",
now what's the meaning of this figures?
Regards,
Gernot Wolf
Am 20.10.11 19:22, schrieb Michael Stapleton:
Hi Gernot,
You have a high context switch rate.
try
#dtrace -n 'sched:::off-cpu { @[execname]=count()}'
For a few seconds to see if you can get the name of and executable.
Mike
On Thu, 2011-10-20 at 18:44 +0200, Gernot Wolf wrote:
Hello all,
I have a machine here at my home running OpenIndiana oi_151a, which
serves as a NAS on my home network. The original install was
OpenSolaris
2009.6 which was later upgraded to snv_134b, and recently to oi_151a.
So far this OSOL (now OI) box has performed excellently, with one major
exception: Sometimes, after a reboot, the cpu load was about 50-60%,
although the system was doing nothing. Until recently, another reboot
solved the issue.
This does not work any longer. The system has always a cpu load of
50-60% when idle (and higher of course when there is actually some work
to do).
I've already googled the symptoms. This didn't turn up very much useful
info, and the few things I found didn't apply to my problem. Most
noticably was this problem which could be solved by disabling cpupm in
/etc/power.conf, but trying that didn't show any effect on my system.
So I'm finally out of my depth. I have to admit that my knowledge of
Unix is superficial at best, so I decided to try looking for help here.
I've run several diagnostic commands like top, powertop, lockstat etc.
and attached the results to this email (I've zipped the results of
kstat
because they were>1MB).
One important thing is that when I boot into the oi_151a live dvd
instead of booting into the installed system, I also get the high cpu
load. I mention this because I have installed several things on my OI
box like vsftpd, svn, netstat etc. I first thought that this problem
might be caused by some of this extra stuff, but getting the same
system
when booting the live dvd ruled that out (I think).
The machine is a custom build medium tower:
S-775 Intel DG965WHMKR ATX mainbord
Intel Core 2 Duo E4300 CPU 1.8GHz
1x IDE DVD recorder
1x IDE HD 200GB (serves as system drive)
6x SATA II 1.5TB HD (configured as zfs raidz2 array)
I have to solve this problem. Although the system runs fine and
absolutely serves it's purpose, having the cpu at 50-60% load
constantly
is a waste of energy and surely a rather unhealthy stress on the
hardware.
Anyone any ideas...?
Regards,
Gernot Wolf
_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss
_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss
_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss
_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss
_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss