I've installed OpenBSD onto this box from 4.6 through 5.0 to compare wait
times for simple operations. I don't expect miracles from this relatively
cheap raid controller, but, I expect it to be at least as quick as a regular
sata drive!

So, I'm dd'ing 10GB of zeros to a file, sleeping for a second then timing
how long it takes ls to list the directory contents...

To summarise, write speeds were quickest in 5.0 but system response times
were
worst. Everything was pretty respectable in 4.6 but still a lot slower than
a single disk.

My test was a country mile from scientific so if there's a better way to come
up with results that might help reveal what the problem is I'd be glad to run
more tests...

Here's what I've been doing:

# dd if=/dev/zero of=./testfile bs=1024k count=10000 & sleep 1; time ls -la >
/dev/null;
... followed by a few more...
# time ls -la > /dev/null;

And the results where the ls time is a subjective average:

The other server I've got in the office...
OpenBSD 4.6 i386 on a single SATA drive:
ls: 0.000u 0.020s 0:00.03 66.6%     0+0k 0+0io 0pf+0w
dd: 10485760000 bytes transferred in 94.775 secs (110637306 bytes/sec)

OpenBSD 5.0 amd64 RAID 5
ls: 0m5.80s real     0m0.00s user     0m0.13s system
dd: 10485760000 bytes transferred in 53.736 secs (195132964 bytes/sec)
Remarks: Terribly slow!

OpenBSD 4.9 amd64 RAID 5
ls: 0m5.95s real     0m0.00s user     0m0.06s system
dd: 10485760000 bytes transferred in 75.058 secs (139700269 bytes/sec)
Remarks: No better than 5.0

OpenBSD 4.8 amd64 RAID 5
ls: 0m5.72s real     0m0.00s user     0m0.04s system
dd: 10485760000 bytes transferred in 103.893 secs (100927877 bytes/sec)
Remarks: A bit quicker, got some really quick response times

OpenBSD 4.7 amd64 RAID 5
ls: 0m4.79s real     0m0.00s user     0m0.04s system
dd: 10485760000 bytes transferred in 95.476 secs (109825323 bytes/sec)
Remarks: A little quicker than 4.8

OpenBSD 4.6 amd64 RAID 5
ls: 0m1.90s real     0m0.00s user     0m0.02s system
dd: 10485760000 bytes transferred in 64.263 secs (163166944 bytes/sec)
Remarks: Consistently around the 2 second mark


> George Steel [li...@netglue.co] wrote:
>> I've been testing and comparing between servers using dd -if /dev/zero and
then performing simple tasks like ls.
>> On a 4.6 server with a single SATA disk, ls spits out the listing
immediately, on this RAID 5 box, the terminal hangs for as much as 12 seconds
then begrudgingly spits out the dir listing line by line.
>> I expect the system to become slower whilst writing 10GB of zeros to a
file, but it seems to me that something is going on with this RAID box because
the wait is unbelievable compared to a much lower spec machine.
>> Perhaps this is to be expected with a relatively cheap RAID controller? and
I'd be better off just attaching separate disks and doing softraid.
>> If I cat the 10GB file to /dev/null and perform the same type of
operations, everything is as quick as you'd expect.
>>
>> On 10 Jan 2012, at 17:48, Chris Cappuccio wrote:
>>
>>> George Steel [li...@netglue.co] wrote:
>>>> Yeah, I did start up top before hand on another terminal and biowait was
all I saw with a 1 sec delay.
>>>> I repeated the test several times and never saw anything other than
biowait
>>>> I also had a look with ps but couldn't really interpret what I saw other
than ps reported state as "D" for both processes.
>>>> I'm also not much good at interpreting systat but to my untrained eye, I
couldn't see much difference between the idle machine and a heavy write other
than lots of disk IO
>>>> There's nothing in any logs and I've also tried the RAID card in
different slots.
>>>> I also installed i386 and had the same problem
>>>>
>>>
>>> what activity is tying your disks up like this?
>
> --

Reply via email to