Were you able to deduce anything from the test results?
On 11 May 2014 23:02, "Juergen Sauermann" <juergen.sauerm...@t-online.de>
wrote:

>  Hi Elias,
>
> thanks, already interesting. If you could loop around the core count:
>
> *for ((i=1; $i<=80; ++i)); do*
> * ./Parallel $i*
> * ./Parallel_OMP $i*
> *done*
>
> then I could understand the data better. Also not sure if something
> is wrong with the benchmark program. On my new 4-core with OMP I get
> fluctuations from:
>
> eedjsa@server65 ~/apl-1.3/tools $ ./Parallel_OMP 4
> Pass 0: 4 cores/threads, 8229949 cycles total
> Pass 1: 4 cores/threads, 8262 cycles total
> Pass 2: 4 cores/threads, 4035 cycles total
> Pass 3: 4 cores/threads, 4126 cycles total
> Pass 4: 4 cores/threads, 4179 cycles total
>
> to:
>
> eedjsa@server65 ~/apl-1.3/tools $ ./Parallel_OMP 4
> Pass 0: 4 cores/threads, 11368032 cycles total
> Pass 1: 4 cores/threads, 4042228 cycles total
> Pass 2: 4 cores/threads, 7251419 cycles total
> Pass 3: 4 cores/threads, 3846 cycles total
> Pass 4: 4 cores/threads, 2725 cycles total
>
> The fluctuations with the manual parallel for are smaller:
>
> Pass 0: 4 cores/threads, 87225 cycles total
> Pass 1: 4 cores/threads, 245046 cycles total
> Pass 2: 4 cores/threads, 84632 cycles total
> Pass 3: 4 cores/threads, 63619 cycles total
> Pass 4: 4 cores/threads, 93437 cycles total
>
> but still considerable. The picture so far suggests that OMP fluctuates
> much
> more (in the start-up + sync time) than manual with the highest OMP
> start-up above manual
> and the lowest far below. One change on my  TODO list is to use futexes
> instead of mutexes
> (like OMP does), probably not an issue under Solaris sunce futextes are
> linux-specific.
>
> /// Jürgen
>
>
> On 05/11/2014 04:23 AM, Elias Mårtenson wrote:
>
> Here are the files that I promised earlier.
>
>  Regards,
> Elias
>
>
>

Reply via email to