Anyway, after compiling my code with icc/11.1.069, the job is running
without stuck or that sigv which it occurred before when using icc/12.1.0
module.
Also I have to point that when I was using icc/12.1.0 I was getting strange
outputs or stuck, and I solved them by changing the name of parameters
it is a good question I asked it myself at the first but then I said it
should be correct but anyway I want to confirm that:
her is the code snippet of the program:
...
int ranks[size];
for(i=0; i < size; ++i)
{
ranks[i] = i;
}
...
for(p=8; p <= (size); p+=4)
{
MPI_Barrier(MP
This may be a dumb question, but are you 100% sure that the input values are
correct?
On Jan 10, 2012, at 8:16 AM, Anas Al-Trad wrote:
> Hi Ralph, I changed the intel icc module from 12.1.0 to 11.1.069, the
> previous default one used at a Neolith Cluster. I submitted the job and I
> still wa
Hi Ralph, I changed the intel icc module from 12.1.0 to 11.1.069, the
previous default one used at a Neolith Cluster. I submitted the job and I
still waiting for the result. Here is the message of the segmentation fault:
[n764:29867] *** Process received signal ***
[n764:29867] Signal: Floating p
Have you tried the suggested fix from the email thread Paul cited? Sounds to me
like the most likely cause of the problem, assuming it comes from inside OMPI.
Have you looked at the backtrace to see if it is indeed inside OMPI vs your
code?
On Jan 10, 2012, at 6:13 AM, Anas Al-Trad wrote:
>
>
Thanks Paul,
yes I use Intel 12.1.0, and this error is intermittent, not always produced
but most of the times it occurs.
My program is large and contains many files that are related to each other,
I don't think it will help if I take the snippet of the code. The program
run parallel matrix multipl
A blind guess: did you use Intel compiler?
If so, there is/was an error leading to SIGSEGV _in Open MPI itselv_.
http://www.open-mpi.org/community/lists/users/2012/01/18091.php
If the SIGSEGV arise not in OpenMPI but in application itself it may be
a programming issue.. In any case, more precis
Dear people,
In my application, I have the segmentation fault of Integer
Divide-by-zero when calling MPI_cart_sub routine. My program is as follows,
I have 128 ranks, I make a new communicator of the first 96 ranks via
MPI_Comm_creat. Then I create a grid of 8X12 by calling MPI_Cart_