Hello Andy,

the program is not mine. I have got it from a group upon request. It might be 
program related because I run other codes such as quantum espresso and work 
perfectly fine although it is the cluster people who compiled it. Since I have 
compiled the program I am having problems with, I am thinking that it might be 
"compilation" related. This is why i wanted some experts' opinion on this



________________________________
From: users <users-boun...@open-mpi.org> on behalf of Andy Riebs 
<andy.ri...@hpe.com>
Sent: Saturday, April 23, 2016 12:49 PM
To: us...@open-mpi.org
Subject: Re: [OMPI users] MPIRUN SEGMENTATION FAULT

The challenge for the MPI experts here (of which I am NOT one!) is that the 
problem appears to be in your program; MPI is simply reporting that your 
program failed. If you got the program from someone else, you will need to 
solicit their help. If you wrote it, well, it is never a bad time to learn to 
use gdb!

Best regards
Andy

On 04/23/2016 10:41 AM, Elio Physics wrote:

I am not really an expert with gdb. What is the core file? and how to use gdb? 
I have got three files as an output when the executable is used. One is the 
actual output which stops and the other two are error files (from which I knew 
about the segmentation fault).



thanks


________________________________
From: users <users-boun...@open-mpi.org><mailto:users-boun...@open-mpi.org> on 
behalf of Ralph Castain <r...@open-mpi.org><mailto:r...@open-mpi.org>
Sent: Saturday, April 23, 2016 11:39 AM
To: Open MPI Users
Subject: Re: [OMPI users] MPIRUN SEGMENTATION FAULT

valgrind isn’t going to help here - there are multiple reasons why your 
application could be segfaulting. Take a look at the core file with gdb and 
find out where it is failing.

On Apr 22, 2016, at 10:20 PM, Elio Physics 
<elio-phys...@live.com<mailto:elio-phys...@live.com>> wrote:

One more thing i forgot to mention in my previous e-mail. In the output file I 
get the following message:


2 total processes killed (some possibly by mpirun during cleanup)

Thanks



________________________________
From: users <users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>> on 
behalf of Elio Physics 
<<mailto:elio-phys...@live.com>elio-phys...@live.com<mailto:elio-phys...@live.com>>
Sent: Saturday, April 23, 2016 3:07 AM
To: Open MPI Users
Subject: Re: [OMPI users] MPIRUN SEGMENTATION FAULT

I have used valgrind and this is what i got:

valgrind mpirun ~/Elie/SPRKKR/bin/kkrscf6.3MPI Fe_SCF.inp > 
scf-51551.jlborges.fisica.ufmg.br.out
==8135== Memcheck, a memory error detector
==8135== Copyright (C) 2002-2012, and GNU GPL'd, by Julian Seward et al.
==8135== Using Valgrind-3.8.1 and LibVEX; rerun with -h for copyright info
==8135== Command: mpirun /home/emoujaes/Elie/SPRKKR/bin/kkrscf6.3MPI Fe_SCF.inp
==8135==
--------------------------------------------------------------------------
mpirun noticed that process rank 0 with PID 8147 on node 
jlborges.fisica.ufmg.br<http://jlborges.fisica.ufmg.br/> exited on signal 11 
(Segmentation fault).
--------------------------------------------------------------------------
==8135==
==8135== HEAP SUMMARY:
==8135==     in use at exit: 485,683 bytes in 1,899 blocks
==8135==   total heap usage: 7,723 allocs, 5,824 frees, 12,185,660 bytes 
allocated
==8135==
==8135== LEAK SUMMARY:
==8135==    definitely lost: 34,944 bytes in 34 blocks
==8135==    indirectly lost: 26,613 bytes in 58 blocks
==8135==      possibly lost: 0 bytes in 0 blocks
==8135==    still reachable: 424,126 bytes in 1,807 blocks
==8135==         suppressed: 0 bytes in 0 blocks
==8135== Rerun with --leak-check=full to see details of leaked memory
==8135==
==8135== For counts of detected and suppressed errors, rerun with: -v
==8135== ERROR SUMMARY: 0 errors from 0 contexts (suppressed: 6 from 6)

What does that supposed to mean?

Regards
________________________________
From: users <users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>> on 
behalf of Ralph Castain 
<<mailto:r...@open-mpi.org>r...@open-mpi.org<mailto:r...@open-mpi.org>>
Sent: Saturday, April 23, 2016 1:36:50 AM
To: Open MPI Users
Subject: Re: [OMPI users] MPIRUN SEGMENTATION FAULT

All I can say is that your program segfault’d during execution - you might want 
to look at the core file using a debugger like gdb to see why it failed.


On Apr 22, 2016, at 8:32 PM, Elio Physics 
<<mailto:elio-phys...@live.com>elio-phys...@live.com<mailto:elio-phys...@live.com>>
 wrote:

Dear all,

I have successfully compiled a code where the executable have been produced. 
However when I started using the executable with mpirun, the code stopped with 
the following error:

"mpirun noticed that process rank 0 with PID 570 on node compute-1-9.local 
exited on signal 11 (Segmentation fault)."

What is that error due to? and How can i solve it?

I will post the make.inc compilation file:

BUILD_TYPE ?=
#BUILD_TYPE := debug

VERSION = 6.3

ifeq ($(BUILD_TYPE), debug)
 VERSION := $(VERSION)$(BUILD_TYPE)
endif

BIN =~/Elie/SPRKKR/bin
#BIN=~/bin
#BIN=/tmp/$(USER)

LIB =  -L/opt/intel/mkl/lib/intel64/libmkl_blas95_ilp64 
-L/opt/intel/mkl/lib/intel64/libmkl_lapack95_ilp64 -L/opt/intel/mkl/lib/intel64 
-lmkl_scalapack_ilp64 -lmkl_intel_ilp64 -lmkl_core -lmkl_sequential 
-lmkl_blacs_i
ntelmpi_ilp64 -lpthread -lm -ldl

#Include mpif.h
INCLUDE = -I/opt/intel/mkl/include/intel64/ilp64  -I/opt/intel/mkl/lib/include

#FFLAGS
FFLAGS = -O2

FC   = mpif90 -c $(FFLAGS) $(INCLUDE)
LINK = mpif90   $(FFLAGS) $(INCLUDE)

MPI=MPI

Thanks in advance

Elio
University of Rondonia, brazil

_______________________________________________
users mailing list
us...@open-mpi.org<mailto:us...@open-mpi.org>
Subscription: <http://www.open-mpi.org/mailman/listinfo.cgi/users> 
http://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post: 
<http://www.open-mpi.org/community/lists/users/2016/04/29000.php> 
http://www.open-mpi.org/community/lists/users/2016/04/29000.php

_______________________________________________
users mailing list
us...@open-mpi.org<mailto:us...@open-mpi.org>
Subscription: <http://www.open-mpi.org/mailman/listinfo.cgi/users> 
http://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post: 
<http://www.open-mpi.org/community/lists/users/2016/04/29003.php> 
http://www.open-mpi.org/community/lists/users/2016/04/29003.php




_______________________________________________
users mailing list
us...@open-mpi.org<mailto:us...@open-mpi.org>
Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post: 
http://www.open-mpi.org/community/lists/users/2016/04/29005.php

Reply via email to