valgrind isn’t going to help here - there are multiple reasons why your application could be segfaulting. Take a look at the core file with gdb and find out where it is failing.
> On Apr 22, 2016, at 10:20 PM, Elio Physics <elio-phys...@live.com> wrote: > > One more thing i forgot to mention in my previous e-mail. In the output file > I get the following message: > > > 2 total processes killed (some possibly by mpirun during cleanup) > > Thanks > > > > From: users <users-boun...@open-mpi.org <mailto:users-boun...@open-mpi.org>> > on behalf of Elio Physics <elio-phys...@live.com > <mailto:elio-phys...@live.com>> > Sent: Saturday, April 23, 2016 3:07 AM > To: Open MPI Users > Subject: Re: [OMPI users] MPIRUN SEGMENTATION FAULT > > I have used valgrind and this is what i got: > > valgrind mpirun ~/Elie/SPRKKR/bin/kkrscf6.3MPI Fe_SCF.inp > > scf-51551.jlborges.fisica.ufmg.br.out > ==8135== Memcheck, a memory error detector > ==8135== Copyright (C) 2002-2012, and GNU GPL'd, by Julian Seward et al. > ==8135== Using Valgrind-3.8.1 and LibVEX; rerun with -h for copyright info > ==8135== Command: mpirun /home/emoujaes/Elie/SPRKKR/bin/kkrscf6.3MPI > Fe_SCF.inp > ==8135== > -------------------------------------------------------------------------- > mpirun noticed that process rank 0 with PID 8147 on node > jlborges.fisica.ufmg.br <http://jlborges.fisica.ufmg.br/> exited on signal 11 > (Segmentation fault). > -------------------------------------------------------------------------- > ==8135== > ==8135== HEAP SUMMARY: > ==8135== in use at exit: 485,683 bytes in 1,899 blocks > ==8135== total heap usage: 7,723 allocs, 5,824 frees, 12,185,660 bytes > allocated > ==8135== > ==8135== LEAK SUMMARY: > ==8135== definitely lost: 34,944 bytes in 34 blocks > ==8135== indirectly lost: 26,613 bytes in 58 blocks > ==8135== possibly lost: 0 bytes in 0 blocks > ==8135== still reachable: 424,126 bytes in 1,807 blocks > ==8135== suppressed: 0 bytes in 0 blocks > ==8135== Rerun with --leak-check=full to see details of leaked memory > ==8135== > ==8135== For counts of detected and suppressed errors, rerun with: -v > ==8135== ERROR SUMMARY: 0 errors from 0 contexts (suppressed: 6 from 6) > > What does that supposed to mean? > > Regards > From: users <users-boun...@open-mpi.org <mailto:users-boun...@open-mpi.org>> > on behalf of Ralph Castain <r...@open-mpi.org <mailto:r...@open-mpi.org>> > Sent: Saturday, April 23, 2016 1:36:50 AM > To: Open MPI Users > Subject: Re: [OMPI users] MPIRUN SEGMENTATION FAULT > > All I can say is that your program segfault’d during execution - you might > want to look at the core file using a debugger like gdb to see why it failed. > > >> On Apr 22, 2016, at 8:32 PM, Elio Physics <elio-phys...@live.com >> <mailto:elio-phys...@live.com>> wrote: >> >> Dear all, >> >> I have successfully compiled a code where the executable have been produced. >> However when I started using the executable with mpirun, the code stopped >> with the following error: >> >> "mpirun noticed that process rank 0 with PID 570 on node compute-1-9.local >> exited on signal 11 (Segmentation fault)." >> >> What is that error due to? and How can i solve it? >> >> I will post the make.inc compilation file: >> >> BUILD_TYPE ?= >> #BUILD_TYPE := debug >> >> VERSION = 6.3 >> >> ifeq ($(BUILD_TYPE), debug) >> VERSION := $(VERSION)$(BUILD_TYPE) >> endif >> >> BIN =~/Elie/SPRKKR/bin >> #BIN=~/bin >> #BIN=/tmp/$(USER) >> >> LIB = -L/opt/intel/mkl/lib/intel64/libmkl_blas95_ilp64 >> -L/opt/intel/mkl/lib/intel64/libmkl_lapack95_ilp64 >> -L/opt/intel/mkl/lib/intel64 -lmkl_scalapack_ilp64 -lmkl_intel_ilp64 >> -lmkl_core -lmkl_sequential -lmkl_blacs_i >> ntelmpi_ilp64 -lpthread -lm -ldl >> >> #Include mpif.h >> INCLUDE = -I/opt/intel/mkl/include/intel64/ilp64 >> -I/opt/intel/mkl/lib/include >> >> #FFLAGS >> FFLAGS = -O2 >> >> FC = mpif90 -c $(FFLAGS) $(INCLUDE) >> LINK = mpif90 $(FFLAGS) $(INCLUDE) >> >> MPI=MPI >> >> Thanks in advance >> >> Elio >> University of Rondonia, brazil >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org <mailto:us...@open-mpi.org> >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >> <http://www.open-mpi.org/mailman/listinfo.cgi/users> >> Link to this post: >> http://www.open-mpi.org/community/lists/users/2016/04/29000.php >> <http://www.open-mpi.org/community/lists/users/2016/04/29000.php> > _______________________________________________ > users mailing list > us...@open-mpi.org <mailto:us...@open-mpi.org> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > <http://www.open-mpi.org/mailman/listinfo.cgi/users> > Link to this post: > http://www.open-mpi.org/community/lists/users/2016/04/29003.php > <http://www.open-mpi.org/community/lists/users/2016/04/29003.php>