valgrind isn’t going to help here - there are multiple reasons why your 
application could be segfaulting. Take a look at the core file with gdb and 
find out where it is failing.

> On Apr 22, 2016, at 10:20 PM, Elio Physics <elio-phys...@live.com> wrote:
> 
> One more thing i forgot to mention in my previous e-mail. In the output file 
> I get the following message:
> 
> 
> 2 total processes killed (some possibly by mpirun during cleanup)
> 
> Thanks
> 
> 
> 
> From: users <users-boun...@open-mpi.org <mailto:users-boun...@open-mpi.org>> 
> on behalf of Elio Physics <elio-phys...@live.com 
> <mailto:elio-phys...@live.com>>
> Sent: Saturday, April 23, 2016 3:07 AM
> To: Open MPI Users
> Subject: Re: [OMPI users] MPIRUN SEGMENTATION FAULT
>  
> I have used valgrind and this is what i got:
> 
> valgrind mpirun ~/Elie/SPRKKR/bin/kkrscf6.3MPI Fe_SCF.inp > 
> scf-51551.jlborges.fisica.ufmg.br.out
> ==8135== Memcheck, a memory error detector
> ==8135== Copyright (C) 2002-2012, and GNU GPL'd, by Julian Seward et al.
> ==8135== Using Valgrind-3.8.1 and LibVEX; rerun with -h for copyright info
> ==8135== Command: mpirun /home/emoujaes/Elie/SPRKKR/bin/kkrscf6.3MPI 
> Fe_SCF.inp
> ==8135==
> --------------------------------------------------------------------------
> mpirun noticed that process rank 0 with PID 8147 on node 
> jlborges.fisica.ufmg.br <http://jlborges.fisica.ufmg.br/> exited on signal 11 
> (Segmentation fault).
> --------------------------------------------------------------------------
> ==8135==
> ==8135== HEAP SUMMARY:
> ==8135==     in use at exit: 485,683 bytes in 1,899 blocks
> ==8135==   total heap usage: 7,723 allocs, 5,824 frees, 12,185,660 bytes 
> allocated
> ==8135==
> ==8135== LEAK SUMMARY:
> ==8135==    definitely lost: 34,944 bytes in 34 blocks
> ==8135==    indirectly lost: 26,613 bytes in 58 blocks
> ==8135==      possibly lost: 0 bytes in 0 blocks
> ==8135==    still reachable: 424,126 bytes in 1,807 blocks
> ==8135==         suppressed: 0 bytes in 0 blocks
> ==8135== Rerun with --leak-check=full to see details of leaked memory
> ==8135==
> ==8135== For counts of detected and suppressed errors, rerun with: -v
> ==8135== ERROR SUMMARY: 0 errors from 0 contexts (suppressed: 6 from 6)
> 
> What does that supposed to mean?
> 
> Regards
> From: users <users-boun...@open-mpi.org <mailto:users-boun...@open-mpi.org>> 
> on behalf of Ralph Castain <r...@open-mpi.org <mailto:r...@open-mpi.org>>
> Sent: Saturday, April 23, 2016 1:36:50 AM
> To: Open MPI Users
> Subject: Re: [OMPI users] MPIRUN SEGMENTATION FAULT
>  
> All I can say is that your program segfault’d during execution - you might 
> want to look at the core file using a debugger like gdb to see why it failed.
> 
> 
>> On Apr 22, 2016, at 8:32 PM, Elio Physics <elio-phys...@live.com 
>> <mailto:elio-phys...@live.com>> wrote:
>> 
>> Dear all,
>> 
>> I have successfully compiled a code where the executable have been produced. 
>> However when I started using the executable with mpirun, the code stopped 
>> with the following error:
>> 
>> "mpirun noticed that process rank 0 with PID 570 on node compute-1-9.local 
>> exited on signal 11 (Segmentation fault)."
>> 
>> What is that error due to? and How can i solve it?
>> 
>> I will post the make.inc compilation file:
>> 
>> BUILD_TYPE ?=
>> #BUILD_TYPE := debug
>> 
>> VERSION = 6.3
>> 
>> ifeq ($(BUILD_TYPE), debug)
>>  VERSION := $(VERSION)$(BUILD_TYPE)
>> endif
>> 
>> BIN =~/Elie/SPRKKR/bin
>> #BIN=~/bin
>> #BIN=/tmp/$(USER)
>> 
>> LIB =  -L/opt/intel/mkl/lib/intel64/libmkl_blas95_ilp64 
>> -L/opt/intel/mkl/lib/intel64/libmkl_lapack95_ilp64 
>> -L/opt/intel/mkl/lib/intel64 -lmkl_scalapack_ilp64 -lmkl_intel_ilp64 
>> -lmkl_core -lmkl_sequential -lmkl_blacs_i
>> ntelmpi_ilp64 -lpthread -lm -ldl
>> 
>> #Include mpif.h
>> INCLUDE = -I/opt/intel/mkl/include/intel64/ilp64  
>> -I/opt/intel/mkl/lib/include
>> 
>> #FFLAGS
>> FFLAGS = -O2
>> 
>> FC   = mpif90 -c $(FFLAGS) $(INCLUDE)
>> LINK = mpif90   $(FFLAGS) $(INCLUDE)
>> 
>> MPI=MPI
>> 
>> Thanks in advance
>> 
>> Elio
>> University of Rondonia, brazil
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/users/2016/04/29000.php 
>> <http://www.open-mpi.org/community/lists/users/2016/04/29000.php>
> _______________________________________________
> users mailing list
> us...@open-mpi.org <mailto:us...@open-mpi.org>
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2016/04/29003.php 
> <http://www.open-mpi.org/community/lists/users/2016/04/29003.php>

Reply via email to