George, It turns out I didn't have libunwind either, but didn't notice since mpiP compiled/linked without it (OK, so I should have checked the config log). However, once I got it it wouldn't compile on my RHEL system.
So, following this thread: http://www.mail-archive.com/libunwind-devel@nongnu.org/msg00067.html I had to download an alpha version of libunwind: http://download.savannah.nongnu.org/releases/libunwind/libunwind-snap-070224 .tar.gz ... And build it with: CFLAGS=-fPIC ./configure make CFLAGS=-fPIC LDFLAGS=-fPIC shared make CFLAGS=-fPIC LDFLAGS=-fPIC install After that, everything went as you described. The "strange readings" in the output did list the Parent_Funct's though: --------------------------------------------------------------------------- @--- Callsites: 5 --------------------------------------------------------- --------------------------------------------------------------------------- ID Lev File/Address Line Parent_Funct MPI_Call 1 0 0x000000000041341d RecvData Recv 2 0 0x00000000004133c7 SendData Send 3 0 0x00000000004134b9 SendRepeat Send 4 0 0x0000000000413315 Sync Barrier 5 0 0x00000000004134ef RecvRepeat Recv Thanks for the help! Todd On 3/29/07 5:48 PM, "George Bosilca" <bosi...@cs.utk.edu> wrote: > I used it on a IA64 platform, so I supposed x86_64 is supported, but > I never use it on an AMD 64. On the mpiP webpage they claim they > support the Cray XT3, which as far as I know are based on AMD Opteron > 64 bits. So, there is at least a spark of hope in the dark ... > > I decide to give it a try on my x86_64 AMD box (Debian based system). > First problem, my box didn't have the libunwind. Not a big deal, it's > freely available on HP website (http://www.hpl.hp.com/research/linux/ > libunwind/download.php4). Few minutes later, the libunwind was > installed in /lib64. Now, time to focus on mpiP ... For some obscure > reason the configure script was unable to detect my g77 compiler > (whatever!!!) nor the installation of libunwind. Moreover, it keep > trying to use the clock_gettime call. Fortunately (which make me > think I'm not the only one having trouble with this), mpiP provide > configure options for all these. The final configure line was: ./ > configure --prefix=/opt/ --without-f77 --with-wtime --with-include=-I/ > include --with-lib=-L/lib64. Then a quick "make shared" followed by > "make install", complete the work. So, at least mpiP can compile on a > x86_64 box. > > Now, I modify the makefile of NetPIPE, and add the "-lmpiP -lunwind", > compile NetPIPE and run it. The mpiP headers showed up, the > application run to completion and my human readable output was there. > > @ mpiP > @ Command : ./NPmpi > @ Version : 3.1.0 > @ MPIP Build date : Mar 29 2007, 13:35:47 > @ Start time : 2007 03 29 13:43:40 > @ Stop time : 2007 03 29 13:44:42 > @ Timer Used : PMPI_Wtime > @ MPIP env var : [null] > @ Collector Rank : 0 > @ Collector PID : 22838 > @ Final Output Dir : . > @ Report generation : Single collector task > @ MPI Task Assignment : 0 dancer > @ MPI Task Assignment : 1 dancer > > However, I got some strange reading inside the output. > ------------------------------------------------------------------------ > --- > @--- Callsites: 5 > --------------------------------------------------------- > ------------------------------------------------------------------------ > --- > ID Lev File/Address Line Parent_Funct MPI_Call > 1 0 0x0000000000402ffb [unknown] Barrier > 2 0 0x0000000000403103 [unknown] Recv > 3 0 0x00000000004030ad [unknown] Send > 4 0 0x000000000040319f [unknown] Send > 5 0 0x00000000004031d5 [unknown] Recv > > I didn't dig further to see why. But, this prove that for at least a > basic usage (general statistics gathering) mpiP works on x86_64 > platforms. > > Have fun, > george. > > On Mar 29, 2007, at 11:32 AM, Heywood, Todd wrote: > >> George, >> >> Any other simple, small, text-based (!) suggestions? mpiP seg >> faults on >> x86_64, and indeed its web page doesn't list x86_64 Linux as a >> supported >> platform. >> >> Todd >> >> >> On 3/28/07 10:39 AM, "George Bosilca" <bosi...@cs.utk.edu> wrote: >> >>> Stephen, >>> >>> There are a huge number of MPI profiling tools out there. My >>> preference will be something small, fast and where the output is in >>> human readable text format (and not fancy graphics). The tools I'm >>> talking about is called mpiP (http://mpip.sourceforge.net/). It's not >>> Open MPI specific, but it's really simple to use. >>> >>> george. >>> >>> On Mar 28, 2007, at 10:10 AM, stephen mulcahy wrote: >>> >>>> Hi, >>>> >>>> What is the best way of getting statistics on the size of MPI >>>> messages >>>> being sent/received by my OpenMPI-using application? I'm guessing >>>> MPE is >>>> one route but is there anything built into OpenMPI that will give me >>>> this specific statistic? >>>> >>>> Thanks, >>>> >>>> -stephen >>>> >>>> -- >>>> Stephen Mulcahy, Applepie Solutions Ltd, Innovation in Business >>>> Center, >>>> GMIT, Dublin Rd, Galway, Ireland. http://www.aplpi.com >>>> _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > "Half of what I say is meaningless; but I say it so that the other > half may reach you" > Kahlil Gibran > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users