David, our platform files disable dlopen. That is why you are not seeing any component files. coll/ml is built into libmpi.so.
-Nathan On Thu, Aug 13, 2015 at 09:23:09AM -0600, David Shrader wrote: > Hey Jeff, > > I'm actually not able to find coll_ml related files at that location. All I > see are the following files: > > [dshrader@zo-fe1 openmpi]$ ls > /usr/projects/hpcsoft/toss2/zorrillo/openmpi/1.8.8-gcc-4.4/lib/openmpi/ > libompi_dbg_msgq.a libompi_dbg_msgq.la libompi_dbg_msgq.so > > In this particular build, I am using platform files instead of the stripped > down debug builds I was doing before. Could something in the platform files > move or combine with something else the coll_ml related files? > > Thanks, > David > > On 08/13/2015 04:02 AM, Jeff Squyres (jsquyres) wrote: > >Note that this will require you to have fairly recent GNU Autotools > >installed. > > > >Another workaround for avoiding the coll ml module would be to install Open > >MPI as normal, and then rm the following files after installation: > > > > rm $prefix/lib/openmpi/mca_coll_ml* > > > >This will physically remove the coll ml plugin from the Open MPI > >installation, and therefore it won't/can't be used (or interfere with the > >hcoll plugin). > > > > > >>On Aug 13, 2015, at 2:03 AM, Gilles Gouaillardet <gil...@rist.or.jp> wrote: > >> > >>David, > >> > >>i guess you do not want to use the ml coll module at all in openmpi 1.8.8 > >> > >>you can simply do > >>touch ompi/mca/coll/ml/.ompi_ignore > >>./autogen.pl > >>./configure ... > >>make && make install > >> > >>so the ml component is not even built > >> > >>Cheers, > >> > >>Gilles > >> > >>On 8/13/2015 7:30 AM, David Shrader wrote: > >>>I remember seeing those, but forgot about them. I am curious, though, why > >>>using '-mca coll ^ml' wouldn't work for me. > >>> > >>>We'll watch for the next HPCX release. Is there an ETA on when that > >>>release may happen? Thank you for the help! > >>>David > >>> > >>>On 08/12/2015 04:04 PM, Deva wrote: > >>>>David, > >>>> > >>>>This is because of hcoll symbols conflict with ml coll module inside > >>>>OMPI. HCOLL is derived from ml module. This issue is fixed in hcoll > >>>>library and will be available in next HPCX release. > >>>> > >>>>Some earlier discussion on this issue: > >>>>http://www.open-mpi.org/community/lists/users/2015/06/27154.php > >>>>http://www.open-mpi.org/community/lists/devel/2015/06/17562.php > >>>> > >>>>-Devendar > >>>> > >>>>On Wed, Aug 12, 2015 at 2:52 PM, David Shrader <dshra...@lanl.gov> wrote: > >>>>Interesting... the seg faults went away: > >>>> > >>>>[dshrader@zo-fe1 tests]$ export LD_PRELOAD=/usr/lib64/libhcoll.so > >>>>[dshrader@zo-fe1 tests]$ mpirun -n 2 -mca coll ^ml ./a.out > >>>>App launch reported: 1 (out of 1) daemons - 2 (out of 2) procs > >>>>[1439416182.732720] [zo-fe1:14690:0] shm.c:65 MXM WARN Could > >>>>not open the KNEM device file at /dev/knem : No such file or direc > >>>>tory. Won't use knem. > >>>>[1439416182.733640] [zo-fe1:14689:0] shm.c:65 MXM WARN Could > >>>>not open the KNEM device file at /dev/knem : No such file or direc > >>>>tory. Won't use knem. > >>>>0: Running on host zo-fe1.lanl.gov > >>>>0: We have 2 processors > >>>>0: Hello 1! Processor 1 on host zo-fe1.lanl.gov reporting for duty > >>>> > >>>>This implies to me that some other library is being used instead of > >>>>/usr/lib64/libhcoll.so, but I am not sure how that could be... > >>>> > >>>>Thanks, > >>>>David > >>>> > >>>>On 08/12/2015 03:30 PM, Deva wrote: > >>>>>Hi David, > >>>>> > >>>>>I tried same tarball on OFED-1.5.4.1 and I could not reproduce the > >>>>>issue. Can you do one more quick test with seeing LD_PRELOAD to hcoll > >>>>>lib? > >>>>> > >>>>>$LD_PRELOAD=<path/to/hcoll/lib/libhcoll.so> mpirun -n 2 -mca coll ^ml > >>>>>./a.out > >>>>> > >>>>>-Devendar > >>>>> > >>>>>On Wed, Aug 12, 2015 at 12:52 PM, David Shrader <dshra...@lanl.gov> > >>>>>wrote: > >>>>>The admin that rolled the hcoll rpm that we're using (and got it in > >>>>>system space) said that she got it from > >>>>>hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64.tar. > >>>>> > >>>>>Thanks, > >>>>>David > >>>>> > >>>>> > >>>>>On 08/12/2015 10:51 AM, Deva wrote: > >>>>>> From where did you grab this HCOLL lib? MOFED or HPCX? what version? > >>>>>> > >>>>>>On Wed, Aug 12, 2015 at 9:47 AM, David Shrader <dshra...@lanl.gov> > >>>>>>wrote: > >>>>>>Hey Devendar, > >>>>>> > >>>>>>It looks like I still get the error: > >>>>>> > >>>>>>[dshrader@zo-fe1 tests]$ mpirun -n 2 -mca coll ^ml ./a.out > >>>>>>App launch reported: 1 (out of 1) daemons - 2 (out of 2) procs > >>>>>>[1439397957.351764] [zo-fe1:14678:0] shm.c:65 MXM WARN > >>>>>>Could not open the KNEM device file at /dev/knem : No such file or direc > >>>>>>tory. Won't use knem. > >>>>>>[1439397957.352704] [zo-fe1:14677:0] shm.c:65 MXM WARN > >>>>>>Could not open the KNEM device file at /dev/knem : No such file or direc > >>>>>>tory. Won't use knem. > >>>>>>[zo-fe1:14677:0] Caught signal 11 (Segmentation fault) > >>>>>>[zo-fe1:14678:0] Caught signal 11 (Segmentation fault) > >>>>>>==== backtrace ==== > >>>>>>2 0x0000000000056cdc mxm_handle_error() > >>>>>>/scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h > >>>>>>pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641 > >>>>>>3 0x0000000000056e4c mxm_error_signal_handler() > >>>>>>/scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro > >>>>>>ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616 > >>>>>>4 0x00000000000326a0 killpg() ??:0 > >>>>>>5 0x00000000000b82cb base_bcol_basesmuma_setup_library_buffers() ??:0 > >>>>>>6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query() ??:0 > >>>>>>7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery() > >>>>>>coll_ml_module.c:0 > >>>>>>8 0x000000000002fda2 hmca_coll_ml_comm_query() ??:0 > >>>>>>9 0x000000000006ace9 hcoll_create_context() ??:0 > >>>>>>10 0x00000000000f9706 mca_coll_hcoll_comm_query() ??:0 > >>>>>>11 0x00000000000f684e mca_coll_base_comm_select() ??:0 > >>>>>>12 0x0000000000073fc4 ompi_mpi_init() ??:0 > >>>>>>13 0x0000000000092ea0 PMPI_Init() ??:0 > >>>>>>14 0x00000000004009b6 main() ??:0 > >>>>>>15 0x000000000001ed5d __libc_start_main() ??:0 > >>>>>>16 0x00000000004008c9 _start() ??:0 > >>>>>>=================== > >>>>>>==== backtrace ==== > >>>>>>2 0x0000000000056cdc mxm_handle_error() > >>>>>>/scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h > >>>>>>pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641 > >>>>>>3 0x0000000000056e4c mxm_error_signal_handler() > >>>>>>/scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro > >>>>>>ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616 > >>>>>>4 0x00000000000326a0 killpg() ??:0 > >>>>>>5 0x00000000000b82cb base_bcol_basesmuma_setup_library_buffers() ??:0 > >>>>>>6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query() ??:0 > >>>>>>7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery() > >>>>>>coll_ml_module.c:0 > >>>>>>8 0x000000000002fda2 hmca_coll_ml_comm_query() ??:0 > >>>>>>9 0x000000000006ace9 hcoll_create_context() ??:0 > >>>>>>10 0x00000000000f9706 mca_coll_hcoll_comm_query() ??:0 > >>>>>>11 0x00000000000f684e mca_coll_base_comm_select() ??:0 > >>>>>>12 0x0000000000073fc4 ompi_mpi_init() ??:0 > >>>>>>13 0x0000000000092ea0 PMPI_Init() ??:0 > >>>>>>14 0x00000000004009b6 main() ??:0 > >>>>>>15 0x000000000001ed5d __libc_start_main() ??:0 > >>>>>>16 0x00000000004008c9 _start() ??:0 > >>>>>>=================== > >>>>>>-------------------------------------------------------------------------- > >>>>>>mpirun noticed that process rank 1 with PID 14678 on node zo-fe1 exited > >>>>>>on signal 11 (Segmentation fault). > >>>>>>-------------------------------------------------------------------------- > >>>>>> > >>>>>>Thanks, > >>>>>>David > >>>>>> > >>>>>>On 08/12/2015 10:42 AM, Deva wrote: > >>>>>>>Hi David, > >>>>>>> > >>>>>>>This issue is from hcoll library. This could be because of symbol > >>>>>>>conflict with ml module. This is fixed recently in HCOLL. Can you > >>>>>>>try with "-mca coll ^ml" and see if this workaround works in your > >>>>>>>setup? > >>>>>>> > >>>>>>>-Devendar > >>>>>>> > >>>>>>>On Wed, Aug 12, 2015 at 9:30 AM, David Shrader <dshra...@lanl.gov> > >>>>>>>wrote: > >>>>>>>Hello Gilles, > >>>>>>> > >>>>>>>Thank you very much for the patch! It is much more complete than mine. > >>>>>>>Using that patch and re-running autogen.pl, I am able to build 1.8.8 > >>>>>>>with './configure --with-hcoll' without errors. > >>>>>>> > >>>>>>>I do have issues when it comes to running 1.8.8 with hcoll built in, > >>>>>>>however. In my quick sanity test of running a basic parallel hello > >>>>>>>world C program, I get the following: > >>>>>>> > >>>>>>>[dshrader@zo-fe1 tests]$ mpirun -n 2 ./a.out > >>>>>>>App launch reported: 1 (out of 1) daemons - 2 (out of 2) procs > >>>>>>>[1439390789.039197] [zo-fe1:31354:0] shm.c:65 MXM WARN > >>>>>>>Could not open the KNEM device file at /dev/knem : No such file or > >>>>>>>direc > >>>>>>>tory. Won't use knem. > >>>>>>>[1439390789.040265] [zo-fe1:31353:0] shm.c:65 MXM WARN > >>>>>>>Could not open the KNEM device file at /dev/knem : No such file or > >>>>>>>direc > >>>>>>>tory. Won't use knem. > >>>>>>>[zo-fe1:31353:0] Caught signal 11 (Segmentation fault) > >>>>>>>[zo-fe1:31354:0] Caught signal 11 (Segmentation fault) > >>>>>>>==== backtrace ==== > >>>>>>>2 0x0000000000056cdc mxm_handle_error() > >>>>>>>/scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h > >>>>>>>pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641 > >>>>>>>3 0x0000000000056e4c mxm_error_signal_handler() > >>>>>>>/scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro > >>>>>>>ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616 > >>>>>>>4 0x00000000000326a0 killpg() ??:0 > >>>>>>>5 0x00000000000b91eb base_bcol_basesmuma_setup_library_buffers() ??:0 > >>>>>>>6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query() ??:0 > >>>>>>>7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery() > >>>>>>>coll_ml_module.c:0 > >>>>>>>8 0x000000000002fda2 hmca_coll_ml_comm_query() ??:0 > >>>>>>>9 0x000000000006ace9 hcoll_create_context() ??:0 > >>>>>>>10 0x00000000000fa626 mca_coll_hcoll_comm_query() ??:0 > >>>>>>>11 0x00000000000f776e mca_coll_base_comm_select() ??:0 > >>>>>>>12 0x0000000000074ee4 ompi_mpi_init() ??:0 > >>>>>>>13 0x0000000000093dc0 PMPI_Init() ??:0 > >>>>>>>14 0x00000000004009b6 main() ??:0 > >>>>>>>15 0x000000000001ed5d __libc_start_main() ??:0 > >>>>>>>16 0x00000000004008c9 _start() ??:0 > >>>>>>>=================== > >>>>>>>==== backtrace ==== > >>>>>>>2 0x0000000000056cdc mxm_handle_error() > >>>>>>>/scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h > >>>>>>>pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641 > >>>>>>>3 0x0000000000056e4c mxm_error_signal_handler() > >>>>>>>/scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro > >>>>>>>ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616 > >>>>>>>4 0x00000000000326a0 killpg() ??:0 > >>>>>>>5 0x00000000000b91eb base_bcol_basesmuma_setup_library_buffers() ??:0 > >>>>>>>6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query() ??:0 > >>>>>>>7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery() > >>>>>>>coll_ml_module.c:0 > >>>>>>>8 0x000000000002fda2 hmca_coll_ml_comm_query() ??:0 > >>>>>>>9 0x000000000006ace9 hcoll_create_context() ??:0 > >>>>>>>10 0x00000000000fa626 mca_coll_hcoll_comm_query() ??:0 > >>>>>>>11 0x00000000000f776e mca_coll_base_comm_select() ??:0 > >>>>>>>12 0x0000000000074ee4 ompi_mpi_init() ??:0 > >>>>>>>13 0x0000000000093dc0 PMPI_Init() ??:0 > >>>>>>>14 0x00000000004009b6 main() ??:0 > >>>>>>>15 0x000000000001ed5d __libc_start_main() ??:0 > >>>>>>>16 0x00000000004008c9 _start() ??:0 > >>>>>>>=================== > >>>>>>>-------------------------------------------------------------------------- > >>>>>>>mpirun noticed that process rank 0 with PID 31353 on node zo-fe1 > >>>>>>>exited on signal 11 (Segmentation > >>>>>>> fault). > >>>>>>>-------------------------------------------------------------------------- > >>>>>>> > >>>>>>>I do not get this message with only 1 process. > >>>>>>> > >>>>>>>I am using hcoll 3.2.748. Could this be an issue with hcoll itself or > >>>>>>>something with my ompi build? > >>>>>>> > >>>>>>>Thanks, > >>>>>>>David > >>>>>>> > >>>>>>>On 08/12/2015 12:26 AM, Gilles Gouaillardet wrote: > >>>>>>>>Thanks David, > >>>>>>>> > >>>>>>>>i made a PR for the v1.8 branch at > >>>>>>>>https://github.com/open-mpi/ompi-release/pull/492 > >>>>>>>> > >>>>>>>>the patch is attached (it required some back-porting) > >>>>>>>> > >>>>>>>>Cheers, > >>>>>>>> > >>>>>>>>Gilles > >>>>>>>> > >>>>>>>>On 8/12/2015 4:01 AM, David Shrader wrote: > >>>>>>>>>I have cloned Gilles' topic/hcoll_config branch and, after running > >>>>>>>>>autogen.pl, have found that './configure --with-hcoll' does indeed > >>>>>>>>>work now. I used Gilles' branch as I wasn't sure how best to get the > >>>>>>>>>pull request changes in to my own clone of master. It looks like the > >>>>>>>>>proper checks are happening, too: > >>>>>>>>> > >>>>>>>>>--- MCA component coll:hcoll (m4 configuration macro) > >>>>>>>>>checking for MCA component coll:hcoll compile mode... dso > >>>>>>>>>checking --with-hcoll value... simple ok (unspecified) > >>>>>>>>>checking hcoll/api/hcoll_api.h usability... yes > >>>>>>>>>checking hcoll/api/hcoll_api.h presence... yes > >>>>>>>>>checking for hcoll/api/hcoll_api.h... yes > >>>>>>>>>looking for library without search path > >>>>>>>>>checking for library containing hcoll_get_version... -lhcoll > >>>>>>>>>checking if MCA component coll:hcoll can compile... yes > >>>>>>>>> > >>>>>>>>>I haven't checked whether or not Open MPI builds successfully as I > >>>>>>>>>don't have much experience running off of the latest source. For > >>>>>>>>>now, I think I will try to generate a patch to the 1.8.8 configure > >>>>>>>>>script and see if that works as expected. > >>>>>>>>> > >>>>>>>>>Thanks, > >>>>>>>>>David > >>>>>>>>> > >>>>>>>>>On 08/11/2015 06:34 AM, Jeff Squyres (jsquyres) wrote: > >>>>>>>>>>On Aug 11, 2015, at 1:39 AM, Åke Sandgren > >>>>>>>>>><ake.sandg...@hpc2n.umu.se> > >>>>>>>>>> wrote: > >>>>>>>>>> > >>>>>>>>>>>Please fix the hcoll test (and code) to be correct. > >>>>>>>>>>> > >>>>>>>>>>>Any configure test that adds /usr/lib and/or /usr/include to any > >>>>>>>>>>>compile flags is broken. > >>>>>>>>>>> > >>>>>>>>>>+1 > >>>>>>>>>> > >>>>>>>>>>Gilles filed > >>>>>>>>>>https://github.com/open-mpi/ompi/pull/796 > >>>>>>>>>>; I just added some comments to it. > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>-- > >>>>>>>>>David Shrader > >>>>>>>>>HPC-3 High Performance Computer Systems > >>>>>>>>>Los Alamos National Lab > >>>>>>>>>Email: dshrader <at> > >>>>>>>>>lanl.gov > >>>>>>>>> > >>>>>>>>> > >>>>>>>>>_______________________________________________ > >>>>>>>>>users mailing list > >>>>>>>>> > >>>>>>>>>us...@open-mpi.org > >>>>>>>>> > >>>>>>>>>Subscription: > >>>>>>>>>http://www.open-mpi.org/mailman/listinfo.cgi/users > >>>>>>>>> > >>>>>>>>>Link to this post: > >>>>>>>>>http://www.open-mpi.org/community/lists/users/2015/08/27432.php > >>>>>>>> > >>>>>>>> > >>>>>>>>_______________________________________________ > >>>>>>>>users mailing list > >>>>>>>> > >>>>>>>>us...@open-mpi.org > >>>>>>>> > >>>>>>>>Subscription: > >>>>>>>>http://www.open-mpi.org/mailman/listinfo.cgi/users > >>>>>>>> > >>>>>>>>Link to this post: > >>>>>>>>http://www.open-mpi.org/community/lists/users/2015/08/27434.php > >>>>>>>-- > >>>>>>>David Shrader > >>>>>>>HPC-3 High Performance Computer Systems > >>>>>>>Los Alamos National Lab > >>>>>>>Email: dshrader <at> > >>>>>>>lanl.gov > >>>>>>> > >>>>>>>_______________________________________________ > >>>>>>>users mailing list > >>>>>>>us...@open-mpi.org > >>>>>>>Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > >>>>>>>Link to this post: > >>>>>>>http://www.open-mpi.org/community/lists/users/2015/08/27438.php > >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>>-- > >>>>>>> > >>>>>>> > >>>>>>>-Devendar > >>>>>>> > >>>>>>> > >>>>>>>_______________________________________________ > >>>>>>>users mailing list > >>>>>>> > >>>>>>>us...@open-mpi.org > >>>>>>> > >>>>>>>Subscription: > >>>>>>>http://www.open-mpi.org/mailman/listinfo.cgi/users > >>>>>>> > >>>>>>>Link to this post: > >>>>>>>http://www.open-mpi.org/community/lists/users/2015/08/27439.php > >>>>>>-- > >>>>>>David Shrader > >>>>>>HPC-3 High Performance Computer Systems > >>>>>>Los Alamos National Lab > >>>>>>Email: dshrader <at> > >>>>>>lanl.gov > >>>>>> > >>>>>>_______________________________________________ > >>>>>>users mailing list > >>>>>>us...@open-mpi.org > >>>>>>Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > >>>>>>Link to this post: > >>>>>>http://www.open-mpi.org/community/lists/users/2015/08/27440.php > >>>>>> > >>>>>> > >>>>>> > >>>>>>-- > >>>>>> > >>>>>> > >>>>>>-Devendar > >>>>>> > >>>>>> > >>>>>>_______________________________________________ > >>>>>>users mailing list > >>>>>> > >>>>>>us...@open-mpi.org > >>>>>> > >>>>>>Subscription: > >>>>>>http://www.open-mpi.org/mailman/listinfo.cgi/users > >>>>>> > >>>>>>Link to this post: > >>>>>>http://www.open-mpi.org/community/lists/users/2015/08/27441.php > >>>>>-- > >>>>>David Shrader > >>>>>HPC-3 High Performance Computer Systems > >>>>>Los Alamos National Lab > >>>>>Email: dshrader <at> > >>>>>lanl.gov > >>>>> > >>>>>_______________________________________________ > >>>>>users mailing list > >>>>>us...@open-mpi.org > >>>>>Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > >>>>>Link to this post: > >>>>>http://www.open-mpi.org/community/lists/users/2015/08/27445.php > >>>>> > >>>>> > >>>>> > >>>>>-- > >>>>> > >>>>> > >>>>>-Devendar > >>>>-- > >>>>David Shrader > >>>>HPC-3 High Performance Computer Systems > >>>>Los Alamos National Lab > >>>>Email: dshrader <at> > >>>>lanl.gov > >>>> > >>>> > >>>> > >>>>-- > >>>> > >>>> > >>>>-Devendar > >>>-- > >>>David Shrader > >>>HPC-3 High Performance Computer Systems > >>>Los Alamos National Lab > >>>Email: dshrader <at> lanl.gov > >>> > >>> > >>> > >>>_______________________________________________ > >>>users mailing list > >>> > >>>us...@open-mpi.org > >>> > >>>Subscription: > >>>http://www.open-mpi.org/mailman/listinfo.cgi/users > >>> > >>>Link to this post: > >>>http://www.open-mpi.org/community/lists/users/2015/08/27448.php > >>_______________________________________________ > >>users mailing list > >>us...@open-mpi.org > >>Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > >>Link to this post: > >>http://www.open-mpi.org/community/lists/users/2015/08/27453.php > > > > -- > David Shrader > HPC-3 High Performance Computer Systems > Los Alamos National Lab > Email: dshrader <at> lanl.gov > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/08/27457.php
pgp1EIJk5ce7c.pgp
Description: PGP signature