Just for clarity: does the BIOS on the LGA2011 system have HT enabled?
> On Apr 8, 2015, at 10:55 AM, Lane, William <william.l...@cshs.org> wrote: > > Ralph, > > I added one of the newer LGA2011 nodes to my hostfile and > re-ran the benchmark successfully and saw some strange results WRT the > binding directives. Why are hyperthreading cores being used > on the LGA2011 system but not any of other systems which > are mostly hyperthreaded Westmeres)? Isn't the --use-hwthread-cpus > switch supposed to prevent OpenMPI from using hyperthreaded > cores? > > OpenMPI LAPACK invocation: > > $MPI_DIR/bin/mpirun -np $NSLOTS --report-bindings --hostfile hostfile-single > --mca btl_tcp_if_include eth0 --hetero-nodes --use-hwthread-cpus --prefix > $MPI_DIR $BENCH_DIR/$APP_DIR/$APP_BIN > > Where NSLOTS=72 > > hostfile: > csclprd3-6-1 slots=4 max-slots=4 > csclprd3-6-5 slots=4 max-slots=4 > csclprd3-0-0 slots=12 max-slots=24 > csclprd3-0-1 slots=6 max-slots=12 > csclprd3-0-2 slots=6 max-slots=12 > csclprd3-0-3 slots=6 max-slots=12 > csclprd3-0-4 slots=6 max-slots=12 > csclprd3-0-5 slots=6 max-slots=12 > csclprd3-0-6 slots=6 max-slots=12 > #total number of successfully tested non-hyperthreaded computes slots at this > point is 56 > csclprd3-0-7 slots=16 max-slots=32 > > LGA1366 Westmere w/two Intel Xeon X5675 6-core/12-hyperthread CPU's > > [csclprd3-0-0:11848] MCW rank 11 bound to socket 1[core 7[hwt 0]]: > [./././././.][./B/./././.] > [csclprd3-0-0:11848] MCW rank 12 bound to socket 0[core 2[hwt 0]]: > [././B/././.][./././././.] > [csclprd3-0-0:11848] MCW rank 13 bound to socket 1[core 8[hwt 0]]: > [./././././.][././B/././.] > [csclprd3-0-0:11848] MCW rank 14 bound to socket 0[core 3[hwt 0]]: > [./././B/./.][./././././.] > [csclprd3-0-0:11848] MCW rank 15 bound to socket 1[core 9[hwt 0]]: > [./././././.][./././B/./.] > [csclprd3-0-0:11848] MCW rank 16 bound to socket 0[core 4[hwt 0]]: > [././././B/.][./././././.] > [csclprd3-0-0:11848] MCW rank 17 bound to socket 1[core 10[hwt 0]]: > [./././././.][././././B/.] > [csclprd3-0-0:11848] MCW rank 18 bound to socket 0[core 5[hwt 0]]: > [./././././B][./././././.] > [csclprd3-0-0:11848] MCW rank 19 bound to socket 1[core 11[hwt 0]]: > [./././././.][./././././B] > [csclprd3-0-0:11848] MCW rank 8 bound to socket 0[core 0[hwt 0]]: > [B/././././.][./././././.] > [csclprd3-0-0:11848] MCW rank 9 bound to socket 1[core 6[hwt 0]]: > [./././././.][B/././././.] > [csclprd3-0-0:11848] MCW rank 10 bound to socket 0[core 1[hwt 0]]: > [./B/./././.][./././././.] > > but for the LGA2011 system w/two 8-core/16-hyperthread CPU's > > [csclprd3-0-7:30876] MCW rank 60 bound to socket 0[core 2[hwt 0-1]]: > [../../BB/../../../../..][../../../../../../../..] > [csclprd3-0-7:30876] MCW rank 61 bound to socket 1[core 10[hwt 0-1]]: > [../../../../../../../..][../../BB/../../../../..] > [csclprd3-0-7:30876] MCW rank 62 bound to socket 0[core 3[hwt 0-1]]: > [../../../BB/../../../..][../../../../../../../..] > [csclprd3-0-7:30876] MCW rank 63 bound to socket 1[core 11[hwt 0-1]]: > [../../../../../../../..][../../../BB/../../../..] > [csclprd3-0-7:30876] MCW rank 64 bound to socket 0[core 4[hwt 0-1]]: > [../../../../BB/../../..][../../../../../../../..] > [csclprd3-0-7:30876] MCW rank 65 bound to socket 1[core 12[hwt 0-1]]: > [../../../../../../../..][../../../../BB/../../..] > [csclprd3-0-7:30876] MCW rank 66 bound to socket 0[core 5[hwt 0-1]]: > [../../../../../BB/../..][../../../../../../../..] > [csclprd3-0-7:30876] MCW rank 67 bound to socket 1[core 13[hwt 0-1]]: > [../../../../../../../..][../../../../../BB/../..] > [csclprd3-0-7:30876] MCW rank 68 bound to socket 0[core 6[hwt 0-1]]: > [../../../../../../BB/..][../../../../../../../..] > [csclprd3-0-7:30876] MCW rank 69 bound to socket 1[core 14[hwt 0-1]]: > [../../../../../../../..][../../../../../../BB/..] > [csclprd3-0-7:30876] MCW rank 70 bound to socket 0[core 7[hwt 0-1]]: > [../../../../../../../BB][../../../../../../../..] > [csclprd3-0-7:30876] MCW rank 71 bound to socket 1[core 15[hwt 0-1]]: > [../../../../../../../..][../../../../../../../BB] > [csclprd3-0-7:30876] MCW rank 56 bound to socket 0[core 0[hwt 0-1]]: > [BB/../../../../../../..][../../../../../../../..] > [csclprd3-0-7:30876] MCW rank 57 bound to socket 1[core 8[hwt 0-1]]: > [../../../../../../../..][BB/../../../../../../..] > [csclprd3-0-7:30876] MCW rank 58 bound to socket 0[core 1[hwt 0-1]]: > [../BB/../../../../../..][../../../../../../../..] > [csclprd3-0-7:30876] MCW rank 59 bound to socket 1[core 9[hwt 0-1]]: > [../../../../../../../..][../BB/../../../../../..] > > > > > From: users [users-boun...@open-mpi.org] on behalf of Ralph Castain > [r...@open-mpi.org] > Sent: Wednesday, April 08, 2015 10:26 AM > To: Open MPI Users > Subject: Re: [OMPI users] OpenMPI 1.8.2 problems on CentOS 6.3 > > >> On Apr 8, 2015, at 9:29 AM, Lane, William <william.l...@cshs.org >> <mailto:william.l...@cshs.org>> wrote: >> >> Ralph, >> >> Thanks for YOUR help, I never >> would've managed to get the LAPACK >> benchmark running on more than one >> node in our cluster without your help. >> >> Ralph, is hyperthreading more of a curse >> than an advantage for HPC applications? > > Wow, you’ll get a lot of argument over that issue! From what I can see, it is > very application dependent. Some apps appear to benefit, while others can > even suffer from it. > > I think we should support a mix of nodes in this usage, so I’ll try to come > up with a way to do so. > >> >> I'm going to go through all the OpenMPI >> articles on hyperthreading and NUMA to >> see if that will shed any light on these >> issues. >> >> -Bill L. >> >> >> From: users [users-boun...@open-mpi.org <mailto:users-boun...@open-mpi.org>] >> on behalf of Ralph Castain [r...@open-mpi.org <mailto:r...@open-mpi.org>] >> Sent: Tuesday, April 07, 2015 7:32 PM >> To: Open MPI Users >> Subject: Re: [OMPI users] OpenMPI 1.8.2 problems on CentOS 6.3 >> >> I’m not sure our man pages are good enough to answer your question, but here >> is the URL >> >> http://www.open-mpi.org/doc/v1.8/ <http://www.open-mpi.org/doc/v1.8/> >> >> I’m a tad tied up right now, but I’ll try to address this prior to 1.8.5 >> release. Thanks for all that debug effort! Helps a bunch. >> >>> On Apr 7, 2015, at 1:17 PM, Lane, William <william.l...@cshs.org >>> <mailto:william.l...@cshs.org>> wrote: >>> >>> Ralph, >>> >>> I've finally had some luck using the following: >>> $MPI_DIR/bin/mpirun -np $NSLOTS --report-bindings --hostfile >>> hostfile-single --mca btl_tcp_if_include eth0 --hetero-nodes >>> --use-hwthread-cpus --prefix $MPI_DIR $BENCH_DIR/$APP_DIR/$APP_BIN >>> >>> Where $NSLOTS was 56 and my hostfile hostfile-single is: >>> >>> csclprd3-0-0 slots=12 max-slots=24 >>> csclprd3-0-1 slots=6 max-slots=12 >>> csclprd3-0-2 slots=6 max-slots=12 >>> csclprd3-0-3 slots=6 max-slots=12 >>> csclprd3-0-4 slots=6 max-slots=12 >>> csclprd3-0-5 slots=6 max-slots=12 >>> csclprd3-0-6 slots=6 max-slots=12 >>> csclprd3-6-1 slots=4 max-slots=4 >>> csclprd3-6-5 slots=4 max-slots=4 >>> >>> The max-slots differs from slots on some nodes >>> because I include the hyperthreaded cores in >>> the max-slots, the last two nodes have CPU's that >>> don't support hyperthreading at all. >>> >>> Does --use-hwthread-cpus prevent slots from >>> being assigned to hyperthreading cores? >>> >>> For some reason the manpage for OpenMPI 1.8.2 >>> isn't installed on our CentOS 6.3 systems is there a >>> URL I can I find a copy of the manpages for OpenMPI 1.8.2? >>> >>> Thanks for your help, >>> >>> -Bill Lane >>> >>> From: users [users-boun...@open-mpi.org >>> <mailto:users-boun...@open-mpi.org>] on behalf of Ralph Castain >>> [r...@open-mpi.org <mailto:r...@open-mpi.org>] >>> Sent: Monday, April 06, 2015 1:39 PM >>> To: Open MPI Users >>> Subject: Re: [OMPI users] OpenMPI 1.8.2 problems on CentOS 6.3 >>> >>> Hmmm…well, that shouldn’t be the issue. To check, try running it with >>> “bind-to none”. If you can get a backtrace telling us where it is crashing, >>> that would also help. >>> >>> >>>> On Apr 6, 2015, at 12:24 PM, Lane, William <william.l...@cshs.org >>>> <mailto:william.l...@cshs.org>> wrote: >>>> >>>> Ralph, >>>> >>>> For the following two different commandline invocations of the LAPACK >>>> benchmark >>>> >>>> $MPI_DIR/bin/mpirun -np $NSLOTS --report-bindings --hostfile >>>> hostfile-no_slots --mca btl_tcp_if_include eth0 --hetero-nodes >>>> --use-hwthread-cpus --bind-to hwthread --prefix $MPI_DIR >>>> $BENCH_DIR/$APP_DIR/$APP_BIN >>>> >>>> $MPI_DIR/bin/mpirun -np $NSLOTS --report-bindings --hostfile >>>> hostfile-no_slots --mca btl_tcp_if_include eth0 --hetero-nodes >>>> --bind-to-core --prefix $MPI_DIR $BENCH_DIR/$APP_DIR/$APP_BIN >>>> >>>> I'm receiving the same kinds of OpenMPI error messages (but for different >>>> nodes in the ring): >>>> >>>> [csclprd3-0-16:25940] *** Process received signal *** >>>> [csclprd3-0-16:25940] Signal: Bus error (7) >>>> [csclprd3-0-16:25940] Signal code: Non-existant physical address >>>> (2) >>>> [csclprd3-0-16:25940] Failing at address: 0x7f8b1b5a2600 >>>> >>>> >>>> -------------------------------------------------------------------------- >>>> mpirun noticed that process rank 82 with PID 25936 on node >>>> csclprd3-0-16 exited on signal 7 (Bus error). >>>> >>>> -------------------------------------------------------------------------- >>>> 16 total processes killed (some possibly by mpirun during cleanup) >>>> >>>> It seems to occur on systems that have more than one, physical CPU >>>> installed. Could >>>> this be due to a lack of the correct NUMA libraries being installed? >>>> >>>> -Bill L. >>>> >>>> From: users [users-boun...@open-mpi.org >>>> <mailto:users-boun...@open-mpi.org>] on behalf of Ralph Castain >>>> [r...@open-mpi.org <mailto:r...@open-mpi.org>] >>>> Sent: Sunday, April 05, 2015 6:09 PM >>>> To: Open MPI Users >>>> Subject: Re: [OMPI users] OpenMPI 1.8.2 problems on CentOS 6.3 >>>> >>>> >>>>> On Apr 5, 2015, at 5:58 PM, Lane, William <william.l...@cshs.org >>>>> <mailto:william.l...@cshs.org>> wrote: >>>>> >>>>> I think some of the Intel Blade systems in the cluster are >>>>> dual core, but don't support hyperthreading. Maybe it >>>>> would be better to exclude hyperthreading altogether >>>>> from submitted OpenMPI jobs? >>>> >>>> Yes - or you can add "--hetero-nodes -use-hwthread-cpus --bind-to >>>> hwthread" to the cmd line. This tells mpirun that the nodes aren't all the >>>> same, and so it has to look at each node's topology instead of taking the >>>> first node as the template for everything. The second tells it to use the >>>> HTs as independent cpus where they are supported. >>>> >>>> I'm not entirely sure the suggestion will work - if we hit a place where >>>> HT isn't supported, we may balk at being asked to bind to HTs. I can >>>> probably make a change that supports this kind of hetero arrangement >>>> (perhaps something like bind-to pu) - might make it into 1.8.5 (we are >>>> just starting the release process on it now). >>>> >>>>> >>>>> OpenMPI doesn't crash, but it doesn't run the LAPACK >>>>> benchmark either. >>>>> >>>>> Thanks again Ralph. >>>>> >>>>> Bill L. >>>>> >>>>> From: users [users-boun...@open-mpi.org >>>>> <mailto:users-boun...@open-mpi.org>] on behalf of Ralph Castain >>>>> [r...@open-mpi.org <mailto:r...@open-mpi.org>] >>>>> Sent: Wednesday, April 01, 2015 8:40 AM >>>>> To: Open MPI Users >>>>> Subject: Re: [OMPI users] OpenMPI 1.8.2 problems on CentOS 6.3 >>>>> >>>>> Bingo - you said the magic word. This is a terminology issue. When we say >>>>> "core", we mean the old definition of "core", not "hyperthreads". If you >>>>> want to use HTs as your base processing unit and bind to them, then you >>>>> need to specify --bind-to hwthread. That warning should then go away. >>>>> >>>>> We don't require a swap region be mounted - I didn't see anything in your >>>>> original message indicating that OMPI had actually crashed, but just >>>>> wasn't launching due to the above issue. Were you actually seeing crashes >>>>> as well? >>>>> >>>>> >>>>> On Wed, Apr 1, 2015 at 8:31 AM, Lane, William <william.l...@cshs.org >>>>> <mailto:william.l...@cshs.org>> wrote: >>>>> Ralph, >>>>> >>>>> Here's the associated hostfile: >>>>> >>>>> #openMPI hostfile for csclprd3 >>>>> #max slots prevents oversubscribing csclprd3-0-9 >>>>> csclprd3-0-0 slots=12 max-slots=12 >>>>> csclprd3-0-1 slots=6 max-slots=6 >>>>> csclprd3-0-2 slots=6 max-slots=6 >>>>> csclprd3-0-3 slots=6 max-slots=6 >>>>> csclprd3-0-4 slots=6 max-slots=6 >>>>> csclprd3-0-5 slots=6 max-slots=6 >>>>> csclprd3-0-6 slots=6 max-slots=6 >>>>> csclprd3-0-7 slots=32 max-slots=32 >>>>> csclprd3-0-8 slots=32 max-slots=32 >>>>> csclprd3-0-9 slots=32 max-slots=32 >>>>> csclprd3-0-10 slots=32 max-slots=32 >>>>> csclprd3-0-11 slots=32 max-slots=32 >>>>> csclprd3-0-12 slots=12 max-slots=12 >>>>> csclprd3-0-13 slots=24 max-slots=24 >>>>> csclprd3-0-14 slots=16 max-slots=16 >>>>> csclprd3-0-15 slots=16 max-slots=16 >>>>> csclprd3-0-16 slots=24 max-slots=24 >>>>> csclprd3-0-17 slots=24 max-slots=24 >>>>> csclprd3-6-1 slots=4 max-slots=4 >>>>> csclprd3-6-5 slots=4 max-slots=4 >>>>> >>>>> The number of slots also includes hyperthreading >>>>> cores. >>>>> >>>>> One more question, would not having defined swap >>>>> partitions on all the nodes in the ring cause OpenMPI >>>>> to crash? Because no swap partitions are defined >>>>> for any of the above systems. >>>>> >>>>> -Bill L. >>>>> >>>>> >>>>> From: users [users-boun...@open-mpi.org >>>>> <mailto:users-boun...@open-mpi.org>] on behalf of Ralph Castain >>>>> [r...@open-mpi.org <mailto:r...@open-mpi.org>] >>>>> Sent: Wednesday, April 01, 2015 5:04 AM >>>>> To: Open MPI Users >>>>> Subject: Re: [OMPI users] OpenMPI 1.8.2 problems on CentOS 6.3 >>>>> >>>>> The warning about binding to memory is due to not having numactl-devel >>>>> installed on the system. The job would still run, but we are warning you >>>>> that we cannot bind memory to the same domain as the core where we bind >>>>> the process. Can cause poor performance, but not fatal. I forget the name >>>>> of the param, but you can tell us to "shut up" :-) >>>>> >>>>> The other warning/error indicates that we aren't seeing enough cores on >>>>> the allocation you gave us via the hostile to support one proc/core - >>>>> i.e., we didn't at least 128 cores in the sum of the nodes you told us >>>>> about. I take it you were expecting that there were that many or more? >>>>> >>>>> Ralph >>>>> >>>>> >>>>> On Wed, Apr 1, 2015 at 12:54 AM, Lane, William <william.l...@cshs.org >>>>> <mailto:william.l...@cshs.org>> wrote: >>>>> I'm having problems running OpenMPI jobs >>>>> (using a hostfile) on an HPC cluster running >>>>> ROCKS on CentOS 6.3. I'm running OpenMPI >>>>> outside of Sun Grid Engine (i.e. it is not submitted >>>>> as a job to SGE). The program being run is a LAPACK >>>>> benchmark. The commandline parameter I'm >>>>> using to run the jobs is: >>>>> >>>>> $MPI_DIR/bin/mpirun -np $NSLOTS -bind-to-core -report-bindings --hostfile >>>>> hostfile --mca btl_tcp_if_include eth0 --prefix $MPI_DIR >>>>> $BENCH_DIR/$APP_DIR/$APP_BIN >>>>> >>>>> Where MPI_DIR=/hpc/apps/mpi/openmpi/1.8.2/ >>>>> NSLOTS=128 >>>>> >>>>> I'm getting errors of the form and OpenMPI never runs the LAPACK >>>>> benchmark: >>>>> >>>>> >>>>> -------------------------------------------------------------------------- >>>>> WARNING: a request was made to bind a process. While the system >>>>> supports binding the process itself, at least one node does NOT >>>>> support binding memory to the process location. >>>>> >>>>> Node: csclprd3-0-11 >>>>> >>>>> This usually is due to not having the required NUMA support installed >>>>> on the node. In some Linux distributions, the required support is >>>>> contained in the libnumactl and libnumactl-devel packages. >>>>> This is a warning only; your job will continue, though performance may >>>>> be degraded. >>>>> >>>>> -------------------------------------------------------------------------- >>>>> >>>>> >>>>> -------------------------------------------------------------------------- >>>>> A request was made to bind to that would result in binding more >>>>> processes than cpus on a resource: >>>>> >>>>> Bind to: CORE >>>>> Node: csclprd3-0-11 >>>>> #processes: 2 >>>>> #cpus: 1 >>>>> >>>>> You can override this protection by adding the "overload-allowed" >>>>> option to your binding directive. >>>>> >>>>> -------------------------------------------------------------------------- >>>>> >>>>> The only installed numa packages are: >>>>> numactl.x86_64 2.0.7-3.el6 >>>>> @centos6.3-x86_64-0/$ >>>>> >>>>> When I search for the available NUMA packages I find: >>>>> >>>>> yum search numa | less >>>>> >>>>> Loaded plugins: fastestmirror >>>>> Loading mirror speeds from cached hostfile >>>>> ============================== N/S Matched: numa >>>>> =============================== >>>>> numactl-devel.i686 : Development package for building >>>>> Applications that use numa >>>>> numactl-devel.x86_64 : Development package for building >>>>> Applications that use >>>>> : numa >>>>> numad.x86_64 : NUMA user daemon >>>>> numactl.i686 : Library for tuning for Non Uniform Memory Access >>>>> machines >>>>> numactl.x86_64 : Library for tuning for Non Uniform Memory Access >>>>> machines >>>>> >>>>> Do I need to install additional and/or different NUMA packages in order >>>>> to get OpenMPI to work >>>>> on this cluster? >>>>> >>>>> -Bill Lane >>>>> IMPORTANT WARNING: This message is intended for the use of the person or >>>>> entity to which it is addressed and may contain information that is >>>>> privileged and confidential, the disclosure of which is governed by >>>>> applicable law. If the reader of this message is not the intended >>>>> recipient, or the employee or agent responsible for delivering it to the >>>>> intended recipient, you are hereby notified that any dissemination, >>>>> distribution or copying of this information is strictly prohibited. Thank >>>>> you for your cooperation. >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> us...@open-mpi.org <mailto:us...@open-mpi.org> >>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users> >>>>> Searchable archives: >>>>> http://www.open-mpi.org/community/lists/users/2015/04/index.php >>>>> <http://www.open-mpi.org/community/lists/users/2015/04/index.php> >>>>> >>>>> IMPORTANT WARNING: This message is intended for the use of the person or >>>>> entity to which it is addressed and may contain information that is >>>>> privileged and confidential, the disclosure of which is governed by >>>>> applicable law. If the reader of this message is not the intended >>>>> recipient, or the employee or agent responsible for delivering it to the >>>>> intended recipient, you are hereby notified that any dissemination, >>>>> distribution or copying of this information is strictly prohibited. Thank >>>>> you for your cooperation. >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> us...@open-mpi.org <mailto:us...@open-mpi.org> >>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users> >>>>> Link to this post: >>>>> http://www.open-mpi.org/community/lists/users/2015/04/26589.php >>>>> <http://www.open-mpi.org/community/lists/users/2015/04/26589.php> >>>>> >>>>> IMPORTANT WARNING: This message is intended for the use of the person or >>>>> entity to which it is addressed and may contain information that is >>>>> privileged and confidential, the disclosure of which is governed by >>>>> applicable law. If the reader of this message is not the intended >>>>> recipient, or the employee or agent responsible for delivering it to the >>>>> intended recipient, you are hereby notified that any dissemination, >>>>> distribution or copying of this information is strictly prohibited. Thank >>>>> you for your cooperation. _______________________________________________ >>>>> users mailing list >>>>> us...@open-mpi.org <mailto:us...@open-mpi.org> >>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users> >>>>> Link to this post: >>>>> http://www.open-mpi.org/community/lists/users/2015/04/26611.php >>>>> <http://www.open-mpi.org/community/lists/users/2015/04/26611.php> >>>> IMPORTANT WARNING: This message is intended for the use of the person or >>>> entity to which it is addressed and may contain information that is >>>> privileged and confidential, the disclosure of which is governed by >>>> applicable law. If the reader of this message is not the intended >>>> recipient, or the employee or agent responsible for delivering it to the >>>> intended recipient, you are hereby notified that any dissemination, >>>> distribution or copying of this information is strictly prohibited. Thank >>>> you for your cooperation. _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org <mailto:us...@open-mpi.org> >>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users> >>>> Link to this post: >>>> http://www.open-mpi.org/community/lists/users/2015/04/26618.php >>>> <http://www.open-mpi.org/community/lists/users/2015/04/26618.php> >>> IMPORTANT WARNING: This message is intended for the use of the person or >>> entity to which it is addressed and may contain information that is >>> privileged and confidential, the disclosure of which is governed by >>> applicable law. If the reader of this message is not the intended >>> recipient, or the employee or agent responsible for delivering it to the >>> intended recipient, you are hereby notified that any dissemination, >>> distribution or copying of this information is strictly prohibited. Thank >>> you for your cooperation. _______________________________________________ >>> users mailing list >>> us...@open-mpi.org <mailto:us...@open-mpi.org> >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>> <http://www.open-mpi.org/mailman/listinfo.cgi/users> >>> Link to this post: >>> http://www.open-mpi.org/community/lists/users/2015/04/26643.php >>> <http://www.open-mpi.org/community/lists/users/2015/04/26643.php> >> IMPORTANT WARNING: This message is intended for the use of the person or >> entity to which it is addressed and may contain information that is >> privileged and confidential, the disclosure of which is governed by >> applicable law. If the reader of this message is not the intended recipient, >> or the employee or agent responsible for delivering it to the intended >> recipient, you are hereby notified that any dissemination, distribution or >> copying of this information is strictly prohibited. Thank you for your >> cooperation. _______________________________________________ >> users mailing list >> us...@open-mpi.org <mailto:us...@open-mpi.org> >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >> <http://www.open-mpi.org/mailman/listinfo.cgi/users> >> Link to this post: >> http://www.open-mpi.org/community/lists/users/2015/04/26655.php >> <http://www.open-mpi.org/community/lists/users/2015/04/26655.php> > IMPORTANT WARNING: This message is intended for the use of the person or > entity to which it is addressed and may contain information that is > privileged and confidential, the disclosure of which is governed by > applicable law. If the reader of this message is not the intended recipient, > or the employee or agent responsible for delivering it to the intended > recipient, you are hereby notified that any dissemination, distribution or > copying of this information is strictly prohibited. Thank you for your > cooperation. _______________________________________________ > users mailing list > us...@open-mpi.org <mailto:us...@open-mpi.org> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > <http://www.open-mpi.org/mailman/listinfo.cgi/users> > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/04/26659.php > <http://www.open-mpi.org/community/lists/users/2015/04/26659.php>