Thank you for the clarification, Gilles. That has saved me trying to debug
something that is not expected to work.
Can this be added to the OpenMPI documentation at
https://docs.open-mpi.org/en/main/launching-apps/slurm.html please? I feel
like this should be called out as a reason to use srun ins
Ok, this is a good / consistent output. That being said, I don't grok what is
happening here: it says it finds 2 slots, but then it tells you it doesn't have
enough slots.
Let me dig deeper and get back to you...
--
Jeff Squyres
jsquy...@cisco.com
From: timesir
Thanks for the output.
I'm seeing inconsistent output between your different outputs, however. For
example, one of your outputs seems to ignore the hostfile and only show slots
on the local host, but another output shows 2 hosts with 1 slot each. But I
don't know what was in the hosts file fo
The netmask of both is 255.255.255.0
在 2022/11/15 12:32, timesir 写道:
*(py3.9) ➜ /share mpirun -n 2 --machinefile hosts --mca
rmaps_base_verbose 100 --mca ras_base_verbose 100 which mpirun*
[computer01:39342] mca: base: component_find: searching NULL for ras
components
[computer01:39342] m
*(py3.9) ➜ /share ompi_info --version*
Open MPI v5.0.0rc9
https://www.open-mpi.org/community/help/
*(py3.9) ➜ /share cat hosts*
192.168.180.48 slots=1
192.168.60.203 slots=1
*(py3.9) ➜** /share mpirun -n 2 --machinefile hosts --mca
plm_base_verbose 100 --mca rmaps_base_verbose 100 --m
Do you receive this email?
在 2022年11月23日星期三,timesir 写道:
>
> *1. This command now runs correctly *
>
> *(py3.9) ➜ /share mpirun -n 2 --machinefile hosts --mca plm_base_verbose
> 100 --mca rmaps_base_verbose 100 --mca ras_base_verbose 100 uptime*
>
>
>
> *2. But this command gets stuck. It seem
*(py3.9) ➜ /share ompi_info --version*
Open MPI v5.0.0rc9
https://www.open-mpi.org/community/help/
*(py3.9) ➜ /share cat hosts*
192.168.180.48 slots=1
192.168.60.203 slots=1
*(py3.9) ➜** /share mpirun -n 2 --machinefile hosts --mca
plm_base_verbose 100 --mca rmaps_base_verbose 100 --m
Yes, Gilles responded within a few hours:
https://www.mail-archive.com/users@lists.open-mpi.org/msg35057.html
Looking closer, we should still be seeing more output compared to what you
posted. It's almost like you have a busted Open MPI installation -- perhaps
it's missing the "hostfile" compo
I see 2 config.log files -- can you also send the other information requested
on that page? I.e, the version you're using (I think you said in a prior
email that it was 5.0rc9, but I'm not 100% sure), and the output from ompi_info
--all.
--
Jeff Squyres
jsquy...@cisco.com
do you receive my email?
timesir 于2022年11月15日周二 12:33写道:
> *(py3.9) ➜ /share mpirun -n 2 --machinefile hosts --mca
> rmaps_base_verbose 100 --mca ras_base_verbose 100 which mpirun*
> [computer01:39342] mca: base: component_find: searching NULL for ras
> components
> [computer01:39342] mca: b
FWIW, I cannot reproduce this error.
What is the netmask on both hosts?
Cheers,
Gilles
On 11/15/2022 1:32 PM, timesir via users wrote:
*(py3.9) ➜ /share mpirun -n 2 --machinefile hosts --mca
rmaps_base_verbose 100 --mca ras_base_verbose 100 which mpirun*
[computer01:39342] mca: base
*(py3.9) ➜ /share mpirun -n 2 --machinefile hosts --mca
rmaps_base_verbose 100 --mca ras_base_verbose 100 which mpirun*
[computer01:39342] mca: base: component_find: searching NULL for ras
components
[computer01:39342] mca: base: find_dyn_components: checking NULL for ras
components
[compute
I'm using an infiniband interconnect. Am I guaranteed asynchronous progress
without doing anything else? For example, to get asynchronous progress with
Intel MPI, you need to set the environment variable I_MPI_ASYN_PROGRESS.
Bruce
-Original Message-
From: users [mailto:users-boun...@lis
Hi,
there is likely something wrong in Open MPI (i will follow up in the
devel ML)
meanwhile, you can
mpirun --mca opal_set_max_sys_limits core:unlimited ...
Cheers,
Gilles
On 3/3/2017 1:01 PM, gzzh...@buaa.edu.cn wrote:
Hi Jeff:
Thanks for your suggestions.
1. I have execute
Hi Jeff:
Thanks for your suggestions.
1. I have execute the command " find / -name core*" in each node, no
coredump file was found. and no coredump files in the /home, /core and pwd file
path as well.
2. I have changed core_pattern follow your advice, still no expected
coredump file
Hi Jason,
No my main problem is in regards to running mpirun --np any #. I am unable
to run any command on multiple processors using mpi4py and openmpi
together. Whenever I type in a command that uses both, I receive no output.
I was wondering whether it was a compatibility issue
On Thu, Oct 6,
Sam,
openmpi-devel-1.7.3-1.fc20 rpm provides
/usr/lib64/openmpi/bin/mpicc
this is the mpicc you want to use to build mpi4py
of course, you can download and install a recent Open MPI version from
open-mpi.org.
if you decide to go this way, i recommend you download 1.10.3 from
https://www.o
To dave, from the installation guide I found, it seemed I couldnt just
directly download it from the package list, but rather Id need to use the
mpicc wrapper to compile and install. I also wanted to see if I could build
it from the installation guide, sorta learn how the whole process worked.
To
c.stuttgart
> >
> > ------
> >
> >
> >
> >
> >
> > ___
> > users mailing list
> > us...@open-mpi.org
> > Subscription: https://www.open
s mailing list
> us...@open-mpi.org
> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post:
> http://www.open-mpi.org/community/lists/users/2016/06/29340.php
--
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to:
http://www.cisco.com/web/ab
;
>
>
> Today's Topics:
>
> 1. Re: users Digest, Vol 3510, Issue 2 (Jeff Squyres (jsquyres))
> 2. Re: segmentation fault for slot-list and openmpi-1.10.3rc2
> (Siegmar Gross)
> 3. OpenMPI virtualization aware (Marco D'Amico)
> 4. Re: OpenMPI virtualization aware (Ralph Castain)
>
>
>
1
List-Post: users@lists.open-mpi.org
Date: Thu, 26 May 2016 23:28:17 +
From: "Jeff Squyres (jsquyres)"
To: Megdich Islem , "Open MPI User's List"
Cc: Dave Love
Subject: Re: [OMPI users] users Digest, Vol 3510, Issue 2
Message-ID: <441f803d-fdbb-443d-82aa-74f
You're still intermingling your Open MPI and MPICH installations.
You need to ensure to use the wrapper compilers and mpirun/mpiexec from the
same MPI implementation.
For example, if you use mpicc/mpifort from Open MPI to build your program, then
you must use Open MPI's mpirun/mpiexec.
If you
Thank you all for your suggestions !!
I found an answer to a similar case in Open MPI FAQ (Question 15)FAQ: Running
MPI jobs
| |
| | | | | | | |
| FAQ: Running MPI jobsTable of contents: What pre-requisites are necessary for
running an Open MPI job? What ABI guarantees does Open M
I wrote:
> You could wrap one (set of) program(s) in a script to set the
> appropriate environment before invoking the real program.
I realize I should have said something like "program invocations",
i.e. if you have no control over something invoking mpirun for programs
using different MPIs,
Most commercial applications, i.e., Ansys Fluent, Abaqus, NASTRAN, and
PAM-CRASH is IBM Platform MPI bundled with each application and is the default
MPI when running parallel simulations. Depending on which Abaqus release
you're using your choices are IBM Platform MPI or Intel MPI. I don't r
Just to clarify, as this is a frequent misconception: the statement that the
absolute path will setup your remote environment is only true when using the
rsh/ssh launcher. It is not true when running under a resource manager (e.g.,
SLURM, LSF, PBSPro, etc.). In those cases, it is up to the RM co
Doesn't Abaqus do its own environment setup? I.e., I'm *guessing* that you
should be able to set your environment startup files (e.g., $HOME/.bashrc) to
point your PATH / LD_LIBRARY_PATH to point to whichever MPI implementation you
want, and Abaqus will do whatever it needs to a) be independent
Megdich Islem writes:
> Yes, Empire does the fluid structure coupling. It couples OpenFoam (fluid
> analysis) and Abaqus (structural analysis).
> Does all the software need to have the same MPI architecture in order to
> communicate ?
I doubt it's doing that, and presumably you have no control
Yes, Empire does the fluid structure coupling. It couples OpenFoam (fluid
analysis) and Abaqus (structural analysis).
Does all the software need to have the same MPI architecture in order to
communicate ?
Regards,Islem
Le Mardi 24 mai 2016 1h02, Gilles Gouaillardet a écrit :
what do y
what do you mean by coupling ?
does Empire and OpenFoam communicate via MPI ?
wouldn't it be much easier if you rebuild OpenFoam with mpich or intelmpi ?
Cheers,
Gilles
On 5/24/2016 8:44 AM, Megdich Islem wrote:
"Open MPI does not work when MPICH or intel MPI are installed"
Thank you for
"Open MPI does not work when MPICH or intel MPI are installed"
Thank you for your suggestion. But I need to run OpenFoam and Empire at the
same time. In fact, Empire couples OpenFoam with another software.
Is there any solution for this case ?
Regards,Islem
Le Lundi 23 mai 2016 17h00, "user
Its not really a good idea to mix active and passive synchronization (we
may actually explicitly forbid it in the future). You can remove the
calls to MPI_Win_fence () and still have correct synchronization. That
said, you did find a bug in my bad synchronization detection because
this is legal:
Gilles,
I downloaded and built openmpi-2.0.0rc2 and used that for the test. I get a
crash on more than 1 processor for the lock/unlock protocol with the error
message
[node005:29916] *** An error occurred in MPI_Win_lock
[node005:29916] *** reported by process [3736862721,6]
[node005:29916] ***
Thanks Rowena,
greg
-Original Message-
From: users [mailto:users-boun...@open-mpi.org] On Behalf Of
users-requ...@open-mpi.org
Sent: Thursday, 28 April 2016 11:34 AM
To: us...@open-mpi.org
Subject: users Digest, Vol 3486, Issue 3
Send users mailing list submissions to
us...@open-
-
Message: 1
Date: Tue, 6 May 2014 17:45:09 -0700
From: Ralph Castain mailto:r...@open-mpi.org>>
To: Open MPI Users mailto:us...@open-mpi.org>>
Subject: Re: [OMPI users] users Digest, Vol 2881, Issue 2
Subject line so it is more specific
> than "Re: Contents of users digest..."
>
>
> Today's Topics:
>
>1. Re: users Digest, Vol 2881, Issue 2 (Ralph Castain)
>
>
> --
>
> Message: 1
> Date: Tue, 6 May 2014 17:45:09 -0700
> From: Ral
-
>
> Message: 1
> Date: Tue, 6 May 2014 17:45:09 -0700
> From: Ralph Castain
> To: Open MPI Users
> Subject: Re: [OMPI users] users Digest, Vol 2881, Issue 2
> Message-ID: <4b207e61-952a-4744-9a7b-0704c4b0d...@open-mpi.o
Today's Topics:
>
>1. Re: users Digest, Vol 2881, Issue 1 (Clay Kirkland)
>2. Re: users Digest, Vol 2881, Issue 1 (Clay Kirkland)
>
>
> --
>
> Message: 1
> Date: Tue, 6 May 2014 18:28:59
's Topics:
>
>1. Re: users Digest, Vol 2881, Issue 1 (Clay Kirkland)
>2. Re: users Digest, Vol 2881, Issue 1 (Clay Kirkland)
>
>
> --
>
> Message: 1
> Date: Tue, 6 May 2014 18:28:59 -0500
&g
t
> >>
> >> http://www.open-mpi.org/faq/?category=building#install-overwrite
> >>
> >> that I should completely uninstall my previous version.
> >
> > Yes, that is best. OR: you can install into a whole separate tree and
> > ignore the first
gt;
>> On May 6, 2014, at 9:32 AM, Imran Ali
>> wrote:
>>
>> > I will attempt that than. I read at
>> >
>> > http://www.open-mpi.org/faq/?category=building#install-overwrite
>> >
>> > that I should completely uninstall my previous version
lding#install-overwrite
> >>
> >> that I should completely uninstall my previous version.
> >
> > Yes, that is best. OR: you can install into a whole separate tree and
> ignore the first installation.
> >
> >> Could you recommend to me how I can go abou
Are you using TCP as the MPI transport?
If so, another thing to try is to limit the IP interfaces that MPI uses for its
traffic to see if there's some kind of problem with specific networks.
For example:
mpirun --mca btl_tcp_if_include eth0 ...
If that works, then try adding in any/all othe
I am configuring with all defaults. Just doing a ./configure and then
make and make install. I have used open mpi on several kinds of
unix systems this way and have had no trouble before. I believe I
last had success on a redhat version of linux.
On Sat, May 3, 2014 at 11:00 AM, wrote:
l have a Fortran module error inopenmpi-1.9a1r29341
>> (Siegmar Gross)
>> 3. Re: I still have a Fortran module errorin
>> openmpi-1.9a1r29341 (Jeff Squyres (jsquyres))
>>
>>
>> --
>>
>> Message: 1
>> Date: Thu, 3 Oct
ross)
> 3. Re: I still have a Fortran module errorin
> openmpi-1.9a1r29341 (Jeff Squyres (jsquyres))
>
>
> --
>
> Message: 1
> Date: Thu, 3 Oct 2013 12:11:17 -0500
> From: Jeff Hammond
You try Google Scholar yet?
Always exhaust all nonhuman resources before requesting human
assistance. The human brain is a terrible resource to waste when a
computer can do the job.
Jeff
Sent from my iPhone
On Oct 3, 2013, at 10:18 AM, Yin Zhao wrote:
> Hi all,
>
> Does anybody have done expe
Hi all,
Does anybody have done experiments comparing the speed between mpich and
openmpi?
Best regards,
Yin Zhao
> 在 2013年10月3日,0:00,users-requ...@open-mpi.org 写道:
>
> Send users mailing list submissions to
>us...@open-mpi.org
>
> To subscribe or unsubscribe via the World Wide Web, visi
Sirs,
I did follow the link provided in the output and checked my settings. I am
certain my system is properly confiugured but the message persists. Is there
any way to outright disable the message?.. I didn't see any way to do it.
Thanks,
-Original Message-
From: Bryan, Clifton W
-Original Message-
From: users [mailto:users-boun...@open-mpi.org] On Behalf Of
users-requ...@open-mpi.org
Sent: Wednesday, September 25, 2013 11:00 AM
To: us...@open-mpi.org
Subject: users Digest, Vol 2689, Issue 1
Send users mailing list submissions to
us...@open-mpi.org
To s
I proclaimed implicit definition of type to add 'implicit node'.
And I change include 'mpif.h' to use mpi.
But Unfortunately I couldn't get more information.
Error occuring on only slave node is to feel uneasy.
But I have no idea except this.
(2013/05/15 1:11), Andrea Negri wrote:
I'm not an
I'm not an expert of MPI, but I stronly encourage you to use
use mpi
implicit none
This can save a LOT of time in the debug.
On 14 May 2013 18:00, wrote:
> Send users mailing list submissions to
> us...@open-mpi.org
>
> To subscribe or unsubscribe via the World Wide Web, visit
>
thank you for your reply and I apologize for my English and for my
mistake, in every way I have had configured OPMI with this /usr/local/
prefix.
2012/12/12
> Send users mailing list submissions to
> us...@open-mpi.org
>
> To subscribe or unsubscribe via the World Wide Web, visit
>
Hi all,
I have a MPI_Isend/MPI_Recv problem in a multi-thread program.
In the program:
The first machine has one thread does some computation and call MPI_Isend
to send buffer to the second machine, and another thread is always trying to
MPI_Recv data from the second machine. And the
Hi, Gus and John,
my code (zeusmp2) is a F77 code ported in F95, and the very first time
I have launched it, the processed disappears almost immediately.
I checked the code with valgrind and some unallocated arrays were
passed wrongly to some subroutines.
After having corrected this bug, the code
Users im pretty sure this is going to be possible for as long as the internet
is around things are falling into place http://t.co/j8cZZZdC
I have mentioned it two times that by doing this we got good results and the
question was just in terms of process migration .. I have explained the
current work as well.
regards,
Mudassar
From: "users-requ...@open-mpi.org"
To: us...@open-mpi.org
Sent: T
Sorry for the late reply, but, as I just say, we are attempting
to recover the full operation of part of our cluster
Yes, it was a typo, I use to add the "sm" flag to the "--mca btl"
option. However I think this is not mandatory, as I suppose
openmpi use the the so-called "Law of Least Astonishme
"MPI can get through your firewall, right?"
As far as I can tell the firewall is not the problem - have tried it with
firewalls disabled, automatic fw polices based on port requests from MPI, and
with manual exception policies.
> From: users-requ...@open-mpi.org
> Subject: users Digest, Vol
Message: 2
Date: Tue, 23 Nov 2010 10:27:37 -0800
From: Eugene Loh
Subject: Re: [OMPI users] tool for measuring the ping with accuracy
To: Open MPI Users
Message-ID: <4cec0799.60...@oracle.com>
Content-Type: text/plain; charset=ISO-8859-1; format=flowed
George Markomanolis wrote:
Dear Eug
I don't think it's really a fault -- it's just how we designed and implemented
it.
On Sep 6, 2010, at 7:40 AM, lyb wrote:
> Thanks for your answer, but I test with MPICH2, it doesn't have this fault.
> Why?
>> Message: 9
>> Date: Wed, 1 Sep 2010 20:14:44 -0600
>> From: Ralph Castain
>> Subject
Thanks for your answer, but I test with MPICH2, it doesn't have this
fault. Why?
Message: 9
Date: Wed, 1 Sep 2010 20:14:44 -0600
From: Ralph Castain
Subject: Re: [OMPI users] MPI_Comm_accept and MPI_Comm_connect both
use 100%one cpu core. Is it a bug?
To: Open MPI Users
Message-ID
I probably won't have an opportunity to work on reproducing this on the 1.4.2.
The trunk has a bunch of bug fixes that probably will not be backported to the
1.4 series (things have changed too much since that branch). So I would suggest
trying the 1.5 series.
-- Josh
On Aug 13, 2010, at 10:12
Josh
I am having problems compiling the sources from the latest trunk. It
complains of libgomp.spec missing even though that file exists on my
system. I will see if I have to change any other environment variables
to have a successful compilation. I will keep you posted.
BTW, were you successful
FWIW, I took your code compiled it on a linux system using OMPI 1.4
r22761 and Solaris Studio C compilers. Then I ran it with "mpirun -np 4
a.out" and it seems to work for me:
Hello MPI World From process 0: Num processes: 4
Hello MPI World from process 1!
Hello MPI World from process 2!
Hello
Thanks for all your inputs.
It is good to know this initial latency is an expected behavior and the
workaround by using one dummy iteration before timing is started.
I did not notice this because my older parallel CFD code runs a large number
of time steps and the initial latency was compensated
It looks like your executable is explicitly calling MPI_ABORT in the
CmiAbort function -- perhaps in response to something happening in the
namd or CmiHandleMessage functions. The next logical step would
likely be to look in those routines and see why MPI_ABORT/CmiAbort
would be invoked.
Yes.. The executables run initially and then gives the mentioned error in
the first message!
i.e.
./mpirun -hostfile machines executable
--
MPI_ABORT was invoked on rank 2 in communicator MPI_COMM_WORLD
with errorcode 1.
NOTE
Patrick Geoffray wrote:
> Hi Oskar,
>
> Oskar Enoksson wrote:
>> The reason in this case was that cl120 had some kind of hardware
>> problem, perhaps memory error or myrinet NIC hardware error. The system
>> hung.
>>
>> I will try MX_ZOMBIE_SEND=0, thanks for the hint!
>
> I would not recommen
Yes, I am using the right one. I've installed the freshly compiled
openmpi into /opt/openmpi/1.3.3-g95-32. If I edit the mpif.h file by
hand and put "error!" in the first line I get:
zamblap:sandbox zamb$ edit /opt/openmpi/1.3.3-g95-32/include/mpif.h
zamblap:sandbox zamb$ mpif77 inplace_test
This mailing list supports the Open MPI implementation of the MPI
standard. If you have concerns about Intel MPI you should contact
their support group.
The ompi_checkpoint/ompi_restart routines are designed to work with
Open MPI, and will certainly fail when used with other MPI
implement
DEAR SIR/MADAM
kindly tell the commands for checkpointing and restarting of mpi programs
using intel mpi
i tried the following commands they did not work
ompi_checkpoint
ompi_restart file name of global snap shot
with regards
mallikarjuna shastry
Today's Topics:
1. Re: btl_openib_connect_oob.c:459:qp_create_one: errorcreating
qp (Jeff Squyres)
2. Re: [OMPI users]
btl_openib_connect_oob.c:459:qp_create_one:errorcreating qp
(Jeff Squyres)
-- Message: 2 Date: Wed, 1 Jul 2009 08:56
Michael Kuklik
Sent: Wednesday, May 27, 2009 10:04 PM
To: us...@open-mpi.org
Subject: Re: [OMPI users] users Digest, Vol 1242, Issue 5
Joe,
I don't know why I didn't think about it. It works with the whole path.
I put the intel env script in user .bash_login file.
Do you think I
On May 28, 2009, at 1:04 AM, Michael Kuklik wrote:
I don't know why I didn't think about it. It works with the whole
path.
I put the intel env script in user .bash_login file.
Do you think I should put the intel env script in the global shell
config file like /etc/profile in order for libto
Joe,
I don't know why I didn't think about it. It works with the whole path.
I put the intel env script in user .bash_login file.
Do you think I should put the intel env script in the global shell config file
like /etc/profile in order for libtool to see icc?
Thanks for help,
mike
___
Thanks Jeff. We solve this problem finally. Download the newest
OFED-1.4.1-rc6.tgz, and reinstall all node's infiniband drivers and
utilities. Everythings looks good, and I have my own coffee time now. Thanks
again.
Best Regards,
Gloria Jan
Wavelink Technology Inc
I don't think the speed of
I don't think the speed of the down port matters; port_down means that
there's no cable connected, so the values are probably fairly random.
On May 7, 2009, at 10:38 PM, jan wrote:
Anyone can help me to find out problem or bug in my cluster? The
output of "ibv_devinfo -v" from Dell blade wi
Anyone can help me to find out problem or bug in my cluster? The output of
"ibv_devinfo -v" from Dell blade with infiniband module look very strange. The
phys_port_cnt is 2, one active, and another down. The active port is 20x speed,
the down port is 10x speed. We are using Dell PowerEdge M600 B
Jan,
I guess that you have OFED driver installed on you machines. You may do
basic network verification with ibdiagnet utility
(http://linux.die.net/man/1/ibdiagnet) that is part of OFED installation.
Regards,
Pasha
Jeff Squyres wrote:
On May 4, 2009, at 9:50 AM, jan wrote:
Thank you Jef
On May 4, 2009, at 9:50 AM, jan wrote:
Thank you Jeff. I have passed the mail to the IB vendor Dell
company(the
blade was ordered from Dell Taiwan), but he todl me that he didn't
understand "layer 0 diagnostics". Coluld you help us to get more
information of "layer 0 diagnostics". Thanks agai
Thank you Jeff. I have passed the mail to the IB vendor Dell company(the
blade was ordered from Dell Taiwan), but he todl me that he didn't
understand "layer 0 diagnostics". Coluld you help us to get more
information of "layer 0 diagnostics". Thanks again.
Regards,
Gloria Jan
Wavelink Technol
Thank you Jeff. I have passed the mail to the IB vendor Dell company(the
blade was ordered from Dell Taiwan), but he todl me that he didn't
understand "layer 0 diagnostics". Coluld you help us to get more
information of "layer 0 diagnostics". Thanks again.
Regards,
Gloria Jan
Wavelink Technol
As I've indicated a few times in this thread:
>> Have you run layer 0 diagnostics to know that your fabric is clean?
>> You might want to contact your IB vendor to find out how to do that.
On May 4, 2009, at 4:34 AM, jan wrote:
Hi Jeff,
I have updated the firmware of Infiniband module on De
On Apr 30, 2009, at 5:33 AM, jan wrote:
Thank You Jeff Squyres. Could you suggest the method to run layer 0
diagnostics to know that if the fabric is clean. I have contacted Dell
local(Taiwan). I don't think they are familiar with Openmpi even the
infiniband module.
Note that the layer 0 diag
Thank You Jeff Squyres. Could you suggest the method to run layer 0
diagnostics to know that if the fabric is clean. I have contacted Dell
local(Taiwan). I don't think they are familiar with Openmpi even the
infiniband module. Does anyone have the IB stack hangs problem with Mellanox
ConnectX p
On Apr 27, 2009, at 10:22 PM, jan wrote:
Thank You Jeff Squyres.
I have checked out the web page
http://www.open-mpi.org/community/lists/announce/2009/03/0029.php,
then the
page https://svn.open-mpi.org/trac/ompi/ticket/1853 , but the web page
svn.open-mpi.org seems crash.
Try that ticket
Thank You Jeff Squyres.
I have checked out the web page
http://www.open-mpi.org/community/lists/announce/2009/03/0029.php, then the
page https://svn.open-mpi.org/trac/ompi/ticket/1853 , but the web page
svn.open-mpi.org seems crash.
Then I tried OpenMpi V1.3.2 for many different configuratio
Dear Jeff,
Thankyou for your help. I have tried the OpenMPI v1.3.2 on Sunday, but the
problems till occured.
Regards, Gloria
Wavelink Technology Inc.
Per http://www.open-mpi.org/community/lists/announce/2009/03/0029.php,
can you try upgrading to Open MPI v1.3.2?
On Apr 24, 2009, at 5:21
Thankyou for that. I have not yet been able to get the IT department to help me
with disabling the firewalls, but hopefully that is the problem. Sorry for the
late response, I was hoping the IT department would be faster.
Robertson
Message: 2
List-Post: users@lists.open-mpi.org
Date: Fri, 6 Feb
The problem is that you didn't specify or allocate any nodes for the
job. At the least, you need to tell us what nodes to use via a hostfile.
Alternatively, are you using a resource manager to assign the nodes?
OMPI didn't see anything from one, but it could be that we just didn't
see the r
Thanks a lot Ralph!
I corrected the no_local to nolocal and now when I try to execute the
script step1 (pls find it attached)
[rchaud@helios amber10]$ ./step1
[helios.structure.uic.edu:16335] [0,0,0] ORTE_ERROR_LOG: Not available
in file ras_bjs.c at line 247
---
system thinks it should be building.
Thanks
Ralph
On Oct 31, 2008, at 4:54 PM, Allan Menezes wrote:
Date: Fri, 31 Oct 2008 09:34:52 -0600
From: Ralph Castain
Subject: Re: [OMPI users] users Digest, Vol 1052, Issue 1
To: Open MPI Users
Message-ID: <0cf28492-b13e-4f82-ac43-c1580f079...@lanl.g
properly
detecting that it exists.
Can you attach a copy of your config.log? It will tell us what the
system thinks it should be building.
Thanks
Ralph
On Oct 31, 2008, at 4:54 PM, Allan Menezes wrote:
Date: Fri, 31 Oct 2008 09:34:52 -0600
From: Ralph Castain
Subject: Re: [OMPI users
It looks like the daemon isn't seeing the other interface address on
host x2. Can you ssh to x2 and send the contents of ifconfig -a?
Ralph
On Oct 31, 2008, at 9:18 AM, Allan Menezes wrote:
users-requ...@open-mpi.org wrote:
Send users mailing list submissions to
us...@open-mpi.org
users-requ...@open-mpi.org wrote:
Send users mailing list submissions to
us...@open-mpi.org
To subscribe or unsubscribe via the World Wide Web, visit
http://www.open-mpi.org/mailman/listinfo.cgi/users
or, via email, send a message with subject or body 'help' to
users-req
Hi
I think it's correct. what I want to to is to send a 3d array from the
process 1 to process 0 =root):
call MPI_Send(toroot,3,MPI_DOUBLE_PRECISION,root,n,MPI_COMM_WORLD
in some other part of the code process 0 acts on the 3d array and
turns it into a 4d one and sends it back to process 1, which
On Sep 9, 2008, at 3:05 PM, Christopher Tanner wrote:
I think I've found the problem / solution. With Ubuntu, there's a
program called 'ldconfig' that updates the dynamic linker run-time
bindings. Since Open MPI was compiled to use dynamic linking, these
have to be updated. Thus, these comm
On clusters where I'm using the Intel compilers and OpenMPI, I setup
the compiler directory (usually /opt/intel) as a NFS export. The
computation nodes then mount that export. Next, I add the following
lines to the ld.so.conf file and distribute it to the computation
nodes:
/opt/intel/cce/version_n
1 - 100 of 123 matches
Mail list logo