Hi Jody,
I tried your suggestion but it still failed. Attached is the modified code.
If your machine has fortran compiler as well, you can try it.
BTW, how many processors did you use for testing your C code?
Thanks,
Kan
On Thu, Feb 25, 2010 at 3:35 AM, jody wrote:
> Hi
> Just wanted to l
Hi All,
Thanks a lot for your support. It was a big help. I found a race condition
in my code and now the problem is solved.
Regards,
Amr
On Fri, Feb 26, 2010 at 7:45 AM, Prentice Bisbal wrote:
> Amr Hassan wrote:
> > Thanks alot for your reply,
> >
> > I'm using blocking Send and Receive. All
Amr Hassan wrote:
> Thanks alot for your reply,
>
> I'm using blocking Send and Receive. All the clients are sending data
> and the server is receive the messages from the clients with
> MPI_ANY_SOURCE as the sender. Do you think there is a race condition
> near this pattern?
>
> I searched a
I was getting the same error a few weeks ago. In my case the error
message was spot on. I was trying to put too much data in a buffer using
MPI_Pack.
I was able to track down the problem using valgrind. Have you tried that
yet? You need to install valgrind first and then compile OpenMPI with
valgr
On Feb 10, 2010, at 9:45 AM, Addepalli, Srirangam V wrote:
> I am trying to test orte-checkpoint with a MPI JOB. It how ever hangs for all
> jobs. This is how i submit the job is started
> mpirun -np 8 -mca ft-enable cr /apps/nwchem-5.1.1/bin/LINUX64/nwchem
> siosi6.nw
This might be the prob
Anton,
I don't know if there usual or typical way of initiating a checkpoint amongst
various resource managers. I know that the BLCR folks (I believe Eric Roman is
heading this effort - CC'ed) have been investigating a tighter integration of
Open MPI, BLCR and Torque. He might be able to give y
Yes but only one thread at each client is allowed to use MPI. Also, there is
a semaphore on the MPI usage.
On Fri, Feb 26, 2010 at 1:09 AM, Brian Budge wrote:
> Is your code multithreaded?
>
> On Feb 25, 2010 12:56 AM, "Amr Hassan" wrote:
>
> Thanks alot for your reply,
>
> I'm using blocking
On Feb 25, 2010, at 3:56 AM, Amr Hassan wrote:
> Thanks alot for your reply,
>
> I'm using blocking Send and Receive. All the clients are sending data and the
> server is receive the messages from the clients with MPI_ANY_SOURCE as the
> sender. Do you think there is a race condition near this
Is your code multithreaded?
On Feb 25, 2010 12:56 AM, "Amr Hassan" wrote:
Thanks alot for your reply,
I'm using blocking Send and Receive. All the clients are sending data and
the server is receive the messages from the clients with MPI_ANY_SOURCE as
the sender. Do you think there is a race con
This typically means that either libmpi.so does not exist on the machine that
you are trying to run it on, or it cannot be found. You may need to extend the
value of the LD_LIBRARY_PATH environment variable with the lib directory of
your Open MPI installation (don't just overwrite it -- check f
Respected Users
I have installed openmpi successfully also compiled the hello_world program
with mpicc. But when I am running the exectable with the command mpirun -np 2
hello_mpi(hello_mpi is the executable) in my desktop pc(dual-core processor) is
giving an error like this:
hello_mpi: error
Thanks alot for your reply,
I'm using blocking Send and Receive. All the clients are sending data and
the server is receive the messages from the clients with MPI_ANY_SOURCE as
the sender. Do you think there is a race condition near this pattern?
I searched a lot and used totalview but I couldn't
Hi
Just wanted to let you know:
I translated your program to C ran it, and it crashed at MPI_FILE_SET_VIEW
in a similar way than yours did.
then i added an if-clause to prevent the call of MPI_FILE_WRITE with
the undefined value.
if (myid == 0) {
MPI_File_write(fh, temp, count, MPI_DOU
We've seen similar things in our code. In our case it is probably due to a
race condition. Try running the segv'ing process in a debugger, and it will
likely show you a bug in your code
On Feb 24, 2010 9:36 PM, "Amr Hassan" wrote:
Hi All,
I'm facing a strange problem with OpenMPI.
I'm develo
Hi All,
I'm facing a strange problem with OpenMPI.
I'm developing an application which is required to send a message from each
client (1 MB each) to a server node for around 10 times per second (it's a
distributed render application and I'm trying to reach a higher frame rate
). The problem is t
15 matches
Mail list logo