Hello Community,
I just finished building an OpenMPI version with custom configuration
(1.4.4) on a server. I do not have root privileges on the server.
The server already had a "default" version on OpenMPI, which seemed to be
single-threaded:
ompi_info | grep Thread yields Thread support:
Excellent. The bug fix will be in 1.6.1, too.
On May 16, 2012, at 1:26 PM, Ricardo Reis wrote:
>
> all problems gone, thanks for the input and assistance.
>
> cheers,
>
> Ricardo Reis
>
> 'Non Serviam'
>
> PhD/MSc Mechanical Engineering | Lic. Aerospace Engineering
>
> Computational Fluid
all problems gone, thanks for the input and assistance.
cheers,
Ricardo Reis
'Non Serviam'
PhD/MSc Mechanical Engineering | Lic. Aerospace Engineering
Computational Fluid Dynamics, High Performance Computing, Turbulence
http://www.lasef.ist.utl.pt
Cultural Instigator @ Rádio Zero
h
We've had a few reports of this - it looks like someone made a change to R that
can cause problems. Basically, the open fabrics driver for Infiniband doesn't
support "fork" operations - it can lead to memory corruption issues if you
inadvertently do the "wrong thing" at some point after the fork
Looks like you have a firewall between hosts tik34x and tik33x - you might
check to ensure all firewalls are disabled. The error is saying it can't open a
TCP socket between the two nodes, so there is no communication path between
those two processes.
On May 16, 2012, at 4:22 AM, devendra rai
I'm getting the following error with a new version of R, using Rmpi and
a few other modules. I've already had a couple of good suggestions from
this group about how to diagnose the cause of the fork error using
"strace" but we don't have it on our LSF Linux cluster. This is my
first use of R/
I have following structure of MPI code -
int main(){
MPI_INIT.
//initialize MPI
data = malloc(sizeof(int)*200);
//initialize data
/*Master-*/
if(taskid == 0){
//send data to each slave
MPI_SEND
}
/*Slaves---*/
if(taskid > 0){
//accep
Hello All,
I am trying to run an OpenMPI application across two physical machines.
I get an error "Returned "Unreachable" (-12) instead of "Success" (0)", and
looking through the logs (attached), I cannot seem to find out the cause, and
how to fix it.
I see lot of (communication) components
Dear all
I think I find the culprit.
I was calculating my offset using
offset = my_rank*dim
where dim is the array size. Both my_rank and dim are normal integers and
here lies the rub.
Fortran (or should I say gfortran?) multiplies my_rank*dim in integer*4
and then converts to integ