It sounds to me like TCP communication isn't getting through for some
reason. Try the following:
mpirun --mca plm_base_verbose 5 --hostfile myh3 -pernode hostname
You should see output from the receipt of a daemon callback for each
daemon, the the sending of the launch command. My guess is t
Many thanks. The firewall is the issue.
On Feb 9, 2009, at 5:56 AM, Ralph Castain wrote:
It sounds to me like TCP communication isn't getting through for
some reason. Try the following:
mpirun --mca plm_base_verbose 5 --hostfile myh3 -pernode hostname
black@ccn3:~/Documents/mp> mpirun --mca
I'm trying to run a small "proof of concept" program using OpenMPI 1.3. I
am using Solaris 8 with Sparc processors across 2 nodes. It appears that
the MPI_Reduce function is hanging. If I run the same program with only 4
instances on 1 node , or 2 instances on 2 nodes, it works fine. The proble
OK, I think I got things sorted out. Thanx for your inputs.
Ted
--- On Fri, 2/6/09, Jeff Squyres wrote:
From: Jeff Squyres
Subject: Re: [OMPI users] Global Communicator
To: "Open MPI Users"
Cc: ted...@wag.caltech.edu
List-Post: users@lists.open-mpi.org
Date: Friday, February 6, 2009, 2:46 PM