On Apr 23, 2009, at 3:59 PM, Luis Vitorio Cargnini wrote:
I'm using NFS, my home dir is the same in all nodes the problem is
when generating the key it is been generated for a specific machine
end of the key is the user@host, the system is consulting id_dsa in
each machine.
That's ok. I have a similar setup: svbu-mpi is my cluster "head node"
and that's where I generated my DSA key. So my id_dsa.pub file looks
like this:
[13:05] svbu-mpi:~/hg % cat ~/.ssh/id_dsa.pub
ssh-dss
AAAAB3NzaC1kc3MAAACBAPhFvzoDPw1da2aYf2PCW9sQfOT4SYmvI5EYfJvJXyyVLs7C
+ETY5Zma7js2PCfk4kgHUVJQgglP5V/Dp9uBjgP/zpNdOWbP
+chULEXaz0HKOV3NZM5BH6oBRTSGTZh4DhqnQjotQsp6gi9LZ+GGl00tzc
+
EzlfqIfSuKHQjSTADAAAAFQCM1AbE8Z7+mcCzFpNUAa7eLBFOhQAAAIEAjMEiDNceRdvMjf
+Of1nwaMb8ndx/w4ltEH67P0g2xn8PfJP56rYn7ffiEuB5Ndu
+iLskII5CkDwLZOmv4nP32gNzxxyo23Qbnd88a+BYe+j9yu35czqvPzxHBKlP5t0zaeZQt/
fXr/VKd1P9OhZKMVmGZm1m2Yn5M21d16V1j4QAAACBALe2hbtgzqSMSVyX7ED31MfJsYxW/
y01VH9f7Ot+WfJrpTsTRTWMYb6x1jTAozC/DvZlx/KPKiekQH
+ApkfL1e6TSlug1Y5Kv9zCvXwEAbgwHEwUoWvTT
+IpBwD318AjraZtJXlIb03tkX7l2gZNncwOmzFbwqGwypD3YtHAY3j1 jsquyres@svbu-
mpi
[13:05] svbu-mpi:~/hg %
And that same $HOME/.ssh/id_dsa.pub (and corresponding $HOME/.ssh/
id_dsa) file is available on all my nodes via NFS. The email address
at the end is not really part of the key; it's just there for human
reference for you to remember where it came from. It doesn't affect
the authentication at all.
so to fix the problem since my applications are launch from node srv0
I just create the keys in node 0 and that is it start to work in to
connect in the others node, the problem is the reverse path I can't
access from srv1 srv0 for example.
Why not? If you copy your id_dsa.pub file to authorized_keys, it
should Just Work (assuming the permissions are all set correctly:
- $HOME/.ssh needs, owned by you, 0700
- $HOME/.ssh/authorized_keys owned by you, 0600
- $HOME/.ssh/id_dsa.pub owned by you, 0644
- $HOME/.ssh/id_dsa owned by you, 0600
The SSH setup HOWTOs and recipes sent in this thread (I assume) must
talk about such things..?
The point is working from node0, the connections trough ssh. Now the
execution it start but do not stop, like keep running ad infinitum,
any ideas ?
mpirun -d -v -hostfile chosts -np 35 ~/mpi/hello
[cluster-srv0:29466] procdir: /tmp/openmpi-sessions-
lvcargnini@cluster-
srv0_0/44411/0/0
[snipped]
Are you able to run non-MPI apps through mpirun? For example:
mpirun -d -v -hostfile chosts hostname | sort
If that works, then did you compile "hello" correctly (e.g., with
mpicc)? I assume this is a simple "hello world" kind of MPI program
-- calls MPI_INIT, maybe MPI_COMM_RANK and MPI_COMM_SIZE, and
MPI_FINALIZE?
Do you have TCP firewalling disabled on all of your cluster nodes?
--
Jeff Squyres
Cisco Systems