The PATH variable contains /home/hx019035/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/ bin:/usr/games:/usr/local/maui/bin/:
/home/hx019035/bin contains the local installation of OMPI 1.3.3 The LD_LIBRARY_PATH variable contains /home/hx019035/lib: These variables are being set in the .profile file on the hpcc00 node. Would there be a change anywhere else? From: Ralph Castain [mailto:rhc.open...@gmail.com] On Behalf Of Ralph Castain Sent: 07 October 2009 13:32 To: Blesson Varghese Subject: Re: [OMPI users] MPI_Comm_accept()/connect() errors Yes, it does. But the error message indicates a 1.2 version is running on hpcc00. On Oct 7, 2009, at 5:46 AM, Blesson Varghese wrote: Just a quick question. Would mpirun -version give me the version of the mpirun being executed? I am getting the result of that as 1.3.3. From: Ralph Castain [mailto:rhc.open...@gmail.com] On Behalf Of Ralph Castain Sent: 07 October 2009 11:58 To: Blesson Varghese Subject: Re: [OMPI users] MPI_Comm_accept()/connect() errors Hate to tell you this, but your output clearly indicates you are NOT running 1.3.3 - that is an output from a 1.2.x version of OMPI. Check you path and ld_library_path - you're still picking up the 1.2.5 version somewhere. On Oct 7, 2009, at 4:05 AM, Blesson Varghese wrote: Hi, Please refer to the emails below. I have made an upgrade to Open MPI 1.3.3 as suggested. The necessary environment variables have all been set. Attaching the output of ompi_info -all. However, the errors continue to persist. [hpcc00:31864] [0,0,0] ORTE_ERROR_LOG: Not found in file dss/dss_unpack.c at line 209 [hpcc00:31864] [0,0,0] ORTE_ERROR_LOG: Not found in file communicator/comm_dyn.c at line 186 [hpcc00:31864] *** An error occurred in MPI_Comm_connect [hpcc00:31864] *** on communicator MPI_COMM_WORLD [hpcc00:31864] *** MPI_ERR_INTERN: internal error [hpcc00:31864] *** MPI_ERRORS_ARE_FATAL (goodbye) The server program is as follows: #include <mpi.h> #include <stdio.h> #include <stdlib.h> int main( int argc, char **argv ) { MPI_Comm client; MPI_Status status; char port_name[MPI_MAX_PORT_NAME]; int buf; int size, again; MPI_Info portInfo; MPI_Init( &argc, &argv ); MPI_Comm_size(MPI_COMM_WORLD, &size); MPI_Open_port(MPI_INFO_NULL, port_name); printf("server available at %s\n",port_name); MPI_Comm_accept(port_name, MPI_INFO_NULL, 0, MPI_COMM_WORLD, &client ); MPI_Recv(&buf, 1, MPI_INT, MPI_ANY_SOURCE, MPI_ANY_TAG, client, &status ); MPI_Comm_disconnect( &client ); } The client program is as follows: #include <mpi.h> #include <stdlib.h> #include <stdio.h> #include <string.h> int main( int argc, char **argv ) { MPI_Comm server; int buf = 8; char port_name[MPI_MAX_PORT_NAME]; MPI_Info portInfo; MPI_Init( &argc, &argv ); strcpy(port_name, "0.0.0:2000"); //The port name is hardcoded since 0.0.0:2000 is generated by the server program MPI_Comm_connect(port_name, MPI_INFO_NULL, 0, MPI_COMM_WORLD, &server ); MPI_Send(&buf, 1, MPI_INT, 0, 1, server ); MPI_Comm_disconnect( &server ); MPI_Finalize(); return 0; } Would you please advise? Regards, Blesson. -----Original Message----- From: Blesson Varghese [mailto:hx019...@reading.ac.uk] Sent: 03 October 2009 12:20 To: 'Jeff Squyres' Subject: RE: [OMPI users] MPI_Comm_accept()/connect() errors Thank you. I shall try the upgrade very soon. -----Original Message----- From: Jeff Squyres [mailto:jsquy...@cisco.com] Sent: 03 October 2009 12:18 To: Blesson Varghese Subject: Re: [OMPI users] MPI_Comm_accept()/connect() errors On Oct 3, 2009, at 7:14 AM, Blesson Varghese wrote: > Thanks for your reply Jeff. Since, it is a teaching cluster of the > University, I am quite unsure if I would be able to upgrade it very > soon. > > Do you reckon that the error is due to the Open MPI version? > You can always install your own version of Open MPI under your $HOME or somesuch -- there is no requirement that Open MPI is installed by root in a central location. That being said, you might want to check with your administrator to ensure that this is ok with local policies -- see if they did any special setup for Open MPI, etc. But yes, we made a bunch of COMM_SPAWN improvements since the 1.2 series. -- Jeff Squyres jsquy...@cisco.com From: Blesson Varghese [mailto:hx019...@reading.ac.uk] Sent: 01 October 2009 12:01 To: 'Open MPI Users'; 'Ralph Castain' Subject: RE: [OMPI users] MPI_Comm_accept()/connect() errors The following is the information regarding the error. I am running Open MPI 1.2.5 on Ubuntu 4.2.4, kernel version 2.6.24 I ran the server program as mpirun -np 1 server. This program gave me the output port as 0.1.0:2000. I used this port name value as the command line argument for the client program: mpirun -np 1 client 0.1.1:2000. - The output of the "ompi_info --all" is attached with the email - PATH Variable: /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr /local/maui/bin/: - LD_LIBRARY_PATH variable was empty - The following is the output of ifconfig on hpcc00 from where the error has been generated: eth0 Link encap:Ethernet HWaddr 00:12:3f:4c:2d:78 inet addr:134.225.200.100 Bcast:134.225.200.255 Mask:255.255.255.0 inet6 addr: fe80::212:3fff:fe4c:2d78/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:15912728 errors:0 dropped:0 overruns:0 frame:0 TX packets:15312376 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:2951880321 (2.7 GB) TX bytes:2788249498 (2.5 GB) Interrupt:16 lo Link encap:Local Loopback inet addr:127.0.0.1 Mask:255.0.0.0 inet6 addr: ::1/128 Scope:Host UP LOOPBACK RUNNING MTU:16436 Metric:1 RX packets:3507489 errors:0 dropped:0 overruns:0 frame:0 TX packets:3507489 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:1794266658 (1.6 GB) TX bytes:1794266658 (1.6 GB) Regards, Blesson. From: users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] On Behalf Of Ralph Castain Sent: 29 September 2009 23:59 To: Open MPI Users Subject: Re: [OMPI users] MPI_Comm_accept()/connect() errors I will ask the obvious - what version of Open MPI are you running? In what environment? What was your command line? :-) On Sep 29, 2009, at 3:50 PM, Blesson Varghese wrote: Hi, I have been trying to execute the server.c and client.c program provided in http://www.mpi-forum.org/docs/mpi21-report/node213.htm#Node213, using accept() and connect() function in MPI. However, the following errors are generated. [hpcc00:16522] *** An error occurred in MPI_Comm_connect [hpcc00:16522] *** on communicator MPI_COMM_WORLD [hpcc00:16522] *** MPI_ERR_INTERN: internal error [hpcc00:16522] *** MPI_ERRORS_ARE_FATAL (goodbye) Could anybody please help me? Many thanks, Blesson. _______________________________________________ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users <outputompi.txt>