i also made machine file which contain ip adresses of all compute nodes + .ncbirc file for path to mpiblast and shared ,local storage path.... Sir I ran the same command of mpirun on my college supercomputer 8 nodes each having 24 processors but it just running....gave no result uptill 3 hours...
On Thu, Apr 3, 2014 at 10:39 PM, Nisha Dhankher -M.Tech(CSE) < nishadhankher-coaese...@pau.edu> wrote: > i first formatted my database with mpiformatdb command then i ran command : > mpirun -np 64 -machinefile mf mpiblast -d all.fas -p blastn -i query.fas > -o output.txt > but then it gave this error 113 from some hosts and continue to run for > other but with results even after 2 hours lapsed.....on rocks 6.0 cluster > with 12 virtual nodes on pc's ...2 on each using virt-manger , 1 gb ram to > each > > > > On Thu, Apr 3, 2014 at 8:37 PM, Ralph Castain <r...@open-mpi.org> wrote: > >> I'm having trouble understanding your note, so perhaps I am getting this >> wrong. Let's see if I can figure out what you said: >> >> * your perl command fails with "no route to host" - but I don't see any >> host in your cmd. Maybe I'm just missing something. >> >> * you tried running a couple of "mpirun", but the mpirun command wasn't >> recognized? Is that correct? >> >> * you then ran mpiblast and it sounds like it successfully started the >> processes, but then one aborted? Was there an error message beyond just the >> -1 return status? >> >> >> On Apr 2, 2014, at 11:17 PM, Nisha Dhankher -M.Tech(CSE) < >> nishadhankher-coaese...@pau.edu> wrote: >> >> error btl_tcp_endpint.c: 638 connection failed due to error >> 113<http://biosupport.se/questions/696/error-btl_tcp_endpintc-638-connection-failed-due-to-error-113> >> >> In openmpi: this error came when i run my mpiblast program on rocks >> cluster.Connect to hosts failed on ip 10.1.255.236,10.1.255.244 . And when >> i run following command linux_shell$ perl -e 'die$!=113' this msg comes: >> "No route to host at -e line 1." shell$ mpirun --mca btl ^tcp shell$ mpirun >> --mca btl_tcp_if_include eth1,eth2 shell$ mpirun --mca btl_tcp_if_include >> 10.1.255.244 was also executed but it did nt recognized these >> commands....nd aborted.... what should i do...? When i run my mpiblast >> program for the frst time then it give mpi_abort error...bailing out of >> signal -1 on rank 2 processor...then i removed my public ethernet >> cable....and then give btl_tcp endpint error 113.... >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> > >