Note that p4_error messages are typically associated with the MPICH implementation of MPI, which is a totally different code base and set of developers. We cannot answer questions about MPICH on this list; you probably want to post your question to the MPICH support list.
Good luck. On 10/9/06 5:45 PM, "Vadivelan Ranjith" <achillesve...@yahoo.co.in> wrote: > Hi > I thank you for helping to all. > Today i got a error message by sumbitting job. First i > ran the code > using explict method. I got result accurately, and no > problem occured when > i sumbit job. Now i changed my code to implict method. > I got error when > i sumbit job. > I checked correctly, it reading all files and > iteration starts. after > one iteration it gives the following error. The same > code is running on > other machine, giving result correctly. So please help > me how to fix > it. > > Advance thanks > Velan > > ---------------------------------------------------------------- > job.e file: > p4_error: latest msg from perror: Bad file > descriptor > p4_error: latest msg from perror: Bad file > descriptor > p4_error: latest msg from perror: Bad file > descriptor > p4_error: latest msg from perror: Bad file > descriptor > ----------------------------------------------------------------- > job.o file: > 3 > node18.local > node19.local > node17.local > # Allocating 5 nodes to block 1 > # Allocating 1 nodes to block 2 > # Require mxb >= 97 > # Require mxa >= 26 mya >= 97 and mza >= 75 > # Maximum load imbalance = 71.69% > # Navier-Stokes Simulation > # Implicit Full Matrix DP-LUR > # Reading restart files...( 0.34 seconds) > # Freestream Mach Number = 6.50 > > 1 0.3670E+01 0.7803E+05 16 15 7 2 > 0.1222E-08 > p5_2609: p4_error: interrupt SIGx: 13 > bm_list_17559: (3.666982) wakeup_slave: unable to > interrupt slave 0 pid > 17542 > rm_l_1_18696: (2.738297) net_send: could not write to > fd=6, errno = 9 > rm_l_1_18696: p4_error: net_send write: -1 > rm_l_2_2605: (2.614927) net_send: could not write to > fd=6, errno = 9 > rm_l_4_18718: (2.373120) net_send: could not write to > fd=6, errno = 9 > rm_l_4_18718: p4_error: net_send write: -1 > rm_l_2_2605: p4_error: net_send write: -1 > rm_l_3_17584: (2.496277) net_send: could not write to > fd=6, errno = 9 > rm_l_3_17584: p4_error: net_send write: -1 > rm_l_5_2626: (2.249144) net_send: could not write to > fd=5, errno = 32 > p5_2609: (2.251356) net_send: could not write to fd=5, > errno = 32 > ------------------------------------------------------------------- > job file: > #!/bin/bash > #PBS -l nodes=3:ppn=1 > > cd $PBS_O_WORKDIR > n=`/usr/local/bin/pbs.py $PBS_NODEFILE hosts` > echo $n > cat hosts > /opt/mpich/intel/bin/mpirun -nolocal -machinefile > hosts -np 6 pg3d.exe > ------------------------------------------------------------------- > Machine configuration: > CPU: Intel(R) Dual Processor Xeon(R) CPU 3.2GHz > Installation using rocks4.1 > > > > > > __________________________________________________________ > Yahoo! India Answers: Share what you know. Learn something new > http://in.answers.yahoo.com/ > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres Server Virtualization Business Unit Cisco Systems