Note that p4_error messages are typically associated with the MPICH
implementation of MPI, which is a totally different code base and set of
developers.  We cannot answer questions about MPICH on this list; you
probably want to post your question to the MPICH support list.

Good luck.


On 10/9/06 5:45 PM, "Vadivelan Ranjith" <achillesve...@yahoo.co.in> wrote:

> Hi
> I thank you for helping to all.
> Today i got a error message by sumbitting job. First i
> ran the code 
> using explict method. I got result accurately, and no
> problem occured when
> i sumbit job. Now i changed my code to implict method.
> I got error when 
> i sumbit job.
> I checked correctly, it reading all files and
> iteration starts. after
> one iteration it gives the following error. The same
> code is running on
> other machine, giving result correctly. So please help
> me how to fix 
> it.
> 
> Advance thanks
> Velan
> 
> ----------------------------------------------------------------
> job.e file:
>     p4_error: latest msg from perror: Bad file
> descriptor
>     p4_error: latest msg from perror: Bad file
> descriptor
>     p4_error: latest msg from perror: Bad file
> descriptor
>     p4_error: latest msg from perror: Bad file
> descriptor
> -----------------------------------------------------------------
> job.o file:
> 3
> node18.local
> node19.local
> node17.local
> # Allocating   5 nodes to block  1
> # Allocating   1 nodes to block  2
> # Require mxb >=   97
> # Require mxa >=   26 mya >=   97 and mza >=   75
> # Maximum load imbalance =  71.69%
> # Navier-Stokes Simulation
> # Implicit Full Matrix DP-LUR
> # Reading restart files...( 0.34 seconds)
> # Freestream Mach Number =  6.50
> 
>  1   0.3670E+01   0.7803E+05   16   15    7    2
> 0.1222E-08
> p5_2609:  p4_error: interrupt SIGx: 13
> bm_list_17559: (3.666982) wakeup_slave: unable to
> interrupt slave 0 pid
> 17542
> rm_l_1_18696: (2.738297) net_send: could not write to
> fd=6, errno = 9
> rm_l_1_18696:  p4_error: net_send write: -1
> rm_l_2_2605: (2.614927) net_send: could not write to
> fd=6, errno = 9
> rm_l_4_18718: (2.373120) net_send: could not write to
> fd=6, errno = 9
> rm_l_4_18718:  p4_error: net_send write: -1
> rm_l_2_2605:  p4_error: net_send write: -1
> rm_l_3_17584: (2.496277) net_send: could not write to
> fd=6, errno = 9
> rm_l_3_17584:  p4_error: net_send write: -1
> rm_l_5_2626: (2.249144) net_send: could not write to
> fd=5, errno = 32
> p5_2609: (2.251356) net_send: could not write to fd=5,
> errno = 32
> -------------------------------------------------------------------
> job file:
> #!/bin/bash
> #PBS -l nodes=3:ppn=1
> 
> cd $PBS_O_WORKDIR
> n=`/usr/local/bin/pbs.py $PBS_NODEFILE hosts`
> echo $n
> cat hosts
> /opt/mpich/intel/bin/mpirun -nolocal -machinefile
> hosts -np 6 pg3d.exe
> -------------------------------------------------------------------
> Machine configuration:
>  CPU: Intel(R) Dual Processor Xeon(R) CPU 3.2GHz
> Installation using rocks4.1
> 
> 
> 
> 
> 
> __________________________________________________________
> Yahoo! India Answers: Share what you know. Learn something new
> http://in.answers.yahoo.com/
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


-- 
Jeff Squyres
Server Virtualization Business Unit
Cisco Systems

Reply via email to