Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

From: Jeff Squyres (jsquyres_at_[hidden])
Date: 2006-10-10 07:40:44


Note that p4_error messages are typically associated with the MPICH
implementation of MPI, which is a totally different code base and set of
developers. We cannot answer questions about MPICH on this list; you
probably want to post your question to the MPICH support list.

Good luck.

On 10/9/06 5:45 PM, "Vadivelan Ranjith" <achillesvelan_at_[hidden]> wrote:

> Hi
> I thank you for helping to all.
> Today i got a error message by sumbitting job. First i
> ran the code
> using explict method. I got result accurately, and no
> problem occured when
> i sumbit job. Now i changed my code to implict method.
> I got error when
> i sumbit job.
> I checked correctly, it reading all files and
> iteration starts. after
> one iteration it gives the following error. The same
> code is running on
> other machine, giving result correctly. So please help
> me how to fix
> it.
>
> Advance thanks
> Velan
>
> ----------------------------------------------------------------
> job.e file:
> p4_error: latest msg from perror: Bad file
> descriptor
> p4_error: latest msg from perror: Bad file
> descriptor
> p4_error: latest msg from perror: Bad file
> descriptor
> p4_error: latest msg from perror: Bad file
> descriptor
> -----------------------------------------------------------------
> job.o file:
> 3
> node18.local
> node19.local
> node17.local
> # Allocating 5 nodes to block 1
> # Allocating 1 nodes to block 2
> # Require mxb >= 97
> # Require mxa >= 26 mya >= 97 and mza >= 75
> # Maximum load imbalance = 71.69%
> # Navier-Stokes Simulation
> # Implicit Full Matrix DP-LUR
> # Reading restart files...( 0.34 seconds)
> # Freestream Mach Number = 6.50
>
> 1 0.3670E+01 0.7803E+05 16 15 7 2
> 0.1222E-08
> p5_2609: p4_error: interrupt SIGx: 13
> bm_list_17559: (3.666982) wakeup_slave: unable to
> interrupt slave 0 pid
> 17542
> rm_l_1_18696: (2.738297) net_send: could not write to
> fd=6, errno = 9
> rm_l_1_18696: p4_error: net_send write: -1
> rm_l_2_2605: (2.614927) net_send: could not write to
> fd=6, errno = 9
> rm_l_4_18718: (2.373120) net_send: could not write to
> fd=6, errno = 9
> rm_l_4_18718: p4_error: net_send write: -1
> rm_l_2_2605: p4_error: net_send write: -1
> rm_l_3_17584: (2.496277) net_send: could not write to
> fd=6, errno = 9
> rm_l_3_17584: p4_error: net_send write: -1
> rm_l_5_2626: (2.249144) net_send: could not write to
> fd=5, errno = 32
> p5_2609: (2.251356) net_send: could not write to fd=5,
> errno = 32
> -------------------------------------------------------------------
> job file:
> #!/bin/bash
> #PBS -l nodes=3:ppn=1
>
> cd $PBS_O_WORKDIR
> n=`/usr/local/bin/pbs.py $PBS_NODEFILE hosts`
> echo $n
> cat hosts
> /opt/mpich/intel/bin/mpirun -nolocal -machinefile
> hosts -np 6 pg3d.exe
> -------------------------------------------------------------------
> Machine configuration:
> CPU: Intel(R) Dual Processor Xeon(R) CPU 3.2GHz
> Installation using rocks4.1
>
>
>
>
>
> __________________________________________________________
> Yahoo! India Answers: Share what you know. Learn something new
> http://in.answers.yahoo.com/
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users

-- 
Jeff Squyres
Server Virtualization Business Unit
Cisco Systems