This web mail archive is frozen.
This page is part of a frozen web archive of this mailing list.
You can still navigate around this archive, but know that no new mails
have been added to it since July of 2016.
Click here to be taken to the new web archives of this list; it includes all the mails that are in this frozen archive plus all new mails that have been sent to the list since it was migrated to the new archives.
I am running into a problem with a simple program (which performs
several MPI_Bcast operations) hanging. Most processes hang in
MPI_Finalize, the others hang in MPI_Bcast. Interestingly enough,
this only happens when I oversubscribe the nodes. For instance, using
IU's Odin cluster, I take 4 nodes (each has two Opteron processors)
and run 9 processes:
mpirun -np 9 ./a.out
The backtrace from 7/9 of the processes shows that they're in
#0 0x0000003d1b92e813 in sigprocmask () from /lib64/tls/libc.so.6
#1 0x0000002a9598f55f in poll_dispatch ()
#2 0x0000002a9598e3f3 in opal_event_loop ()
#3 0x0000002a960487c4 in mca_oob_tcp_msg_wait ()
#4 0x0000002a9604ca13 in mca_oob_tcp_recv ()
#5 0x0000002a9585d833 in mca_oob_recv_packed ()
#6 0x0000002a9585dd37 in mca_oob_xcast ()
#7 0x0000002a956cbfb0 in ompi_mpi_finalize ()
#8 0x000000000040bd3e in main ()
The other two processes are in MPI_Bcast:
#0 0x0000002a97c2cbe3 in mca_btl_mvapi_component_progress ()
#1 0x0000002a97b21072 in mca_bml_r2_progress ()
#2 0x0000002a95988a4a in opal_progress ()
#3 0x0000002a97a13fe7 in mca_pml_ob1_recv ()
#4 0x0000002a9846d0aa in ompi_coll_tuned_bcast_intra_chain ()
#5 0x0000002a9846d100 in ompi_coll_tuned_bcast_intra_pipeline ()
#6 0x0000002a9846a3d7 in ompi_coll_tuned_bcast_intra_dec_fixed ()
#7 0x0000002a956deae3 in PMPI_Bcast ()
#8 0x000000000040bcc7 in main ()
Other random information:
- The two processes stuck in MPI_Bcast are not on the same node.
This has been the case both times I've gone through the backtraces,
but I can't conclude that it's a necessary condition.
- If I force the use of the "basic" MCA for collectives, this
problem does not occur.
- If I don't oversubscribe the nodes, things seem to work properly.
- The C++ program source and result of ompi_info are attached
This should be easy to reproduce for anyone with access to Odin. I'm
using Open MPI 1.1 configured with no special options. It is
available as the module "mpi/openmpi-1.1-gcc" on the cluster. I'm
using SLURM interactively to allocate the nodes before executing mpirun:
srun -A -N 4