From: Mohamad Chaarawi (mschaara_at_[hidden])
Date: 2007-09-21 13:30:32


Hey all,

I'm trying to execute the collective bakeoff tests for OMPI, MPICH2, and
MVAPICH2. OMPI and MPICH2 are working out fine, However when MVAPICH2 is
configuring, it gives an error with the C compiler, pasted at the end..

Note that I get the error when im running the mtt client. When i go in
the scratch directory to the MVAPICH2 sources and configure it myself,
with the same configure arguments that it did from config.log, it works
out fine..

Ive been banging my head a while now to figure this out, but i got
nowhere. Probably it's some environment settings being messed up
somewhere, but i don't know..
If anyone has stumbled upon this before, let me know..
I attached my ini file..

OUT:Configuring MVAPICH2...
OUT:Configuring MPICH2 version MVAPICH2-0.9.8 with
   --prefix=/home/mschaara/mtt-testing/scratch-coll/installs/NhjQ/install
   --with-device=osu_ch3:mrail --with-rdma=gen2 --with-pm=mpd
   --disable-romio --without-mpe
OUT:sourcing
   /home/mschaara/mtt-testing/scratch-coll/installs/NhjQ/src/mvapich2-0.9.8
   p3/src/pm/mpd/setup_pm
OUT:checking for gcc... OUT:gcc
OUT:checking for C compiler default output file name... OUT:configure:
error: C compiler cannot create executables
   See `config.log' for more details.
OUT:Configuring MPICH2 version MVAPICH2-0.9.8 with
   --prefix=/home/mschaara/mtt-testing/scratch-coll/installs/NhjQ/install
   --with-device=osu_ch3:mrail --with-rdma=gen2 --with-pm=mpd
   --disable-romio --without-mpe
   sourcing
   /home/mschaara/mtt-testing/scratch-coll/installs/NhjQ/src/mvapich2-0.9.8
   p3/src/pm/mpd/setup_pm
   checking for gcc... gcc
   checking for C compiler default output file name... configure: error: C
   compiler cannot create executables
   See `config.log' for more details.
OUT:Failure in configuration.
   Please file an error report to mvapich-discuss_at_[hidden] with
   all your log files.
Command complete, exit status: 1

-- 
Mohamad Chaarawi
Instructional Assistant		  http://www.cs.uh.edu/~mschaara
Department of Computer Science	  University of Houston
4800 Calhoun, PGH Room 526        Houston, TX 77204, USA

#======================================================================
# Generic OMPI core performance testing template configuration
#======================================================================

[MTT]
# Leave this string so that we can identify this data subset in the
# database
# OMPI Core: Use a "test" label until we're ready to run real results
description = [testbake]
#description = [2007 collective performance bakeoff]
# OMPI Core: Use the "trial" flag until we're ready to run real results
trial = 1

# Put other values here as relevant to your environment.

#----------------------------------------------------------------------

[Lock]
# Put values here as relevant to your environment.

#======================================================================
# MPI get phase
#======================================================================

[MPI get: MVAPICH2]
mpi_details = MVAPICH2

module = Download
download_url = http://mvapich.cse.ohio-state.edu/download/mvapich2/mvapich2-0.9.8p3.tar.gz

#======================================================================
# Install MPI phase
#======================================================================

[MPI install: MVAPICH2]
mpi_get = mvapich2
save_stdout_on_success = 1
merge_stdout_stderr = 0
# Adjust this for your site (this is what works at Cisco). Needed to
# launch in SLURM; adding this to LD_LIBRARY_PATH here propagates this
# all the way through the test run phases that use this MPI install,
# where the test executables will need to have this set.
prepend_path = LD_LIBRARY_PATH /opt/SLURM/lib

module = MVAPICH2
# Adjust this to be where your OFED is installed
mvapich2_setenv = OPEN_IB_HOME /usr/local/ofed
mvapich2_build_script = make.mvapich2.ofa
mvapich2_compiler_name = gnu
mvapich2_compiler_version = &get_gcc_version()
# These are needed to launch through SLURM; adjust as appropriate.
mvapich2_additional_wrapper_ldflags = -L/opt/SLURM/lib
mvapich2_additional_wrapper_libs = -lpmi

#======================================================================
# MPI run details
#======================================================================

[MPI Details: MVAPICH2]

# Launching through SLURM. If you use mpdboot instead, you need to
# ensure that multiple mpd's on the same node don't conflict (or never
# happen).
exec = srun @alloc@ -n &test_np() &test_executable() &test_argv()

# If not using SLURM, you'll need something like this (not tested).
# You may need different hostfiles for running by slot/by node.
#exec = mpiexec -np &test_np() --host &hostlist() &test_executable()

network = loopback,verbs,shmem

# In this example, each node has 4 CPUs, so to launch "by node", just
# specify that each MPI process will use 4 CPUs.
alloc = &if(&eq(&test_alloc(), "node"), "-c 2", "")

#======================================================================
# Test get phase
#======================================================================

[Test get: skampi]
module = SVN
svn_url = https://svn.open-mpi.org/svn/ompi-tests/trunk/skampi-5.0.1

#======================================================================
# Test build phase
#======================================================================

[Test build: skampi]
test_get = skampi
save_stdout_on_success = 1
merge_stdout_stderr = 1
stderr_save_lines = 100

module = Shell
# Set EVERYONE_CAN_MPI_IO for HP MPI
shell_build_command = <<EOT
make CFLAGS="-O2 -DPRODUCE_SPARSE_OUTPUT -DEVERYONE_CAN_MPI_IO"
EOT

#======================================================================
# Test Run phase
#======================================================================

[Test run: skampi]
test_build = skampi
pass = &and(&cmd_wifexited(), &eq(&cmd_wexitstatus(), 0))
# Timeout hueristic: 10 minutes
timeout = 10:00
save_stdout_on_pass = 1
# Ensure to leave this value as "-1", or performance results could be lost!
stdout_save_lines = -1
merge_stdout_stderr = 1
np = &env_max_procs()

argv = -i &find("mtt_.+.ski", "input_files_bakeoff")

specify_module = Simple
analyze_module = SKaMPI

simple_pass:tests = skampi

#======================================================================
# Reporter phase
#======================================================================

[Reporter: IU database]
module = MTTDatabase

mttdatabase_realm = OMPI
mttdatabase_url = https://www.open-mpi.org/mtt/submit/
# Change this to be the username and password for your submit user.
# Get this from the OMPI MTT administrator.
mttdatabase_username = uh
mttdatabase_password = &stringify(&cat("/home/mschaara/mtt-testing/mtt-db-password.txt"))
# Change this to be some short string identifying your cluster.
mttdatabase_platform = shark

mttdatabase_debug_filename = mttdb_debug_file_perf
mttdatabase_keep_debug_files = 1

#----------------------------------------------------------------------

# This is a backup reporter; it also writes results to a local text
# file

[Reporter: text file backup]
module = TextFile

textfile_filename = $phase-$section-$mpi_name-$mpi_version.txt

textfile_summary_header = <<EOT
Hostname: &shell("hostname")
uname: &shell("uname -a")
Username: &shell("who am i")
EOT

textfile_summary_footer =
textfile_detail_header =
textfile_detail_footer =

textfile_textwrap = 78