Open MPI logo

MTT Users Mailing List Archives

  |   Home   |   Support   |   FAQ   |  

This web mail archive is frozen.

This page is part of a frozen web archive of this mailing list.

You can still navigate around this archive, but know that no new mails have been added to it since July of 2016.

Click here to be taken to the new web archives of this list; it includes all the mails that are in this frozen archive plus all new mails that have been sent to the list since it was migrated to the new archives.

Subject: Re: [MTT users] MTT:Error while executing ompi-core-perf-testing.ini
From: Jeff Squyres (jsquyres_at_[hidden])
Date: 2010-01-04 11:36:58


Vishal --

We'd love to have Chelsio's results included in our database. Contact me off-list if you'd like a username/password for the database (I'm 99% sure that a "chelsio" user already exists; I can reset the password if you'd like).

On Jan 4, 2010, at 10:20 AM, Ethan Mallove wrote:

> Hi Vishal,
>
> The debug output shows that your MTT run aborted because you do not
> have a username/passwd for the open-mpi.org/mtt/submit URL. To skip
> the IU Database part, try running with --no-section iu.database, e.g.,
>
> # cat ompi-core-perf-testing.ini | ../client/mtt --no-section iu.database --debug -
>
> Thanks,
> Ethan
>
> On Mon, Jan/04/2010 05:08:36PM, vishal shorghar wrote:
> > Hi All,
> >
> > Today I ran ompi-core-perf-test script as "# cat
> > ompi-core-perf-testing.ini | ../client/mtt --debug -"
> > It gave me following error and exits .
> >
> > [root_at_veera samples]# cat ompi-core-perf-testing.ini | ../client/mtt
> > --debug -
> > Debug is 1, Verbose is 1
> > *** MTT: ../client/mtt --debug -
> > *** Running on veera
> > Chdir ../client
> > Chdir /root/mtt-svn/samples
> > Copying: stdin to /tmp/dw1zvwGZEa.ini
> > Expanding include_file(s) parameters in
> > /tmp/dw1zvwGZEa.ini Reading ini file: stdin
> > Validating INI inifile:
> > /tmp/bH9q7NY7Hl.ini FilterINI: Final
> > list of sections: [mtt]
> >
> > [lock]
> > [mpi get: ompi-nightly-trunk]
> > [mpi get: ompi-nightly-v1.2]
> > [mpi get: ompi-released-v1.2]
> > [mpi get: mpich1]
> > [mpi get: mpich-mx]
> > [mpi get: mpich2]
> > [mpi get: mvapich1]
> > [mpi get: mvapich2]
> > [mpi get: hp mpi]
> > [mpi get: intel mpi]
> > [skip mpi get: scali mpi]
> > [skip mpi get: cray mpi]
> > [mpi install: ompi/gnu-standard]
> > [mpi install: mpich1]
> > [mpi install: mpich2]
> > [mpi install: mvapich1]
> > [mpi install: mvapich2]
> > [mpi install: intel mpi]
> > [mpi install: hp mpi]
> > [mpi details: ompi]
> > [mpi details: mpich1]
> > [mpi details: mpich2]
> > [mpi details: mvapich1]
> > [mpi details: mvapich2]
> > [mpi details: mpich-mx]
> > [mpi details: intel
> > mpi] [mpi details: hp
> > mpi] [test get:
> > netpipe] [test
> > get: osu]
> > [test get: imb]
> > [test get: skampi]
> > [test get: nbcbench]
> > [test build: netpipe]
> > [test build: osu]
> > [test build: imb]
> > [test build: skampi]
> > [test build: nbcbench]
> > [test run: netpipe]
> > [test run: osu]
> > [test run: imb]
> > [test run: skampi]
> > [test run: nbcbench]
> > [reporter: iu database]
> > [reporter: text file backup]
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT
> > scratch Value returning:
> > scratch: .
> > scratch resolved:
> > /root/mtt-svn/samples Chdir
> > /root/mtt-svn/samples
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT fast_scratch
> > Value returning:
> > fast_scratch: /root/mtt-svn/samples
> > fast_scratch resolved: /root/mtt-svn/samples
> > *** Main scratch tree: /root/mtt-svn/samples
> > *** Fast scratch tree: /root/mtt-svn/samples
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT force
> > Value returning:
> > Making dir: /root/mtt-svn/samples/sources (cwd:
> > /root/mtt-svn/samples) Making dir: /root/mtt-svn/samples/installs
> > (cwd: /root/mtt-svn/samples) Making dir:
> > /root/mtt-svn/samples/test_runs (cwd: /root/mtt-svn/samples) Value
> > got: Config::IniFiles=HASH(0xf872ad0) MTT funclet_files
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT hostfile
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT hostlist
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT max_np
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT
> > textwrap Value returning:
> > Value got:
> > Config::IniFiles=HASH(0xf872ad0) MTT drain_timeout Value
> > returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT trim_save_successful
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT trim_save_failed
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT trial
> > Value returning: 1
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT
> > http_proxy Value returning:
> > Value got:
> > Config::IniFiles=HASH(0xf872ad0) MTT https_proxy Value
> > returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT ftp_proxy
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT terminate_files
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT pause_files
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT
> > min_disk_free Value returning:
> > Value got:
> > Config::IniFiles=HASH(0xf872ad0) MTT min_disk_free_wait Value
> > returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT delete_fast_scratch
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT
> > save_fast_scratch_files Value returning:
> > Value got:
> > Config::IniFiles=HASH(0xf872ad0) MTT
> > docommand_timeout_notify_file
> >
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT
> > docommand_timeout_notify_email
> >
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT
> > docommand_timeout_notify_timeout
> >
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT
> > docommand_timeout_backtrace_program
> >
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT before_all_exec
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT
> > before_all_exec_timeout Value returning:
> > Value got:
> > Config::IniFiles=HASH(0xf872ad0) MTT before_all_exec_pass Value
> > returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT before_each_exec
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT
> > before_each_exec_timeout Value returning:
> > Value got:
> > Config::IniFiles=HASH(0xf872ad0) MTT before_each_exec_pass Value
> > returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT after_all_exec
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT
> > after_all_exec_timeout Value returning:
> > Value got:
> > Config::IniFiles=HASH(0xf872ad0) MTT after_all_exec_pass Value
> > returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT after_each_exec
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf872ad0) MTT
> > after_each_exec_timeout Value returning:
> > Value got:
> > Config::IniFiles=HASH(0xf872ad0) MTT after_each_exec_pass Value
> > returning:
> > Value got: Config::IniFiles=HASH(0xf8fb9d0) lock module
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf8fb9d0) mtt log_file logfile
> > Value returning:
> > *** Reporter initializing
> > Value got: Config::IniFiles=HASH(0xf8fb9d0) reporter: iu
> > database module Value returning: MTTDatabase
> > >> Initializing reporter module: MTTDatabase
> > Evaluating: require MTT::Reporter::MTTDatabase
> > Evaluating: $ret =
> > &MTT::Reporter::MTTDatabase::Init(@args) Value got:
> > Config::IniFiles=HASH(0xf8fb9d0) reporter: iu database
> > mttdatabase_username
> >
> > Value returning: you must set this value
> > Value got: Config::IniFiles=HASH(0xf8fb9d0) reporter: iu database
> > mttdatabase_password
> >
> > Value returning: you must set this value
> > Value got: Config::IniFiles=HASH(0xf8fb9d0) reporter: iu database
> > mttdatabase_url
> >
> > Value returning: https://www.open-mpi.org/mtt/submit/
> > Value got: Config::IniFiles=HASH(0xf8fb9d0) reporter: iu database
> > mttdatabase_realm
> >
> > Value returning: OMPI
> > Value got: Config::IniFiles=HASH(0xf8fb9d0) reporter: iu database
> > mttdatabase_email_errors_to
> >
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf8fb9d0) reporter: iu database
> > mttdatabase_debug_filename
> > Value returning: mttdb_debug_file_perf
> > Value got: Config::IniFiles=HASH(0xf8fb9d0) reporter: iu database
> > mttdatabase_keep_debug_files
> > Value returning: 1
> > Value got: Config::IniFiles=HASH(0xf8fb9d0) reporter: iu database
> > mttdatabase_debug_server
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf8fb9d0) reporter: iu database
> > mttdatabase_hostname
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf8fb9d0) mtt local_username
> > Value returning:
> > Value got: Config::IniFiles=HASH(0xf8fb9d0) reporter: iu database
> > mttdatabase_platform
> > Value returning: you must set this value
> > Set HTTP credentials for realm "OMPI"
> > MTTDatabase client getting a client serial number...
> > MTTDatabase client trying proxy: / Default (none)
> > MTTDatabase proxy successful / not 500
> > *** WARNING: >> Failed test ping to MTTDatabase URL:
> > https://www.open-mpi.org:443/mtt/submit/
> > *** WARNING: >> Error was: 401 Authorization Required
> > <!DOCTYPE HTML PUBLIC "-//IETF//DTD HTML 2.0//EN">
> > <html><head>
> > <title>401 Authorization Required</title>
> > </head><body>
> > <h1>Authorization Required</h1>
> > <p>This server could not verify that you
> > are authorized to access the document
> > requested. Either you supplied the wrong
> > credentials (e.g., bad password), or your
> > browser doesn't understand how to supply
> > the credentials required.</p>
> > <hr>
> > <address>Apache Server at www.open-mpi.org Port 443</address>
> > </body></html>
> > *** ERROR: Module aborted: MTT::Reporter::MTTDatabase:Init: *** ERROR: >>
> > Do not want to continue with possible bad submission URL --
> > aborting.
> >
> > Could anyone suggest me if any modification required in
> > ompi-core-perf-testing.ini or any other way to run this test.
> >
> > ompi-core-perf-testing.ini file is attached.
> >
> > Regards,
> >
> > Vishal
> >
>
> > #
> > # Copyright (c) 2006-2009 Cisco Systems, Inc. All rights reserved.
> > #
> > # Note that there are many items in this file that, while they are
> > # good for examples, may not work for random MTT users. For example,
> > # the "ompi-tests" SVN repository that is used in the examples below
> > # is *not* a public repository (there's nothing really secret in this
> > # repository; it contains many publicly-available MPI tests and
> > # benchmarks, but we have never looked into the redistribution rights
> > # of these codes, so we keep the SVN repository "closed" to the
> > # general public and only use it internally in the Open MPI project).
> >
> > #======================================================================
> > # Generic OMPI core performance testing template configuration
> > #======================================================================
> >
> > [MTT]
> > # Leave this string so that we can identify this data subset in the
> > # database
> > # OMPI Core: Use a "test" label until we're ready to run real results
> > description = [testbake]
> > # description = [2007 collective performance bakeoff]
> > # OMPI Core: Use the "trial" flag until we're ready to run real results
> > trial = 1
> >
> > # Put other values here as relevant to your environment.
> >
> > #----------------------------------------------------------------------
> >
> > [Lock]
> > # Put values here as relevant to your environment.
> >
> > #======================================================================
> > # MPI get phase
> > #======================================================================
> >
> > [MPI get: ompi-nightly-trunk]
> > mpi_details = OMPI
> >
> > module = OMPI_Snapshot
> > ompi_snapshot_url = http://www.open-mpi.org/nightly/trunk
> >
> > #----------------------------------------------------------------------
> >
> > [MPI get: ompi-nightly-v1.2]
> > mpi_details = OMPI
> >
> > module = OMPI_Snapshot
> > ompi_snapshot_url = http://www.open-mpi.org/nightly/v1.2
> >
> > #----------------------------------------------------------------------
> >
> > [MPI get: ompi-released-v1.2]
> > mpi_details = OMPI
> >
> > module = OMPI_Snapshot
> > ompi_snapshot_url = http://www.open-mpi.org/software/ompi/v1.2/downloads
> >
> > #----------------------------------------------------------------------
> >
> > [MPI get: MPICH1]
> > mpi_details = MPICH1
> >
> > module = Download
> > # MPICH1 from the Argonne web site
> > #download_url = http://www-unix.mcs.anl.gov/mpi/mpich1/downloads/mpich.tar.gz
> > # If you are using SLURM, use this URL -- it's the same exact
> > # mpich.tar.gz but with the SLURM 1.2.12/etc/mpich1.slurm.patch in it
> > # (which allows native launching under SLURM).
> > download_url = http://www.open-mpi.org/~jsquyres/ompi-coll-bakeoff/mpich-1.2.7p1-patched-for-slurm.tar.gz
> > # This version is fixed/frozen, so it's ok to hard-code it
> > download_version = 1.2.7p1
> >
> > #----------------------------------------------------------------------
> >
> > [MPI get: MPICH-MX]
> > mpi_details = MPICH-MX
> >
> > module = Download
> > download_url = http://www.myri.com/ftp/pub/MPICH-MX/mpich-mx_1.2.7..5.tar.gz
> > # You need to obtain the username and password from Myricom
> > download_username = <OBTAIN THIS FROM MYRICOM>
> > download_password = <OBTAIN THIS FROM MYRICOM>
> >
> > #----------------------------------------------------------------------
> >
> > [MPI get: MPICH2]
> > mpi_details = MPICH2
> >
> > module = Download
> > download_url = http://www-unix.mcs.anl.gov/mpi/mpich2/downloads/mpich2-1.0.5p4.tar.gz
> >
> > #----------------------------------------------------------------------
> >
> > [MPI get: MVAPICH1]
> > mpi_details = MVAPICH1
> >
> > module = Download
> > download_url = http://mvapich.cse.ohio-state.edu/download/mvapich/mvapich-0.9.9.tar.gz
> >
> > #----------------------------------------------------------------------
> >
> > [MPI get: MVAPICH2]
> > mpi_details = MVAPICH2
> >
> > module = Download
> > download_url = http://mvapich.cse.ohio-state.edu/download/mvapich2/mvapich2-0.9.8p3.tar.gz
> >
> > #----------------------------------------------------------------------
> >
> > [MPI get: HP MPI]
> > mpi_details = HP MPI
> >
> > # You need to have HP MPI already installed somewhere
> > module = AlreadyInstalled
> > # Fill this in with the version of your HP MPI
> > alreadyinstalled_version = 2.2.5.1b1
> >
> > #----------------------------------------------------------------------
> >
> > [MPI get: Intel MPI]
> > mpi_details = Intel MPI
> >
> > # You need to have Intel MPI already installed somewhere
> > module = AlreadyInstalled
> > # Fill this in with the version of your Intel MPI
> > alreadyinstalled_version = 3.0
> >
> > #----------------------------------------------------------------------
> >
> > [SKIP MPI get: Scali MPI]
> > mpi_details = Scali MPI
> >
> > # You need to have Scali MPI already installed somewhere
> > module = AlreadyInstalled
> > # Fill this in with the version of your Scali MPI
> > alreadyinstalled_version = ???
> >
> > #----------------------------------------------------------------------
> >
> > [SKIP MPI get: Cray MPI]
> > mpi_details = Cray MPI
> >
> > # You need to have Cray MPI already installed somewhere
> > module = AlreadyInstalled
> > # Fill this in with the version of your Cray MPI
> > alreadyinstalled_version = ???
> >
> > #======================================================================
> > # Install MPI phase
> > #======================================================================
> >
> > # All flavors of Open MPI
> > [MPI install: OMPI/GNU-standard]
> > mpi_get = ompi-nightly-trunk, ompi-nightly-v1.2, ompi-released-v1.2
> > save_stdout_on_success = 1
> > merge_stdout_stderr = 0
> >
> > module = OMPI
> > ompi_make_all_arguments = -j 8
> > ompi_compiler_name = gnu
> > ompi_compiler_version = &get_gcc_version()
> > # Adjust these configure flags for your site
> > ompi_configure_arguments = CFLAGS=-O3 --with-openib --enable-mpirun-prefix-by-default --enable-branch-probabilities --disable-heterogeneous --without-mpi-param-check
> >
> > #----------------------------------------------------------------------
> >
> > [MPI install: MPICH1]
> > mpi_get = mpich1
> > save_stdout_on_success = 1
> > merge_stdout_stderr = 0
> > # Ensure that MPICH allocates enough shared memory (32MB seems to be
> > # enough for ppn=4; went to 64MB to give it plenty of room)
> > setenv = P4_GLOBMEMSIZE 67108864
> >
> > module = MPICH2
> > mpich2_use_all_target = 0
> > mpich2_apply_slurm_patch = 1
> > mpich2_compiler_name = gnu
> > mpich2_compiler_version = &get_gcc_version()
> > mpich2_configure_arguments = -cflags=-O3 -rsh=ssh --with-device=ch_p4 --with-comm=shared
> >
> > #----------------------------------------------------------------------
> >
> > [MPI install: MPICH2]
> > mpi_get = mpich2
> > save_stdout_on_success = 1
> > merge_stdout_stderr = 0
> > # Adjust this for your site (this is what works at Cisco). Needed to
> > # launch in SLURM; adding this to LD_LIBRARY_PATH here propagates this
> > # all the way through the test run phases that use this MPI install,
> > # where the test executables will need to have this set.
> > prepend_path = LD_LIBRARY_PATH /opt/slurm/current/lib
> >
> > module = MPICH2
> > mpich2_compiler_name = gnu
> > mpich2_compiler_version = &get_gcc_version()
> > mpich2_configure_arguments = --disable-f90 CFLAGS=-O3 --enable-fast --with-device=ch3:nemesis
> > # These are needed to launch through SLURM; adjust as appropriate.
> > mpich2_additional_wrapper_ldflags = -L/opt/slurm/current/lib
> > mpich2_additional_wrapper_libs = -lpmi
> >
> > #----------------------------------------------------------------------
> >
> > [MPI install: MVAPICH1]
> > mpi_get = mvapich1
> > save_stdout_on_success = 1
> > merge_stdout_stderr = 0
> > # Setting this makes MVAPICH assume that the same HCAs are on all
> > # hosts and therefore it makes some optimizations
> > setenv = VIADEV_USE_COMPAT_MODE 0
> >
> > module = MVAPICH2
> > # Adjust this to be where your OFED is installed
> > mvapich2_setenv = IBHOME /usr
> > # Leave this set (COMPAT); if you don't, their script asks questions,
> > # causing MTT to hang
> > mvapich2_setenv = COMPAT AUTO_DETECT
> > mvapich2_build_script = make.mvapich.gen2
> > mvapich2_compiler_name = gnu
> > mvapich2_compiler_version = &get_gcc_version()
> >
> > #----------------------------------------------------------------------
> >
> > [MPI install: MVAPICH2]
> > mpi_get = mvapich2
> > save_stdout_on_success = 1
> > merge_stdout_stderr = 0
> > # Adjust this for your site (this is what works at Cisco). Needed to
> > # launch in SLURM; adding this to LD_LIBRARY_PATH here propagates this
> > # all the way through the test run phases that use this MPI install,
> > # where the test executables will need to have this set.
> > prepend_path = LD_LIBRARY_PATH /opt/slurm/current/lib
> >
> > module = MVAPICH2
> > # Adjust this to be where your OFED is installed
> > mvapich2_setenv = OPEN_IB_HOME /usr
> > mvapich2_build_script = make.mvapich2.ofa
> > mvapich2_compiler_name = gnu
> > mvapich2_compiler_version = &get_gcc_version()
> > # These are needed to launch through SLURM; adjust as appropriate.
> > mvapich2_additional_wrapper_ldflags = -L/opt/slurm/current/lib
> > mvapich2_additional_wrapper_libs = -lpmi
> >
> > #----------------------------------------------------------------------
> >
> > [MPI install: Intel MPI]
> > mpi_get = Intel MPI
> > # Adjust this if you need to. It guarantees that multiple MPD's
> > # running on the same host will not collide. If you'll ever have
> > # multi-job-on-the-same-host conflicts, you may want to adjust this to
> > # reflect some unique identifier (e.g., a resource manager ID).
> > setenv = MPD_CON_EXT mtt-unique-mpd.&getenv("SLURM_JOBID")
> >
> > module = Analyze::IntelMPI
> >
> > #----------------------------------------------------------------------
> >
> > [MPI install: HP MPI]
> > mpi_get = HP MPI
> >
> > module = Analyze::HPMPI
> >
> > #======================================================================
> > # MPI run details
> > #======================================================================
> >
> > [MPI Details: OMPI]
> > # Check &test_alloc() for byslot or bynode
> > exec = mpirun @alloc@ -np &test_np() @mca@ &test_executable() &test_argv()
> > parameters = &MPI::OMPI::find_mpirun_params(&test_command_line(), \
> > &test_executable())
> > network = &MPI::OMPI::find_network(&test_command_line(), &test_executable())
> >
> > alloc = &if(&eq(&test_alloc(), "node"), "--bynode", "--byslot")
> > mca = &enumerate( \
> > "--mca btl sm,tcp,self " . @common_params@, \
> > "--mca btl tcp,self " . @common_params@, \
> > "--mca btl sm,openib,self " . @common_params@, \
> > "--mca btl sm,openib,self --mca mpi_leave_pinned 1 " . @common_params@, \
> > "--mca btl openib,self " . @common_params@, \
> > "--mca btl openib,self --mca mpi_leave_pinned 1 " . @common_params@, \
> > "--mca btl openib,self --mca mpi_leave_pinned_pipeline 1 " . @common_params@, \
> > "--mca btl openib,self --mca btl_openib_use_srq 1 " . @common_params@)
> >
> > # v1.2 has a problem with striping across heterogeneous ports right now:
> > # https://svn.open-mpi.org/trac/ompi/ticket/1125. Also keep the coll
> > # bakeoff tests on DDR only.
> > common_params = "--mca btl_tcp_if_include ib0 --mca oob_tcp_if_include ib0 --mca btl_openib_if_include mthca0 --mca mpi_paffinity_alone 1" . \
> > &if(&or(&eq(&mpi_get_name(), "ompi-nightly-v1.2"), \
> > &eq(&mpi_get_name(), "ompi-released-v1.2")), \
> > "--mca btl_openib_max_btls 1", "")
> >
> > # It is important that the after_each_exec step is a single
> > # command/line so that MTT will launch it directly (instead of via a
> > # temporary script). This is because the "srun" command is
> > # (intentionally) difficult to kill in some cases. See
> > # https://svn.open-mpi.org/trac/mtt/changeset/657 for details.
> >
> > after_each_exec = &if(&ne("", &getenv("SLURM_NNODES")), "srun -N " . &getenv("SLURM_NNODES")) /home/mpiteam/svn/ompi-tests/cisco/mtt/after_each_exec.pl
> >
> > #----------------------------------------------------------------------
> >
> > [MPI Details: MPICH1]
> >
> > # Launching through SLURM. From
> > # http://www.llnl.gov/linux/slurm/quickstart.html.
> > exec = srun @alloc@ -n &test_np() --mpi=mpich1_p4 &test_executable() &test_argv()
> >
> > # If not using SLURM, you'll need something like this (not tested).
> > # You may need different hostfiles for running by slot/by node.
> > #exec = mpirun -np &test_np() -machinefile &hostfile() &test_executable()
> >
> > network = loopback,ethernet
> >
> > # In this SLURM example, if each node has 4 CPUs, telling SLURM to
> > # launching "by node" means specifying that each MPI process will use 4
> > # CPUs.
> > alloc = &if(&eq(&test_alloc(), "node"), "-c 4", "")
> >
> > #----------------------------------------------------------------------
> >
> > [MPI Details: MPICH2]
> >
> > # Launching through SLURM. If you use mpdboot instead, you need to
> > # ensure that multiple mpd's on the same node don't conflict (or never
> > # happen).
> > exec = srun @alloc@ -n &test_np() &test_executable() &test_argv()
> >
> > # If not using SLURM, you'll need something like this (not tested).
> > # You may need different hostfiles for running by slot/by node.
> > #exec = mpiexec -np &test_np() --host &hostlist() &test_executable()
> >
> > network = loopback,ethernet,shmem
> >
> > # In this SLURM example, if each node has 4 CPUs, telling SLURM to
> > # launching "by node" means specifying that each MPI process will use 4
> > # CPUs.
> > alloc = &if(&eq(&test_alloc(), "node"), "-c 4", "")
> >
> > #----------------------------------------------------------------------
> >
> > [MPI Details: MVAPICH1]
> >
> > # Launching through SLURM. From
> > # http://www.llnl.gov/linux/slurm/quickstart.html.
> > exec = srun @alloc@ -n &test_np() --mpi=mvapich &test_executable() &test_argv()
> >
> > # If not using SLURM, you'll need something like this (not tested).
> > # You may need different hostfiles for running by slot/by node.
> > #exec = mpirun -np &test_np() --machinefile &hostfile() &test_executable()
> >
> > network = loopback,verbs,shmem
> >
> > # In this example, each node has 4 CPUs, so to launch "by node", just
> > # specify that each MPI process will use 4 CPUs.
> > alloc = &if(&eq(&test_alloc(), "node"), "-c 4", "")
> >
> > #----------------------------------------------------------------------
> >
> > [MPI Details: MVAPICH2]
> >
> > # Launching through SLURM. If you use mpdboot instead, you need to
> > # ensure that multiple mpd's on the same node don't conflict (or never
> > # happen).
> > exec = srun @alloc@ -n &test_np() &test_executable() &test_argv()
> >
> > # If not using SLURM, you'll need something like this (not tested).
> > # You may need different hostfiles for running by slot/by node.
> > #exec = mpiexec -np &test_np() --host &hostlist() &test_executable()
> >
> > network = loopback,verbs,shmem
> >
> > # In this example, each node has 4 CPUs, so to launch "by node", just
> > # specify that each MPI process will use 4 CPUs.
> > alloc = &if(&eq(&test_alloc(), "node"), "-c 4", "")
> >
> > #----------------------------------------------------------------------
> >
> > [MPI Details: MPICH-MX]
> >
> > # Launching through SLURM. From
> > # http://www.llnl.gov/linux/slurm/quickstart.html.
> > exec = srun @alloc@ -n &test_np() --mpi=mpichgm &test_executable() &test_argv()
> > network = mx
> >
> > # If not using SLURM, you'll need something like this (not tested).
> > # You may need different hostfiles for running by slot/by node.
> > #exec = mpirun -np &test_np() --machinefile &hostfile() &test_executable()
> >
> > # In this example, each node has 4 CPUs, so to launch "by node", just
> > # specify that each MPI process will use 4 CPUs.
> > alloc = &if(&eq(&test_alloc(), "node"), "-c 4", "")
> >
> > #----------------------------------------------------------------------
> >
> > [MPI Details: Intel MPI]
> >
> > # Need a before_any_exec step to launch MPDs
> > before_any_exec = <<EOF
> > h=`hostname`
> > file=mtt-hostlist.$$
> > rm -f $file
> > # If we're allocating by node, get each hostname once. Otherwise, get
> > # each hostname as many times as we have slots on that node.
> > srun hostname | uniq > $file
> > # Add localhost if it's not in there (e.g., srun -A)
> > local=`grep $h $file`
> > touch /tmp/mtt-mpiexec-options.$SLURM_JOBID
> > if test "$local" = ""; then
> > echo $h >> $file
> > echo -nolocal > /tmp/mpiexec-options.$SLURM_JOBID
> > fi
> > num=`wc -l $file | awk '{ print $1 }'`
> > mpdboot -n $num -r ssh --verbose --file=$file
> > mpdtrace
> > rm -f $file
> > EOF
> >
> > # Intel MPI seems to default to by-node allocation and I can't figure
> > # out how to override it. Sigh.
> > exec = mpiexec @options@ -n &test_np() ./&test_executable() &test_argv()
> > network = loopback,verbs,shmem
> >
> > # Test both the "normal" collective algorithms and Intel's "fast"
> > # collective algorithms (their docs state that the "fast" algorithms
> > # may not be MPI conformant, and may not give the same results between
> > # multiple runs, assumedly if the process layout is different).
> > options = &stringify(&cat("/tmp/mpiexec-options." . &getenv("SLURM_JOBID"))) \
> > &enumerate("-genv I_MPI_DEVICE rdssm", \
> > "-genv I_MPI_DEVICE rdssm -genv I_MPI_FAST_COLLECTIVES 1")
> >
> > after_all_exec = <<EOT
> > rm -f /tmp/mpiexec-options.$SLURM_JOBID
> > mpdallexit
> > EOT
> >
> > #----------------------------------------------------------------------
> >
> > [MPI Details: HP MPI]
> >
> > # I use MPI_IBV_NO_FORK_SAFE=1 because I'm using RHEL4U4, which
> > # doesn't have IBV fork() support. I also have multiple active HCA
> > # ports and therefore need to give HP MPI a clue on the scheduling of
> > # ports via MPI_IB_CARD_ORDER. I got this information via e-mailing
> > # HP MPI support.
> > #
> > # In SLURM, HP MPI seems to schedule first by node and then by slot.
> > # So if you have 2 quad-core nodes in your SLURM alloc, if you mpirun
> > # -np 2, you'll get one proc on each node. If you mpirun -np 4,
> > # you'll get MCW ranks 0 and 1 on the first node and MCA ranks 2 and 3
> > # on the second node. This is pretty much exactly what we want, so we
> > # don't need to check &test_alloc() here.
> > exec = mpirun -IBV -e MPI_IBV_NO_FORK_SAFE=1 -e MPI_IB_CARD_ORDER=0:0 -srun -n&test_np() ./&test_executable() &test_argv()
> > network = loopback,verbs,shmem
> >
> > #======================================================================
> > # Test get phase
> > #======================================================================
> >
> > [Test get: netpipe]
> > module = Download
> > download_url = http://www.scl.ameslab.gov/netpipe/code/NetPIPE_3.6.2.tar.gz
> >
> > #----------------------------------------------------------------------
> >
> > [Test get: osu]
> > module = SVN
> > svn_url = https://svn.open-mpi.org/svn/ompi-tests/trunk/osu
> >
> > #----------------------------------------------------------------------
> >
> > [Test get: imb]
> > module = SVN
> > svn_url = https://svn.open-mpi.org/svn/ompi-tests/trunk/IMB_2.3
> >
> > #----------------------------------------------------------------------
> >
> > [Test get: skampi]
> > module = SVN
> > svn_url = https://svn.open-mpi.org/svn/ompi-tests/trunk/skampi-5.0.1
> >
> > #----------------------------------------------------------------------
> >
> > [Test get: nbcbench]
> > module = SVN
> > svn_url = https://svn.open-mpi.org/svn/ompi-tests/trunk/nbcbench
> >
> > #======================================================================
> > # Test build phase
> > #======================================================================
> >
> > [Test build: netpipe]
> > test_get = netpipe
> > save_stdout_on_success = 1
> > merge_stdout_stderr = 1
> > stderr_save_lines = 100
> >
> > module = Shell
> > shell_build_command = <<EOT
> > make mpi
> > EOT
> >
> > #----------------------------------------------------------------------
> >
> > [Test build: osu]
> > test_get = osu
> > save_stdout_on_success = 1
> > merge_stdout_stderr = 1
> > stderr_save_lines = 100
> >
> > module = Shell
> > shell_build_command = <<EOT
> > make osu_latency osu_bw osu_bibw
> > EOT
> >
> > #----------------------------------------------------------------------
> >
> > [Test build: imb]
> > test_get = imb
> > save_stdout_on_success = 1
> > merge_stdout_stderr = 1
> >
> > module = Shell
> > shell_build_command = <<EOT
> > cd src
> > make clean IMB-MPI1
> > EOT
> >
> > #----------------------------------------------------------------------
> >
> > [Test build: skampi]
> > test_get = skampi
> > save_stdout_on_success = 1
> > merge_stdout_stderr = 1
> > stderr_save_lines = 100
> >
> > module = Shell
> > # Set EVERYONE_CAN_MPI_IO for HP MPI
> > shell_build_command = <<EOT
> > make CFLAGS="-O2 -DPRODUCE_SPARSE_OUTPUT -DEVERYONE_CAN_MPI_IO"
> > EOT
> >
> > #----------------------------------------------------------------------
> >
> > [Test build: nbcbench]
> > test_get = nbcbench
> > save_stdout_on_success = 1
> > merge_stdout_stderr = 1
> > stderr_save_lines = 100
> >
> > module = Shell
> > shell_build_command = <<EOT
> > make
> > EOT
> >
> > #======================================================================
> > # Test Run phase
> > #======================================================================
> >
> > [Test run: netpipe]
> > test_build = netpipe
> > pass = &and(&cmd_wifexited(), &eq(&cmd_wexitstatus(), 0))
> > # Timeout hueristic: 10 minutes
> > timeout = 10:00
> > save_stdout_on_pass = 1
> > # Ensure to leave this value as "-1", or performance results could be lost!
> > stdout_save_lines = -1
> > merge_stdout_stderr = 1
> > np = 2
> > alloc = node
> >
> > specify_module = Simple
> > analyze_module = NetPipe
> >
> > simple_pass:tests = NPmpi
> >
> > #----------------------------------------------------------------------
> >
> > [Test run: osu]
> > test_build = osu
> > pass = &and(&cmd_wifexited(), &eq(&cmd_wexitstatus(), 0))
> > # Timeout hueristic: 10 minutes
> > timeout = 10:00
> > save_stdout_on_pass = 1
> > # Ensure to leave this value as "-1", or performance results could be lost!
> > stdout_save_lines = -1
> > merge_stdout_stderr = 1
> > np = 2
> > alloc = node
> >
> > specify_module = Simple
> > analyze_module = OSU
> >
> > simple_pass:tests = osu_bw osu_latency osu_bibw
> >
> > #----------------------------------------------------------------------
> >
> > [Test run: imb]
> > test_build = imb
> > pass = &and(&cmd_wifexited(), &eq(&cmd_wexitstatus(), 0))
> > # Timeout hueristic: 10 minutes
> > timeout = 10:00
> > save_stdout_on_pass = 1
> > # Ensure to leave this value as "-1", or performance results could be lost!
> > stdout_save_lines = -1
> > merge_stdout_stderr = 1
> > np = &env_max_procs()
> >
> > argv = -npmin &test_np() &enumerate("PingPong", "PingPing", "Sendrecv", "Exchange", "Allreduce", "Reduce", "Reduce_scatter", "Allgather", "Allgatherv", "Alltoall", "Bcast", "Barrier")
> >
> > specify_module = Simple
> > analyze_module = IMB
> >
> > simple_pass:tests = src/IMB-MPI1
> >
> > #----------------------------------------------------------------------
> >
> > [Test run: skampi]
> > test_build = skampi
> > pass = &and(&cmd_wifexited(), &eq(&cmd_wexitstatus(), 0))
> > # Timeout hueristic: 10 minutes
> > timeout = 10:00
> > save_stdout_on_pass = 1
> > # Ensure to leave this value as "-1", or performance results could be lost!
> > stdout_save_lines = -1
> > merge_stdout_stderr = 1
> > np = &env_max_procs()
> >
> > argv = -i &find("mtt_.+.ski", "input_files_bakeoff")
> >
> > specify_module = Simple
> > analyze_module = SKaMPI
> >
> > simple_pass:tests = skampi
> >
> > #----------------------------------------------------------------------
> >
> > [Test run: nbcbench]
> > test_build = nbcbench
> > pass = &and(&test_wifexited(), &eq(&test_wexitstatus(), 0))
> > timeout = -1
> > save_stdout_on_pass = 1
> > # Ensure to leave this value as "-1", or performance results could be lost!
> > stdout_save_lines = -1
> > merge_stdout_stderr = 1
> >
> > specify_module = Simple
> > analyze_module = NBCBench
> > simple_pass:tests = nbcbench
> >
> > argv = -p &test_np()-&test_np() -s 1-1048576 -v -t \
> > &enumerate("MPI_Allgatherv", "MPI_Allgather", "MPI_Allreduce", \
> > "MPI_Alltoall", "MPI_Alltoallv", "MPI_Barrier", "MPI_Bcast", \
> > "MPI_Gather", "MPI_Gatherv", "MPI_Reduce", "MPI_Reduce_scatter", \
> > "MPI_Scan", "MPI_Scatter", "MPI_Scatterv")
> >
> > #======================================================================
> > # Reporter phase
> > #======================================================================
> >
> > [Reporter: IU database]
> > module = MTTDatabase
> >
> > mttdatabase_realm = OMPI
> > mttdatabase_url = https://www.open-mpi.org/mtt/submit/
> > # Change this to be the username and password for your submit user.
> > # Get this from the OMPI MTT administrator.
> > mttdatabase_username = you must set this value
> > mttdatabase_password = you must set this value
> > # Change this to be some short string identifying your cluster.
> > mttdatabase_platform = you must set this value
> >
> > mttdatabase_debug_filename = mttdb_debug_file_perf
> > mttdatabase_keep_debug_files = 1
> >
> > #----------------------------------------------------------------------
> >
> > # This is a backup reporter; it also writes results to a local text
> > # file
> >
> > [Reporter: text file backup]
> > module = TextFile
> >
> > textfile_filename = $phase-$section-$mpi_name-$mpi_version.txt
> >
> > textfile_summary_header = <<EOT
> > Hostname: &shell("hostname")
> > uname: &shell("uname -a")
> > Username: &shell("who am i")
> > EOT
> >
> > textfile_summary_footer =
> > textfile_detail_header =
> > textfile_detail_footer =
> >
> > textfile_textwrap = 78
>
> > _______________________________________________
> > mtt-users mailing list
> > mtt-users_at_[hidden]
> > http://www.open-mpi.org/mailman/listinfo.cgi/mtt-users
>
> _______________________________________________
> mtt-users mailing list
> mtt-users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/mtt-users
>

-- 
Jeff Squyres
jsquyres_at_[hidden]