Open MPI logo

MTT Devel Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all MTT Users mailing list

From: Andrew Friedley (afriedle_at_[hidden])
Date: 2006-07-20 19:10:44


Ethan Mallove wrote:
> Here's a draft for a nightly test results email format. One
> will receive reports through mtt-results_at_open-mpi.org.
> Note: The following is a faux-report. Just pretend there is
> nothing non-sensical about the data.

Looks better than what I had come up with :) some comments inline below.

Are you planning on one big report per night, or breaking it up somehow
into several smaller reports (ie per site, platform, or some other
criteria)?

> This might be better sent as an attachment, or a link to a
> webpage as lots of email clients have default textwidth
> settings that will make the formatted tables look like
> gobbledygook.

I personally would lean towards inline plain text with minimal fancy
formatting, and just enough information to go dig in the source code
when I see a problem in one of these emails. Then having url's to a
full report would certainly be nice as well.

> We are also planning a webpage frontend (e.g.,
> www.open-mpi.org/mtt-results.php) to the test results with
> more flexibility in generating reports than this static
> email report. In other words, we're leaving out lots of
> bells and whistles here as those will be implemented in the
> webpage frontend.
>
> -Ethan
>
>
> ======
>
> Subject: Test results for <datestamp>

Something Brian requested was that the subject indicate whether there
are any failure reports in the email, so that he can choose not to read
the email when everything works, possibly by using procmail for example.

> Body:
>
> * Failed MPI Install(s)
>
> os_version: SunOS 5.10
> platform_hardware: i86pc
> platform_id: Sun's Cluster
> mpi_name: ompi-nightly-v1.0
> mpi_version: 1.0.3a1r10793
>
> http://www.open-mpi.org/mtt/mpi_install_logs/config.log.>
>
http://www.open-mpi.org/mtt/mpi_install_logs/make.log.>

I forget if MTT submits the full configure/make logs - I'm guessing yes.
  Posting say the last 10-15 lines here in addition to the links might
be useful. Same idea for stdout/stderr in other places as well.

>
> * Failed Test Build(s)
>
> os_version: SunOS 5.10
> platform_hardware: i86pc
> platform_id: Sun's Cluster
> mpi_name: ompi-nightly-v1.0
> mpi_version: 1.0.3a1r10793
>
> +----+------------+---------+------------------------+
> | # | time | cat | name |
> +----+------------+---------+------------------------+
> | 1 | 2006-07-20 | ibm | test_abc |
> | 2 | 2006-07-20 | ibm | test_xyz |
> | 3 | 2006-07-20 | intel | test_123 |

> [snip] ...
>
> [1]

Wasn't sure what these numbers meant at first :) Maybe you could add
the test name or something as well here?

> environment: foo = bar
> compiler_flags: -g -fugly-complex -fno-globals -Wno-globals -Isrc -I.
> stderr: blah blah blah
> stdout: blah blah blah
>
> * Failed Test Run(s)
>
> os_version: SunOS 5.10
> platform_hardware: i86pc
> platform_id: Sun's Cluster
> mpi_name: ompi-nightly-v1.0
> mpi_version: 1.0.3a1r10793
>
> +----+------------+---------+------------------------+
> | # | time | cat | name |
> +----+------------+---------+------------------------+
> | 1 | 2006-07-20 | ibm | zero1 |
> | 2 | 2006-07-20 | ibm | spawn |
> | 3 | 2006-07-20 | ibm | spawn_multiple |
> | 4 | 2006-07-20 | ibm | abort |
> | 5 | 2006-07-20 | ibm | final |
> | 6 | 2006-07-20 | intel | init_thread_funneled |
> | 7 | 2006-07-20 | intel | init_thread_serialized |
> | 8 | 2006-07-20 | intel | init_thread_multiple |
> | 9 | 2006-07-20 | imb | range |
> | 10 | 2006-07-20 | imb | cart |
> | 11 | 2006-07-20 | imb | graph |
> | 12 | 2006-07-20 | imb | sub |
>
> [snip] ...
>
> * Failed Test Run Details
>
> [1]
>
> environment: foo = bar
> test_command: mpirun -np 2 --prefix
> /workspace/em162155/hpc/mtt/trunk/installs/ompi-nightly-v1.0/solaris_i386/install
> ./zero1
> stderr: blah blah blah
> stdout: blah blah blah
>
> [2]
>
> environment: foo = bar
> test_command: mpirun -np 2 --prefix
> /workspace/em162155/hpc/mtt/trunk/installs/ompi-nightly-v1.0/solaris_i386/install
> ./spawn
> stderr: blah blah blah
> stdout: blah blah blah
>
> [snip] ...
>
> * Failed Test(s)
>
> os_version: Linux
> platform_hardware: i86pc
> platform_id: IU Odin Cluster
> mpi_name: ompi-nightly-v1.0
> mpi_version: 1.0.3a1r10793
>
> +----+------------+---------+------------------------+
> | # | time | cat | name |
> +----+------------+---------+------------------------+
> | 1 | 2006-07-20 | ibm | zero1 |
> | 2 | 2006-07-20 | ibm | spawn |
> | 3 | 2006-07-20 | ibm | spawn_multiple |
> | 9 | 2006-07-20 | imb | range |
> | 12 | 2006-07-20 | imb | sub |
>
> [snip] ...
>
> * Failed Test Details
>
> [1]
>
> environment: foo = bar
> test_command: mpirun -np 2 --prefix
> /workspace/em162155/hpc/mtt/trunk/installs/ompi-nightly-v1.0/solaris_i386/install
> ./zero1
> stderr: blah blah blah
> stdout: blah blah blah
>
> [2]
>
> environment: foo = bar
> test_command: mpirun -np 2 --prefix
> /workspace/em162155/hpc/mtt/trunk/installs/ompi-nightly-v1.0/solaris_i386/install
> ./spawn
> stderr: blah blah blah
> stdout: blah blah blah
>
> [snip] ...
>
> ======
>
> * Passed Test Run(s)
>
> os_version: SunOS 5.10
> platform_hardware: i86pc
> platform_id: Sun's Cluster
> mpi_name: ompi-nightly-v1.0
> mpi_version: 1.0.3a1r10793
>
> +----+------------+---------+-------------------------+
> | # | time | cat | name |
> +----+------------+---------+-------------------------+
> | 3 | 2006-07-17 | trivial | f77_hello |
> | 32 | 2006-07-20 | ibm | allgatherv_in_place |
> | 33 | 2006-07-20 | ibm | allreduce |
> | 34 | 2006-07-20 | ibm | allreduce_in_place |
> | 35 | 2006-07-20 | ibm | alltoall |
> | 36 | 2006-07-20 | intel | alltoallw |
> | 37 | 2006-07-20 | intel | barrier |
> | 38 | 2006-07-20 | intel | bcast |
> | 39 | 2006-07-20 | imb | bcast_struct |
> | 40 | 2006-07-20 | imb | exscan |
> | 41 | 2006-07-20 | imb | gather |
> | 42 | 2006-07-20 | imb | gather_in_place |
> | 43 | 2006-07-20 | imb | gatherv |
> | 44 | 2006-07-20 | imb | gatherv_in_place |
>
> [snip] ...
>
> os_version: Linux
> platform_hardware: i86pc
> platform_id: IU Odin Cluster
> mpi_name: ompi-nightly-v1.0
> mpi_version: 1.0.3a1r10793
>
> +----+------------+---------+-------------------------+
> | # | time | cat | name |
> +----+------------+---------+-------------------------+
> | 3 | 2006-07-17 | trivial | f77_hello |
> | 32 | 2006-07-20 | ibm | allgatherv_in_place |
> | 33 | 2006-07-20 | ibm | allreduce |
> | 34 | 2006-07-20 | ibm | allreduce_in_place |
> | 35 | 2006-07-20 | ibm | alltoall |
> | 36 | 2006-07-20 | intel | alltoallw |
> | 37 | 2006-07-20 | intel | barrier |
> | 38 | 2006-07-20 | intel | bcast |
> | 39 | 2006-07-20 | imb | bcast_struct |
> | 40 | 2006-07-20 | imb | exscan |
> | 41 | 2006-07-20 | imb | gather |
> | 42 | 2006-07-20 | imb | gather_in_place |
> | 43 | 2006-07-20 | imb | gatherv |
> | 44 | 2006-07-20 | imb | gatherv_in_place |
>
> [snip] ...
>
> _______________________________________________
> mtt-results mailing list
> mtt-results_at_[hidden]
>
http://www.open-mpi.org/mailman/listinfo.cgi/mtt-results
>
> _______________________________________________
> mtt-users mailing list
> mtt-users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/mtt-users