Open MPI logo

MTT Devel Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all MTT Devel mailing list

Subject: Re: [MTT devel] MTT Visualization
From: Ethan Mallove (ethan.mallove_at_[hidden])
Date: 2008-01-11 13:29:30


On Fri, Jan/11/2008 12:49:50PM, Jeff Squyres wrote:
> On Jan 10, 2008, at 10:29 AM, Josh Hursey wrote:
>
> > I met with Joseph Cottam (Grad student in my lab at IU) yesterday
> > about MTT visualization. He is working on some new visualization
> > techniques and wants to apply them to the MTT dataset.
>
> Awesome.
>
> > Since we are ramping up to a v1.3 release we want to visualization to
> > support this effort. So we want to make sure that the visualization
> > will meet the development community's needs. We should probably ask
> > the devel-core list, but I thought I would start some of the
> > discussion here to make sure I am asking the right questions of the
> > group.
>
> Sounds reasonable.
>
> After a first go-round here, we might want to have a conversation with
> the OMPI RM's to get their input - that would still be a small group
> to get targeted feedback on these questions.
>
> > To start I have some basic questions:
> > - How does Open MPI determine that it is stable enough to release?
>
> I personally have a Magic 8 Ball on my desk that I consult frequently
> for questions like this. ;-)
>
> It's a mix of many different metrics, actually:
>
> - stuff unrelated to MTT results:
> - how many trac tickets are open against that release and do we care
> - how urgent are the bug fixes that are included
> - external requirements (e.g., get an OMPI release out to meet the
> OFED release schedule)
> - ...and probably others
> - related to MTT results
> - "good" coverage on platforms (where "platform" = host arch, OS,
> OS version, compiler, compiler version, MCA params, interconnect, and
> scheduler -- note that some of these are orthogonal from each other...)
> - the only failures and timeouts we have are a) repeatable, b)
> consistent across multiple organizations (if relevant), and deemed to
> be acceptable
>
> > - What dimensions of testing are most/least important (i.e.,
> > platforms, compilers, feature sets, scale, ...)?
>
> This is a hard question. :-\ I listed several dimensions above:
>
> - host architecture
> - OS
> - OS version
> - compiler
> - compiler version
> - MCA parameters used
> - interconnect
> - scheduler
>
> Here's some more:
>
> - number of processes tested
> - layout of processes (by node, by proc, ...etc.)
>
> I don't quite know how to order those in terms of priority. :-\
>
> > - What other questions would be useful to answer with regard to
> > testing (thinking completely outside of the box)?
> > * Example: Are we testing a specific platform/configuration set
> > too much/too little?
>
> This is a great question.
>
> I would love to be able to configure this question -- e.g., are we
> testing some MCA params too much/too little.
>
> The performance stuff can always be visualized better, especially over
> time. One idea is expressed in https://svn.open-mpi.org/trac/mtt/ticket/330
> .
>
> I also very much like the ideas in https://svn.open-mpi.org/trac/mtt/ticket/236
> and https://svn.open-mpi.org/trac/mtt/ticket/302 (302 is not
> expressed as a visualization issue, but it could be -- you can imagine
> a tree-based display showing the relationships between phase results,
> perhaps even incorporated with a timeline -- that would be awesome).
>
> Here's a whacky idea -- can our MTT data be combined with SCM data
> (SVN, in this case) to answer questions like:
>
> - what parts of the code are the most troublesome? i.e., when this
> part of the code changes, these tests tend to break
>
> - what tests seem to be related to what parts of the OMPI code base?
>
> - who / what SVN commit(s) seemed to cause specific tests to break?
>
> (this seems like a longer-term set of questions, but I thought I'd
> bring it up...)

I like this idea :-)

A level of redirection missing to do this is keying SVN r
numbers to files modified. We also need to be able to
somehow track *new* failures (see
https://svn.open-mpi.org/trac/mtt/ticket/70). E.g., "was it
*this* revision that broke test xyz or was it an older one?"

-Ethan

>
> > - Other questions you think we should pose to the group?
> >
> > We are currently feeling out the domain of possibilities, but hope to
> > start doing some sketching some ideas in another week or so. This work
> > should proceed fairly quickly since we are targeting a paper about
> > this for the ACM Symposium on Software Visualization (http://www.softvis.org/
> > ) which is due in early April. How is that for expecting success :)
>
> Awesome.
>
> --
> Jeff Squyres
> Cisco Systems
>
> _______________________________________________
> mtt-devel mailing list
> mtt-devel_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/mtt-devel