Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

From: Terry Frankcombe (terry_at_[hidden])
Date: 2007-06-11 04:17:36


Hi Victor

I'd suggest 3 seconds of CPU time is far, far to small a problem to do
scaling tests with. Even with only 2 CPUs, I wouldn't go below 100
times that.

On Mon, 2007-06-11 at 01:10 -0700, victor marian wrote:
> Hi Jeff
>
> I ran the NAS Parallel Bechmark and it gives for me
> -bash%/export/home/vmarian/fortran/benchmarks/NPB3.2/NPB3.2-MPI/bin$
> mpirun -np 1 cg.A.1
> --------------------------------------------------------------------------
> [0,1,0]: uDAPL on host SERVSOLARIS was unable to find
> any NICs.
> Another transport will be used instead, although this
> may result in
> lower performance.
> --------------------------------------------------------------------------
> NAS Parallel Benchmarks 3.2 -- CG Benchmark
>
> Size: 14000
> Iterations: 15
> Number of active processes: 1
> Number of nonzeroes per row: 11
> Eigenvalue shift: .200E+02
> Benchmark completed
> VERIFICATION SUCCESSFUL
> Zeta is 0.171302350540E+02
> Error is 0.512264003323E-13
>
>
> CG Benchmark Completed.
> Class = A
> Size = 14000
> Iterations = 15
> Time in seconds = 3.02
> Total processes = 1
> Compiled procs = 1
> Mop/s total = 495.93
> Mop/s/process = 495.93
> Operation type = floating point
> Verification = SUCCESSFUL
> Version = 3.2
> Compile date = 11 Jun 2007
>
>
> -bash%/export/home/vmarian/fortran/benchmarks/NPB3.2/NPB3.2-MPI/bin$
> mpirun -np 2 cg.A.2
> --------------------------------------------------------------------------
> [0,1,0]: uDAPL on host SERVSOLARIS was unable to find
> any NICs.
> Another transport will be used instead, although this
> may result in
> lower performance.
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> [0,1,1]: uDAPL on host SERVSOLARIS was unable to find
> any NICs.
> Another transport will be used instead, although this
> may result in
> lower performance.
> --------------------------------------------------------------------------
>
>
> NAS Parallel Benchmarks 3.2 -- CG Benchmark
>
> Size: 14000
> Iterations: 15
> Number of active processes: 2
> Number of nonzeroes per row: 11
> Eigenvalue shift: .200E+02
>
> Benchmark completed
> VERIFICATION SUCCESSFUL
> Zeta is 0.171302350540E+02
> Error is 0.522633719989E-13
>
>
> CG Benchmark Completed.
> Class = A
> Size = 14000
> Iterations = 15
> Time in seconds = 2.47
> Total processes = 2
> Compiled procs = 2
> Mop/s total = 606.32
> Mop/s/process = 303.16
> Operation type = floating point
> Verification = SUCCESSFUL
> Version = 3.2
> Compile date = 11 Jun 2007
>
>
> You can remark that the scalling is not so good
> like yours. Maibe I am having comunications problems
> between processors.
> You can also remark that I am faster on one process
> concared to your processor.
>
> Victor
>
>
>
>
>
> --- Jeff Pummill <jpummil_at_[hidden]> wrote:
>
> > Perfect! Thanks Jeff!
> >
> > The NAS Parallel Benchmark on a dual core AMD
> > machine now returns this...
> > [jpummil_at_localhost bin]$ mpirun -np 1 cg.A.1
> > NAS Parallel Benchmarks 3.2 -- CG Benchmark
> > CG Benchmark Completed.
> > Class = A
> > Size = 14000
> > Iterations = 15
> > Time in seconds = 4.75
> > Total processes = 1
> > Compiled procs = 1
> > Mop/s total = 315.32
> >
> > ...and...
> >
> > [jpummil_at_localhost bin]$ mpirun -np 2 cg.A.2
> > NAS Parallel Benchmarks 3.2 -- CG Benchmark
> > CG Benchmark Completed.
> > Class = A
> > Size = 14000
> > Iterations = 15
> > Time in seconds = 2.48
> > Total processes = 2
> > Compiled procs = 2
> > Mop/s total = 604.46
> >
> > Not quite linear, but one must account for all of
> > the OS traffic that
> > one core or the other must deal with.
> >
> >
> > Jeff F. Pummill
> > Senior Linux Cluster Administrator
> > University of Arkansas
> > Fayetteville, Arkansas 72701
> > (479) 575 - 4590
> > http://hpc.uark.edu
> >
> > "A supercomputer is a device for turning
> > compute-bound
> > problems into I/O-bound problems." -Seymour Cray
> >
> >
> > Jeff Squyres wrote:
> > > Just remove the -L and -l arguments -- OMPI's
> > "mpif90" (and other
> > > wrapper compilers) will do all that magic for you.
> > >
> > > Many -L/-l arguments in MPI application Makefiles
> > are throwbacks to
> > > older versions of MPICH wrapper compilers that
> > didn't always work
> > > properly. Those days are long gone; most (all?)
> > MPI wrapper
> > > compilers do not need you to specify -L/-l these
> > days.
> > >
> > >
> > >
> > > On Jun 10, 2007, at 3:08 PM, Jeff Pummill wrote:
> > >
> > >
> > >> Maybe the "dumb question" of the week, but here
> > goes...
> > >>
> > >> I am trying to compile a piece of code (NPB)
> > under OpenMPI and I am
> > >> having a problem with specifying the right
> > library. Possibly
> > >> something I need to define in a LD_LIBRARY_PATH
> > statement?
> > >>
> > >> Using Gnu mpich, the line looked like this...
> > >>
> > >> FMPI_LIB = -L/opt/mpich/gnu/lib/ -lmpich
> > >>
> > >> I tried to replace this with...
> > >>
> > >> FMPI_LIB = -L/usr/lib/openmpi/ -llibmpi
> > >>
> > >> to which the make responded...
> > >> mpif90 -O -o ../bin/cg.A.2 cg.o
> > ../common/print_results.o ../common/
> > >> randdp.o ../common/timers.o -L/usr/lib/openmpi/
> > -llibmpi
> > >> /usr/bin/ld: cannot find -llibmpi
> > >> collect2: ld returned 1 exit status
> > >>
> > >> Wrong library file? Setup or path issue?
> > >>
> > >> --
> > >> Jeff F. Pummill
> > >> Senior Linux Cluster Administrator
> > >> University of Arkansas
> > >>
> > >> _______________________________________________
> > >> users mailing list
> > >> users_at_[hidden]
> > >>
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > >>
> > >
> > >
> > >
> > > _______________________________________________
> > users mailing list
> > users_at_[hidden]
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
>
>
> ____________________________________________________________________________________
> Get your own web address.
> Have a HUGE year through Yahoo! Small Business.
> http://smallbusiness.yahoo.com/domains/?p=BESTDEAL
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users