Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

From: victor marian (vctrmarian_at_[hidden])
Date: 2007-06-11 04:10:40


Hi Jeff

I ran the NAS Parallel Bechmark and it gives for me
-bash%/export/home/vmarian/fortran/benchmarks/NPB3.2/NPB3.2-MPI/bin$
mpirun -np 1 cg.A.1
--------------------------------------------------------------------------
[0,1,0]: uDAPL on host SERVSOLARIS was unable to find
any NICs.
Another transport will be used instead, although this
may result in
lower performance.
--------------------------------------------------------------------------
 NAS Parallel Benchmarks 3.2 -- CG Benchmark

 Size: 14000
 Iterations: 15
 Number of active processes: 1
 Number of nonzeroes per row: 11
 Eigenvalue shift: .200E+02
 Benchmark completed
 VERIFICATION SUCCESSFUL
 Zeta is 0.171302350540E+02
 Error is 0.512264003323E-13

 CG Benchmark Completed.
 Class = A
 Size = 14000
 Iterations = 15
 Time in seconds = 3.02
 Total processes = 1
 Compiled procs = 1
 Mop/s total = 495.93
 Mop/s/process = 495.93
 Operation type = floating point
 Verification = SUCCESSFUL
 Version = 3.2
 Compile date = 11 Jun 2007

-bash%/export/home/vmarian/fortran/benchmarks/NPB3.2/NPB3.2-MPI/bin$
mpirun -np 2 cg.A.2
--------------------------------------------------------------------------
[0,1,0]: uDAPL on host SERVSOLARIS was unable to find
any NICs.
Another transport will be used instead, although this
may result in
lower performance.
--------------------------------------------------------------------------
--------------------------------------------------------------------------
[0,1,1]: uDAPL on host SERVSOLARIS was unable to find
any NICs.
Another transport will be used instead, although this
may result in
lower performance.
--------------------------------------------------------------------------

 NAS Parallel Benchmarks 3.2 -- CG Benchmark

 Size: 14000
 Iterations: 15
 Number of active processes: 2
 Number of nonzeroes per row: 11
 Eigenvalue shift: .200E+02

 Benchmark completed
 VERIFICATION SUCCESSFUL
 Zeta is 0.171302350540E+02
 Error is 0.522633719989E-13

 CG Benchmark Completed.
 Class = A
 Size = 14000
 Iterations = 15
 Time in seconds = 2.47
 Total processes = 2
 Compiled procs = 2
 Mop/s total = 606.32
 Mop/s/process = 303.16
 Operation type = floating point
 Verification = SUCCESSFUL
 Version = 3.2
 Compile date = 11 Jun 2007

    You can remark that the scalling is not so good
like yours. Maibe I am having comunications problems
between processors.
   You can also remark that I am faster on one process
concared to your processor.

                                       Victor

--- Jeff Pummill <jpummil_at_[hidden]> wrote:

> Perfect! Thanks Jeff!
>
> The NAS Parallel Benchmark on a dual core AMD
> machine now returns this...
> [jpummil_at_localhost bin]$ mpirun -np 1 cg.A.1
> NAS Parallel Benchmarks 3.2 -- CG Benchmark
> CG Benchmark Completed.
> Class = A
> Size = 14000
> Iterations = 15
> Time in seconds = 4.75
> Total processes = 1
> Compiled procs = 1
> Mop/s total = 315.32
>
> ...and...
>
> [jpummil_at_localhost bin]$ mpirun -np 2 cg.A.2
> NAS Parallel Benchmarks 3.2 -- CG Benchmark
> CG Benchmark Completed.
> Class = A
> Size = 14000
> Iterations = 15
> Time in seconds = 2.48
> Total processes = 2
> Compiled procs = 2
> Mop/s total = 604.46
>
> Not quite linear, but one must account for all of
> the OS traffic that
> one core or the other must deal with.
>
>
> Jeff F. Pummill
> Senior Linux Cluster Administrator
> University of Arkansas
> Fayetteville, Arkansas 72701
> (479) 575 - 4590
> http://hpc.uark.edu
>
> "A supercomputer is a device for turning
> compute-bound
> problems into I/O-bound problems." -Seymour Cray
>
>
> Jeff Squyres wrote:
> > Just remove the -L and -l arguments -- OMPI's
> "mpif90" (and other
> > wrapper compilers) will do all that magic for you.
> >
> > Many -L/-l arguments in MPI application Makefiles
> are throwbacks to
> > older versions of MPICH wrapper compilers that
> didn't always work
> > properly. Those days are long gone; most (all?)
> MPI wrapper
> > compilers do not need you to specify -L/-l these
> days.
> >
> >
> >
> > On Jun 10, 2007, at 3:08 PM, Jeff Pummill wrote:
> >
> >
> >> Maybe the "dumb question" of the week, but here
> goes...
> >>
> >> I am trying to compile a piece of code (NPB)
> under OpenMPI and I am
> >> having a problem with specifying the right
> library. Possibly
> >> something I need to define in a LD_LIBRARY_PATH
> statement?
> >>
> >> Using Gnu mpich, the line looked like this...
> >>
> >> FMPI_LIB = -L/opt/mpich/gnu/lib/ -lmpich
> >>
> >> I tried to replace this with...
> >>
> >> FMPI_LIB = -L/usr/lib/openmpi/ -llibmpi
> >>
> >> to which the make responded...
> >> mpif90 -O -o ../bin/cg.A.2 cg.o
> ../common/print_results.o ../common/
> >> randdp.o ../common/timers.o -L/usr/lib/openmpi/
> -llibmpi
> >> /usr/bin/ld: cannot find -llibmpi
> >> collect2: ld returned 1 exit status
> >>
> >> Wrong library file? Setup or path issue?
> >>
> >> --
> >> Jeff F. Pummill
> >> Senior Linux Cluster Administrator
> >> University of Arkansas
> >>
> >> _______________________________________________
> >> users mailing list
> >> users_at_[hidden]
> >>
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>
> >
> >
> >
> > _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users

 
____________________________________________________________________________________
Get your own web address.
Have a HUGE year through Yahoo! Small Business.
http://smallbusiness.yahoo.com/domains/?p=BESTDEAL