HPL_INCLUDES = -I$(INCdir) -I$(INCdir)/$(ARCH) $(LAinc) $(MPinc) # - HPL includes / libraries / specifics. # Cray fcd (fortran character descriptor) forį2CDEFS = -DAdd_ -DF77_INTEGER=int -DStringSunStyle # -DStringCrayStyle : Special option for Cray machines, which uses # 77 string, and the structure is of the form: # -DStringStructVal : A structure is passed by value for each Fortran # Fortran 77 string, and the structure is of the # -DStringStructPtr : The address of a structure is passed by a # passed as an F77_INTEGER after all explicit # tion on the stack, and the string length is then # -DStringSunStyle : The string address is passed at the string loca. # -DF77_INTEGER=short : Fortran 77 INTEGER is a C short. # -DF77_INTEGER=long : Fortran 77 INTEGER is a C long, # -DF77_INTEGER=int : Fortran 77 INTEGER is a C int, # -DAdd_ : the FORTRAN compiler in use is f2c. # -DNoChange : all lower case (IBM RS6000), # -DAdd_ : all lower case and a suffixed underscore (Suns, # 1) name space (How C calls a Fortran 77 routine) **One and only one** option should be chosen in **each** of # necessary to fill out the F2CDEFS variable with the appropriate # a BLAS library featuring a Fortran 77 interface. # You can skip this section if and only if you are not planning to use LAlib = $(LAdir)/libcblas.a $(LAdir)/libatlas.a LAdir = /home/pi/builds/ATLAS/newBuild/lib #LAdir = /mnt/nfs/jahanzeb/bench/atlas/original/ATLAS2/buildDir/lib #LAlib = $(LAdir)/libcblas.a $(LAdir)/libatlas.a The variable LAdir is only used for defining LAinc and LAlib. # header files, LAlib is defined to be the name of the library to be # LAinc tells the C compiler where to find the Linear Algebra library # - Linear Algebra library (BLAS or VSIPL). #MPdir = /mnt/nfs/install/openmpi-install The variable MPdir is only used for defining MPinc and MPlib. # header files, MPlib is defined to be the name of the library to be # MPinc tells the C compiler where to find the Message Passing library # - HPL Directory Structure / HPL library. I'll attach my HPL.dat file to if that helps Has anyone seen this before and would you be able to point me in the right direction, I'm doing this as part of a University project and would really appreciate the help. Mpiexec detected that one or more processes exited with non-zero status, thus causing Per user-direction, the job has been aborted. Primary job terminated normally, but 1 process returnedĪ non-zero exit code. > Need at least 2 processes for these tests > Illegal input in file HPL.dat. HPL ERROR from process # 0, on line 419 of function HPL_pdinfo: What distribution and setup do you have, maybe I am missing something obvious.Ĭode: Select all mpiexec -n 2 -host 192.168.100.50,192.168.100.51. Kind of confused myself now, will go back and check my HPL benchmark on 1 node and see if I can get similar performance as ber0tech. Theoretical peak performance = 0.35 GflopsĪctual Performance = 0.2761 Gflops (2.761e-01) Theoretical peak performance = 700MHz / 2 = 350Mflops = 0.35 GflopsĪctual Performance = 41047 Kflops = 41.047Mflops = 0.04147 GflopsĪnd with ber0tech results HPL benchmark (1 node 700Mhz) Theoretical peak performance = 4 nodes * 900MHz / 2 = 1800Mflops = 1.8 GflopsĮfficiency = Actual Performance GFLOPS / Theoretical Peak Performance GFLOPS = 0.08117 / 1.8 = 4.5%ĭoing the same calculations from the "standard" RPi benchmark (1 node 700Mhz) Using that with my bramble (4 nodes 900Mhz) In the worst case where in your algorithm the result of the current operation is required for the next operation, ie pipelining can't be used, it takes 8 cycles for one operation and we end up with 700MHz/8 = 87.5Mflops. So in the best case it still takes 2 cycles for one operation and then 700MHz/2 = 350 Mflops. Computational tests pass if scaled residuals are less than 16.0 The relative machine precision (eps) is taken to be 1.110223e-16 The following scaled residual check will be computed: The matrix A is randomly generated for each test. The following parameter values will be used: Gflops : Rate of execution for solving the linear system. Time : Time in seconds to solve the linear system. N : The order of the coefficient matrix A. Modified by Julien Langou, University of Colorado DenverĪn explanation of the input/output parameters follows: Modified by Piotr Luszczek, Innovative Computing Laboratory, UTK Clint Whaley, Innovative Computing Laboratory, UTK HPLinpack 2.0 - High-Performance Linpack benchmark - September 10, 2008