Tests
Suites
Latest Results
Search
Register
Login
Popular Tests
Flexible IO Tester
Timed Linux Kernel Compilation
Blender
7-Zip Compression
SVT-AV1
FFmpeg
Newest Tests
OpenVINO GenAI
Rustls
LiteRT
WarpX
Epoch
Valkey
Recently Updated Tests
srsRAN Project
VVenC
x265
RELION
Llamafile
Llama.cpp
New & Recently Updated Tests
Recently Updated Suites
Machine Learning
Server Motherboard
HPC - High Performance Computing
New & Recently Updated Suites
Component Benchmarks
CPUs / Processors
GPUs / Graphics
OpenGL
Disks / Storage
Motherboards
File-Systems
Operating Systems
OpenBenchmarking.org
Corporate / Organization Info
Bug Reports / Feature Requests
HPC Challenge 0.3.2
pts/hpcc-0.3.2
- 06 November 2013 -
Set LIB on latency, LA reporting.
downloads.xml
<?xml version="1.0"?> <!--Phoronix Test Suite v4.8.3 (Sokndal)--> <PhoronixTestSuite> <Downloads> <Package> <URL>http://icl.cs.utk.edu/projectsfiles/hpcc/download/hpcc-1.4.3.tar.gz</URL> <MD5>dd4c2b6b3ec766c469c59d8310778f49</MD5> <SHA256>57cd4fd00065ebb63f5b298e54510be84aecba4b3c4788a0fb8b9fa79c752c79</SHA256> <FileSize>654049</FileSize> </Package> </Downloads> </PhoronixTestSuite>
install.sh
#!/bin/sh tar -zxvf hpcc-1.4.3.tar.gz cd hpcc-1.4.3 # Find MPI To Use if [ ! "X$MPI_PATH" = "X" ] && [ -d $MPI_PATH ] && [ -d $MPI_INCLUDE ] && [ -x $MPI_CC ] && [ -e $MPI_LIBS ] then # PRE-SET MPI echo "Using pre-set environment variables." elif [ -d /usr/lib/openmpi/include ] then # OpenMPI MPI_PATH=/usr/lib/openmpi MPI_INCLUDE=/usr/lib/openmpi/include MPI_LIBS=/usr/lib/openmpi/lib/libmpi.so MPI_CC=/usr/bin/mpicc.openmpi MPI_VERSION=`$MPI_CC -showme:version 2>&1 | grep MPI | cut -d "(" -f1 | cut -d ":" -f2` elif [ -d /usr/lib/mpich/include ] then # MPICH MPI_PATH=/usr/lib/mpich MPI_INCLUDE=/usr/lib/mpich/include MPI_LIBS=/usr/lib/libmpich.so.1.0 MPI_CC=/usr/lib/mpich/bin/mpicc.mpich MPI_VERSION=`$MPI_CC -v 2>&1 | grep "MPICH version"` elif [ -d /usr/include/mpich2 ] then # MPICH2 MPI_PATH=/usr/include/mpich2 MPI_INCLUDE=/usr/include/mpich2 MPI_LIBS=/usr/lib/mpich2/lib/libmpich.so MPI_CC=/usr/bin/mpicc.mpich2 MPI_VERSION=`$MPI_CC -v 2>&1 | grep "MPICH2 version"` fi # Find Linear Algebra Package To Use if [ ! "X$LA_PATH" = "X" ] && [ -d $LA_PATH ] && [ -d $LA_INCLUDE ] && [ -e $LA_LIBS ] then # PRE-SET MPI echo "Using pre-set environment variables." elif [ -d /usr/lib/libblas ] then # libblas LA_PATH=/usr/lib LA_INCLUDE=/usr/include LA_LIBS="-lblas" LA_VERSION="BLAS" elif [ -d /usr/lib/openblas-base ] then # OpenBLAS LA_PATH=/usr/lib/openblas-base LA_INCLUDE=/usr/include LA_LIBS=/usr/lib/openblas-base/libopenblas.so.0 LA_VERSION="OpenBLAS" elif [ -d /usr/lib/atlas-base ] then # ATLAS LA_PATH=/usr/lib/atlas-base LA_INCLUDE=/usr/include LA_LIBS="-llapack -lf77blas -lcblas -latlas" LA_VERSION="ATLAS" fi if [ ! "X$MPI_VERSION" = "X" ] then VERSION_INFO=$MPI_VERSION if [ ! "X$LA_VERSION" = "X" ] then VERSION_INFO="$LA_VERSION + $VERSION_INFO" fi echo $VERSION_INFO > ~/install-footnote fi if [ "X$CFLAGS" = "X" ] then CFLAGS="-O3 -march=native" fi # Make.pts generation echo " SHELL = /bin/sh CD = cd CP = cp LN_S = ln -s MKDIR = mkdir RM = /bin/rm -f TOUCH = touch ARCH = \$(arch) TOPdir = ../../.. INCdir = \$(TOPdir)/include BINdir = \$(TOPdir)/bin/\$(ARCH) LIBdir = \$(TOPdir)/lib/\$(ARCH) HPLlib = \$(LIBdir)/libhpl.a # MPI MPdir = $MPI_PATH MPinc = -I$MPI_INCLUDE MPlib = $MPI_LIBS # BLAS or VSIPL LAdir = $LA_PATH LAinc = -I$LA_INCLUDE LAlib = $LA_LIBS # F77 / C interface F2CDEFS = # HPL includes / libraries / specifics HPL_INCLUDES = -I\$(INCdir) -I\$(INCdir)/\$(ARCH) \$(LAinc) \$(MPinc) HPL_LIBS = \$(HPLlib) \$(LAlib) \$(MPlib) -lm HPL_OPTS = -DHPL_CALL_CBLAS HPL_DEFS = \$(F2CDEFS) \$(HPL_OPTS) \$(HPL_INCLUDES) CC = $MPI_CC CCNOOPT = \$(HPL_DEFS) CCFLAGS = \$(HPL_DEFS) -fomit-frame-pointer $CFLAGS -funroll-loops LINKER = $MPI_CC LINKFLAGS = \$(CCFLAGS) ARCHIVER = ar ARFLAGS = r RANLIB = echo " > hpl/Make.pts cd hpl/ make arch=pts cd .. make arch=pts echo $? > ~/install-exit-status cd ~ echo "#!/bin/sh cd hpcc-1.4.3 if [ \"X\$OMP_NUM_THREADS\" = \"X\" ] then OMP_NUM_THREADS=\$NUM_CPU_CORES fi if [ ! \"X\$HOSTFILE\" = \"X\" ] && [ -f \$HOSTFILE ] then \$HOSTFILE=\"--hostfile \$HOSTFILE\" elif [ -f /etc/hostfile ] then \$HOSTFILE=\"--hostfile /etc/hostfile\" else \$HOSTFILE=\"\" fi # HPL.dat generation N=\$(echo \"scale=0;sqrt(\${OMP_NUM_THREADS}*\${SYS_MEMORY}*128)\" |bc -l) PQ=0 P=\$(echo \"scale=0;sqrt(\$OMP_NUM_THREADS)\" |bc -l) Q=\$P PQ=\$((\$P*\$Q)) while [ \$PQ -ne \$OMP_NUM_THREADS ]; do Q=\$((\$OMP_NUM_THREADS/\$P)) PQ=\$((\$P*\$Q)) if [ \$PQ -ne \$OMP_NUM_THREADS ] && [ \$P -gt 1 ]; then P=\$((\$P-1)); fi done NB=\$((256 - 256 % \$OMP_NUM_THREADS)) echo \"HPLinpack benchmark input file Innovative Computing Laboratory, University of Tennessee HPL.out output file name (if any) 6 device out (6=stdout,7=stderr,file) 1 # of problems sizes (N) \$N 1 # of NBs \$NB NBs 0 PMAP process mapping (0=Row-,1=Column-major) 1 # of process grids (P x Q) \$P Ps \$Q Qs 16.0 threshold 1 # of panel fact 2 PFACTs (0=left, 1=Crout, 2=Right) 1 # of recursive stopping criterium 4 NBMINs (>= 1) 1 # of panels in recursion 2 NDIVs 1 # of recursive panel fact. 2 RFACTs (0=left, 1=Crout, 2=Right) 1 # of broadcast 1 BCASTs (0=1rg,1=1rM,2=2rg,3=2rM,4=Lng,5=LnM) 1 # of lookahead depth 0 DEPTHs (>=0) 1 SWAP (0=bin-exch,1=long,2=mix) 64 swapping threshold 0 L1 in (0=transposed,1=no-transposed) form 0 U in (0=transposed,1=no-transposed) form 1 Equilibration (0=no,1=yes) 8 memory alignment in double (> 0) ##### This line (no. 32) is ignored (it serves as a separator). ###### 0 Number of additional problem sizes for PTRANS 1200 10000 30000 values of N 0 number of additional blocking sizes for PTRANS 40 9 8 13 13 20 16 32 64 values of NB \" > HPL.dat cp HPL.dat hpccinf.txt mpirun -np \$OMP_NUM_THREADS \$HOSTFILE hpcc echo \$? > ~/test-exit-status cat hpccoutf.txt > \$LOG_FILE" > hpcc chmod +x hpcc
results-definition.xml
<?xml version="1.0"?> <!--Phoronix Test Suite v4.8.3 (Sokndal)--> <PhoronixTestSuite> <ResultsParser> <OutputTemplate>HPL_Tflops=#_RESULT_#</OutputTemplate> <MatchToTestArguments>HPL</MatchToTestArguments> <LineHint>HPL_Tflops</LineHint> <MultiplyResultBy>1000</MultiplyResultBy> <ResultScale>GFLOPS</ResultScale> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>PTRANS_GBs=#_RESULT_#</OutputTemplate> <MatchToTestArguments>PTRANS</MatchToTestArguments> <LineHint>PTRANS_GBs</LineHint> <ResultScale>GB/s</ResultScale> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>MPIRandomAccess_GUPs=#_RESULT_#</OutputTemplate> <MatchToTestArguments>MPIRANDOMACCESS</MatchToTestArguments> <LineHint>MPIRandomAccess_GUPs</LineHint> <ResultScale>GUP/s</ResultScale> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>MPIFFT_Gflops=#_RESULT_#</OutputTemplate> <MatchToTestArguments>MPIFFT</MatchToTestArguments> <LineHint>MPIFFT_Gflops</LineHint> <ResultScale>GFLOP/s</ResultScale> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>Randomly Ordered Ring Latency: #_RESULT_# msecs</OutputTemplate> <MatchToTestArguments>RRINGLATENCY</MatchToTestArguments> <LineHint>Randomly Ordered Ring Latency</LineHint> <ResultScale>msecs</ResultScale> <ResultProportion>LIB</ResultProportion> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>Randomly Ordered Ring Bandwidth: #_RESULT_# MB/s</OutputTemplate> <MatchToTestArguments>RRINGBANDWIDTH</MatchToTestArguments> <LineHint>Randomly Ordered Ring Bandwidth</LineHint> <ResultScale>MB/s</ResultScale> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>Bandwidth min / avg / max: 9068.765 / 10192.595 / #_RESULT_# MByte/s</OutputTemplate> <MatchToTestArguments>MAXPPBANDWIDTH</MatchToTestArguments> <LineHint>Bandwidth min / avg / max</LineHint> <ResultScale>MB/s</ResultScale> <ResultPrecision>3</ResultPrecision> </ResultsParser> </PhoronixTestSuite>
test-definition.xml
<?xml version="1.0"?> <!--Phoronix Test Suite v4.8.3 (Sokndal)--> <PhoronixTestSuite> <TestInformation> <Title>HPC Challenge</Title> <AppVersion>1.4.3</AppVersion> <Description>HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAm, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Most of this HPCC super-computing testing has been done on Ubuntu and Red Hat Enterprise Linux installations.</Description> <Proportion>HIB</Proportion> <TimesToRun>3</TimesToRun> </TestInformation> <TestProfile> <Version>0.3.2</Version> <SupportedPlatforms>Linux</SupportedPlatforms> <SoftwareType>Benchmark</SoftwareType> <TestType>Processor</TestType> <License>Free</License> <Status>Unverified</Status> <ExternalDependencies>build-utilities, fortran-compiler, openmpi-development, blas-development, atlas-development</ExternalDependencies> <EnvironmentSize>9</EnvironmentSize> <ProjectURL>http://icl.cs.utk.edu/hpcc/index.html</ProjectURL> <InternalTags>SMP, MPI</InternalTags> <Maintainer>Michael Larabel</Maintainer> </TestProfile> <TestSettings> <Default> <AllowCacheShare>TRUE</AllowCacheShare> </Default> <Option> <DisplayName>Test / Class</DisplayName> <Identifier>run-test</Identifier> <Menu> <Entry> <Name>HPL</Name> <Value>HPL</Value> </Entry> <Entry> <Name>PTRANS</Name> <Value>PTRANS</Value> </Entry> <Entry> <Name>MPIRandomAccess</Name> <Value>MPIRANDOMACCESS</Value> </Entry> <Entry> <Name>MPI FFT</Name> <Value>MPIFFT</Value> </Entry> <Entry> <Name>Random Ring Latency</Name> <Value>RRINGLATENCY</Value> </Entry> <Entry> <Name>Random Ring Bandwidth</Name> <Value>RRINGBANDWIDTH</Value> </Entry> <Entry> <Name>Max Ping Pong Bandwidth</Name> <Value>MAXPPBANDWIDTH</Value> </Entry> </Menu> </Option> </TestSettings> </PhoronixTestSuite>