Tests
Suites
Latest Results
Search
Register
Login
Popular Tests
Timed Linux Kernel Compilation
7-Zip Compression
SVT-AV1
Stockfish
FFmpeg
PostgreSQL
Newest Tests
WarpX
Epoch
Valkey
Whisperfile
XNNPACK
GROMACS
Recently Updated Tests
Apache CouchDB
PostgreSQL
NAMD
Apache Cassandra
Opus Codec Encoding
ParaView
New & Recently Updated Tests
Recently Updated Suites
Database Test Suite
Machine Learning
Steam
New & Recently Updated Suites
Component Benchmarks
CPUs / Processors
GPUs / Graphics
OpenGL
Disks / Storage
Motherboards
File-Systems
Operating Systems
OpenBenchmarking.org
Corporate / Organization Info
Bug Reports / Feature Requests
HPC Challenge 1.1.5
pts/hpcc-1.1.5
- 03 December 2014 -
Fix for building HPCC on Fedora
downloads.xml
<?xml version="1.0"?> <!--Phoronix Test Suite v5.4.0m7 (Lipki)--> <PhoronixTestSuite> <Downloads> <Package> <URL>http://icl.cs.utk.edu/projectsfiles/hpcc/download/hpcc-1.4.3.tar.gz</URL> <MD5>dd4c2b6b3ec766c469c59d8310778f49</MD5> <SHA256>57cd4fd00065ebb63f5b298e54510be84aecba4b3c4788a0fb8b9fa79c752c79</SHA256> <FileSize>654049</FileSize> </Package> </Downloads> </PhoronixTestSuite>
install.sh
#!/bin/sh tar -zxvf hpcc-1.4.3.tar.gz cd hpcc-1.4.3 # Find MPI To Use if [ ! "X$MPI_PATH" = "X" ] && [ -d $MPI_PATH ] && [ -d $MPI_INCLUDE ] && [ -x $MPI_CC ] && [ -e $MPI_LIBS ] then # PRE-SET MPI echo "Using pre-set environment variables." elif [ -d /usr/lib/openmpi/include ] then # OpenMPI On Ubuntu MPI_PATH=/usr/lib/openmpi MPI_INCLUDE=/usr/lib/openmpi/include MPI_LIBS=/usr/lib/openmpi/lib/libmpi.so MPI_CC=/usr/bin/mpicc.openmpi MPI_VERSION=`$MPI_CC -showme:version 2>&1 | grep MPI | cut -d "(" -f1 | cut -d ":" -f2` elif [ -d /usr/lib64/openmpi ] then # OpenMPI On RHEL MPI_PATH=/usr/lib64/openmpi MPI_INCLUDE=/usr/include/openmpi-x86_64/ MPI_LIBS=/usr/lib64/openmpi/lib/libmpi.so MPI_CC=/usr/lib64/openmpi/bin/mpicc MPI_VERSION=`$MPI_CC -showme:version 2>&1 | grep MPI | cut -d "(" -f1 | cut -d ":" -f2` elif [ -d /usr/lib/mpich/include ] then # MPICH MPI_PATH=/usr/lib/mpich MPI_INCLUDE=/usr/lib/mpich/include MPI_LIBS=/usr/lib/libmpich.so.1.0 MPI_CC=/usr/lib/mpich/bin/mpicc.mpich MPI_VERSION=`$MPI_CC -v 2>&1 | grep "MPICH version"` elif [ -d /usr/include/mpich2 ] then # MPICH2 MPI_PATH=/usr/include/mpich2 MPI_INCLUDE=/usr/include/mpich2 MPI_LIBS=/usr/lib/mpich2/lib/libmpich.so MPI_CC=/usr/bin/mpicc.mpich2 MPI_VERSION=`$MPI_CC -v 2>&1 | grep "MPICH2 version"` elif [ -d /usr/include/mpich2-x86_64 ] then # MPICH2 MPI_PATH=/usr/include/mpich2-x86_64 MPI_INCLUDE=/usr/include/mpich2-x86_64 MPI_LIBS=/usr/lib64/mpich2/lib/libmpich.so MPI_CC=/usr/bin/mpicc MPI_VERSION=`$MPI_CC -v 2>&1 | grep "MPICH2 version"` fi # Find Linear Algebra Package To Use if [ ! "X$LA_PATH" = "X" ] && [ -d $LA_PATH ] && [ -d $LA_INCLUDE ] && [ -e $LA_LIBS ] then # PRE-SET MPI echo "Using pre-set environment variables." elif [ -d /usr/lib/libblas ] then # libblas LA_PATH=/usr/lib LA_INCLUDE=/usr/include LA_LIBS="-lblas" LA_VERSION="BLAS" elif [ -d /usr/lib/openblas-base ] then # OpenBLAS LA_PATH=/usr/lib/openblas-base LA_INCLUDE=/usr/include LA_LIBS=/usr/lib/openblas-base/libopenblas.so.0 LA_VERSION="OpenBLAS" elif [ -d /usr/lib/atlas-base ] then # ATLAS LA_PATH=/usr/lib/atlas-base LA_INCLUDE=/usr/include LA_LIBS="-llapack -lf77blas -lcblas -latlas" LA_VERSION="ATLAS" elif [ -d /usr/lib64/atlas ] then # ATLAS LA_PATH=/usr/lib64/atlas LA_INCLUDE=/usr/include LA_LIBS="-L$LA_PATH -lblas" LA_VERSION="ATLAS" fi if [ ! "X$MPI_VERSION" = "X" ] then VERSION_INFO=$MPI_VERSION if [ ! "X$LA_VERSION" = "X" ] then VERSION_INFO="$LA_VERSION + $VERSION_INFO" fi echo $VERSION_INFO > ~/install-footnote fi if [ "X$CFLAGS" = "X" ] then CFLAGS="-O3 -march=native" fi if [ "X$MPI_LD" = "X" ] then MPI_LD=$MPI_CC fi # Make.pts generation echo " SHELL = /bin/sh CD = cd CP = cp LN_S = ln -s MKDIR = mkdir RM = /bin/rm -f TOUCH = touch ARCH = \$(arch) TOPdir = ../../.. INCdir = \$(TOPdir)/include BINdir = \$(TOPdir)/bin/\$(ARCH) LIBdir = \$(TOPdir)/lib/\$(ARCH) HPLlib = \$(LIBdir)/libhpl.a # MPI MPdir = $MPI_PATH MPinc = -I$MPI_INCLUDE MPlib = $MPI_LIBS # BLAS or VSIPL LAdir = $LA_PATH LAinc = -I$LA_INCLUDE LAlib = $LA_LIBS # F77 / C interface F2CDEFS = # HPL includes / libraries / specifics HPL_INCLUDES = -I\$(INCdir) -I\$(INCdir)/\$(ARCH) \$(LAinc) \$(MPinc) HPL_LIBS = \$(HPLlib) \$(LAlib) \$(MPlib) -lm #HPL_OPTS = -DHPL_CALL_CBLAS HPL_DEFS = \$(F2CDEFS) \$(HPL_OPTS) \$(HPL_INCLUDES) CC = $MPI_CC CCNOOPT = \$(HPL_DEFS) CCFLAGS = \$(HPL_DEFS) -fomit-frame-pointer $CFLAGS -funroll-loops LINKER = $MPI_LD LINKFLAGS = $LDFLAGS ARCHIVER = ar ARFLAGS = r RANLIB = echo " > hpl/Make.pts cd hpl/ make arch=pts cd .. make arch=pts echo $? > ~/install-exit-status cd ~ echo "#!/bin/sh cd hpcc-1.4.3 if [ \"X\$MPI_NUM_THREADS\" = \"X\" ] then MPI_NUM_THREADS=\$NUM_CPU_CORES fi if [ ! \"X\$HOSTFILE\" = \"X\" ] && [ -f \$HOSTFILE ] then \$HOSTFILE=\"--hostfile \$HOSTFILE\" elif [ -f /etc/hostfile ] then \$HOSTFILE=\"--hostfile /etc/hostfile\" fi # HPL.dat generation # http://pic.dhe.ibm.com/infocenter/lnxinfo/v3r0m0/index.jsp?topic=%2Fliaai.hpctune%2Fbaselinehpcc_gccatlas.htm PQ=0 P=\$(echo \"scale=0;sqrt(\$MPI_NUM_THREADS)\" |bc -l) Q=\$P PQ=\$((\$P*\$Q)) while [ \$PQ -ne \$MPI_NUM_THREADS ]; do Q=\$((\$MPI_NUM_THREADS/\$P)) PQ=\$((\$P*\$Q)) if [ \$PQ -ne \$MPI_NUM_THREADS ] && [ \$P -gt 1 ]; then P=\$((\$P-1)); fi done if [ \"X\$N\" = \"X\" ] || [ \"X\$NB\" = \"X\" ] then # SYS_MEMORY * about .62% of that, go from MB to bytes and divide by 8 N=\$(echo \"scale=0;sqrt(\${SYS_MEMORY}*0.62*1048576/8)\" |bc -l) NB=\$((256 - 256 % \$MPI_NUM_THREADS)) N=\$((\$N - \$N % \$NB)) fi echo \"HPLinpack benchmark input file Innovative Computing Laboratory, University of Tennessee HPL.out output file name (if any) 6 device out (6=stdout,7=stderr,file) 1 # of problems sizes (N) \$N 1 # of NBs \$NB NBs 0 PMAP process mapping (0=Row-,1=Column-major) 1 # of process grids (P x Q) \$P Ps \$Q Qs 16.0 threshold 1 # of panel fact 2 PFACTs (0=left, 1=Crout, 2=Right) 1 # of recursive stopping criterium 4 NBMINs (>= 1) 1 # of panels in recursion 2 NDIVs 1 # of recursive panel fact. 2 RFACTs (0=left, 1=Crout, 2=Right) 1 # of broadcast 1 BCASTs (0=1rg,1=1rM,2=2rg,3=2rM,4=Lng,5=LnM) 1 # of lookahead depth 0 DEPTHs (>=0) 1 SWAP (0=bin-exch,1=long,2=mix) 64 swapping threshold 0 L1 in (0=transposed,1=no-transposed) form 0 U in (0=transposed,1=no-transposed) form 1 Equilibration (0=no,1=yes) 8 memory alignment in double (> 0) ##### This line (no. 32) is ignored (it serves as a separator). ###### 0 Number of additional problem sizes for PTRANS 1200 10000 30000 values of N 0 number of additional blocking sizes for PTRANS 40 9 8 13 13 20 16 32 64 values of NB \" > HPL.dat cp HPL.dat hpccinf.txt PATH=\$PATH:$MPI_PATH/bin LD_PRELOAD=$MPI_LIBS mpirun -np \$MPI_NUM_THREADS \$HOSTFILE hpcc echo \$? > ~/test-exit-status cat hpccoutf.txt > \$LOG_FILE" > hpcc chmod +x hpcc
results-definition.xml
<?xml version="1.0"?> <!--Phoronix Test Suite v5.4.0m7 (Lipki)--> <PhoronixTestSuite> <ResultsParser> <OutputTemplate>HPL_Tflops=#_RESULT_#</OutputTemplate> <MatchToTestArguments>HPL</MatchToTestArguments> <MultiplyResultBy>1000</MultiplyResultBy> <ResultScale>GFLOPS</ResultScale> <ResultProportion>HIB</ResultProportion> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>PTRANS_GBs=#_RESULT_#</OutputTemplate> <MatchToTestArguments>PTRANS</MatchToTestArguments> <ResultScale>GB/s</ResultScale> <ResultProportion>HIB</ResultProportion> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>MPIRandomAccess_GUPs=#_RESULT_#</OutputTemplate> <MatchToTestArguments>MPIRANDOMACCESS</MatchToTestArguments> <ResultScale>GUP/s</ResultScale> <ResultProportion>HIB</ResultProportion> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>MPIFFT_Gflops=#_RESULT_#</OutputTemplate> <MatchToTestArguments>MPIFFT</MatchToTestArguments> <ResultScale>GFLOPS</ResultScale> <ResultProportion>HIB</ResultProportion> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>MPIFFT_Gflops=#_RESULT_#</OutputTemplate> <MatchToTestArguments>MPIFFT</MatchToTestArguments> <ResultScale>GFLOP/s</ResultScale> <ResultProportion>HIB</ResultProportion> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>StarSTREAM_Triad=#_RESULT_#</OutputTemplate> <MatchToTestArguments>STARSTREAMTRIAD</MatchToTestArguments> <ResultScale>GB/s</ResultScale> <ResultProportion>HIB</ResultProportion> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>StarDGEMM_Gflops=#_RESULT_#</OutputTemplate> <MatchToTestArguments>STARDGEMMFLOPS</MatchToTestArguments> <ResultScale>GFLOPS</ResultScale> <ResultProportion>HIB</ResultProportion> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>RandomlyOrderedRingLatency_usec=#_RESULT_#</OutputTemplate> <MatchToTestArguments>RRINGLATENCY</MatchToTestArguments> <ResultScale>usecs</ResultScale> <ResultProportion>LIB</ResultProportion> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>RandomlyOrderedRingBandwidth_GBytes=#_RESULT_#</OutputTemplate> <MatchToTestArguments>RRINGBANDWIDTH</MatchToTestArguments> <ResultScale>GB/s</ResultScale> <ResultProportion>HIB</ResultProportion> <ResultPrecision>5</ResultPrecision> </ResultsParser> <ResultsParser> <OutputTemplate>Bandwidth min / avg / max: 9068.765 / 10192.595 / #_RESULT_# MByte/s</OutputTemplate> <MatchToTestArguments>MAXPPBANDWIDTH</MatchToTestArguments> <ResultScale>MB/s</ResultScale> <ResultProportion>HIB</ResultProportion> <ResultPrecision>3</ResultPrecision> </ResultsParser> </PhoronixTestSuite>
test-definition.xml
<?xml version="1.0"?> <!--Phoronix Test Suite v5.4.0m7 (Lipki)--> <PhoronixTestSuite> <TestInformation> <Title>HPC Challenge</Title> <AppVersion>1.4.3</AppVersion> <Description>HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified.</Description> <Proportion>HIB</Proportion> <TimesToRun>3</TimesToRun> <PreInstallMessage>Supported install-time optional variables include $MPI_PATH, $MPI_INCLUDE, $MPI_CC, $MPI_LIBS, $LA_PATH, $LA_INCLUDE, $LA_LIBS, $CFLAGS, $LD_FLAGS, and $MPI_LD</PreInstallMessage> <PostInstallMessage>Supported run-time optional environment variables include $N, $NB, $MPI_NUM_THREADS, $HOSTFILE</PostInstallMessage> </TestInformation> <TestProfile> <Version>1.1.5</Version> <SupportedPlatforms>Linux</SupportedPlatforms> <SoftwareType>Benchmark</SoftwareType> <TestType>Processor</TestType> <License>Free</License> <Status>Verified</Status> <ExternalDependencies>build-utilities, fortran-compiler, openmpi-development, blas-development, atlas-development, lapack-development</ExternalDependencies> <EnvironmentSize>9</EnvironmentSize> <ProjectURL>http://icl.cs.utk.edu/hpcc/index.html</ProjectURL> <InternalTags>SMP, MPI</InternalTags> <Maintainer>Michael Larabel</Maintainer> </TestProfile> <TestSettings> <Default> <AllowCacheShare>TRUE</AllowCacheShare> </Default> <Option> <DisplayName>Test / Class</DisplayName> <Identifier>run-test</Identifier> <DefaultEntry>0</DefaultEntry> <Menu> <Entry> <Name>G-HPL</Name> <Value>HPL</Value> </Entry> <Entry> <Name>G-Ptrans</Name> <Value>PTRANS</Value> </Entry> <Entry> <Name>G-Random Access</Name> <Value>MPIRANDOMACCESS</Value> </Entry> <Entry> <Name>G-Ffte</Name> <Value>MPIFFT</Value> </Entry> <Entry> <Name>EP-STREAM Triad</Name> <Value>STARSTREAMTRIAD</Value> </Entry> <Entry> <Name>EP-DGEMM</Name> <Value>STARDGEMMFLOPS</Value> </Entry> <Entry> <Name>Random Ring Latency</Name> <Value>RRINGLATENCY</Value> </Entry> <Entry> <Name>Random Ring Bandwidth</Name> <Value>RRINGBANDWIDTH</Value> </Entry> <Entry> <Name>Max Ping Pong Bandwidth</Name> <Value>MAXPPBANDWIDTH</Value> </Entry> </Menu> </Option> </TestSettings> </PhoronixTestSuite>