HPCC EPYC 7642

AMD EPYC 7642 48-Core testing with a ASRockRack EPYCD8 (P2.10 BIOS) and llvmpipe 126GB on Ubuntu 19.10 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 1911180-HU-HPCCEPYC762
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
EPYC 7642
November 18 2019
  2 Hours, 2 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


HPCC EPYC 7642OpenBenchmarking.orgPhoronix Test SuiteAMD EPYC 7642 48-Core @ 2.30GHz (48 Cores / 96 Threads)ASRockRack EPYCD8 (P2.10 BIOS)AMD Starship/Matisse129024MB280GB INTEL SSDPED1D280GAllvmpipe 126GBAMD Starship/Matisse2 x Intel I350Ubuntu 19.105.3.0-050300-generic (x86_64)GNOME Shell 3.34.1X Server 1.20.5modesetting 1.20.5GCC 9.2.1 20191008ext41024x768ProcessorMotherboardChipsetMemoryDiskGraphicsAudioNetworkOSKernelDesktopDisplay ServerDisplay DriverCompilerFile-SystemScreen ResolutionHPCC EPYC 7642 BenchmarksSystem Logs- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-offload-targets=nvptx-none,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: acpi-cpufreq ondemand - CPU Microcode: 0x830101c- l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional IBRS_FW STIBP: always-on RSB filling

HPCC EPYC 7642hpcc: G-Ptranshpcc: EP-STREAM Triadhpcc: Rand Ring Bandwidthhpcc: G-Fftehpcc: G-HPLhpcc: G-Fftehpcc: EP-DGEMMhpcc: G-Rand Accesshpcc: Max Ping Pong Bandwidthhpcc: Rand Ring LatencyEPYC 764211.868102.125651.5125316.92980112.5130016.9298021.256700.1816515507.4221.12956OpenBenchmarking.org

HPC Challenge

HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGB/s, More Is BetterHPC Challenge 1.5.0Test / Class: G-PtransEPYC 7642369121511.871. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. OpenBLAS + Open MPI 3.1.3

OpenBenchmarking.orgGB/s, More Is BetterHPC Challenge 1.5.0Test / Class: EP-STREAM TriadEPYC 76420.47830.95661.43491.91322.39152.125651. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. OpenBLAS + Open MPI 3.1.3

OpenBenchmarking.orgGB/s, More Is BetterHPC Challenge 1.5.0Test / Class: Random Ring BandwidthEPYC 76420.34030.68061.02091.36121.70151.512531. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. OpenBLAS + Open MPI 3.1.3

OpenBenchmarking.orgGFLOP/s, More Is BetterHPC Challenge 1.5.0Test / Class: G-FfteEPYC 76424812162016.931. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. OpenBLAS + Open MPI 3.1.3

OpenBenchmarking.orgGFLOPS, More Is BetterHPC Challenge 1.5.0Test / Class: G-HPLEPYC 7642306090120150112.511. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. OpenBLAS + Open MPI 3.1.3

OpenBenchmarking.orgGFLOPS, More Is BetterHPC Challenge 1.5.0Test / Class: G-FfteEPYC 76424812162016.931. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. OpenBLAS + Open MPI 3.1.3

OpenBenchmarking.orgGFLOPS, More Is BetterHPC Challenge 1.5.0Test / Class: EP-DGEMMEPYC 764251015202521.261. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. OpenBLAS + Open MPI 3.1.3

OpenBenchmarking.orgGUP/s, More Is BetterHPC Challenge 1.5.0Test / Class: G-Random AccessEPYC 76420.04090.08180.12270.16360.20450.181651. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. OpenBLAS + Open MPI 3.1.3

OpenBenchmarking.orgMB/s, More Is BetterHPC Challenge 1.5.0Test / Class: Max Ping Pong BandwidthEPYC 76423K6K9K12K15K15507.421. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. OpenBLAS + Open MPI 3.1.3

OpenBenchmarking.orgusecs, Fewer Is BetterHPC Challenge 1.5.0Test / Class: Random Ring LatencyEPYC 76420.25420.50840.76261.01681.2711.129561. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. OpenBLAS + Open MPI 3.1.3