hpc-run-1 KVM testing on Ubuntu 18.04 via the Phoronix Test Suite.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 2007171-NE-HPCRUN16802 optimised vm run 1 Processor: Intel Core (11 Cores), Motherboard: RDO OpenStack Compute (1.11.0-2.el7 BIOS), Chipset: Intel 82G33/G31/P35/P31 + ICH9, Memory: 90GB, Disk: 21GB QEMU HDD + 172GB QEMU HDD, Graphics: Red Hat Virtio GPU, Network: 2 x Red Hat Virtio device
OS: Ubuntu 18.04, Kernel: 4.15.0-111-generic (x86_64), Compiler: GCC 7.5.0, File-System: ext4, System Layer: KVM
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -vProcessor Notes: CPU Microcode: 0x1Python Notes: Python 3.6.9Security Notes: itlb_multihit: KVM: Vulnerable + l1tf: Mitigation of PTE Inversion + mds: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown + meltdown: Mitigation of PTI + spec_store_bypass: Vulnerable + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline STIBP: disabled RSB filling + srbds: Unknown: Dependent on hypervisor status + tsx_async_abort: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown
hpc-run-1 OpenBenchmarking.org Phoronix Test Suite Intel Core (11 Cores) RDO OpenStack Compute (1.11.0-2.el7 BIOS) Intel 82G33/G31/P35/P31 + ICH9 90GB 21GB QEMU HDD + 172GB QEMU HDD Red Hat Virtio GPU 2 x Red Hat Virtio device Ubuntu 18.04 4.15.0-111-generic (x86_64) GCC 7.5.0 ext4 KVM Processor Motherboard Chipset Memory Disk Graphics Network OS Kernel Compiler File-System System Layer Hpc-run-1 Benchmarks System Logs - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v - CPU Microcode: 0x1 - Python 3.6.9 - itlb_multihit: KVM: Vulnerable + l1tf: Mitigation of PTE Inversion + mds: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown + meltdown: Mitigation of PTI + spec_store_bypass: Vulnerable + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline STIBP: disabled RSB filling + srbds: Unknown: Dependent on hypervisor status + tsx_async_abort: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown
hpc-run-1 hpcc: G-HPL hpcc: G-Ffte hpcc: EP-DGEMM hpcc: G-Ptrans hpcc: EP-STREAM Triad hpcc: G-Rand Access hpcc: Rand Ring Latency hpcc: Rand Ring Bandwidth hpcc: Max Ping Pong Bandwidth npb: BT.C npb: EP.C npb: EP.D npb: FT.C npb: MG.C npb: SP.B rodinia: OpenMP LavaMD rodinia: OpenMP Leukocyte rodinia: OpenMP CFD Solver rodinia: OpenMP Streamcluster intel-mpi: IMB-P2P PingPong intel-mpi: IMB-MPI1 Exchange intel-mpi: IMB-MPI1 Exchange intel-mpi: IMB-MPI1 PingPong intel-mpi: IMB-MPI1 Sendrecv intel-mpi: IMB-MPI1 Sendrecv optimised vm run 1 212.03333 3.33202 25.08453 2.94095 4.13247 0.04281 0.34193 2.21744 11983.660 20876.02 407.20 409.39 9575.53 18666.08 10600.78 1215.651 215.149 30.222 35.544 1000000 4072.93 201.27 3196.17 2968.55 155.73 OpenBenchmarking.org
HPC Challenge HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org GFLOPS, More Is Better HPC Challenge 1.5.0 Test / Class: G-HPL optimised vm run 1 50 100 150 200 250 SE +/- 0.38, N = 3 212.03 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org GFLOPS, More Is Better HPC Challenge 1.5.0 Test / Class: G-Ffte optimised vm run 1 0.7497 1.4994 2.2491 2.9988 3.7485 SE +/- 0.00911, N = 3 3.33202 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org GFLOPS, More Is Better HPC Challenge 1.5.0 Test / Class: EP-DGEMM optimised vm run 1 6 12 18 24 30 SE +/- 0.08, N = 3 25.08 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org GB/s, More Is Better HPC Challenge 1.5.0 Test / Class: G-Ptrans optimised vm run 1 0.6617 1.3234 1.9851 2.6468 3.3085 SE +/- 0.01490, N = 3 2.94095 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org GB/s, More Is Better HPC Challenge 1.5.0 Test / Class: EP-STREAM Triad optimised vm run 1 0.9298 1.8596 2.7894 3.7192 4.649 SE +/- 0.00058, N = 3 4.13247 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org GUP/s, More Is Better HPC Challenge 1.5.0 Test / Class: G-Random Access optimised vm run 1 0.0096 0.0192 0.0288 0.0384 0.048 SE +/- 0.00021, N = 3 0.04281 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org usecs, Fewer Is Better HPC Challenge 1.5.0 Test / Class: Random Ring Latency optimised vm run 1 0.0769 0.1538 0.2307 0.3076 0.3845 SE +/- 0.00144, N = 3 0.34193 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org GB/s, More Is Better HPC Challenge 1.5.0 Test / Class: Random Ring Bandwidth optimised vm run 1 0.4989 0.9978 1.4967 1.9956 2.4945 SE +/- 0.01993, N = 3 2.21744 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org MB/s, More Is Better HPC Challenge 1.5.0 Test / Class: Max Ping Pong Bandwidth optimised vm run 1 3K 6K 9K 12K 15K SE +/- 17.86, N = 3 11983.66 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
NAS Parallel Benchmarks NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: BT.C optimised vm run 1 4K 8K 12K 16K 20K SE +/- 16.20, N = 3 20876.02 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: EP.C optimised vm run 1 90 180 270 360 450 SE +/- 1.40, N = 3 407.20 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: EP.D optimised vm run 1 90 180 270 360 450 SE +/- 0.25, N = 3 409.39 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: FT.C optimised vm run 1 2K 4K 6K 8K 10K SE +/- 127.65, N = 12 9575.53 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: MG.C optimised vm run 1 4K 8K 12K 16K 20K SE +/- 119.53, N = 3 18666.08 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: SP.B optimised vm run 1 2K 4K 6K 8K 10K SE +/- 15.83, N = 3 10600.78 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
Rodinia Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes select OpenCL, NVIDIA CUDA and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP LavaMD optimised vm run 1 300 600 900 1200 1500 SE +/- 1.98, N = 3 1215.65 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP Leukocyte optimised vm run 1 50 100 150 200 250 SE +/- 0.60, N = 3 215.15 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP CFD Solver optimised vm run 1 7 14 21 28 35 SE +/- 0.41, N = 3 30.22 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP Streamcluster optimised vm run 1 8 16 24 32 40 SE +/- 0.43, N = 15 35.54 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Average Mbytes/sec, More Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Exchange optimised vm run 1 900 1800 2700 3600 4500 SE +/- 9.47, N = 3 4072.93 MAX: 13289.54 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
OpenBenchmarking.org Average usec, Fewer Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Exchange optimised vm run 1 40 80 120 160 200 SE +/- 1.97, N = 3 201.27 MIN: 0.61 / MAX: 4035.45 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
OpenBenchmarking.org Average Mbytes/sec, More Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 PingPong optimised vm run 1 700 1400 2100 2800 3500 SE +/- 47.70, N = 3 3196.17 MIN: 206.78 / MAX: 6638.16 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
OpenBenchmarking.org Average Mbytes/sec, More Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Sendrecv optimised vm run 1 600 1200 1800 2400 3000 SE +/- 23.29, N = 3 2968.55 MAX: 12422.68 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
OpenBenchmarking.org Average usec, Fewer Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Sendrecv optimised vm run 1 30 60 90 120 150 SE +/- 0.22, N = 3 155.73 MIN: 0.34 / MAX: 2590.7 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
optimised vm run 1 Processor: Intel Core (11 Cores), Motherboard: RDO OpenStack Compute (1.11.0-2.el7 BIOS), Chipset: Intel 82G33/G31/P35/P31 + ICH9, Memory: 90GB, Disk: 21GB QEMU HDD + 172GB QEMU HDD, Graphics: Red Hat Virtio GPU, Network: 2 x Red Hat Virtio device
OS: Ubuntu 18.04, Kernel: 4.15.0-111-generic (x86_64), Compiler: GCC 7.5.0, File-System: ext4, System Layer: KVM
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -vProcessor Notes: CPU Microcode: 0x1Python Notes: Python 3.6.9Security Notes: itlb_multihit: KVM: Vulnerable + l1tf: Mitigation of PTE Inversion + mds: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown + meltdown: Mitigation of PTI + spec_store_bypass: Vulnerable + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline STIBP: disabled RSB filling + srbds: Unknown: Dependent on hypervisor status + tsx_async_abort: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown
Testing initiated at 17 July 2020 10:40 by user ubuntu.