hpc-run-1 KVM testing on Ubuntu 18.04 via the Phoronix Test Suite.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 2007269-NE-HPCRUN10961 vm.optimized-flavor Processor: Intel Core (8 Cores), Motherboard: RDO OpenStack Compute (1.11.0-2.el7 BIOS), Chipset: Intel 440FX 82441FX PMC, Memory: 26GB, Disk: 20GB, Graphics: Cirrus Logic GD 5446, Network: Red Hat Virtio device
OS: Ubuntu 18.04, Kernel: 4.15.0-111-generic (x86_64), Compiler: GCC 7.5.0, File-System: ext4, System Layer: KVM
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -vProcessor Notes: CPU Microcode: 0x1Python Notes: Python 3.6.9Security Notes: itlb_multihit: KVM: Vulnerable + l1tf: Mitigation of PTE Inversion + mds: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown + meltdown: Mitigation of PTI + spec_store_bypass: Vulnerable + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline STIBP: disabled RSB filling + srbds: Unknown: Dependent on hypervisor status + tsx_async_abort: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown
hpc-run-1 OpenBenchmarking.org Phoronix Test Suite Intel Core (8 Cores) RDO OpenStack Compute (1.11.0-2.el7 BIOS) Intel 440FX 82441FX PMC 26GB 20GB Cirrus Logic GD 5446 Red Hat Virtio device Ubuntu 18.04 4.15.0-111-generic (x86_64) GCC 7.5.0 ext4 KVM Processor Motherboard Chipset Memory Disk Graphics Network OS Kernel Compiler File-System System Layer Hpc-run-1 Benchmarks System Logs - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v - CPU Microcode: 0x1 - Python 3.6.9 - itlb_multihit: KVM: Vulnerable + l1tf: Mitigation of PTE Inversion + mds: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown + meltdown: Mitigation of PTI + spec_store_bypass: Vulnerable + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline STIBP: disabled RSB filling + srbds: Unknown: Dependent on hypervisor status + tsx_async_abort: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown
hpc-run-1 hpcg: npb: BT.C npb: CG.C npb: EP.C npb: EP.D npb: FT.C npb: IS.D npb: LU.C npb: MG.C npb: SP.B hpcc: G-HPL hpcc: G-Ffte hpcc: EP-DGEMM hpcc: G-Ptrans hpcc: EP-STREAM Triad hpcc: G-Rand Access hpcc: Rand Ring Latency hpcc: Rand Ring Bandwidth hpcc: Max Ping Pong Bandwidth rodinia: OpenMP LavaMD rodinia: OpenMP Leukocyte rodinia: OpenMP CFD Solver rodinia: OpenMP Streamcluster intel-mpi: IMB-P2P PingPong intel-mpi: IMB-MPI1 Exchange intel-mpi: IMB-MPI1 Exchange intel-mpi: IMB-MPI1 PingPong intel-mpi: IMB-MPI1 Sendrecv intel-mpi: IMB-MPI1 Sendrecv vm.optimized-flavor 7.43482 10674.52 3876.64 316.26 316.84 10121.14 432.50 23891.04 20130.11 6183.01 80.04833 3.57392 25.59557 1.41594 5.45644 0.03992 0.32910 3.51222 12304.259 1572.148 243.156 34.749 26.836 3797622.5185185 6939.47 128.24 5326.92 5705.88 105.24 OpenBenchmarking.org
NAS Parallel Benchmarks NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: BT.C vm.optimized-flavor 2K 4K 6K 8K 10K SE +/- 6.22, N = 3 10674.52 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: CG.C vm.optimized-flavor 800 1600 2400 3200 4000 SE +/- 1.59, N = 3 3876.64 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: EP.C vm.optimized-flavor 70 140 210 280 350 SE +/- 0.14, N = 3 316.26 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: EP.D vm.optimized-flavor 70 140 210 280 350 SE +/- 0.16, N = 3 316.84 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: FT.C vm.optimized-flavor 2K 4K 6K 8K 10K SE +/- 38.17, N = 3 10121.14 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: IS.D vm.optimized-flavor 90 180 270 360 450 SE +/- 0.35, N = 3 432.50 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: LU.C vm.optimized-flavor 5K 10K 15K 20K 25K SE +/- 32.02, N = 3 23891.04 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: MG.C vm.optimized-flavor 4K 8K 12K 16K 20K SE +/- 14.44, N = 3 20130.11 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: SP.B vm.optimized-flavor 1300 2600 3900 5200 6500 SE +/- 19.14, N = 3 6183.01 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
HPC Challenge HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org GFLOPS, More Is Better HPC Challenge 1.5.0 Test / Class: G-HPL vm.optimized-flavor 20 40 60 80 100 SE +/- 0.82, N = 3 80.05 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org GFLOPS, More Is Better HPC Challenge 1.5.0 Test / Class: G-Ffte vm.optimized-flavor 0.8041 1.6082 2.4123 3.2164 4.0205 SE +/- 0.01696, N = 3 3.57392 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org GFLOPS, More Is Better HPC Challenge 1.5.0 Test / Class: EP-DGEMM vm.optimized-flavor 6 12 18 24 30 SE +/- 0.08, N = 3 25.60 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org GB/s, More Is Better HPC Challenge 1.5.0 Test / Class: G-Ptrans vm.optimized-flavor 0.3186 0.6372 0.9558 1.2744 1.593 SE +/- 0.00269, N = 3 1.41594 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org GB/s, More Is Better HPC Challenge 1.5.0 Test / Class: EP-STREAM Triad vm.optimized-flavor 1.2277 2.4554 3.6831 4.9108 6.1385 SE +/- 0.00189, N = 3 5.45644 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org GUP/s, More Is Better HPC Challenge 1.5.0 Test / Class: G-Random Access vm.optimized-flavor 0.009 0.018 0.027 0.036 0.045 SE +/- 0.00037, N = 3 0.03992 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org usecs, Fewer Is Better HPC Challenge 1.5.0 Test / Class: Random Ring Latency vm.optimized-flavor 0.074 0.148 0.222 0.296 0.37 SE +/- 0.00136, N = 3 0.32910 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org GB/s, More Is Better HPC Challenge 1.5.0 Test / Class: Random Ring Bandwidth vm.optimized-flavor 0.7902 1.5804 2.3706 3.1608 3.951 SE +/- 0.02509, N = 3 3.51222 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
OpenBenchmarking.org MB/s, More Is Better HPC Challenge 1.5.0 Test / Class: Max Ping Pong Bandwidth vm.optimized-flavor 3K 6K 9K 12K 15K SE +/- 26.34, N = 3 12304.26 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
Rodinia Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes select OpenCL, NVIDIA CUDA and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP LavaMD vm.optimized-flavor 300 600 900 1200 1500 SE +/- 1.49, N = 3 1572.15 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP Leukocyte vm.optimized-flavor 50 100 150 200 250 SE +/- 0.47, N = 3 243.16 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP CFD Solver vm.optimized-flavor 8 16 24 32 40 SE +/- 0.12, N = 3 34.75 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP Streamcluster vm.optimized-flavor 6 12 18 24 30 SE +/- 0.04, N = 3 26.84 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Average Mbytes/sec, More Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Exchange vm.optimized-flavor 1500 3000 4500 6000 7500 SE +/- 38.56, N = 3 6939.47 MAX: 19308.99 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
OpenBenchmarking.org Average usec, Fewer Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Exchange vm.optimized-flavor 30 60 90 120 150 SE +/- 1.12, N = 3 128.24 MIN: 0.6 / MAX: 2973.73 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
OpenBenchmarking.org Average Mbytes/sec, More Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 PingPong vm.optimized-flavor 1100 2200 3300 4400 5500 SE +/- 43.94, N = 3 5326.92 MIN: 356.36 / MAX: 10052.7 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
OpenBenchmarking.org Average Mbytes/sec, More Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Sendrecv vm.optimized-flavor 1200 2400 3600 4800 6000 SE +/- 42.36, N = 3 5705.88 MAX: 19813.98 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
OpenBenchmarking.org Average usec, Fewer Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Sendrecv vm.optimized-flavor 20 40 60 80 100 SE +/- 0.34, N = 3 105.24 MIN: 0.32 / MAX: 2133.08 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
vm.optimized-flavor Processor: Intel Core (8 Cores), Motherboard: RDO OpenStack Compute (1.11.0-2.el7 BIOS), Chipset: Intel 440FX 82441FX PMC, Memory: 26GB, Disk: 20GB, Graphics: Cirrus Logic GD 5446, Network: Red Hat Virtio device
OS: Ubuntu 18.04, Kernel: 4.15.0-111-generic (x86_64), Compiler: GCC 7.5.0, File-System: ext4, System Layer: KVM
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -vProcessor Notes: CPU Microcode: 0x1Python Notes: Python 3.6.9Security Notes: itlb_multihit: KVM: Vulnerable + l1tf: Mitigation of PTE Inversion + mds: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown + meltdown: Mitigation of PTI + spec_store_bypass: Vulnerable + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline STIBP: disabled RSB filling + srbds: Unknown: Dependent on hypervisor status + tsx_async_abort: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown
Testing initiated at 26 July 2020 13:08 by user ubuntu.