KVM testing on Ubuntu 18.04 via the Phoronix Test Suite.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 2007224-NE-HPCRUN12663 hpc-run-1 - Phoronix Test Suite hpc-run-1 KVM testing on Ubuntu 18.04 via the Phoronix Test Suite.
HTML result view exported from: https://openbenchmarking.org/result/2007224-NE-HPCRUN12663 .
hpc-run-1 Processor Motherboard Chipset Memory Disk Graphics Network OS Kernel Compiler File-System System Layer optimized.v1.xlarge optimized.vm.xlarge 2 x Intel Core (Broadwell) (30 Cores) RDO OpenStack Compute (1.11.0-2.el7 BIOS) Intel 82G33/G31/P35/P31 + ICH9 100GB 21GB QEMU HDD + 365GB QEMU HDD Red Hat Virtio GPU Red Hat Virtio device Ubuntu 18.04 4.15.0-111-generic (x86_64) GCC 7.5.0 ext4 KVM OpenBenchmarking.org Compiler Details - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v Processor Details - CPU Microcode: 0x1 Python Details - Python 3.6.9 Security Details - itlb_multihit: KVM: Vulnerable + l1tf: Mitigation of PTE Inversion + mds: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown + meltdown: Mitigation of PTI + spec_store_bypass: Vulnerable + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline STIBP: disabled RSB filling + srbds: Unknown: Dependent on hypervisor status + tsx_async_abort: Vulnerable: Clear buffers attempted no microcode; SMT Host state unknown
hpc-run-1 hpcg: npb: BT.C npb: EP.C npb: EP.D npb: FT.C npb: LU.C npb: MG.C npb: SP.B hpcc: G-HPL hpcc: G-Ffte hpcc: EP-DGEMM hpcc: G-Ptrans hpcc: EP-STREAM Triad hpcc: G-Rand Access hpcc: Rand Ring Latency hpcc: Rand Ring Bandwidth hpcc: Max Ping Pong Bandwidth rodinia: OpenMP LavaMD rodinia: OpenMP Leukocyte rodinia: OpenMP CFD Solver rodinia: OpenMP Streamcluster intel-mpi: IMB-P2P PingPong intel-mpi: IMB-MPI1 Exchange intel-mpi: IMB-MPI1 Exchange intel-mpi: IMB-MPI1 PingPong intel-mpi: IMB-MPI1 Sendrecv intel-mpi: IMB-MPI1 Sendrecv optimized.v1.xlarge optimized.vm.xlarge 16.3233 52854.17 1180.31 1182.31 18375.88 69523.79 37114.36 30909.75 16.3123 52891.94 1137.29 1177.90 18151.56 69684.82 36875.38 30770.19 123.58000 6.32367 22.76757 6.34966 3.21699 0.08061 0.61320 1.40170 12056.515 431.919 107.427 12.797 17.993 10016813.8412697 3655.83 387.07 3962.4622 2633.68 221.30 OpenBenchmarking.org
High Performance Conjugate Gradient OpenBenchmarking.org GFLOP/s, More Is Better High Performance Conjugate Gradient 3.1 optimized.v1.xlarge optimized.vm.xlarge 4 8 12 16 20 SE +/- 0.03, N = 3 SE +/- 0.04, N = 3 16.32 16.31 1. (CXX) g++ options: -O3 -ffast-math -ftree-vectorize -pthread -lmpi_cxx -lmpi
NAS Parallel Benchmarks Test / Class: BT.C OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: BT.C optimized.v1.xlarge optimized.vm.xlarge 11K 22K 33K 44K 55K SE +/- 38.65, N = 3 SE +/- 57.16, N = 3 52854.17 52891.94 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
NAS Parallel Benchmarks Test / Class: EP.C OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: EP.C optimized.v1.xlarge optimized.vm.xlarge 300 600 900 1200 1500 SE +/- 2.04, N = 3 SE +/- 18.46, N = 15 1180.31 1137.29 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
NAS Parallel Benchmarks Test / Class: EP.D OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: EP.D optimized.v1.xlarge optimized.vm.xlarge 300 600 900 1200 1500 SE +/- 0.73, N = 3 SE +/- 0.82, N = 3 1182.31 1177.90 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
NAS Parallel Benchmarks Test / Class: FT.C OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: FT.C optimized.v1.xlarge optimized.vm.xlarge 4K 8K 12K 16K 20K SE +/- 262.25, N = 4 SE +/- 310.47, N = 3 18375.88 18151.56 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
NAS Parallel Benchmarks Test / Class: LU.C OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: LU.C optimized.v1.xlarge optimized.vm.xlarge 15K 30K 45K 60K 75K SE +/- 392.59, N = 3 SE +/- 56.09, N = 3 69523.79 69684.82 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
NAS Parallel Benchmarks Test / Class: MG.C OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: MG.C optimized.v1.xlarge optimized.vm.xlarge 8K 16K 24K 32K 40K SE +/- 274.84, N = 3 SE +/- 134.98, N = 3 37114.36 36875.38 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
NAS Parallel Benchmarks Test / Class: SP.B OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: SP.B optimized.v1.xlarge optimized.vm.xlarge 7K 14K 21K 28K 35K SE +/- 241.19, N = 3 SE +/- 72.57, N = 3 30909.75 30770.19 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 2.1.1
HPC Challenge Test / Class: G-HPL OpenBenchmarking.org GFLOPS, More Is Better HPC Challenge 1.5.0 Test / Class: G-HPL optimized.vm.xlarge 30 60 90 120 150 SE +/- 0.11, N = 3 123.58 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
HPC Challenge Test / Class: G-Ffte OpenBenchmarking.org GFLOPS, More Is Better HPC Challenge 1.5.0 Test / Class: G-Ffte optimized.vm.xlarge 2 4 6 8 10 SE +/- 0.00551, N = 3 6.32367 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
HPC Challenge Test / Class: EP-DGEMM OpenBenchmarking.org GFLOPS, More Is Better HPC Challenge 1.5.0 Test / Class: EP-DGEMM optimized.vm.xlarge 5 10 15 20 25 SE +/- 0.29, N = 3 22.77 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
HPC Challenge Test / Class: G-Ptrans OpenBenchmarking.org GB/s, More Is Better HPC Challenge 1.5.0 Test / Class: G-Ptrans optimized.vm.xlarge 2 4 6 8 10 SE +/- 0.02493, N = 3 6.34966 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
HPC Challenge Test / Class: EP-STREAM Triad OpenBenchmarking.org GB/s, More Is Better HPC Challenge 1.5.0 Test / Class: EP-STREAM Triad optimized.vm.xlarge 0.7238 1.4476 2.1714 2.8952 3.619 SE +/- 0.01061, N = 3 3.21699 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
HPC Challenge Test / Class: G-Random Access OpenBenchmarking.org GUP/s, More Is Better HPC Challenge 1.5.0 Test / Class: G-Random Access optimized.vm.xlarge 0.0181 0.0362 0.0543 0.0724 0.0905 SE +/- 0.00369, N = 3 0.08061 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
HPC Challenge Test / Class: Random Ring Latency OpenBenchmarking.org usecs, Fewer Is Better HPC Challenge 1.5.0 Test / Class: Random Ring Latency optimized.vm.xlarge 0.138 0.276 0.414 0.552 0.69 SE +/- 0.00124, N = 3 0.61320 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
HPC Challenge Test / Class: Random Ring Bandwidth OpenBenchmarking.org GB/s, More Is Better HPC Challenge 1.5.0 Test / Class: Random Ring Bandwidth optimized.vm.xlarge 0.3154 0.6308 0.9462 1.2616 1.577 SE +/- 0.01062, N = 3 1.40170 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
HPC Challenge Test / Class: Max Ping Pong Bandwidth OpenBenchmarking.org MB/s, More Is Better HPC Challenge 1.5.0 Test / Class: Max Ping Pong Bandwidth optimized.vm.xlarge 3K 6K 9K 12K 15K SE +/- 16.04, N = 3 12056.52 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. ATLAS + Open MPI 2.1.1
Rodinia Test: OpenMP LavaMD OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP LavaMD optimized.vm.xlarge 90 180 270 360 450 SE +/- 0.29, N = 3 431.92 1. (CXX) g++ options: -O2 -lOpenCL
Rodinia Test: OpenMP Leukocyte OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP Leukocyte optimized.vm.xlarge 20 40 60 80 100 SE +/- 0.37, N = 3 107.43 1. (CXX) g++ options: -O2 -lOpenCL
Rodinia Test: OpenMP CFD Solver OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP CFD Solver optimized.vm.xlarge 3 6 9 12 15 SE +/- 0.05, N = 3 12.80 1. (CXX) g++ options: -O2 -lOpenCL
Rodinia Test: OpenMP Streamcluster OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP Streamcluster optimized.vm.xlarge 4 8 12 16 20 SE +/- 0.18, N = 3 17.99 1. (CXX) g++ options: -O2 -lOpenCL
Intel MPI Benchmarks Test: IMB-P2P PingPong OpenBenchmarking.org Average Msg/sec, More Is Better Intel MPI Benchmarks 2019.3 Test: IMB-P2P PingPong optimized.vm.xlarge 2M 4M 6M 8M 10M SE +/- 29013.76, N = 3 10016813.84 MIN: 2965 / MAX: 31040839 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
Intel MPI Benchmarks Test: IMB-MPI1 Exchange OpenBenchmarking.org Average Mbytes/sec, More Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Exchange optimized.vm.xlarge 800 1600 2400 3200 4000 SE +/- 46.29, N = 3 3655.83 MIN: 3.44 / MAX: 15331.79 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
Intel MPI Benchmarks Test: IMB-MPI1 Exchange OpenBenchmarking.org Average usec, Fewer Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Exchange optimized.vm.xlarge 80 160 240 320 400 SE +/- 3.78, N = 3 387.07 MIN: 1.11 / MAX: 5678.55 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
Intel MPI Benchmarks Test: IMB-MPI1 PingPong OpenBenchmarking.org Average Mbytes/sec, More Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 PingPong optimized.vm.xlarge 800 1600 2400 3200 4000 SE +/- 45.67, N = 3 3962.46 MIN: 30.9 / MAX: 9652.94 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
Intel MPI Benchmarks Test: IMB-MPI1 Sendrecv OpenBenchmarking.org Average Mbytes/sec, More Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Sendrecv optimized.vm.xlarge 600 1200 1800 2400 3000 SE +/- 33.86, N = 3 2633.68 MIN: 2.78 / MAX: 9181.98 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
Intel MPI Benchmarks Test: IMB-MPI1 Sendrecv OpenBenchmarking.org Average usec, Fewer Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Sendrecv optimized.vm.xlarge 50 100 150 200 250 SE +/- 4.06, N = 3 221.30 MIN: 0.72 / MAX: 3343.92 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
Phoronix Test Suite v10.8.4