2 x Intel Xeon E5-2680 v2 testing with a Supermicro X9DRL-3F/iF v1.01 (3.0a BIOS) and Matrox MGA G200eW WPCM450 on Ubuntu 18.04 via the Phoronix Test Suite.
2 x Intel Xeon E5-2680 v2 Processor: 2 x Intel Xeon E5-2680 v2 @ 3.60GHz (20 Cores / 40 Threads), Motherboard: Supermicro X9DRL-3F/iF v1.01 (3.0a BIOS), Chipset: Intel Xeon E7 v2/Xeon, Memory: 4 x 16384 MB DDR3-1866MT/s Samsung M393B2G70DB0-, Disk: 120GB KINGSTON SA400S3, Graphics: Matrox MGA G200eW WPCM450, Monitor: SE2717H/HX, Network: 2 x Intel 82574L
OS: Ubuntu 18.04, Kernel: 4.15.0-47-generic (x86_64), Display Server: X Server 1.19.6, Display Driver: modesetting 1.19.6, Compiler: GCC 7.3.0, File-System: ext4, Screen Resolution: 1280x1024
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate performanceSecurity Notes: KPTI + __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp + PTE Inversion; VMX: conditional cache flushes SMT vulnerable
SciMark This test runs the ANSI C version of SciMark 2.0, which is a benchmark for scientific and numerical computing developed by programmers at the National Institute of Standards and Technology. This test is made up of Fast Foruier Transform, Jacobi Successive Over-relaxation, Monte Carlo, Sparse Matrix Multiply, and dense LU matrix factorization benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Mflops, More Is Better SciMark 2.0 Computational Test: Composite 2 x Intel Xeon E5-2680 v2 120 240 360 480 600 SE +/- 1.69, N = 3 554.68 1. (CC) gcc options: -lm
OpenBenchmarking.org Mflops, More Is Better SciMark 2.0 Computational Test: Monte Carlo 2 x Intel Xeon E5-2680 v2 30 60 90 120 150 SE +/- 0.16, N = 3 119.68 1. (CC) gcc options: -lm
OpenBenchmarking.org Mflops, More Is Better SciMark 2.0 Computational Test: Fast Fourier Transform 2 x Intel Xeon E5-2680 v2 60 120 180 240 300 SE +/- 1.13, N = 3 256.24 1. (CC) gcc options: -lm
OpenBenchmarking.org Mflops, More Is Better SciMark 2.0 Computational Test: Sparse Matrix Multiply 2 x Intel Xeon E5-2680 v2 120 240 360 480 600 SE +/- 8.82, N = 3 573.99 1. (CC) gcc options: -lm
OpenBenchmarking.org Mflops, More Is Better SciMark 2.0 Computational Test: Dense LU Matrix Factorization 2 x Intel Xeon E5-2680 v2 170 340 510 680 850 SE +/- 0.61, N = 3 769.04 1. (CC) gcc options: -lm
OpenBenchmarking.org Mflops, More Is Better SciMark 2.0 Computational Test: Jacobi Successive Over-Relaxation 2 x Intel Xeon E5-2680 v2 200 400 600 800 1000 SE +/- 0.10, N = 3 1054.43 1. (CC) gcc options: -lm
OpenBenchmarking.org Nanoseconds/Operation, Fewer Is Better Go Benchmarks Test: build 2 x Intel Xeon E5-2680 v2 4000M 8000M 12000M 16000M 20000M SE +/- 243000790.27, N = 3 20045335959
CLOMP CLOMP is the C version of the Livermore OpenMP benchmark developed to measure OpenMP overheads and other performance impacts due to threading in order to influence future system designs. This particular test profile configuration is currently set to look at the OpenMP static schedule speed-up across all available CPU cores using the recommended test configuration. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Speedup, More Is Better CLOMP 3.3 Static OMP Speedup 2 x Intel Xeon E5-2680 v2 0.3983 0.7966 1.1949 1.5932 1.9915 SE +/- 0.02, N = 6 1.77 1. (CC) gcc options: --openmp -O3 -lm
OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 2 - Type: Thread 2 x Intel Xeon E5-2680 v2 0.9608 1.9216 2.8824 3.8432 4.804 SE +/- 0.03, N = 3 4.27 1. (CC) gcc options: -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 4 - Type: Thread 2 x Intel Xeon E5-2680 v2 2 4 6 8 10 SE +/- 0.01, N = 3 7.38 1. (CC) gcc options: -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 8 - Type: Thread 2 x Intel Xeon E5-2680 v2 4 8 12 16 20 SE +/- 0.14, N = 3 13.62 1. (CC) gcc options: -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 1 - Type: Process 2 x Intel Xeon E5-2680 v2 0.846 1.692 2.538 3.384 4.23 SE +/- 0.01, N = 3 3.76 1. (CC) gcc options: -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 16 - Type: Thread 2 x Intel Xeon E5-2680 v2 6 12 18 24 30 SE +/- 0.48, N = 15 25.34 1. (CC) gcc options: -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 2 - Type: Process 2 x Intel Xeon E5-2680 v2 0.8933 1.7866 2.6799 3.5732 4.4665 SE +/- 0.01, N = 3 3.97 1. (CC) gcc options: -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 4 - Type: Process 2 x Intel Xeon E5-2680 v2 2 4 6 8 10 SE +/- 0.02, N = 3 7.02 1. (CC) gcc options: -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 8 - Type: Process 2 x Intel Xeon E5-2680 v2 3 6 9 12 15 SE +/- 0.10, N = 3 12.82 1. (CC) gcc options: -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 16 - Type: Process 2 x Intel Xeon E5-2680 v2 6 12 18 24 30 SE +/- 0.35, N = 3 24.11 1. (CC) gcc options: -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 32 - Type: Process 2 x Intel Xeon E5-2680 v2 11 22 33 44 55 SE +/- 0.75, N = 3 46.89 1. (CC) gcc options: -lpthread
Minion Minion is an open-source constraint solver that is designed to be very scalable. This test profile uses Minion's integrated benchmarking problems to solve. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Minion 1.8 Benchmark: Graceful 2 x Intel Xeon E5-2680 v2 15 30 45 60 75 SE +/- 0.81, N = 3 69.45 1. (CXX) g++ options: -std=gnu++11 -O3 -fomit-frame-pointer -rdynamic
OpenBenchmarking.org Seconds, Fewer Is Better Minion 1.8 Benchmark: Solitaire 2 x Intel Xeon E5-2680 v2 20 40 60 80 100 SE +/- 0.34, N = 3 94.27 1. (CXX) g++ options: -std=gnu++11 -O3 -fomit-frame-pointer -rdynamic
OpenBenchmarking.org Seconds, Fewer Is Better Minion 1.8 Benchmark: Quasigroup 2 x Intel Xeon E5-2680 v2 30 60 90 120 150 SE +/- 1.36, N = 3 152.47 1. (CXX) g++ options: -std=gnu++11 -O3 -fomit-frame-pointer -rdynamic
OpenBenchmarking.org MB/s, More Is Better lzbench 2017-08-08 Test: XZ 0 - Process: Decompression 2 x Intel Xeon E5-2680 v2 20 40 60 80 100 87 1. (CXX) g++ options: -lrt -static -lpthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.org MB/s, More Is Better lzbench 2017-08-08 Test: Zstd 1 - Process: Compression 2 x Intel Xeon E5-2680 v2 70 140 210 280 350 SE +/- 1.15, N = 3 304 1. (CXX) g++ options: -lrt -static -lpthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.org MB/s, More Is Better lzbench 2017-08-08 Test: Zstd 1 - Process: Decompression 2 x Intel Xeon E5-2680 v2 200 400 600 800 1000 SE +/- 3.67, N = 3 880 1. (CXX) g++ options: -lrt -static -lpthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.org MB/s, More Is Better lzbench 2017-08-08 Test: Brotli 0 - Process: Compression 2 x Intel Xeon E5-2680 v2 70 140 210 280 350 SE +/- 1.15, N = 3 340 1. (CXX) g++ options: -lrt -static -lpthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.org MB/s, More Is Better lzbench 2017-08-08 Test: Brotli 0 - Process: Decompression 2 x Intel Xeon E5-2680 v2 90 180 270 360 450 SE +/- 1.20, N = 3 418 1. (CXX) g++ options: -lrt -static -lpthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.org MB/s, More Is Better lzbench 2017-08-08 Test: Libdeflate 1 - Process: Compression 2 x Intel Xeon E5-2680 v2 40 80 120 160 200 SE +/- 0.33, N = 3 165 1. (CXX) g++ options: -lrt -static -lpthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.org MB/s, More Is Better lzbench 2017-08-08 Test: Libdeflate 1 - Process: Decompression 2 x Intel Xeon E5-2680 v2 200 400 600 800 1000 801 1. (CXX) g++ options: -lrt -static -lpthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.org MiB/s, More Is Better Botan 2.8.0 Test: KASUMI - Decrypt 2 x Intel Xeon E5-2680 v2 16 32 48 64 80 SE +/- 0.18, N = 3 71.60 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
OpenBenchmarking.org MiB/s, More Is Better Botan 2.8.0 Test: AES-256 - Encrypt 2 x Intel Xeon E5-2680 v2 500 1000 1500 2000 2500 SE +/- 10.00, N = 3 2218.78 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
OpenBenchmarking.org MiB/s, More Is Better Botan 2.8.0 Test: AES-256 - Decrypt 2 x Intel Xeon E5-2680 v2 500 1000 1500 2000 2500 SE +/- 30.37, N = 3 2178.77 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
OpenBenchmarking.org MiB/s, More Is Better Botan 2.8.0 Test: Twofish - Encrypt 2 x Intel Xeon E5-2680 v2 60 120 180 240 300 SE +/- 0.45, N = 3 267.14 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
OpenBenchmarking.org MiB/s, More Is Better Botan 2.8.0 Test: Twofish - Decrypt 2 x Intel Xeon E5-2680 v2 60 120 180 240 300 SE +/- 0.89, N = 3 267.00 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
OpenBenchmarking.org MiB/s, More Is Better Botan 2.8.0 Test: Blowfish - Encrypt 2 x Intel Xeon E5-2680 v2 50 100 150 200 250 SE +/- 2.20, N = 3 206.88 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
OpenBenchmarking.org MiB/s, More Is Better Botan 2.8.0 Test: Blowfish - Decrypt 2 x Intel Xeon E5-2680 v2 50 100 150 200 250 SE +/- 1.11, N = 3 207.97 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
OpenBenchmarking.org MiB/s, More Is Better Botan 2.8.0 Test: CAST-256 - Encrypt 2 x Intel Xeon E5-2680 v2 20 40 60 80 100 SE +/- 0.25, N = 3 106.29 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
OpenBenchmarking.org MiB/s, More Is Better Botan 2.8.0 Test: CAST-256 - Decrypt 2 x Intel Xeon E5-2680 v2 20 40 60 80 100 SE +/- 0.43, N = 3 106.22 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
Java SciMark This test runs the Java version of SciMark 2.0, which is a benchmark for scientific and numerical computing developed by programmers at the National Institute of Standards and Technology. This benchmark is made up of Fast Foruier Transform, Jacobi Successive Over-relaxation, Monte Carlo, Sparse Matrix Multiply, and dense LU matrix factorization benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Mflops, More Is Better Java SciMark 2.0 Computational Test: Composite 2 x Intel Xeon E5-2680 v2 400 800 1200 1600 2000 SE +/- 9.26, N = 4 1799.01
OpenBenchmarking.org MB/s, More Is Better CacheBench Test: Write 2 x Intel Xeon E5-2680 v2 5K 10K 15K 20K 25K SE +/- 17.16, N = 3 22392.52 1. (CC) gcc options: -lrt
OpenBenchmarking.org MB/s, More Is Better CacheBench Test: Read / Modify / Write 2 x Intel Xeon E5-2680 v2 6K 12K 18K 24K 30K SE +/- 9.10, N = 3 25860.26 1. (CC) gcc options: -lrt
glibc bench The GNU C Library project provides the core libraries for the GNU system and GNU/Linux systems, as well as many other systems that use Linux as the kernel. These libraries provide critical APIs including ISO C11, POSIX.1-2008, BSD, OS-specific APIs and more.
Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org nanoseconds, Fewer Is Better glibc bench 1.0 Benchmark: cos 2 x Intel Xeon E5-2680 v2 10K 20K 30K 40K 50K SE +/- 353.41, N = 3 47824.23
OpenBenchmarking.org Mflops, More Is Better LuaJIT 2.1-git Test: Monte Carlo 2 x Intel Xeon E5-2680 v2 90 180 270 360 450 SE +/- 4.05, N = 3 416.13 1. (CC) gcc options: -lm -ldl -O2 -fomit-frame-pointer -U_FORTIFY_SOURCE -fno-stack-protector
OpenBenchmarking.org Mflops, More Is Better LuaJIT 2.1-git Test: Fast Fourier Transform 2 x Intel Xeon E5-2680 v2 90 180 270 360 450 SE +/- 1.42, N = 3 407.50 1. (CC) gcc options: -lm -ldl -O2 -fomit-frame-pointer -U_FORTIFY_SOURCE -fno-stack-protector
OpenBenchmarking.org Mflops, More Is Better LuaJIT 2.1-git Test: Sparse Matrix Multiply 2 x Intel Xeon E5-2680 v2 200 400 600 800 1000 SE +/- 0.98, N = 3 1092.50 1. (CC) gcc options: -lm -ldl -O2 -fomit-frame-pointer -U_FORTIFY_SOURCE -fno-stack-protector
OpenBenchmarking.org Mflops, More Is Better LuaJIT 2.1-git Test: Dense LU Matrix Factorization 2 x Intel Xeon E5-2680 v2 500 1000 1500 2000 2500 SE +/- 1.26, N = 3 2164.58 1. (CC) gcc options: -lm -ldl -O2 -fomit-frame-pointer -U_FORTIFY_SOURCE -fno-stack-protector
OpenBenchmarking.org Mflops, More Is Better LuaJIT 2.1-git Test: Jacobi Successive Over-Relaxation 2 x Intel Xeon E5-2680 v2 300 600 900 1200 1500 SE +/- 17.37, N = 3 1526.87 1. (CC) gcc options: -lm -ldl -O2 -fomit-frame-pointer -U_FORTIFY_SOURCE -fno-stack-protector
FFTW FFTW is a C subroutine library for computing the discrete Fourier transform (DFT) in one or more dimensions. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 1D FFT Size 32 2 x Intel Xeon E5-2680 v2 1500 3000 4500 6000 7500 SE +/- 13.41, N = 3 7184.67 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 1D FFT Size 64 2 x Intel Xeon E5-2680 v2 1500 3000 4500 6000 7500 SE +/- 2.82, N = 3 7170.00 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 2D FFT Size 32 2 x Intel Xeon E5-2680 v2 1600 3200 4800 6400 8000 SE +/- 23.17, N = 3 7449.37 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 2D FFT Size 64 2 x Intel Xeon E5-2680 v2 1300 2600 3900 5200 6500 SE +/- 92.19, N = 3 6160.63 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 1D FFT Size 128 2 x Intel Xeon E5-2680 v2 1300 2600 3900 5200 6500 SE +/- 57.47, N = 3 6288.23 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 1D FFT Size 256 2 x Intel Xeon E5-2680 v2 1300 2600 3900 5200 6500 SE +/- 17.28, N = 3 6124.83 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 1D FFT Size 512 2 x Intel Xeon E5-2680 v2 1300 2600 3900 5200 6500 SE +/- 33.17, N = 3 6026.80 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 2D FFT Size 128 2 x Intel Xeon E5-2680 v2 1200 2400 3600 4800 6000 SE +/- 63.51, N = 3 5532.77 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 2D FFT Size 256 2 x Intel Xeon E5-2680 v2 1100 2200 3300 4400 5500 SE +/- 16.42, N = 3 5259.63 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 2D FFT Size 512 2 x Intel Xeon E5-2680 v2 1100 2200 3300 4400 5500 SE +/- 5.11, N = 3 5066.07 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 1D FFT Size 1024 2 x Intel Xeon E5-2680 v2 1300 2600 3900 5200 6500 SE +/- 8.26, N = 3 6040.53 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 1D FFT Size 2048 2 x Intel Xeon E5-2680 v2 1200 2400 3600 4800 6000 SE +/- 56.80, N = 8 5600.11 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 1D FFT Size 4096 2 x Intel Xeon E5-2680 v2 1200 2400 3600 4800 6000 SE +/- 78.60, N = 4 5408.10 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 2D FFT Size 1024 2 x Intel Xeon E5-2680 v2 1100 2200 3300 4400 5500 SE +/- 16.91, N = 3 5168.13 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 2D FFT Size 2048 2 x Intel Xeon E5-2680 v2 900 1800 2700 3600 4500 SE +/- 16.46, N = 3 4303.77 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Stock - Size: 2D FFT Size 4096 2 x Intel Xeon E5-2680 v2 800 1600 2400 3200 4000 SE +/- 15.80, N = 3 3791.80 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 1D FFT Size 32 2 x Intel Xeon E5-2680 v2 2K 4K 6K 8K 10K SE +/- 167.74, N = 3 9714.20 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 1D FFT Size 64 2 x Intel Xeon E5-2680 v2 3K 6K 9K 12K 15K SE +/- 14.15, N = 3 14814 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 2D FFT Size 32 2 x Intel Xeon E5-2680 v2 5K 10K 15K 20K 25K SE +/- 219.89, N = 3 21324 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 2D FFT Size 64 2 x Intel Xeon E5-2680 v2 4K 8K 12K 16K 20K SE +/- 188.17, N = 3 20694 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 1D FFT Size 128 2 x Intel Xeon E5-2680 v2 4K 8K 12K 16K 20K SE +/- 120.39, N = 3 18326 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 1D FFT Size 256 2 x Intel Xeon E5-2680 v2 4K 8K 12K 16K 20K SE +/- 248.31, N = 3 19143 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 1D FFT Size 512 2 x Intel Xeon E5-2680 v2 4K 8K 12K 16K 20K SE +/- 182.68, N = 3 20881 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 2D FFT Size 128 2 x Intel Xeon E5-2680 v2 4K 8K 12K 16K 20K SE +/- 237.75, N = 3 16588 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 2D FFT Size 256 2 x Intel Xeon E5-2680 v2 3K 6K 9K 12K 15K SE +/- 177.79, N = 3 15025 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 2D FFT Size 512 2 x Intel Xeon E5-2680 v2 3K 6K 9K 12K 15K SE +/- 77.80, N = 3 15965 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 1D FFT Size 1024 2 x Intel Xeon E5-2680 v2 5K 10K 15K 20K 25K SE +/- 9.06, N = 3 21251 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 1D FFT Size 2048 2 x Intel Xeon E5-2680 v2 4K 8K 12K 16K 20K SE +/- 182.60, N = 3 20641 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 1D FFT Size 4096 2 x Intel Xeon E5-2680 v2 4K 8K 12K 16K 20K SE +/- 90.12, N = 3 19110 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 2D FFT Size 1024 2 x Intel Xeon E5-2680 v2 3K 6K 9K 12K 15K SE +/- 44.78, N = 3 13972 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 2D FFT Size 2048 2 x Intel Xeon E5-2680 v2 3K 6K 9K 12K 15K SE +/- 145.22, N = 3 12736 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
OpenBenchmarking.org Mflops, More Is Better FFTW 3.3.6 Build: Float + SSE - Size: 2D FFT Size 4096 2 x Intel Xeon E5-2680 v2 2K 4K 6K 8K 10K SE +/- 43.38, N = 3 10283 1. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm
CloverLeaf CloverLeaf is a Lagrangian-Eulerian hydrodynamics benchmark. This test profile currently makes use of CloverLeaf's OpenMP version and benchmarked with the clover_bm8192.in input file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better CloverLeaf Lagrangian-Eulerian Hydrodynamics 2 x Intel Xeon E5-2680 v2 0.5558 1.1116 1.6674 2.2232 2.779 SE +/- 0.01, N = 3 2.47 1. (F9X) gfortran options: -O3 -march=native -funroll-loops -fopenmp
Parboil The Parboil Benchmarks from the IMPACT Research Group at University of Illinois are a set of throughput computing applications for looking at computing architecture and compilers. Parboil test-cases support OpenMP, OpenCL, and CUDA multi-processing environments. However, at this time the test profile is just making use of the OpenMP and OpenCL test workloads. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Parboil 2.5 Test: OpenMP LBM 2 x Intel Xeon E5-2680 v2 20 40 60 80 100 SE +/- 0.75, N = 3 87.12 1. (CXX) g++ options: -lm -lpthread -lgomp -O3 -ffast-math -fopenmp
Parboil The Parboil Benchmarks from the IMPACT Research Group at University of Illinois are a set of throughput computing applications for looking at computing architecture and compilers. Parboil test-cases support OpenMP, OpenCL, and CUDA multi-processing environments. However, at this time the test profile is just making use of the OpenMP and OpenCL test workloads. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Parboil 2.5 Test: OpenMP CUTCP 2 x Intel Xeon E5-2680 v2 0.6233 1.2466 1.8699 2.4932 3.1165 SE +/- 0.02, N = 3 2.77 1. (CXX) g++ options: -lm -lpthread -lgomp -O3 -ffast-math -fopenmp
OpenBenchmarking.org Seconds, Fewer Is Better Parboil 2.5 Test: OpenMP Stencil 2 x Intel Xeon E5-2680 v2 3 6 9 12 15 SE +/- 0.16, N = 15 11.90 1. (CXX) g++ options: -lm -lpthread -lgomp -O3 -ffast-math -fopenmp
OpenBenchmarking.org Seconds, Fewer Is Better Parboil 2.5 Test: OpenMP MRI Gridding 2 x Intel Xeon E5-2680 v2 40 80 120 160 200 SE +/- 1.14, N = 3 173.03 1. (CXX) g++ options: -lm -lpthread -lgomp -O3 -ffast-math -fopenmp
Rodinia Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes the OpenCL and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 2.4 Test: OpenMP LavaMD 2 x Intel Xeon E5-2680 v2 12 24 36 48 60 SE +/- 0.08, N = 3 52.78 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 2.4 Test: OpenMP CFD Solver 2 x Intel Xeon E5-2680 v2 4 8 12 16 20 SE +/- 0.07, N = 3 16.99 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 2.4 Test: OpenMP Streamcluster 2 x Intel Xeon E5-2680 v2 5 10 15 20 25 SE +/- 0.24, N = 3 19.87 1. (CXX) g++ options: -O2 -lOpenCL
NAMD NAMD is a parallel molecular dynamics code designed for high-performance simulation of large biomolecular systems. NAMD was developed by the Theoretical and Computational Biophysics Group in the Beckman Institute for Advanced Science and Technology at the University of Illinois at Urbana-Champaign. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org days/ns, Fewer Is Better NAMD 2.13b1 ATPase Simulation - 327,506 Atoms 2 x Intel Xeon E5-2680 v2 0.2474 0.4948 0.7422 0.9896 1.237 SE +/- 0.00204, N = 3 1.09953
FFTE FFTE is a package by Daisuke Takahashi to compute Discrete Fourier Transforms of 1-, 2- and 3- dimensional sequences of length (2^p)*(3^q)*(5^r). Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MFLOPS, More Is Better FFTE 6.0 Test: N=256, 1D Complex FFT Routine 2 x Intel Xeon E5-2680 v2 1200 2400 3600 4800 6000 SE +/- 1.94, N = 3 5530.72 1. (F9X) gfortran options: -O3 -fomit-frame-pointer -fopenmp -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi
Stockfish This is a test of Stockfish, an advanced C++11 chess benchmark that can scale up to 128 CPU cores. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Nodes Per Second, More Is Better Stockfish 9 Total Time 2 x Intel Xeon E5-2680 v2 9M 18M 27M 36M 45M SE +/- 267329.80, N = 3 39846050 1. (CXX) g++ options: -m64 -lpthread -fno-exceptions -std=c++11 -pedantic -O3 -msse -msse3 -mpopcnt -flto
MKL-DNN This is a test of the Intel MKL-DNN as the Intel Math Kernel Library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Deconvolution Batch deconv_all - Data Type: u8s8u8s32 2 x Intel Xeon E5-2680 v2 5K 10K 15K 20K 25K SE +/- 17.08, N = 3 22813.77 MIN: 21683.7 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Convolution Batch conv_googlenet_v3 - Data Type: u8s8u8s32 2 x Intel Xeon E5-2680 v2 300 600 900 1200 1500 SE +/- 4.37, N = 3 1452.16 MIN: 1257.69 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Deconvolution Batch deconv_3d - Data Type: u8s8u8s32 2 x Intel Xeon E5-2680 v2 1300 2600 3900 5200 6500 SE +/- 37.37, N = 3 6266.18 MIN: 6226.98 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32s32 2 x Intel Xeon E5-2680 v2 800 1600 2400 3200 4000 SE +/- 6.79, N = 3 3686.86 MIN: 3676.18 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Convolution Batch conv_all - Data Type: u8s8u8s32 2 x Intel Xeon E5-2680 v2 6K 12K 18K 24K 30K SE +/- 56.71, N = 3 27549.83 MIN: 26342.8 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Convolution Batch conv_googlenet_v3 - Data Type: f32 2 x Intel Xeon E5-2680 v2 40 80 120 160 200 SE +/- 0.59, N = 3 203.47 MIN: 201.49 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: IP Batch 1D - Data Type: u8s8u8s32 2 x Intel Xeon E5-2680 v2 13 26 39 52 65 SE +/- 0.09, N = 3 59.85 MIN: 48.25 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Convolution Batch conv_3d - Data Type: u8s8u8s32 2 x Intel Xeon E5-2680 v2 1400 2800 4200 5600 7000 SE +/- 17.47, N = 3 6562.83 MIN: 6534.65 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: IP Batch 1D - Data Type: f32 2 x Intel Xeon E5-2680 v2 3 6 9 12 15 SE +/- 0.15, N = 6 12.83 MIN: 8.19 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: IP Batch All - Data Type: f32 2 x Intel Xeon E5-2680 v2 30 60 90 120 150 SE +/- 1.19, N = 3 155.94 MIN: 105.98 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
Rust Mandelbrot This test profile is of the combined time for the serial and parallel Mandelbrot sets written in Rustlang via willi-kappler/mandel-rust. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Rust Mandelbrot Time To Complete Serial/Parallel Mandelbrot 2 x Intel Xeon E5-2680 v2 12 24 36 48 60 SE +/- 0.25, N = 3 53.79 1. (CC) gcc options: -m64 -pie -nodefaultlibs -lutil -ldl -lrt -lpthread -lgcc_s -lc -lm
MKL-DNN This is a test of the Intel MKL-DNN as the Intel Math Kernel Library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: IP Batch 1D - Data Type: u8s8f32s32 2 x Intel Xeon E5-2680 v2 13 26 39 52 65 SE +/- 0.24, N = 3 59.09 MIN: 46.91 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Deconvolution Batch deconv_all - Data Type: f32 2 x Intel Xeon E5-2680 v2 2K 4K 6K 8K 10K SE +/- 16.66, N = 3 8542.30 MIN: 7760.38 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: IP Batch All - Data Type: u8s8u8s32 2 x Intel Xeon E5-2680 v2 140 280 420 560 700 SE +/- 3.38, N = 3 661.20 MIN: 542.19 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Convolution Batch conv_3d - Data Type: u8s8f32s32 2 x Intel Xeon E5-2680 v2 1500 3000 4500 6000 7500 SE +/- 5.27, N = 3 6873.92 MIN: 6857.38 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: IP Batch All - Data Type: u8s8f32s32 2 x Intel Xeon E5-2680 v2 140 280 420 560 700 SE +/- 3.12, N = 3 661.73 MIN: 538.18 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Convolution Batch conv_all - Data Type: u8s8f32s32 2 x Intel Xeon E5-2680 v2 6K 12K 18K 24K 30K SE +/- 59.46, N = 3 27521.03 MIN: 26432.5 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Convolution Batch conv_3d - Data Type: f32 2 x Intel Xeon E5-2680 v2 6 12 18 24 30 SE +/- 0.02, N = 3 23.38 MIN: 23.12 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Deconvolution Batch deconv_1d - Data Type: u8s8u8s32 2 x Intel Xeon E5-2680 v2 800 1600 2400 3200 4000 SE +/- 3.85, N = 3 3709.17 MIN: 3701.38 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Convolution Batch conv_all - Data Type: f32 2 x Intel Xeon E5-2680 v2 800 1600 2400 3200 4000 SE +/- 8.72, N = 3 3761.97 MIN: 3738.45 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Convolution Batch conv_alexnet - Data Type: u8s8u8s32 2 x Intel Xeon E5-2680 v2 400 800 1200 1600 2000 SE +/- 9.59, N = 3 1709.92 MIN: 1678.03 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Deconvolution Batch deconv_1d - Data Type: f32 2 x Intel Xeon E5-2680 v2 2 4 6 8 10 SE +/- 0.01, N = 3 8.80 MIN: 8.68 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32s32 2 x Intel Xeon E5-2680 v2 1300 2600 3900 5200 6500 SE +/- 2.20, N = 3 6206.88 MIN: 6201.71 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Deconvolution Batch deconv_3d - Data Type: f32 2 x Intel Xeon E5-2680 v2 5 10 15 20 25 SE +/- 0.06, N = 3 19.83 MIN: 16.56 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Convolution Batch conv_alexnet - Data Type: u8s8f32s32 2 x Intel Xeon E5-2680 v2 400 800 1200 1600 2000 SE +/- 0.59, N = 3 1665.73 MIN: 1646.75 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Convolution Batch conv_alexnet - Data Type: f32 2 x Intel Xeon E5-2680 v2 120 240 360 480 600 SE +/- 1.42, N = 3 538.08 MIN: 533.96 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN 2019-04-16 Harness: Convolution Batch conv_googlenet_v3 - Data Type: u8s8f32s32 2 x Intel Xeon E5-2680 v2 300 600 900 1200 1500 SE +/- 8.20, N = 3 1449.43 MIN: 1273.24 1. (CXX) g++ options: -std=c++11 -march=native -mtune=native -fPIC -fopenmp -O3 -pie -lmklml_intel -ldl
OpenBenchmarking.org kH/s - Hash Speed, More Is Better Cpuminer-Opt 3.8.8.1 Algorithm: deep 2 x Intel Xeon E5-2680 v2 2K 4K 6K 8K 10K SE +/- 3.07, N = 3 9814.23 1. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.org kH/s - Hash Speed, More Is Better Cpuminer-Opt 3.8.8.1 Algorithm: lbry 2 x Intel Xeon E5-2680 v2 2K 4K 6K 8K 10K SE +/- 63.33, N = 3 10233 1. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.org kH/s - Hash Speed, More Is Better Cpuminer-Opt 3.8.8.1 Algorithm: skein 2 x Intel Xeon E5-2680 v2 6K 12K 18K 24K 30K SE +/- 3523.98, N = 15 28829 1. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.org kH/s - Hash Speed, More Is Better Cpuminer-Opt 3.8.8.1 Algorithm: myr-gr 2 x Intel Xeon E5-2680 v2 2K 4K 6K 8K 10K SE +/- 15.28, N = 3 10580 1. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.org kH/s - Hash Speed, More Is Better Cpuminer-Opt 3.8.8.1 Algorithm: sha256t 2 x Intel Xeon E5-2680 v2 13K 26K 39K 52K 65K SE +/- 529.52, N = 11 59694 1. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.30 Operation: Rotate 2 x Intel Xeon E5-2680 v2 40 80 120 160 200 SE +/- 0.67, N = 3 193 1. (CC) gcc options: -fopenmp -O2 -pthread -lXext -lSM -lICE -lX11 -lz -lm -lgomp -lpthread
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.30 Operation: Sharpen 2 x Intel Xeon E5-2680 v2 30 60 90 120 150 151 1. (CC) gcc options: -fopenmp -O2 -pthread -lXext -lSM -lICE -lX11 -lz -lm -lgomp -lpthread
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.30 Operation: Enhanced 2 x Intel Xeon E5-2680 v2 30 60 90 120 150 155 1. (CC) gcc options: -fopenmp -O2 -pthread -lXext -lSM -lICE -lX11 -lz -lm -lgomp -lpthread
Mencoder This test uses mplayer's mencoder utility and the libavcodec family for testing the system's audio/video encoding performance. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Mencoder 1.3.0 AVI To LAVC 2 x Intel Xeon E5-2680 v2 5 10 15 20 25 SE +/- 0.16, N = 3 22.36 1. (CC) gcc options: -ffast-math -fpie -pie -lrt -lpng -lz -lfreetype -lfontconfig -lpthread -ldl -rdynamic -lm
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.30 Operation: Noise-Gaussian 2 x Intel Xeon E5-2680 v2 30 60 90 120 150 144 1. (CC) gcc options: -fopenmp -O2 -pthread -lXext -lSM -lICE -lX11 -lz -lm -lgomp -lpthread
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.30 Operation: HWB Color Space 2 x Intel Xeon E5-2680 v2 50 100 150 200 250 214 1. (CC) gcc options: -fopenmp -O2 -pthread -lXext -lSM -lICE -lX11 -lz -lm -lgomp -lpthread
x264 This is a simple test of the x264 encoder run on the CPU (OpenCL support disabled) with a sample video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better x264 2018-09-25 H.264 Video Encoding 2 x Intel Xeon E5-2680 v2 20 40 60 80 100 SE +/- 0.46, N = 3 100.20 1. (CC) gcc options: -ldl -lavformat -lavcodec -lavutil -lswscale -m64 -lm -lpthread -O3 -ffast-math -std=gnu99 -fPIC -fomit-frame-pointer -fno-tree-vectorize
SVT-AV1 This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-AV1 CPU-based multi-threaded video encoder for the AV1 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 2019-03-07 1080p 8-bit YUV To AV1 Video Encode 2 x Intel Xeon E5-2680 v2 0.9563 1.9126 2.8689 3.8252 4.7815 SE +/- 0.01, N = 3 4.25 1. (CXX) g++ options: -O3 -pie -lpthread -lm
C-Ray This is a test of C-Ray, a simple raytracer designed to test the floating-point CPU performance. This test is multi-threaded (16 threads per core), will shoot 8 rays per pixel for anti-aliasing, and will generate a 1600 x 1200 image. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better C-Ray 1.1 Total Time - 4K, 16 Rays Per Pixel 2 x Intel Xeon E5-2680 v2 15 30 45 60 75 SE +/- 0.21, N = 3 65.81 1. (CC) gcc options: -lm -lpthread -O3
TTSIOD 3D Renderer A portable GPL 3D software renderer that supports OpenMP and Intel Threading Building Blocks with many different rendering modes. This version does not use OpenGL but is entirely CPU/software based. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org FPS, More Is Better TTSIOD 3D Renderer 2.3b Phong Rendering With Soft-Shadow Mapping 2 x Intel Xeon E5-2680 v2 110 220 330 440 550 SE +/- 8.43, N = 3 487.69 1. (CXX) g++ options: -O3 -fomit-frame-pointer -ffast-math -mtune=native -flto -msse -mrecip -mfpmath=sse -msse2 -mssse3 -lSDL -fopenmp -fwhole-program -lstdc++
SVT-HEVC This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 2019-02-03 1080p 8-bit YUV To HEVC Video Encode 2 x Intel Xeon E5-2680 v2 5 10 15 20 25 SE +/- 0.16, N = 15 20.62 1. (CC) gcc options: -fPIE -fPIC -O2 -flto -fvisibility=hidden -march=native -pie -rdynamic -lpthread -lrt
POV-Ray This is a test of POV-Ray, the Persistence of Vision Raytracer. POV-Ray is used to create 3D graphics using ray-tracing. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better POV-Ray 3.7.0.7 Trace Time 2 x Intel Xeon E5-2680 v2 9 18 27 36 45 SE +/- 0.01, N = 3 37.17 1. (CXX) g++ options: -pipe -O3 -ffast-math -march=native -pthread -lSM -lICE -lX11 -lIlmImf -lImath -lHalf -lIex -lIexMath -lIlmThread -lpthread -ltiff -ljpeg -lpng -lz -lrt -lm -lboost_thread -lboost_system
FFmpeg This test uses FFmpeg for testing the system's audio/video encoding performance. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better FFmpeg 4.0.2 H.264 HD To NTSC DV 2 x Intel Xeon E5-2680 v2 2 4 6 8 10 SE +/- 0.09, N = 7 8.19 1. (CC) gcc options: -lavdevice -lavfilter -lavformat -lavcodec -lswresample -lswscale -lavutil -lm -lxcb -lxcb-shm -lxcb-shape -lxcb-xfixes -pthread -std=c11 -fomit-frame-pointer -fPIC -O3 -fno-math-errno -fno-signed-zeros -fno-tree-vectorize -MMD -MF -MT
Smallpt Smallpt is a C++ global illumination renderer written in less than 100 lines of code. Global illumination is done via unbiased Monte Carlo path tracing and there is multi-threading support via the OpenMP library. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Smallpt 1.0 Global Illumination Renderer; 128 Samples 2 x Intel Xeon E5-2680 v2 2 4 6 8 10 SE +/- 0.10, N = 6 8.71 1. (CXX) g++ options: -fopenmp -O3
OpenSSL OpenSSL is an open-source toolkit that implements SSL (Secure Sockets Layer) and TLS (Transport Layer Security) protocols. This test measures the RSA 4096-bit performance of OpenSSL. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Signs Per Second, More Is Better OpenSSL 1.1.1 RSA 4096-bit Performance 2 x Intel Xeon E5-2680 v2 600 1200 1800 2400 3000 SE +/- 0.51, N = 3 2634.30 1. (CC) gcc options: -pthread -m64 -O3 -lssl -lcrypto -ldl
Fhourstones This integer benchmark solves positions in the game of Connect-4, as played on a vertical 7x6 board. By default, it uses a 64Mb transposition table with the twobig replacement strategy. Positions are represented as 64-bit bitboards, and the hash function is computed using a single 64-bit modulo operation, giving 64-bit machines a slight edge. The alpha-beta searcher sorts moves dynamically based on the history heuristic. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Kpos / sec, More Is Better Fhourstones 3.1 Complex Connect-4 Solving 2 x Intel Xeon E5-2680 v2 3K 6K 9K 12K 15K SE +/- 11.03, N = 3 11887.03 1. (CC) gcc options: -O3
OpenBenchmarking.org Seconds, Fewer Is Better PolyBench-C 4.2 Test: Correlation Computation 2 x Intel Xeon E5-2680 v2 3 6 9 12 15 SE +/- 0.04, N = 3 12.10 1. (CC) gcc options: -O3 -march=native
OpenBenchmarking.org Seconds, Fewer Is Better PolyBench-C 4.2 Test: 3 Matrix Multiplications 2 x Intel Xeon E5-2680 v2 5 10 15 20 25 SE +/- 0.10, N = 3 18.42 1. (CC) gcc options: -O3 -march=native
2 x Intel Xeon E5-2680 v2 Processor: 2 x Intel Xeon E5-2680 v2 @ 3.60GHz (20 Cores / 40 Threads), Motherboard: Supermicro X9DRL-3F/iF v1.01 (3.0a BIOS), Chipset: Intel Xeon E7 v2/Xeon, Memory: 4 x 16384 MB DDR3-1866MT/s Samsung M393B2G70DB0-, Disk: 120GB KINGSTON SA400S3, Graphics: Matrox MGA G200eW WPCM450, Monitor: SE2717H/HX, Network: 2 x Intel 82574L
OS: Ubuntu 18.04, Kernel: 4.15.0-47-generic (x86_64), Display Server: X Server 1.19.6, Display Driver: modesetting 1.19.6, Compiler: GCC 7.3.0, File-System: ext4, Screen Resolution: 1280x1024
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate performanceSecurity Notes: KPTI + __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp + PTE Inversion; VMX: conditional cache flushes SMT vulnerable
Testing initiated at 19 April 2019 14:52 by user root.