Intel Core i9-11900K testing with a ASUS ROG MAXIMUS XIII HERO (1007 BIOS) and ASUS Intel RKL GT1 3GB on Ubuntu 21.10 via the Phoronix Test Suite.
A Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x40 - Thermald 2.4.6Java Notes: OpenJDK Runtime Environment (build 11.0.15+10-Ubuntu-0ubuntu0.21.10.1)Python Notes: Python 3.9.7Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
B C D Processor: Intel Core i9-11900K @ 5.10GHz (8 Cores / 16 Threads), Motherboard: ASUS ROG MAXIMUS XIII HERO (1007 BIOS), Chipset: Intel Tiger Lake-H, Memory: 32GB, Disk: 2000GB Corsair Force MP600, Graphics: ASUS Intel RKL GT1 3GB (1300MHz), Audio: Intel Tiger Lake-H HD Audio, Monitor: MX279, Network: 2 x Intel I225-V + Intel Wi-Fi 6 AX210/AX211/AX411
OS: Ubuntu 21.10, Kernel: 5.15.0-051500rc7daily20211029-generic (x86_64) 20211028, Desktop: GNOME Shell 40.5, Display Server: X Server 1.20.13 + Wayland, OpenGL: 4.6 Mesa 22.0.0-devel (git-f13d486 2021-11-03 impish-oibaf-ppa), Vulkan: 1.2.195, Compiler: GCC 11.2.0, File-System: ext4, Screen Resolution: 1920x1080
AOM AV1 OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 4K A B C D 0.036 0.072 0.108 0.144 0.18 0.16 0.16 0.16 0.16 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 4K C B D A 1.1633 2.3266 3.4899 4.6532 5.8165 5.14 5.15 5.15 5.17 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 6 Realtime - Input: Bosphorus 4K C D B A 4 8 12 16 20 16.03 16.06 16.13 16.21 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 4K B C D A 3 6 9 12 15 9.79 9.79 9.81 9.90 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 8 Realtime - Input: Bosphorus 4K D C A B 11 22 33 44 55 48.04 48.22 48.31 48.44 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 9 Realtime - Input: Bosphorus 4K A D C B 15 30 45 60 75 64.44 67.18 68.35 68.64 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 10 Realtime - Input: Bosphorus 4K C B D A 16 32 48 64 80 62.05 71.70 71.97 72.48 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 1080p A B C D 0.0945 0.189 0.2835 0.378 0.4725 0.42 0.42 0.42 0.42 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 1080p C B D A 3 6 9 12 15 11.25 11.27 11.27 11.32 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 6 Realtime - Input: Bosphorus 1080p C A B D 3 6 9 12 15 11.90 11.93 11.94 12.31 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 1080p D B C A 7 14 21 28 35 30.29 30.30 30.33 30.47 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 8 Realtime - Input: Bosphorus 1080p B C D A 30 60 90 120 150 143.21 143.74 143.85 145.53 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 9 Realtime - Input: Bosphorus 1080p A C D B 40 80 120 160 200 176.81 186.32 187.30 188.18 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 10 Realtime - Input: Bosphorus 1080p D C A B 40 80 120 160 200 197.60 201.74 203.50 203.95 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
Blender Blender is an open-source 3D creation and modeling software project. This test is of Blender's Cycles benchmark with various sample files. GPU computing via NVIDIA OptiX and NVIDIA CUDA is currently supported and HIP for AMD Radeon GPUs. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Blender 3.2 Blend File: BMW27 - Compute: CPU-Only A B C 30 60 90 120 150 131.60 131.44 131.29
OpenBenchmarking.org FPS, More Is Better dav1d 1.0 Video Input: Summer Nature 4K D C B A 40 80 120 160 200 189.52 189.63 194.82 198.82 1. (CC) gcc options: -pthread -lm
OpenBenchmarking.org FPS, More Is Better dav1d 1.0 Video Input: Summer Nature 1080p A C D B 200 400 600 800 1000 893.33 903.95 904.36 904.56 1. (CC) gcc options: -pthread -lm
OpenBenchmarking.org FPS, More Is Better dav1d 1.0 Video Input: Chimera 1080p 10-bit C D B A 110 220 330 440 550 471.44 476.40 490.12 492.46 1. (CC) gcc options: -pthread -lm
Etcpak Etcpack is the self-proclaimed "fastest ETC compressor on the planet" with focused on providing open-source, very fast ETC and S3 texture compression support. The test profile uses a 8K x 8K game texture as a sample input. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Mpx/s, More Is Better Etcpak 1.0 Benchmark: Multi-Threaded - Configuration: ETC2 C B A D 600 1200 1800 2400 3000 2656.20 2656.51 2658.51 2658.62 1. (CXX) g++ options: -O3 -march=native -std=c++11 -lpthread
OpenBenchmarking.org Mpx/s, More Is Better Etcpak 1.0 Benchmark: Single-Threaded - Configuration: ETC2 D C B A 70 140 210 280 350 306.41 306.91 307.01 307.60 1. (CXX) g++ options: -O3 -march=native -std=c++11 -lpthread
OpenBenchmarking.org Op/s, More Is Better Facebook RocksDB 7.0.1 Test: Update Random C B A 150K 300K 450K 600K 750K 700065 703316 703459 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
OpenBenchmarking.org Op/s, More Is Better Facebook RocksDB 7.0.1 Test: Read While Writing A B C 500K 1000K 1500K 2000K 2500K 2152704 2199309 2223231 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
OpenBenchmarking.org Op/s, More Is Better Facebook RocksDB 7.0.1 Test: Read Random Write Random A C B 400K 800K 1200K 1600K 2000K 2027221 2041553 2041642 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
Glibc Benchmarks The GNU C Library project provides the core libraries for the GNU system and GNU/Linux systems, as well as many other systems that use Linux as the kernel. These libraries provide critical APIs including ISO C11, POSIX.1-2008, BSD, OS-specific APIs and more. This test profile makes use of Glibc's "benchtests" integrated benchmark suite. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ns, Fewer Is Better Glibc Benchmarks Benchmark: cos A B C D 12 24 36 48 60 54.45 53.84 53.81 53.80 1. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s
OpenBenchmarking.org ns, Fewer Is Better Glibc Benchmarks Benchmark: ffs B D C A 0.7591 1.5182 2.2773 3.0364 3.7955 3.37358 3.33422 3.33392 3.33365 1. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s
OpenBenchmarking.org ns, Fewer Is Better Glibc Benchmarks Benchmark: modf A D C B 1.0492 2.0984 3.1476 4.1968 5.246 4.66291 4.65987 4.65758 4.65510 1. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s
OpenBenchmarking.org ns, Fewer Is Better Glibc Benchmarks Benchmark: sqrt D A C B 0.8987 1.7974 2.6961 3.5948 4.4935 3.99406 3.99377 3.99370 3.99317 1. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s
OpenBenchmarking.org ns, Fewer Is Better Glibc Benchmarks Benchmark: ffsll C B A D 0.759 1.518 2.277 3.036 3.795 3.37318 3.33819 3.33510 3.33486 1. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s
OpenBenchmarking.org ns, Fewer Is Better Glibc Benchmarks Benchmark: pthread_once C B D A 0.7621 1.5242 2.2863 3.0484 3.8105 3.38731 3.32250 3.32184 3.32140 1. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s
GROMACS The GROMACS (GROningen MAchine for Chemical Simulations) molecular dynamics package testing with the water_GMX50 data. This test profile allows selecting between CPU and GPU-based GROMACS builds. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Ns Per Day, More Is Better GROMACS 2022.1 Implementation: MPI CPU - Input: water_GMX50_bare A B C 0.2259 0.4518 0.6777 0.9036 1.1295 0.995 1.001 1.004 1. (CXX) g++ options: -O3
InfluxDB This is a benchmark of the InfluxDB open-source time-series database optimized for fast, high-availability storage for IoT and other use-cases. The InfluxDB test profile makes use of InfluxDB Inch for facilitating the benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org val/sec, More Is Better InfluxDB 1.8.2 Concurrent Streams: 4 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000 B A C 400K 800K 1200K 1600K 2000K 1804674.6 1806268.7 1815340.8
OpenBenchmarking.org val/sec, More Is Better InfluxDB 1.8.2 Concurrent Streams: 64 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000 B C A 400K 800K 1200K 1600K 2000K 1805397.9 1814759.0 1818333.8
Nettle GNU Nettle is a low-level cryptographic library used by GnuTLS and other software. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Mbyte/s, More Is Better Nettle 3.8 Test: aes256 D C B A 3K 6K 9K 12K 15K 12155.79 12195.87 12205.23 12207.52 MIN: 7981.08 / MAX: 20658.54 MIN: 7973.01 / MAX: 20659.96 MIN: 7983.96 / MAX: 20655.99 MIN: 7986.83 / MAX: 20656.74 1. (CC) gcc options: -O2 -ggdb3 -lnettle -lm -lcrypto
OpenBenchmarking.org Mbyte/s, More Is Better Nettle 3.8 Test: chacha B C A D 300 600 900 1200 1500 1511.89 1517.83 1519.06 1520.61 MIN: 679.88 / MAX: 4633.29 MIN: 679.93 / MAX: 4633.69 MIN: 679.9 / MAX: 4635.71 MIN: 680.19 / MAX: 4644.14 1. (CC) gcc options: -O2 -ggdb3 -lnettle -lm -lcrypto
OpenBenchmarking.org Mbyte/s, More Is Better Nettle 3.8 Test: sha512 B A D C 160 320 480 640 800 721.29 722.49 726.78 747.25 1. (CC) gcc options: -O2 -ggdb3 -lnettle -lm -lcrypto
OpenBenchmarking.org Mbyte/s, More Is Better Nettle 3.8 Test: poly1305-aes B C D A 1000 2000 3000 4000 5000 4442.12 4637.23 4638.26 4641.58 1. (CC) gcc options: -O2 -ggdb3 -lnettle -lm -lcrypto
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU C A B D 0.9056 1.8112 2.7168 3.6224 4.528 4.02480 4.01177 4.00817 4.00116 MIN: 3.92 MIN: 3.9 MIN: 3.9 MIN: 3.91 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU C B D A 3 6 9 12 15 11.12 11.12 11.12 11.11 MIN: 11.07 MIN: 11.06 MIN: 11.05 MIN: 11.05 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU D A C B 0.1635 0.327 0.4905 0.654 0.8175 0.726471 0.720371 0.716538 0.715645 MIN: 0.67 MIN: 0.67 MIN: 0.66 MIN: 0.67 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU D A C B 0.698 1.396 2.094 2.792 3.49 3.10236 3.08878 3.06944 3.06780 MIN: 3.03 MIN: 3.02 MIN: 3.03 MIN: 3.02 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU A D B C 2 4 6 8 10 8.55826 8.55027 8.54762 8.54691 MIN: 8.44 MIN: 8.42 MIN: 8.42 MIN: 8.39 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU C A D B 1.2256 2.4512 3.6768 4.9024 6.128 5.44701 5.38692 5.22850 5.20558 MIN: 4.95 MIN: 4.8 MIN: 4.75 MIN: 4.65 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU A C B D 4 8 12 16 20 14.14 14.13 14.12 14.12 MIN: 14.06 MIN: 14.04 MIN: 14.03 MIN: 14.04 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU A B D C 2 4 6 8 10 8.06688 8.03869 7.72180 7.70230 MIN: 4.81 MIN: 4.84 MIN: 4.81 MIN: 4.77 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU A B C D 0.9761 1.9522 2.9283 3.9044 4.8805 4.33830 4.33567 4.33558 4.33067 MIN: 4.22 MIN: 4.25 MIN: 4.24 MIN: 4.23 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU C A B D 3 6 9 12 15 11.97 11.97 11.96 11.96 MIN: 11.9 MIN: 11.9 MIN: 11.89 MIN: 11.89 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU C D B A 0.186 0.372 0.558 0.744 0.93 0.826686 0.826144 0.826044 0.825884 MIN: 0.82 MIN: 0.82 MIN: 0.81 MIN: 0.81 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU C D A B 0.3405 0.681 1.0215 1.362 1.7025 1.51319 1.50262 1.47653 1.46563 MIN: 1.43 MIN: 1.4 MIN: 1.39 MIN: 1.39 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU B A D C 700 1400 2100 2800 3500 3189.95 3118.35 3106.54 3106.14 MIN: 3078.21 MIN: 3113.21 MIN: 3101.69 MIN: 3100.97 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU C B A D 400 800 1200 1600 2000 1832.56 1831.92 1831.84 1825.78 MIN: 1827.43 MIN: 1827.22 MIN: 1827.31 MIN: 1821.07 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU A C D B 900 1800 2700 3600 4500 4358.46 3113.19 3110.84 3107.86 MIN: 3112.17 MIN: 3107.55 MIN: 3105.51 MIN: 3100.51 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU A C D B 4 8 12 16 20 16.16 16.16 16.15 16.15 MIN: 16.07 MIN: 16.07 MIN: 16.07 MIN: 16.07 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU A C D B 5 10 15 20 25 18.79 18.78 18.78 18.76 MIN: 18.6 MIN: 18.58 MIN: 18.6 MIN: 18.56 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU B C A D 4 8 12 16 20 17.15 17.15 17.15 17.14 MIN: 16.79 MIN: 16.78 MIN: 16.73 MIN: 16.81 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU A C B D 500 1000 1500 2000 2500 2245.91 1839.49 1833.40 1831.95 MIN: 1825.27 MIN: 1826.96 MIN: 1828.11 MIN: 1828.17 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU D C A B 0.729 1.458 2.187 2.916 3.645 3.24005 3.23367 3.23261 3.23137 MIN: 3.19 MIN: 3.18 MIN: 3.18 MIN: 3.18 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU A C D B 700 1400 2100 2800 3500 3115.29 3114.22 3108.58 3104.69 MIN: 3100.97 MIN: 3107.51 MIN: 3103.14 MIN: 3099.04 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU A C B D 400 800 1200 1600 2000 1833.14 1831.34 1831.29 1830.71 MIN: 1828.37 MIN: 1827.21 MIN: 1826.61 MIN: 1826.24 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU D C A B 0.2927 0.5854 0.8781 1.1708 1.4635 1.30110 1.29680 1.29412 1.29350 MIN: 1.25 MIN: 1.24 MIN: 1.25 MIN: 1.25 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU D B A C 0.7757 1.5514 2.3271 3.1028 3.8785 3.44775 3.44465 3.43567 3.43528 MIN: 3.28 MIN: 3.29 MIN: 3.29 MIN: 3.29 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
ONNX Runtime ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: GPT-2 - Device: CPU - Executor: Parallel C A B 1200 2400 3600 4800 6000 5376 5384 5396 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: GPT-2 - Device: CPU - Executor: Standard C A B 1400 2800 4200 5600 7000 6640 6701 6702 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: yolov4 - Device: CPU - Executor: Parallel A C B 70 140 210 280 350 306 306 308 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: yolov4 - Device: CPU - Executor: Standard C A B 110 220 330 440 550 525 526 526 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: bertsquad-12 - Device: CPU - Executor: Parallel B C A 110 220 330 440 550 500 511 512 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: bertsquad-12 - Device: CPU - Executor: Standard C B A 200 400 600 800 1000 642 1010 1015 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel A C B 16 32 48 64 80 70 70 71 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: fcn-resnet101-11 - Device: CPU - Executor: Standard A B C 20 40 60 80 100 104 104 104 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel C A B 300 600 900 1200 1500 1255 1262 1269 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard B C A 400 800 1200 1600 2000 2071 2076 2078 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: super-resolution-10 - Device: CPU - Executor: Parallel B A C 1000 2000 3000 4000 5000 4525 4547 4586 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: super-resolution-10 - Device: CPU - Executor: Standard B C A 1000 2000 3000 4000 5000 4528 4548 4558 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: particle_volume/pathtracer/real_time D A B C 60 120 180 240 300 297.35 297.46 297.50 297.68
OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: gravity_spheres_volume/dim_512/ao/real_time C A B D 0.959 1.918 2.877 3.836 4.795 4.22833 4.24782 4.24824 4.26224
OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: gravity_spheres_volume/dim_512/scivis/real_time B C A D 0.9527 1.9054 2.8581 3.8108 4.7635 4.18832 4.19096 4.20308 4.23414
OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_time B D C A 1.1456 2.2912 3.4368 4.5824 5.728 5.06027 5.08714 5.08966 5.09145
OSPray Studio Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer A D C B 400 800 1200 1600 2000 1918 1912 1911 1907 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer B C D A 400 800 1200 1600 2000 1957 1956 1953 1938 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer B C A 500 1000 1500 2000 2500 2332 2331 2328 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer C B A 7K 14K 21K 28K 35K 30584 30572 30515 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer A B C 14K 28K 42K 56K 70K 64451 64393 64242 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer B C A 7K 14K 21K 28K 35K 31244 31173 30947 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer B A C 14K 28K 42K 56K 70K 65765 65641 65487 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer B C A 8K 16K 24K 32K 40K 37170 37059 36885 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer C B A 17K 34K 51K 68K 85K 77555 77543 77357 1. (CXX) g++ options: -O3 -lm -ldl
perf-bench This test profile is used for running Linux perf-bench, the benchmark support within the Linux kernel's perf tool. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ops/sec, More Is Better perf-bench Benchmark: Epoll Wait A C B D 30K 60K 90K 120K 150K 127515 128524 130239 131078 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
OpenBenchmarking.org ops/sec, More Is Better perf-bench Benchmark: Futex Hash D B C A 1.5M 3M 4.5M 6M 7.5M 6460586 6685401 6810542 6819705 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
OpenBenchmarking.org GB/sec, More Is Better perf-bench Benchmark: Memcpy 1MB D B A C 7 14 21 28 35 31.36 31.36 31.36 31.49 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
OpenBenchmarking.org GB/sec, More Is Better perf-bench Benchmark: Memset 1MB D A C B 13 26 39 52 65 54.12 54.19 55.72 57.45 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
OpenBenchmarking.org ops/sec, More Is Better perf-bench Benchmark: Sched Pipe B C A D 70K 140K 210K 280K 350K 315267 316210 318866 322747 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
OpenBenchmarking.org ops/sec, More Is Better perf-bench Benchmark: Futex Lock-Pi B A C D 200 400 600 800 1000 1076 1082 1087 1089 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
OpenBenchmarking.org ops/sec, More Is Better perf-bench Benchmark: Syscall Basic C B A D 5M 10M 15M 20M 25M 22538675 22542556 22549195 22573195 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
QMCPACK QMCPACK is a modern high-performance open-source Quantum Monte Carlo (QMC) simulation code making use of MPI for this benchmark of the H20 example code. QMCPACK is an open-source production level many-body ab initio Quantum Monte Carlo code for computing the electronic structure of atoms, molecules, and solids. QMCPACK is supported by the U.S. Department of Energy. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Total Execution Time - Seconds, Fewer Is Better QMCPACK 3.13 Input: simple-H2O A B D C 6 12 18 24 30 26.69 22.65 22.64 21.87 1. (CXX) g++ options: -finline-limit=1000 -fstrict-aliasing -funroll-all-loops -ffast-math -march=native -O3 -lm -ldl
Quantum ESPRESSO Quantum ESPRESSO is an integrated suite of Open-Source computer codes for electronic-structure calculations and materials modeling at the nanoscale. It is based on density-functional theory, plane waves, and pseudopotentials. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Quantum ESPRESSO 7.0 Input: AUSURF112 A D C B 300 600 900 1200 1500 1542.31 1536.74 1535.90 1534.64 1. (F9X) gfortran options: -pthread -fopenmp -ldevXlib -lopenblas -lFoX_dom -lFoX_sax -lFoX_wxml -lFoX_common -lFoX_utils -lFoX_fsys -lfftw3_omp -lfftw3 -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Random Forest C A D B 100 200 300 400 500 481.8 480.8 480.4 480.0 MIN: 437.62 / MAX: 607.1 MIN: 434.42 / MAX: 584.45 MIN: 443.53 / MAX: 630.56 MIN: 439.84 / MAX: 608
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: ALS Movie Lens B A C D 3K 6K 9K 12K 15K 13827.4 13808.3 13710.1 13643.6 MIN: 13827.36 / MAX: 15305.31 MIN: 13808.28 / MAX: 15301.71 MIN: 13710.09 / MAX: 15212.77 MAX: 15019.04
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Apache Spark ALS D C B A 4K 8K 12K 16K 20K 20797.9 20526.5 20524.6 20416.4 MIN: 20754.97 / MAX: 20902.04 MIN: 20468.32 / MAX: 20594.78 MIN: 20452.4 / MAX: 20618.01 MIN: 20337.14 / MAX: 20472.95
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Apache Spark Bayes C A B D 300 600 900 1200 1500 1389.8 1380.4 1371.9 1365.4 MIN: 1032.48 / MAX: 1390.69 MIN: 1030.73 MIN: 1025.2 MIN: 1013.57
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Savina Reactors.IO A D B C 2K 4K 6K 8K 10K 7837.9 7760.0 7456.8 7232.3 MIN: 7837.85 / MAX: 11331.76 MAX: 10905.77 MAX: 11038.64 MIN: 6221.45 / MAX: 10912.79
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Apache Spark PageRank B D A C 600 1200 1800 2400 3000 2757.9 2719.0 2693.3 2560.8 MIN: 2482.23 / MAX: 2957.59 MIN: 2491.58 / MAX: 2936.02 MIN: 2463.11 / MAX: 2826.4 MIN: 2382.1 / MAX: 2770.21
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Finagle HTTP Requests A B C D 600 1200 1800 2400 3000 2720.8 2710.6 2707.7 2692.1 MIN: 2537.94 / MAX: 2897.99 MIN: 2527.4 / MAX: 2864.03 MIN: 2487.03 / MAX: 2841.18 MIN: 2507.61 / MAX: 2821.13
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: In-Memory Database Shootout C B D A 500 1000 1500 2000 2500 2381.9 2363.6 2355.9 2355.5 MIN: 2244.67 / MAX: 2504.25 MIN: 2214.57 / MAX: 2486.06 MIN: 2208.8 / MAX: 2429.73 MIN: 2184.59 / MAX: 2554.54
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Akka Unbalanced Cobwebbed Tree C A B D 2K 4K 6K 8K 10K 9803.8 9752.9 9690.2 9650.6 MIN: 7656.59 / MAX: 9803.83 MIN: 7618.53 MIN: 7503.63 MIN: 7528
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Genetic Algorithm Using Jenetics + Futures C A D B 300 600 900 1200 1500 1344.6 1326.7 1308.1 1297.7 MIN: 1327.22 / MAX: 1358.81 MIN: 1307.06 / MAX: 1345.19 MIN: 1265.88 / MAX: 1358.77 MIN: 1270.51 / MAX: 1321.22
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: NUMA C B A 70 140 210 280 350 314.08 316.03 320.08 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Futex A B C 500K 1000K 1500K 2000K 2500K 2019792.94 2102295.18 2247778.96 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: MEMFD B A C 200 400 600 800 1000 1004.36 1012.73 1013.30 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Atomic B A C 80K 160K 240K 320K 400K 344544.20 357591.98 366961.11 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Crypto B C A 3K 6K 9K 12K 15K 13448.13 13507.57 13517.57 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Malloc A B C 2M 4M 6M 8M 10M 9452055.42 9467042.29 9521309.34 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Forking A C B 17K 34K 51K 68K 85K 76531.56 77378.24 77706.13 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: IO_uring A C B 5K 10K 15K 20K 25K 24227.96 25003.00 25422.10 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: SENDFILE B C A 50K 100K 150K 200K 250K 228041.03 228816.34 229138.15 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: CPU Cache C A B 50 100 150 200 250 219.93 221.73 222.46 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: CPU Stress B C A 5K 10K 15K 20K 25K 23119.63 23206.52 23228.30 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Semaphores B A C 300K 600K 900K 1200K 1500K 1286303.11 1287975.24 1288557.18 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Matrix Math B C A 12K 24K 36K 48K 60K 57205.06 57451.66 57452.05 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Vector Math A C B 13K 26K 39K 52K 65K 58868.12 58922.31 58952.20 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: x86_64 RdRand C B A 20K 40K 60K 80K 100K 77859.90 77948.33 77949.10 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Memory Copying B A C 300 600 900 1200 1500 1607.21 1607.49 1610.06 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Socket Activity C B A 2K 4K 6K 8K 10K 9236.06 9245.36 9279.07 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Context Switching A C B 400K 800K 1200K 1600K 2000K 1924100.28 1946582.89 1955167.18 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Glibc C String Functions B A C 400K 800K 1200K 1600K 2000K 1359296.88 1700830.81 1733539.89 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Glibc Qsort Data Sorting A C B 40 80 120 160 200 182.60 183.73 183.97 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: System V Message Passing B C A 2M 4M 6M 8M 10M 11517263.47 11528872.37 11534338.15 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
SVT-AV1 This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 4 - Input: Bosphorus 4K A B D C 0.5157 1.0314 1.5471 2.0628 2.5785 2.261 2.272 2.284 2.292 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 8 - Input: Bosphorus 4K B D A C 7 14 21 28 35 31.36 31.38 31.43 31.62 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 10 - Input: Bosphorus 4K A B C D 20 40 60 80 100 67.48 79.17 79.41 79.62 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 12 - Input: Bosphorus 4K A B D C 20 40 60 80 100 103.41 106.23 106.51 106.92 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 4 - Input: Bosphorus 1080p C A B D 2 4 6 8 10 6.636 6.689 6.740 6.772 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 8 - Input: Bosphorus 1080p B D A C 20 40 60 80 100 109.33 110.00 110.73 111.24 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 10 - Input: Bosphorus 1080p D A C B 50 100 150 200 250 219.67 225.40 226.72 227.35 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 12 - Input: Bosphorus 1080p A C D B 90 180 270 360 450 408.32 419.71 422.85 423.77 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
SVT-HEVC This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 1 - Input: Bosphorus 4K A B C D 0.522 1.044 1.566 2.088 2.61 2.32 2.32 2.32 2.32 1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 7 - Input: Bosphorus 4K C D B A 10 20 30 40 50 42.38 42.54 42.55 43.99 1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 10 - Input: Bosphorus 4K D A C B 20 40 60 80 100 81.04 81.12 81.88 82.05 1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 1 - Input: Bosphorus 1080p A C B D 3 6 9 12 15 9.28 9.29 9.30 9.30 1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 7 - Input: Bosphorus 1080p A B C D 30 60 90 120 150 137.02 137.36 137.39 137.68 1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 10 - Input: Bosphorus 1080p C A B D 60 120 180 240 300 269.66 269.91 272.36 272.73 1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
SVT-VP9 This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample YUV input video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: VMAF Optimized - Input: Bosphorus 4K A C D B 12 24 36 48 60 49.54 53.02 53.77 54.02 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: VMAF Optimized - Input: Bosphorus 1080p A D C B 40 80 120 160 200 180.10 182.34 182.43 182.72 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: PSNR/SSIM Optimized - Input: Bosphorus 4K A B D C 14 28 42 56 70 51.50 60.65 60.66 60.71 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: PSNR/SSIM Optimized - Input: Bosphorus 1080p A C D B 40 80 120 160 200 183.60 196.43 196.51 197.19 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: Visual Quality Optimized - Input: Bosphorus 4K D A C B 11 22 33 44 55 50.30 50.36 50.37 50.41 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: Visual Quality Optimized - Input: Bosphorus 1080p A C D B 40 80 120 160 200 161.18 161.45 161.49 161.53 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
WebP2 Image Encode This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better WebP2 Image Encode 20220422 Encode Settings: Default C B A 0.7578 1.5156 2.2734 3.0312 3.789 3.368 3.295 3.251 1. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl
OpenBenchmarking.org Frames Per Second, More Is Better x264 2022-02-22 Video Input: Bosphorus 1080p C D B A 30 60 90 120 150 125.68 128.89 129.21 129.99 1. (CC) gcc options: -ldl -m64 -lm -lpthread -O3 -flto
A Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x40 - Thermald 2.4.6Java Notes: OpenJDK Runtime Environment (build 11.0.15+10-Ubuntu-0ubuntu0.21.10.1)Python Notes: Python 3.9.7Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 14 June 2022 15:07 by user pts.
B Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x40 - Thermald 2.4.6Java Notes: OpenJDK Runtime Environment (build 11.0.15+10-Ubuntu-0ubuntu0.21.10.1)Python Notes: Python 3.9.7Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 14 June 2022 20:41 by user pts.
C Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x40 - Thermald 2.4.6Java Notes: OpenJDK Runtime Environment (build 11.0.15+10-Ubuntu-0ubuntu0.21.10.1)Python Notes: Python 3.9.7Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 15 June 2022 04:10 by user pts.
D Processor: Intel Core i9-11900K @ 5.10GHz (8 Cores / 16 Threads), Motherboard: ASUS ROG MAXIMUS XIII HERO (1007 BIOS), Chipset: Intel Tiger Lake-H, Memory: 32GB, Disk: 2000GB Corsair Force MP600, Graphics: ASUS Intel RKL GT1 3GB (1300MHz), Audio: Intel Tiger Lake-H HD Audio, Monitor: MX279, Network: 2 x Intel I225-V + Intel Wi-Fi 6 AX210/AX211/AX411
OS: Ubuntu 21.10, Kernel: 5.15.0-051500rc7daily20211029-generic (x86_64) 20211028, Desktop: GNOME Shell 40.5, Display Server: X Server 1.20.13 + Wayland, OpenGL: 4.6 Mesa 22.0.0-devel (git-f13d486 2021-11-03 impish-oibaf-ppa), Vulkan: 1.2.195, Compiler: GCC 11.2.0, File-System: ext4, Screen Resolution: 1920x1080
Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x40 - Thermald 2.4.6Java Notes: OpenJDK Runtime Environment (build 11.0.15+10-Ubuntu-0ubuntu0.21.10.1)Python Notes: Python 3.9.7Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 15 June 2022 09:28 by user pts.