Intel Core i9-11900K testing with a ASUS ROG MAXIMUS XIII HERO (1007 BIOS) and ASUS Intel RKL GT1 3GB on Ubuntu 21.10 via the Phoronix Test Suite.
A Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x40 - Thermald 2.4.6Java Notes: OpenJDK Runtime Environment (build 11.0.15+10-Ubuntu-0ubuntu0.21.10.1)Python Notes: Python 3.9.7Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
B C D Processor: Intel Core i9-11900K @ 5.10GHz (8 Cores / 16 Threads), Motherboard: ASUS ROG MAXIMUS XIII HERO (1007 BIOS), Chipset: Intel Tiger Lake-H, Memory: 32GB, Disk: 2000GB Corsair Force MP600, Graphics: ASUS Intel RKL GT1 3GB (1300MHz), Audio: Intel Tiger Lake-H HD Audio, Monitor: MX279, Network: 2 x Intel I225-V + Intel Wi-Fi 6 AX210/AX211/AX411
OS: Ubuntu 21.10, Kernel: 5.15.0-051500rc7daily20211029-generic (x86_64) 20211028, Desktop: GNOME Shell 40.5, Display Server: X Server 1.20.13 + Wayland, OpenGL: 4.6 Mesa 22.0.0-devel (git-f13d486 2021-11-03 impish-oibaf-ppa), Vulkan: 1.2.195, Compiler: GCC 11.2.0, File-System: ext4, Screen Resolution: 1920x1080
Glibc Benchmarks The GNU C Library project provides the core libraries for the GNU system and GNU/Linux systems, as well as many other systems that use Linux as the kernel. These libraries provide critical APIs including ISO C11, POSIX.1-2008, BSD, OS-specific APIs and more. This test profile makes use of Glibc's "benchtests" integrated benchmark suite. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ns, Fewer Is Better Glibc Benchmarks Benchmark: cos A B C D 12 24 36 48 60 54.45 53.84 53.81 53.80 1. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s
OpenBenchmarking.org ns, Fewer Is Better Glibc Benchmarks Benchmark: ffs A B C D 0.7591 1.5182 2.2773 3.0364 3.7955 3.33365 3.37358 3.33392 3.33422 1. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s
OpenBenchmarking.org ns, Fewer Is Better Glibc Benchmarks Benchmark: modf A B C D 1.0492 2.0984 3.1476 4.1968 5.246 4.66291 4.65510 4.65758 4.65987 1. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s
OpenBenchmarking.org ns, Fewer Is Better Glibc Benchmarks Benchmark: sqrt A B C D 0.8987 1.7974 2.6961 3.5948 4.4935 3.99377 3.99317 3.99370 3.99406 1. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s
OpenBenchmarking.org ns, Fewer Is Better Glibc Benchmarks Benchmark: ffsll A B C D 0.759 1.518 2.277 3.036 3.795 3.33510 3.33819 3.37318 3.33486 1. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s
OpenBenchmarking.org ns, Fewer Is Better Glibc Benchmarks Benchmark: pthread_once A B C D 0.7621 1.5242 2.2863 3.0484 3.8105 3.32140 3.32250 3.38731 3.32184 1. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s
perf-bench This test profile is used for running Linux perf-bench, the benchmark support within the Linux kernel's perf tool. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ops/sec, More Is Better perf-bench Benchmark: Epoll Wait A B C D 30K 60K 90K 120K 150K 127515 130239 128524 131078 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
OpenBenchmarking.org ops/sec, More Is Better perf-bench Benchmark: Futex Hash A B C D 1.5M 3M 4.5M 6M 7.5M 6819705 6685401 6810542 6460586 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
OpenBenchmarking.org GB/sec, More Is Better perf-bench Benchmark: Memcpy 1MB A B C D 7 14 21 28 35 31.36 31.36 31.49 31.36 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
OpenBenchmarking.org GB/sec, More Is Better perf-bench Benchmark: Memset 1MB A B C D 13 26 39 52 65 54.19 57.45 55.72 54.12 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
OpenBenchmarking.org ops/sec, More Is Better perf-bench Benchmark: Sched Pipe A B C D 70K 140K 210K 280K 350K 318866 315267 316210 322747 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
OpenBenchmarking.org ops/sec, More Is Better perf-bench Benchmark: Futex Lock-Pi A B C D 200 400 600 800 1000 1082 1076 1087 1089 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
OpenBenchmarking.org ops/sec, More Is Better perf-bench Benchmark: Syscall Basic A B C D 5M 10M 15M 20M 25M 22549195 22542556 22538675 22573195 1. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lpython3.9 -lcrypt -lutil -lz -lnuma
Etcpak Etcpack is the self-proclaimed "fastest ETC compressor on the planet" with focused on providing open-source, very fast ETC and S3 texture compression support. The test profile uses a 8K x 8K game texture as a sample input. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Mpx/s, More Is Better Etcpak 1.0 Benchmark: Multi-Threaded - Configuration: ETC2 A B C D 600 1200 1800 2400 3000 2658.51 2656.51 2656.20 2658.62 1. (CXX) g++ options: -O3 -march=native -std=c++11 -lpthread
OpenBenchmarking.org Mpx/s, More Is Better Etcpak 1.0 Benchmark: Single-Threaded - Configuration: ETC2 A B C D 70 140 210 280 350 307.60 307.01 306.91 306.41 1. (CXX) g++ options: -O3 -march=native -std=c++11 -lpthread
QMCPACK QMCPACK is a modern high-performance open-source Quantum Monte Carlo (QMC) simulation code making use of MPI for this benchmark of the H20 example code. QMCPACK is an open-source production level many-body ab initio Quantum Monte Carlo code for computing the electronic structure of atoms, molecules, and solids. QMCPACK is supported by the U.S. Department of Energy. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Total Execution Time - Seconds, Fewer Is Better QMCPACK 3.13 Input: simple-H2O A B C D 6 12 18 24 30 26.69 22.65 21.87 22.64 1. (CXX) g++ options: -finline-limit=1000 -fstrict-aliasing -funroll-all-loops -ffast-math -march=native -O3 -lm -ldl
Quantum ESPRESSO Quantum ESPRESSO is an integrated suite of Open-Source computer codes for electronic-structure calculations and materials modeling at the nanoscale. It is based on density-functional theory, plane waves, and pseudopotentials. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Quantum ESPRESSO 7.0 Input: AUSURF112 A B C D 300 600 900 1200 1500 1542.31 1534.64 1535.90 1536.74 1. (F9X) gfortran options: -pthread -fopenmp -ldevXlib -lopenblas -lFoX_dom -lFoX_sax -lFoX_wxml -lFoX_common -lFoX_utils -lFoX_fsys -lfftw3_omp -lfftw3 -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Random Forest A B C D 100 200 300 400 500 480.8 480.0 481.8 480.4 MIN: 434.42 / MAX: 584.45 MIN: 439.84 / MAX: 608 MIN: 437.62 / MAX: 607.1 MIN: 443.53 / MAX: 630.56
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: ALS Movie Lens A B C D 3K 6K 9K 12K 15K 13808.3 13827.4 13710.1 13643.6 MIN: 13808.28 / MAX: 15301.71 MIN: 13827.36 / MAX: 15305.31 MIN: 13710.09 / MAX: 15212.77 MAX: 15019.04
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Apache Spark ALS A B C D 4K 8K 12K 16K 20K 20416.4 20524.6 20526.5 20797.9 MIN: 20337.14 / MAX: 20472.95 MIN: 20452.4 / MAX: 20618.01 MIN: 20468.32 / MAX: 20594.78 MIN: 20754.97 / MAX: 20902.04
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Apache Spark Bayes A B C D 300 600 900 1200 1500 1380.4 1371.9 1389.8 1365.4 MIN: 1030.73 MIN: 1025.2 MIN: 1032.48 / MAX: 1390.69 MIN: 1013.57
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Savina Reactors.IO A B C D 2K 4K 6K 8K 10K 7837.9 7456.8 7232.3 7760.0 MIN: 7837.85 / MAX: 11331.76 MAX: 11038.64 MIN: 6221.45 / MAX: 10912.79 MAX: 10905.77
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Apache Spark PageRank A B C D 600 1200 1800 2400 3000 2693.3 2757.9 2560.8 2719.0 MIN: 2463.11 / MAX: 2826.4 MIN: 2482.23 / MAX: 2957.59 MIN: 2382.1 / MAX: 2770.21 MIN: 2491.58 / MAX: 2936.02
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Finagle HTTP Requests A B C D 600 1200 1800 2400 3000 2720.8 2710.6 2707.7 2692.1 MIN: 2537.94 / MAX: 2897.99 MIN: 2527.4 / MAX: 2864.03 MIN: 2487.03 / MAX: 2841.18 MIN: 2507.61 / MAX: 2821.13
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: In-Memory Database Shootout A B C D 500 1000 1500 2000 2500 2355.5 2363.6 2381.9 2355.9 MIN: 2184.59 / MAX: 2554.54 MIN: 2214.57 / MAX: 2486.06 MIN: 2244.67 / MAX: 2504.25 MIN: 2208.8 / MAX: 2429.73
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Akka Unbalanced Cobwebbed Tree A B C D 2K 4K 6K 8K 10K 9752.9 9690.2 9803.8 9650.6 MIN: 7618.53 MIN: 7503.63 MIN: 7656.59 / MAX: 9803.83 MIN: 7528
OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Genetic Algorithm Using Jenetics + Futures A B C D 300 600 900 1200 1500 1326.7 1297.7 1344.6 1308.1 MIN: 1307.06 / MAX: 1345.19 MIN: 1270.51 / MAX: 1321.22 MIN: 1327.22 / MAX: 1358.81 MIN: 1265.88 / MAX: 1358.77
Nettle GNU Nettle is a low-level cryptographic library used by GnuTLS and other software. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Mbyte/s, More Is Better Nettle 3.8 Test: aes256 A B C D 3K 6K 9K 12K 15K 12207.52 12205.23 12195.87 12155.79 MIN: 7986.83 / MAX: 20656.74 MIN: 7983.96 / MAX: 20655.99 MIN: 7973.01 / MAX: 20659.96 MIN: 7981.08 / MAX: 20658.54 1. (CC) gcc options: -O2 -ggdb3 -lnettle -lm -lcrypto
OpenBenchmarking.org Mbyte/s, More Is Better Nettle 3.8 Test: chacha A B C D 300 600 900 1200 1500 1519.06 1511.89 1517.83 1520.61 MIN: 679.9 / MAX: 4635.71 MIN: 679.88 / MAX: 4633.29 MIN: 679.93 / MAX: 4633.69 MIN: 680.19 / MAX: 4644.14 1. (CC) gcc options: -O2 -ggdb3 -lnettle -lm -lcrypto
OpenBenchmarking.org Mbyte/s, More Is Better Nettle 3.8 Test: sha512 A B C D 160 320 480 640 800 722.49 721.29 747.25 726.78 1. (CC) gcc options: -O2 -ggdb3 -lnettle -lm -lcrypto
OpenBenchmarking.org Mbyte/s, More Is Better Nettle 3.8 Test: poly1305-aes A B C D 1000 2000 3000 4000 5000 4641.58 4442.12 4637.23 4638.26 1. (CC) gcc options: -O2 -ggdb3 -lnettle -lm -lcrypto
OpenBenchmarking.org FPS, More Is Better dav1d 1.0 Video Input: Summer Nature 4K A B C D 40 80 120 160 200 198.82 194.82 189.63 189.52 1. (CC) gcc options: -pthread -lm
OpenBenchmarking.org FPS, More Is Better dav1d 1.0 Video Input: Summer Nature 1080p A B C D 200 400 600 800 1000 893.33 904.56 903.95 904.36 1. (CC) gcc options: -pthread -lm
OpenBenchmarking.org FPS, More Is Better dav1d 1.0 Video Input: Chimera 1080p 10-bit A B C D 110 220 330 440 550 492.46 490.12 471.44 476.40 1. (CC) gcc options: -pthread -lm
AOM AV1 OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 4K A B C D 0.036 0.072 0.108 0.144 0.18 0.16 0.16 0.16 0.16 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 4K A B C D 1.1633 2.3266 3.4899 4.6532 5.8165 5.17 5.15 5.14 5.15 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 6 Realtime - Input: Bosphorus 4K A B C D 4 8 12 16 20 16.21 16.13 16.03 16.06 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 4K A B C D 3 6 9 12 15 9.90 9.79 9.79 9.81 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 8 Realtime - Input: Bosphorus 4K A B C D 11 22 33 44 55 48.31 48.44 48.22 48.04 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 9 Realtime - Input: Bosphorus 4K A B C D 15 30 45 60 75 64.44 68.64 68.35 67.18 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 10 Realtime - Input: Bosphorus 4K A B C D 16 32 48 64 80 72.48 71.70 62.05 71.97 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 1080p A B C D 0.0945 0.189 0.2835 0.378 0.4725 0.42 0.42 0.42 0.42 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 1080p A B C D 3 6 9 12 15 11.32 11.27 11.25 11.27 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 6 Realtime - Input: Bosphorus 1080p A B C D 3 6 9 12 15 11.93 11.94 11.90 12.31 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 1080p A B C D 7 14 21 28 35 30.47 30.30 30.33 30.29 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 8 Realtime - Input: Bosphorus 1080p A B C D 30 60 90 120 150 145.53 143.21 143.74 143.85 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 9 Realtime - Input: Bosphorus 1080p A B C D 40 80 120 160 200 176.81 188.18 186.32 187.30 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 10 Realtime - Input: Bosphorus 1080p A B C D 40 80 120 160 200 203.50 203.95 201.74 197.60 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
SVT-AV1 This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 4 - Input: Bosphorus 4K A B C D 0.5157 1.0314 1.5471 2.0628 2.5785 2.261 2.272 2.292 2.284 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 8 - Input: Bosphorus 4K A B C D 7 14 21 28 35 31.43 31.36 31.62 31.38 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 10 - Input: Bosphorus 4K A B C D 20 40 60 80 100 67.48 79.17 79.41 79.62 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 12 - Input: Bosphorus 4K A B C D 20 40 60 80 100 103.41 106.23 106.92 106.51 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 4 - Input: Bosphorus 1080p A B C D 2 4 6 8 10 6.689 6.740 6.636 6.772 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 8 - Input: Bosphorus 1080p A B C D 20 40 60 80 100 110.73 109.33 111.24 110.00 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 10 - Input: Bosphorus 1080p A B C D 50 100 150 200 250 225.40 227.35 226.72 219.67 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.0 Encoder Mode: Preset 12 - Input: Bosphorus 1080p A B C D 90 180 270 360 450 408.32 423.77 419.71 422.85 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
SVT-HEVC This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 1 - Input: Bosphorus 4K A B C D 0.522 1.044 1.566 2.088 2.61 2.32 2.32 2.32 2.32 1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 7 - Input: Bosphorus 4K A B C D 10 20 30 40 50 43.99 42.55 42.38 42.54 1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 10 - Input: Bosphorus 4K A B C D 20 40 60 80 100 81.12 82.05 81.88 81.04 1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 1 - Input: Bosphorus 1080p A B C D 3 6 9 12 15 9.28 9.30 9.29 9.30 1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 7 - Input: Bosphorus 1080p A B C D 30 60 90 120 150 137.02 137.36 137.39 137.68 1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 10 - Input: Bosphorus 1080p A B C D 60 120 180 240 300 269.91 272.36 269.66 272.73 1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
SVT-VP9 This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample YUV input video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: VMAF Optimized - Input: Bosphorus 4K A B C D 12 24 36 48 60 49.54 54.02 53.02 53.77 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: VMAF Optimized - Input: Bosphorus 1080p A B C D 40 80 120 160 200 180.10 182.72 182.43 182.34 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: PSNR/SSIM Optimized - Input: Bosphorus 4K A B C D 14 28 42 56 70 51.50 60.65 60.71 60.66 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: PSNR/SSIM Optimized - Input: Bosphorus 1080p A B C D 40 80 120 160 200 183.60 197.19 196.43 196.51 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: Visual Quality Optimized - Input: Bosphorus 4K A B C D 11 22 33 44 55 50.36 50.41 50.37 50.30 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: Visual Quality Optimized - Input: Bosphorus 1080p A B C D 40 80 120 160 200 161.18 161.53 161.45 161.49 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Frames Per Second, More Is Better x264 2022-02-22 Video Input: Bosphorus 1080p A B C D 30 60 90 120 150 129.99 129.21 125.68 128.89 1. (CC) gcc options: -ldl -m64 -lm -lpthread -O3 -flto
OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: particle_volume/pathtracer/real_time A B C D 60 120 180 240 300 297.46 297.50 297.68 297.35
OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: gravity_spheres_volume/dim_512/ao/real_time A B C D 0.959 1.918 2.877 3.836 4.795 4.24782 4.24824 4.22833 4.26224
OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: gravity_spheres_volume/dim_512/scivis/real_time A B C D 0.9527 1.9054 2.8581 3.8108 4.7635 4.20308 4.18832 4.19096 4.23414
OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_time A B C D 1.1456 2.2912 3.4368 4.5824 5.728 5.09145 5.06027 5.08966 5.08714
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU A B C D 0.9056 1.8112 2.7168 3.6224 4.528 4.01177 4.00817 4.02480 4.00116 MIN: 3.9 MIN: 3.9 MIN: 3.92 MIN: 3.91 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU A B C D 3 6 9 12 15 11.11 11.12 11.12 11.12 MIN: 11.05 MIN: 11.06 MIN: 11.07 MIN: 11.05 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU A B C D 0.1635 0.327 0.4905 0.654 0.8175 0.720371 0.715645 0.716538 0.726471 MIN: 0.67 MIN: 0.67 MIN: 0.66 MIN: 0.67 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU A B C D 0.698 1.396 2.094 2.792 3.49 3.08878 3.06780 3.06944 3.10236 MIN: 3.02 MIN: 3.02 MIN: 3.03 MIN: 3.03 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU A B C D 2 4 6 8 10 8.55826 8.54762 8.54691 8.55027 MIN: 8.44 MIN: 8.42 MIN: 8.39 MIN: 8.42 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU A B C D 1.2256 2.4512 3.6768 4.9024 6.128 5.38692 5.20558 5.44701 5.22850 MIN: 4.8 MIN: 4.65 MIN: 4.95 MIN: 4.75 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU A B C D 4 8 12 16 20 14.14 14.12 14.13 14.12 MIN: 14.06 MIN: 14.03 MIN: 14.04 MIN: 14.04 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU A B C D 2 4 6 8 10 8.06688 8.03869 7.70230 7.72180 MIN: 4.81 MIN: 4.84 MIN: 4.77 MIN: 4.81 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU A B C D 0.9761 1.9522 2.9283 3.9044 4.8805 4.33830 4.33567 4.33558 4.33067 MIN: 4.22 MIN: 4.25 MIN: 4.24 MIN: 4.23 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU A B C D 3 6 9 12 15 11.97 11.96 11.97 11.96 MIN: 11.9 MIN: 11.89 MIN: 11.9 MIN: 11.89 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU A B C D 0.186 0.372 0.558 0.744 0.93 0.825884 0.826044 0.826686 0.826144 MIN: 0.81 MIN: 0.81 MIN: 0.82 MIN: 0.82 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU A B C D 0.3405 0.681 1.0215 1.362 1.7025 1.47653 1.46563 1.51319 1.50262 MIN: 1.39 MIN: 1.39 MIN: 1.43 MIN: 1.4 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU A B C D 700 1400 2100 2800 3500 3118.35 3189.95 3106.14 3106.54 MIN: 3113.21 MIN: 3078.21 MIN: 3100.97 MIN: 3101.69 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU A B C D 400 800 1200 1600 2000 1831.84 1831.92 1832.56 1825.78 MIN: 1827.31 MIN: 1827.22 MIN: 1827.43 MIN: 1821.07 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU A B C D 900 1800 2700 3600 4500 4358.46 3107.86 3113.19 3110.84 MIN: 3112.17 MIN: 3100.51 MIN: 3107.55 MIN: 3105.51 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU A B C D 4 8 12 16 20 16.16 16.15 16.16 16.15 MIN: 16.07 MIN: 16.07 MIN: 16.07 MIN: 16.07 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU A B C D 5 10 15 20 25 18.79 18.76 18.78 18.78 MIN: 18.6 MIN: 18.56 MIN: 18.58 MIN: 18.6 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU A B C D 4 8 12 16 20 17.15 17.15 17.15 17.14 MIN: 16.73 MIN: 16.79 MIN: 16.78 MIN: 16.81 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU A B C D 500 1000 1500 2000 2500 2245.91 1833.40 1839.49 1831.95 MIN: 1825.27 MIN: 1828.11 MIN: 1826.96 MIN: 1828.17 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU A B C D 0.729 1.458 2.187 2.916 3.645 3.23261 3.23137 3.23367 3.24005 MIN: 3.18 MIN: 3.18 MIN: 3.18 MIN: 3.19 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU A B C D 700 1400 2100 2800 3500 3115.29 3104.69 3114.22 3108.58 MIN: 3100.97 MIN: 3099.04 MIN: 3107.51 MIN: 3103.14 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU A B C D 400 800 1200 1600 2000 1833.14 1831.29 1831.34 1830.71 MIN: 1828.37 MIN: 1826.61 MIN: 1827.21 MIN: 1826.24 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU A B C D 0.2927 0.5854 0.8781 1.1708 1.4635 1.29412 1.29350 1.29680 1.30110 MIN: 1.25 MIN: 1.25 MIN: 1.24 MIN: 1.25 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU A B C D 0.7757 1.5514 2.3271 3.1028 3.8785 3.43567 3.44465 3.43528 3.44775 MIN: 3.29 MIN: 3.29 MIN: 3.29 MIN: 3.28 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread
OSPray Studio Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer A B C D 400 800 1200 1600 2000 1918 1907 1911 1912 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer A B C D 400 800 1200 1600 2000 1938 1957 1956 1953 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer A B C 500 1000 1500 2000 2500 2328 2332 2331 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer A B C 7K 14K 21K 28K 35K 30515 30572 30584 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer A B C 14K 28K 42K 56K 70K 64451 64393 64242 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer A B C 7K 14K 21K 28K 35K 30947 31244 31173 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer A B C 14K 28K 42K 56K 70K 65641 65765 65487 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer A B C 8K 16K 24K 32K 40K 36885 37170 37059 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer A B C 17K 34K 51K 68K 85K 77357 77543 77555 1. (CXX) g++ options: -O3 -lm -ldl
WebP2 Image Encode This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better WebP2 Image Encode 20220422 Encode Settings: Default A B C 0.7578 1.5156 2.2734 3.0312 3.789 3.251 3.295 3.368 1. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl
GROMACS The GROMACS (GROningen MAchine for Chemical Simulations) molecular dynamics package testing with the water_GMX50 data. This test profile allows selecting between CPU and GPU-based GROMACS builds. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Ns Per Day, More Is Better GROMACS 2022.1 Implementation: MPI CPU - Input: water_GMX50_bare A B C 0.2259 0.4518 0.6777 0.9036 1.1295 0.995 1.001 1.004 1. (CXX) g++ options: -O3
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: NUMA A B C 70 140 210 280 350 320.08 316.03 314.08 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Futex A B C 500K 1000K 1500K 2000K 2500K 2019792.94 2102295.18 2247778.96 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: MEMFD A B C 200 400 600 800 1000 1012.73 1004.36 1013.30 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Atomic A B C 80K 160K 240K 320K 400K 357591.98 344544.20 366961.11 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Crypto A B C 3K 6K 9K 12K 15K 13517.57 13448.13 13507.57 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Malloc A B C 2M 4M 6M 8M 10M 9452055.42 9467042.29 9521309.34 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Forking A B C 17K 34K 51K 68K 85K 76531.56 77706.13 77378.24 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: IO_uring A B C 5K 10K 15K 20K 25K 24227.96 25422.10 25003.00 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: SENDFILE A B C 50K 100K 150K 200K 250K 229138.15 228041.03 228816.34 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: CPU Cache A B C 50 100 150 200 250 221.73 222.46 219.93 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: CPU Stress A B C 5K 10K 15K 20K 25K 23228.30 23119.63 23206.52 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Semaphores A B C 300K 600K 900K 1200K 1500K 1287975.24 1286303.11 1288557.18 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Matrix Math A B C 12K 24K 36K 48K 60K 57452.05 57205.06 57451.66 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Vector Math A B C 13K 26K 39K 52K 65K 58868.12 58952.20 58922.31 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: x86_64 RdRand A B C 20K 40K 60K 80K 100K 77949.10 77948.33 77859.90 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Memory Copying A B C 300 600 900 1200 1500 1607.49 1607.21 1610.06 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Socket Activity A B C 2K 4K 6K 8K 10K 9279.07 9245.36 9236.06 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Context Switching A B C 400K 800K 1200K 1600K 2000K 1924100.28 1955167.18 1946582.89 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Glibc C String Functions A B C 400K 800K 1200K 1600K 2000K 1700830.81 1359296.88 1733539.89 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: Glibc Qsort Data Sorting A B C 40 80 120 160 200 182.60 183.97 183.73 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.14 Test: System V Message Passing A B C 2M 4M 6M 8M 10M 11534338.15 11517263.47 11528872.37 1. (CC) gcc options: -O2 -std=gnu99 -lm -lapparmor -latomic -lc -lcrypt -ldl -ljpeg -lrt -lsctp -lz -pthread
Blender Blender is an open-source 3D creation and modeling software project. This test is of Blender's Cycles benchmark with various sample files. GPU computing via NVIDIA OptiX and NVIDIA CUDA is currently supported and HIP for AMD Radeon GPUs. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Blender 3.2 Blend File: BMW27 - Compute: CPU-Only A B C 30 60 90 120 150 131.60 131.44 131.29
OpenBenchmarking.org Op/s, More Is Better Facebook RocksDB 7.0.1 Test: Update Random A B C 150K 300K 450K 600K 750K 703459 703316 700065 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
OpenBenchmarking.org Op/s, More Is Better Facebook RocksDB 7.0.1 Test: Read While Writing A B C 500K 1000K 1500K 2000K 2500K 2152704 2199309 2223231 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
OpenBenchmarking.org Op/s, More Is Better Facebook RocksDB 7.0.1 Test: Read Random Write Random A B C 400K 800K 1200K 1600K 2000K 2027221 2041642 2041553 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
ONNX Runtime ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: GPT-2 - Device: CPU - Executor: Parallel A B C 1200 2400 3600 4800 6000 5384 5396 5376 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: GPT-2 - Device: CPU - Executor: Standard A B C 1400 2800 4200 5600 7000 6701 6702 6640 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: yolov4 - Device: CPU - Executor: Parallel A B C 70 140 210 280 350 306 308 306 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: yolov4 - Device: CPU - Executor: Standard A B C 110 220 330 440 550 526 526 525 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: bertsquad-12 - Device: CPU - Executor: Parallel A B C 110 220 330 440 550 512 500 511 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: bertsquad-12 - Device: CPU - Executor: Standard A B C 200 400 600 800 1000 1015 1010 642 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel A B C 16 32 48 64 80 70 71 70 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: fcn-resnet101-11 - Device: CPU - Executor: Standard A B C 20 40 60 80 100 104 104 104 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel A B C 300 600 900 1200 1500 1262 1269 1255 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard A B C 400 800 1200 1600 2000 2078 2071 2076 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: super-resolution-10 - Device: CPU - Executor: Parallel A B C 1000 2000 3000 4000 5000 4547 4525 4586 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: super-resolution-10 - Device: CPU - Executor: Standard A B C 1000 2000 3000 4000 5000 4558 4528 4548 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
InfluxDB This is a benchmark of the InfluxDB open-source time-series database optimized for fast, high-availability storage for IoT and other use-cases. The InfluxDB test profile makes use of InfluxDB Inch for facilitating the benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org val/sec, More Is Better InfluxDB 1.8.2 Concurrent Streams: 4 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000 A B C 400K 800K 1200K 1600K 2000K 1806268.7 1804674.6 1815340.8
OpenBenchmarking.org val/sec, More Is Better InfluxDB 1.8.2 Concurrent Streams: 64 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000 A B C 400K 800K 1200K 1600K 2000K 1818333.8 1805397.9 1814759.0
A Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x40 - Thermald 2.4.6Java Notes: OpenJDK Runtime Environment (build 11.0.15+10-Ubuntu-0ubuntu0.21.10.1)Python Notes: Python 3.9.7Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 14 June 2022 15:07 by user pts.
B Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x40 - Thermald 2.4.6Java Notes: OpenJDK Runtime Environment (build 11.0.15+10-Ubuntu-0ubuntu0.21.10.1)Python Notes: Python 3.9.7Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 14 June 2022 20:41 by user pts.
C Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x40 - Thermald 2.4.6Java Notes: OpenJDK Runtime Environment (build 11.0.15+10-Ubuntu-0ubuntu0.21.10.1)Python Notes: Python 3.9.7Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 15 June 2022 04:10 by user pts.
D Processor: Intel Core i9-11900K @ 5.10GHz (8 Cores / 16 Threads), Motherboard: ASUS ROG MAXIMUS XIII HERO (1007 BIOS), Chipset: Intel Tiger Lake-H, Memory: 32GB, Disk: 2000GB Corsair Force MP600, Graphics: ASUS Intel RKL GT1 3GB (1300MHz), Audio: Intel Tiger Lake-H HD Audio, Monitor: MX279, Network: 2 x Intel I225-V + Intel Wi-Fi 6 AX210/AX211/AX411
OS: Ubuntu 21.10, Kernel: 5.15.0-051500rc7daily20211029-generic (x86_64) 20211028, Desktop: GNOME Shell 40.5, Display Server: X Server 1.20.13 + Wayland, OpenGL: 4.6 Mesa 22.0.0-devel (git-f13d486 2021-11-03 impish-oibaf-ppa), Vulkan: 1.2.195, Compiler: GCC 11.2.0, File-System: ext4, Screen Resolution: 1920x1080
Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x40 - Thermald 2.4.6Java Notes: OpenJDK Runtime Environment (build 11.0.15+10-Ubuntu-0ubuntu0.21.10.1)Python Notes: Python 3.9.7Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 15 June 2022 09:28 by user pts.