2 x Intel Xeon Max 9480 benchmarks for a future article.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 2404035-NE-XEONMAXLI15 Xeon Max Linux 6.9 Kernel Benchmarks - Phoronix Test Suite Xeon Max Linux 6.9 Kernel Benchmarks 2 x Intel Xeon Max 9480 benchmarks for a future article.
HTML result view exported from: https://openbenchmarking.org/result/2404035-NE-XEONMAXLI15 .
Xeon Max Linux 6.9 Kernel Benchmarks Processor Motherboard Chipset Memory Disk Graphics Network OS Kernel Desktop Display Server Compiler File-System Screen Resolution Linux 6.8 Linux 6.9-rc2 2 x Intel Xeon Max 9480 @ 3.50GHz (112 Cores / 224 Threads) Supermicro SYS-221H-TNR X13DEM v1.10 (1.3 BIOS) Intel Device 1bce 512GB 7682GB INTEL SSDPF2KX076TZ ASPEED 2 x Broadcom BCM57508 NetXtreme-E 10Gb/25Gb/40Gb/50Gb/100Gb/200Gb Ubuntu 23.10 6.8.0-060800-generic (x86_64) GNOME Shell 45.0 X Server 1.21.1.7 GCC 13.2.0 ext4 1024x768 6.9.0-060900rc2-generic (x86_64) OpenBenchmarking.org Kernel Details - Transparent Huge Pages: madvise Compiler Details - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details - Scaling Governor: intel_cpufreq performance - CPU Microcode: 0x2c000290 Python Details - Python 3.11.6 Security Details - Linux 6.8: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Not affected - Linux 6.9-rc2: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + reg_file_data_sampling: Not affected + retbleed: Not affected + spec_rstack_overflow: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Not affected
Xeon Max Linux 6.9 Kernel Benchmarks quicksilver: CORAL2 P1 quicksilver: CORAL2 P2 ffmpeg: libx264 - Live ffmpeg: libx265 - Live ffmpeg: libx264 - Upload ffmpeg: libx265 - Upload ffmpeg: libx264 - Platform ffmpeg: libx265 - Platform ffmpeg: libx264 - Video On Demand ffmpeg: libx265 - Video On Demand xmrig: KawPow - 1M xmrig: Monero - 1M xmrig: Wownero - 1M xmrig: GhostRider - 1M xmrig: CryptoNight-Heavy - 1M xmrig: CryptoNight-Femto UPX2 - 1M cachebench: Read cachebench: Write cachebench: Read / Modify / Write compress-lz4: 1 - Compression Speed compress-lz4: 1 - Decompression Speed compress-lz4: 3 - Compression Speed compress-lz4: 3 - Decompression Speed compress-lz4: 9 - Compression Speed compress-lz4: 9 - Decompression Speed jpegxl: PNG - 80 jpegxl: PNG - 90 jpegxl: JPEG - 80 jpegxl: JPEG - 90 jpegxl: PNG - 100 jpegxl: JPEG - 100 jpegxl-decode: All webp2: Default webp2: Quality 75, Compression Effort 7 webp2: Quality 95, Compression Effort 7 webp2: Quality 100, Compression Effort 5 webp2: Quality 100, Lossless Compression srsran: PDSCH Processor Benchmark, Throughput Total embree: Pathtracer - Crown embree: Pathtracer ISPC - Crown embree: Pathtracer - Asian Dragon embree: Pathtracer - Asian Dragon Obj embree: Pathtracer ISPC - Asian Dragon embree: Pathtracer ISPC - Asian Dragon Obj svt-av1: Preset 4 - Bosphorus 4K svt-av1: Preset 8 - Bosphorus 4K svt-av1: Preset 12 - Bosphorus 4K svt-av1: Preset 13 - Bosphorus 4K svt-av1: Preset 4 - Bosphorus 1080p vvenc: Bosphorus 4K - Fast vvenc: Bosphorus 4K - Faster vvenc: Bosphorus 1080p - Fast vvenc: Bosphorus 1080p - Faster oidn: RT.hdr_alb_nrm.3840x2160 - CPU-Only oidn: RT.ldr_alb_nrm.3840x2160 - CPU-Only oidn: RTLightmap.hdr.4096x4096 - CPU-Only ospray: particle_volume/ao/real_time ospray: particle_volume/scivis/real_time ospray: particle_volume/pathtracer/real_time ospray: gravity_spheres_volume/dim_512/ao/real_time ospray: gravity_spheres_volume/dim_512/scivis/real_time ospray: gravity_spheres_volume/dim_512/pathtracer/real_time stockfish: Chess Benchmark build-ffmpeg: Time To Compile build-linux-kernel: defconfig build-linux-kernel: allmodconfig build-mesa: Time To Compile primesieve: 1e12 primesieve: 1e13 y-cruncher: 1B y-cruncher: 500M onednn: IP Shapes 1D - CPU onednn: IP Shapes 3D - CPU onednn: Convolution Batch Shapes Auto - CPU onednn: Deconvolution Batch shapes_1d - CPU onednn: Deconvolution Batch shapes_3d - CPU onednn: Recurrent Neural Network Training - CPU onednn: Recurrent Neural Network Inference - CPU ospray-studio: 1 - 4K - 1 - Path Tracer - CPU ospray-studio: 2 - 4K - 1 - Path Tracer - CPU ospray-studio: 3 - 4K - 1 - Path Tracer - CPU ospray-studio: 1 - 4K - 16 - Path Tracer - CPU ospray-studio: 1 - 4K - 32 - Path Tracer - CPU ospray-studio: 2 - 4K - 16 - Path Tracer - CPU ospray-studio: 2 - 4K - 32 - Path Tracer - CPU ospray-studio: 3 - 4K - 16 - Path Tracer - CPU ospray-studio: 3 - 4K - 32 - Path Tracer - CPU ospray-studio: 1 - 1080p - 1 - Path Tracer - CPU ospray-studio: 2 - 1080p - 1 - Path Tracer - CPU ospray-studio: 3 - 1080p - 1 - Path Tracer - CPU ospray-studio: 1 - 1080p - 16 - Path Tracer - CPU ospray-studio: 1 - 1080p - 32 - Path Tracer - CPU ospray-studio: 2 - 1080p - 16 - Path Tracer - CPU ospray-studio: 2 - 1080p - 32 - Path Tracer - CPU ospray-studio: 3 - 1080p - 16 - Path Tracer - CPU ospray-studio: 3 - 1080p - 32 - Path Tracer - CPU hackbench: 16 - Thread hackbench: 16 - Process hackbench: 32 - Process openssl: SHA256 openssl: SHA512 openssl: RSA4096 openssl: RSA4096 gromacs: MPI CPU - water_GMX50_bare mysqlslap: 512 mysqlslap: 1024 tensorflow: CPU - 1 - AlexNet tensorflow: CPU - 16 - AlexNet tensorflow: CPU - 32 - AlexNet tensorflow: CPU - 64 - AlexNet tensorflow: CPU - 1 - GoogLeNet tensorflow: CPU - 1 - ResNet-50 tensorflow: CPU - 256 - AlexNet tensorflow: CPU - 16 - GoogLeNet tensorflow: CPU - 16 - ResNet-50 tensorflow: CPU - 32 - GoogLeNet tensorflow: CPU - 32 - ResNet-50 tensorflow: CPU - 64 - GoogLeNet tensorflow: CPU - 64 - ResNet-50 tensorflow: CPU - 256 - GoogLeNet tensorflow: CPU - 256 - ResNet-50 draco: Lion draco: Church Facade stress-ng: Hash stress-ng: MMAP stress-ng: NUMA stress-ng: Pipe stress-ng: Poll stress-ng: Zlib stress-ng: MEMFD stress-ng: Mutex stress-ng: Crypto stress-ng: Malloc stress-ng: Pthread stress-ng: AVL Tree stress-ng: IO_uring stress-ng: SENDFILE stress-ng: CPU Stress stress-ng: Semaphores stress-ng: Matrix Math stress-ng: Floating Point stress-ng: Matrix 3D Math stress-ng: Memory Copying stress-ng: Vector Shuffle stress-ng: Mixed Scheduler stress-ng: Socket Activity stress-ng: Wide Vector Math stress-ng: Context Switching stress-ng: Fused Multiply-Add stress-ng: Vector Floating Point stress-ng: Glibc C String Functions stress-ng: Glibc Qsort Data Sorting stress-ng: System V Message Passing blender: BMW27 - CPU-Only blender: Junkshop - CPU-Only blender: Classroom - CPU-Only blender: Fishy Cat - CPU-Only blender: Barbershop - CPU-Only blender: Pabellon Barcelona - CPU-Only openvino: Face Detection FP16 - CPU openvino: Face Detection FP16 - CPU openvino: Person Detection FP16 - CPU openvino: Person Detection FP16 - CPU openvino: Person Detection FP32 - CPU openvino: Person Detection FP32 - CPU openvino: Vehicle Detection FP16 - CPU openvino: Vehicle Detection FP16 - CPU openvino: Face Detection FP16-INT8 - CPU openvino: Face Detection FP16-INT8 - CPU openvino: Face Detection Retail FP16 - CPU openvino: Face Detection Retail FP16 - CPU openvino: Road Segmentation ADAS FP16 - CPU openvino: Road Segmentation ADAS FP16 - CPU openvino: Vehicle Detection FP16-INT8 - CPU openvino: Vehicle Detection FP16-INT8 - CPU openvino: Weld Porosity Detection FP16 - CPU openvino: Weld Porosity Detection FP16 - CPU openvino: Face Detection Retail FP16-INT8 - CPU openvino: Face Detection Retail FP16-INT8 - CPU openvino: Road Segmentation ADAS FP16-INT8 - CPU openvino: Road Segmentation ADAS FP16-INT8 - CPU openvino: Machine Translation EN To DE FP16 - CPU openvino: Machine Translation EN To DE FP16 - CPU openvino: Weld Porosity Detection FP16-INT8 - CPU openvino: Weld Porosity Detection FP16-INT8 - CPU openvino: Person Vehicle Bike Detection FP16 - CPU openvino: Person Vehicle Bike Detection FP16 - CPU openvino: Noise Suppression Poconet-Like FP16 - CPU openvino: Noise Suppression Poconet-Like FP16 - CPU openvino: Handwritten English Recognition FP16 - CPU openvino: Handwritten English Recognition FP16 - CPU openvino: Person Re-Identification Retail FP16 - CPU openvino: Person Re-Identification Retail FP16 - CPU openvino: Age Gender Recognition Retail 0013 FP16 - CPU openvino: Age Gender Recognition Retail 0013 FP16 - CPU openvino: Handwritten English Recognition FP16-INT8 - CPU openvino: Handwritten English Recognition FP16-INT8 - CPU openvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPU openvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPU rocksdb: Overwrite rocksdb: Rand Read rocksdb: Update Rand rocksdb: Read Rand Write Rand speedb: Rand Fill speedb: Rand Read speedb: Update Rand speedb: Seq Fill speedb: Rand Fill Sync speedb: Read Rand Write Rand onnx: GPT-2 - CPU - Parallel onnx: GPT-2 - CPU - Parallel onnx: yolov4 - CPU - Parallel onnx: yolov4 - CPU - Parallel onnx: yolov4 - CPU - Standard onnx: yolov4 - CPU - Standard onnx: T5 Encoder - CPU - Parallel onnx: T5 Encoder - CPU - Parallel onnx: T5 Encoder - CPU - Standard onnx: T5 Encoder - CPU - Standard onnx: bertsquad-12 - CPU - Parallel onnx: bertsquad-12 - CPU - Parallel onnx: bertsquad-12 - CPU - Standard onnx: bertsquad-12 - CPU - Standard onnx: CaffeNet 12-int8 - CPU - Parallel onnx: CaffeNet 12-int8 - CPU - Parallel onnx: CaffeNet 12-int8 - CPU - Standard onnx: CaffeNet 12-int8 - CPU - Standard onnx: fcn-resnet101-11 - CPU - Parallel onnx: fcn-resnet101-11 - CPU - Parallel onnx: fcn-resnet101-11 - CPU - Standard onnx: fcn-resnet101-11 - CPU - Standard onnx: ArcFace ResNet-100 - CPU - Parallel onnx: ArcFace ResNet-100 - CPU - Parallel onnx: ArcFace ResNet-100 - CPU - Standard onnx: ArcFace ResNet-100 - CPU - Standard onnx: ResNet50 v1-12-int8 - CPU - Parallel onnx: ResNet50 v1-12-int8 - CPU - Parallel onnx: ResNet50 v1-12-int8 - CPU - Standard onnx: ResNet50 v1-12-int8 - CPU - Standard onnx: super-resolution-10 - CPU - Parallel onnx: super-resolution-10 - CPU - Parallel onnx: super-resolution-10 - CPU - Standard onnx: super-resolution-10 - CPU - Standard onnx: Faster R-CNN R-50-FPN-int8 - CPU - Parallel onnx: Faster R-CNN R-50-FPN-int8 - CPU - Parallel onnx: Faster R-CNN R-50-FPN-int8 - CPU - Standard onnx: Faster R-CNN R-50-FPN-int8 - CPU - Standard encode-wavpack: WAV To WavPack llama-cpp: llama-2-13b.Q4_0.gguf llama-cpp: llama-2-70b-chat.Q5_0.gguf llamafile: mistral-7b-instruct-v0.2.Q8_0 - CPU llamafile: wizardcoder-python-34b-v1.0.Q6_K - CPU brl-cad: VGR Performance Metric v-ray: CPU Linux 6.8 Linux 6.9-rc2 6814667 6677333 181.51 112.43 11.50 22.31 42.76 42.74 42.94 44.80 32860.8 32361.2 36096.6 9592.3 32856.3 33002.2 13344.699620 93036.302415 99112.699260 602.69 3601.0 98.43 3364.8 32.61 3537.0 32.477 30.969 34.604 32.562 25.666 26.002 386.232 5.90 0.64 0.34 8.35 0.07 17006.8 36.6011 38.8350 35.4033 39.8177 31.5957 36.4359 5.865 48.943 127.202 123.246 18.662 4.902 9.847 15.913 33.986 2.28 2.32 1.06 29.7336 29.2815 102.9501 10.70212 10.50412 24.2805 61999717 17.245 38.462 307.502 19.511 3.356 39.619 13.416 6.483 4.50055 4.79853 9.00621 18.3377 6.16576 4821.32 748.495 1801 1861 2108 34412 61363 35317 63514 40657 72172 466 501 569 6909 13867 7402 14836 8732 17204 14.679 13.112 24.748 84678420667 29940735740 38204.6 1358616.4 12.085 86 34 17.23 213.35 346.10 495.72 6.05 1.88 723.16 58.50 24.12 96.31 35.30 135.18 47.03 202.37 70.10 6129 7747 17071694.52 6169.36 482.62 30377557.97 10389186.02 9670.37 688.26 28490133.28 143967.24 163551396.40 39599.26 810.19 3468957.59 1798203.12 202073.08 154589461.19 384748.54 29600.02 31442.69 21085.10 576905.92 70407.96 47927.75 5874353.63 20343491.52 204365411.59 234669.74 80643523.24 1866.58 17859304.38 21.82 45.60 53.98 38.09 271.99 76.20 119.92 307.74 451.13 81.90 445.90 82.87 2530.04 14.58 331.23 337.21 11260.98 9.92 1018.47 36.28 5067.40 22.06 16147.57 6.85 15532.54 7.17 1446.42 77.34 671.66 54.98 26732.88 4.05 6336.78 17.52 4621.32 23.68 3626.41 30.85 11638.51 9.59 67789.25 0.78 3777.56 29.61 121702.12 0.38 388010 360498518 359825 2120844 326499 380419389 292180 327637 257446 1181758 101.901 9.80514 7.91321 126.384 8.22351 122.236 185.294 5.41191 353.573 2.89501 10.7817 92.7467 14.2656 70.9566 292.173 3.42357 541.827 1.84807 1.35994 747.084 4.43649 231.284 16.7832 59.6021 23.4613 42.7819 94.2487 10.6098 111.5706 9.01538 152.635 6.55011 205.414 4.88181 23.4276 42.6833 32.4503 30.8209 6.349 1.48 0.36 4.50 0.79 3906839 59852 6916667 6689667 182.12 113.99 11.51 22.24 42.77 40.20 42.88 44.24 33072.5 33015.5 36180.7 9602.7 32826.1 32993.8 13341.467325 93862.763311 98711.076947 602.54 3628.0 98.41 3349.6 32.66 3492.0 33.045 30.128 34.187 33.888 25.607 25.960 381.657 5.65 0.61 0.33 8.45 0.07 16877.6 35.0788 35.7348 35.6507 36.5030 32.8297 37.4335 5.760 48.568 126.187 123.098 18.934 4.936 9.524 16.152 33.094 2.23 2.22 1.04 28.8754 29.0693 98.8690 10.5090 10.8236 25.1712 64443099 17.283 37.698 311.745 19.888 3.299 39.461 13.335 6.202 4.51166 4.84245 8.81043 18.0275 5.78202 10195.75 753.364 1774 1839 2164 34953 61162 35422 63529 39794 73367 478 489 570 6804 13807 7506 15415 8552 16687 14.298 13.357 27.994 81678284176 29665499053 38147.4 1396904.0 11.764 86 36 18.28 210.71 341.97 471.88 5.98 1.90 711.74 58.83 24.80 97.53 34.86 128.71 47.96 203.63 70.53 6054 7762 17199190.09 6427.33 487.60 29814480.67 10232390.41 9446.07 688.76 28934422.16 147978.33 161652700.09 35919.64 808.99 3346567.80 1900682.08 200671.70 150005776.47 392750.09 29139.60 31460.42 20605.39 569593.00 70036.20 46123.32 5908316.03 19390362.90 200468629.32 230503.34 85035507.55 1858.70 18740350.56 22.02 46.56 53.82 39.37 279.74 75.56 121.93 302.77 466.45 79.23 467.79 79.00 2656.42 13.89 330.73 337.78 11295.92 9.89 1051.30 35.14 5052.71 22.12 16349.05 6.76 15589.59 7.14 1442.21 77.57 673.46 54.83 27790.24 3.89 6306.42 17.58 4600.95 23.97 3608.21 31.01 11564.78 9.65 73754.74 0.71 3704.38 30.20 119735.57 0.38 391968 361586083 361502 2068588 322816 375837479 293414 330167 259603 1175562 105.7827 9.46496 8.04142 124.376 8.90936 112.428 196.076 5.10755 402.669 2.48234 10.6006 94.3330 15.4110 65.9960 293.029 3.40972 571.190 1.74973 1.50300 668.699 5.41259 184.785 17.5623 56.9478 24.2328 41.5048 95.1441 10.5119 122.601 8.23121 153.564 6.51127 213.737 4.67811 23.1727 43.1511 34.3512 29.1107 6.340 1.53 0.37 4.58 0.77 3991802 58995 OpenBenchmarking.org
Quicksilver Input: CORAL2 P1 OpenBenchmarking.org Figure Of Merit, More Is Better Quicksilver 20230818 Input: CORAL2 P1 Linux 6.8 Linux 6.9-rc2 1.5M 3M 4.5M 6M 7.5M SE +/- 67589.78, N = 6 SE +/- 46491.34, N = 3 6814667 6916667 1. (CXX) g++ options: -fopenmp -O3 -march=native
Quicksilver Input: CORAL2 P2 OpenBenchmarking.org Figure Of Merit, More Is Better Quicksilver 20230818 Input: CORAL2 P2 Linux 6.8 Linux 6.9-rc2 1.4M 2.8M 4.2M 5.6M 7M SE +/- 74194.56, N = 9 SE +/- 91964.82, N = 9 6677333 6689667 1. (CXX) g++ options: -fopenmp -O3 -march=native
FFmpeg Encoder: libx264 - Scenario: Live OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx264 - Scenario: Live Linux 6.8 Linux 6.9-rc2 40 80 120 160 200 SE +/- 1.29, N = 3 SE +/- 0.70, N = 3 181.51 182.12 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
FFmpeg Encoder: libx265 - Scenario: Live OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx265 - Scenario: Live Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 1.36, N = 15 SE +/- 1.80, N = 15 112.43 113.99 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
FFmpeg Encoder: libx264 - Scenario: Upload OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx264 - Scenario: Upload Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 11.50 11.51 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
FFmpeg Encoder: libx265 - Scenario: Upload OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx265 - Scenario: Upload Linux 6.8 Linux 6.9-rc2 5 10 15 20 25 SE +/- 0.31, N = 3 SE +/- 0.26, N = 4 22.31 22.24 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
FFmpeg Encoder: libx264 - Scenario: Platform OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx264 - Scenario: Platform Linux 6.8 Linux 6.9-rc2 10 20 30 40 50 SE +/- 0.17, N = 3 SE +/- 0.10, N = 3 42.76 42.77 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
FFmpeg Encoder: libx265 - Scenario: Platform OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx265 - Scenario: Platform Linux 6.8 Linux 6.9-rc2 10 20 30 40 50 SE +/- 0.71, N = 12 SE +/- 0.46, N = 12 42.74 40.20 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
FFmpeg Encoder: libx264 - Scenario: Video On Demand OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx264 - Scenario: Video On Demand Linux 6.8 Linux 6.9-rc2 10 20 30 40 50 SE +/- 0.06, N = 3 SE +/- 0.10, N = 3 42.94 42.88 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
FFmpeg Encoder: libx265 - Scenario: Video On Demand OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx265 - Scenario: Video On Demand Linux 6.8 Linux 6.9-rc2 10 20 30 40 50 SE +/- 0.50, N = 12 SE +/- 0.76, N = 9 44.80 44.24 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
Xmrig Variant: KawPow - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: KawPow - Hash Count: 1M Linux 6.8 Linux 6.9-rc2 7K 14K 21K 28K 35K SE +/- 68.36, N = 3 SE +/- 137.96, N = 3 32860.8 33072.5 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
Xmrig Variant: Monero - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: Monero - Hash Count: 1M Linux 6.8 Linux 6.9-rc2 7K 14K 21K 28K 35K SE +/- 320.50, N = 13 SE +/- 181.84, N = 3 32361.2 33015.5 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
Xmrig Variant: Wownero - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: Wownero - Hash Count: 1M Linux 6.8 Linux 6.9-rc2 8K 16K 24K 32K 40K SE +/- 245.88, N = 3 SE +/- 490.33, N = 3 36096.6 36180.7 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
Xmrig Variant: GhostRider - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: GhostRider - Hash Count: 1M Linux 6.8 Linux 6.9-rc2 2K 4K 6K 8K 10K SE +/- 31.05, N = 3 SE +/- 42.79, N = 3 9592.3 9602.7 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
Xmrig Variant: CryptoNight-Heavy - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: CryptoNight-Heavy - Hash Count: 1M Linux 6.8 Linux 6.9-rc2 7K 14K 21K 28K 35K SE +/- 49.86, N = 3 SE +/- 160.32, N = 3 32856.3 32826.1 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
Xmrig Variant: CryptoNight-Femto UPX2 - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: CryptoNight-Femto UPX2 - Hash Count: 1M Linux 6.8 Linux 6.9-rc2 7K 14K 21K 28K 35K SE +/- 110.64, N = 3 SE +/- 170.37, N = 3 33002.2 32993.8 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
CacheBench Test: Read OpenBenchmarking.org MB/s, More Is Better CacheBench Test: Read Linux 6.8 Linux 6.9-rc2 3K 6K 9K 12K 15K SE +/- 0.52, N = 3 SE +/- 0.88, N = 3 13344.70 13341.47 MIN: 13338.06 / MAX: 13346.23 MIN: 13335.31 / MAX: 13343.85 1. (CC) gcc options: -O3 -lrt
CacheBench Test: Write OpenBenchmarking.org MB/s, More Is Better CacheBench Test: Write Linux 6.8 Linux 6.9-rc2 20K 40K 60K 80K 100K SE +/- 36.95, N = 3 SE +/- 65.10, N = 3 93036.30 93862.76 MIN: 54037.87 / MAX: 103869.49 MIN: 54012.69 / MAX: 104121.12 1. (CC) gcc options: -O3 -lrt
CacheBench Test: Read / Modify / Write OpenBenchmarking.org MB/s, More Is Better CacheBench Test: Read / Modify / Write Linux 6.8 Linux 6.9-rc2 20K 40K 60K 80K 100K SE +/- 2.71, N = 3 SE +/- 71.07, N = 3 99112.70 98711.08 MIN: 88982.6 / MAX: 105154.55 MIN: 85707.72 / MAX: 105115.44 1. (CC) gcc options: -O3 -lrt
LZ4 Compression Compression Level: 1 - Compression Speed OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.4 Compression Level: 1 - Compression Speed Linux 6.8 Linux 6.9-rc2 130 260 390 520 650 SE +/- 1.15, N = 3 SE +/- 0.16, N = 3 602.69 602.54 1. (CC) gcc options: -O3
LZ4 Compression Compression Level: 1 - Decompression Speed OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.4 Compression Level: 1 - Decompression Speed Linux 6.8 Linux 6.9-rc2 800 1600 2400 3200 4000 SE +/- 44.31, N = 3 SE +/- 18.57, N = 3 3601.0 3628.0 1. (CC) gcc options: -O3
LZ4 Compression Compression Level: 3 - Compression Speed OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.4 Compression Level: 3 - Compression Speed Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.06, N = 3 SE +/- 0.06, N = 3 98.43 98.41 1. (CC) gcc options: -O3
LZ4 Compression Compression Level: 3 - Decompression Speed OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.4 Compression Level: 3 - Decompression Speed Linux 6.8 Linux 6.9-rc2 700 1400 2100 2800 3500 SE +/- 33.92, N = 3 SE +/- 34.17, N = 3 3364.8 3349.6 1. (CC) gcc options: -O3
LZ4 Compression Compression Level: 9 - Compression Speed OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.4 Compression Level: 9 - Compression Speed Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.08, N = 3 SE +/- 0.07, N = 3 32.61 32.66 1. (CC) gcc options: -O3
LZ4 Compression Compression Level: 9 - Decompression Speed OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.4 Compression Level: 9 - Decompression Speed Linux 6.8 Linux 6.9-rc2 800 1600 2400 3200 4000 SE +/- 16.01, N = 3 SE +/- 57.81, N = 3 3537.0 3492.0 1. (CC) gcc options: -O3
JPEG-XL libjxl Input: PNG - Quality: 80 OpenBenchmarking.org MP/s, More Is Better JPEG-XL libjxl 0.10.1 Input: PNG - Quality: 80 Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.38, N = 3 SE +/- 0.36, N = 15 32.48 33.05 1. (CXX) g++ options: -fno-rtti -O3 -fPIE -pie -lm
JPEG-XL libjxl Input: PNG - Quality: 90 OpenBenchmarking.org MP/s, More Is Better JPEG-XL libjxl 0.10.1 Input: PNG - Quality: 90 Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.28, N = 3 SE +/- 0.18, N = 3 30.97 30.13 1. (CXX) g++ options: -fno-rtti -O3 -fPIE -pie -lm
JPEG-XL libjxl Input: JPEG - Quality: 80 OpenBenchmarking.org MP/s, More Is Better JPEG-XL libjxl 0.10.1 Input: JPEG - Quality: 80 Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.51, N = 12 SE +/- 0.53, N = 15 34.60 34.19 1. (CXX) g++ options: -fno-rtti -O3 -fPIE -pie -lm
JPEG-XL libjxl Input: JPEG - Quality: 90 OpenBenchmarking.org MP/s, More Is Better JPEG-XL libjxl 0.10.1 Input: JPEG - Quality: 90 Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.37, N = 3 SE +/- 0.38, N = 15 32.56 33.89 1. (CXX) g++ options: -fno-rtti -O3 -fPIE -pie -lm
JPEG-XL libjxl Input: PNG - Quality: 100 OpenBenchmarking.org MP/s, More Is Better JPEG-XL libjxl 0.10.1 Input: PNG - Quality: 100 Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.15, N = 3 SE +/- 0.21, N = 3 25.67 25.61 1. (CXX) g++ options: -fno-rtti -O3 -fPIE -pie -lm
JPEG-XL libjxl Input: JPEG - Quality: 100 OpenBenchmarking.org MP/s, More Is Better JPEG-XL libjxl 0.10.1 Input: JPEG - Quality: 100 Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.11, N = 3 SE +/- 0.08, N = 3 26.00 25.96 1. (CXX) g++ options: -fno-rtti -O3 -fPIE -pie -lm
JPEG-XL Decoding libjxl CPU Threads: All OpenBenchmarking.org MP/s, More Is Better JPEG-XL Decoding libjxl 0.10.1 CPU Threads: All Linux 6.8 Linux 6.9-rc2 80 160 240 320 400 SE +/- 4.23, N = 3 SE +/- 2.26, N = 3 386.23 381.66
WebP2 Image Encode Encode Settings: Default OpenBenchmarking.org MP/s, More Is Better WebP2 Image Encode 20220823 Encode Settings: Default Linux 6.8 Linux 6.9-rc2 1.3275 2.655 3.9825 5.31 6.6375 SE +/- 0.11, N = 15 SE +/- 0.09, N = 15 5.90 5.65 1. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl
WebP2 Image Encode Encode Settings: Quality 75, Compression Effort 7 OpenBenchmarking.org MP/s, More Is Better WebP2 Image Encode 20220823 Encode Settings: Quality 75, Compression Effort 7 Linux 6.8 Linux 6.9-rc2 0.144 0.288 0.432 0.576 0.72 SE +/- 0.01, N = 3 SE +/- 0.01, N = 4 0.64 0.61 1. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl
WebP2 Image Encode Encode Settings: Quality 95, Compression Effort 7 OpenBenchmarking.org MP/s, More Is Better WebP2 Image Encode 20220823 Encode Settings: Quality 95, Compression Effort 7 Linux 6.8 Linux 6.9-rc2 0.0765 0.153 0.2295 0.306 0.3825 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 0.34 0.33 1. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl
WebP2 Image Encode Encode Settings: Quality 100, Compression Effort 5 OpenBenchmarking.org MP/s, More Is Better WebP2 Image Encode 20220823 Encode Settings: Quality 100, Compression Effort 5 Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.11, N = 15 SE +/- 0.07, N = 3 8.35 8.45 1. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl
WebP2 Image Encode Encode Settings: Quality 100, Lossless Compression OpenBenchmarking.org MP/s, More Is Better WebP2 Image Encode 20220823 Encode Settings: Quality 100, Lossless Compression Linux 6.8 Linux 6.9-rc2 0.0158 0.0316 0.0474 0.0632 0.079 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 0.07 0.07 1. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl
srsRAN Project Test: PDSCH Processor Benchmark, Throughput Total OpenBenchmarking.org Mbps, More Is Better srsRAN Project 23.10.1-20240219 Test: PDSCH Processor Benchmark, Throughput Total Linux 6.8 Linux 6.9-rc2 4K 8K 12K 16K 20K SE +/- 263.34, N = 15 SE +/- 377.19, N = 12 17006.8 16877.6 1. (CXX) g++ options: -march=native -mavx2 -mavx -msse4.1 -mfma -mavx512f -mavx512cd -mavx512bw -mavx512dq -O3 -fno-trapping-math -fno-math-errno -ldl
Embree Binary: Pathtracer - Model: Crown OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer - Model: Crown Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.59, N = 15 SE +/- 1.07, N = 15 36.60 35.08 MIN: 25.71 / MAX: 54.49 MIN: 22.8 / MAX: 59.06
Embree Binary: Pathtracer ISPC - Model: Crown OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Crown Linux 6.8 Linux 6.9-rc2 9 18 27 36 45 SE +/- 0.47, N = 4 SE +/- 1.44, N = 12 38.84 35.73 MIN: 30.82 / MAX: 52.25 MIN: 20.32 / MAX: 57.51
Embree Binary: Pathtracer - Model: Asian Dragon OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer - Model: Asian Dragon Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.86, N = 15 SE +/- 0.54, N = 15 35.40 35.65 MIN: 23.5 / MAX: 52.61 MIN: 24.54 / MAX: 51.88
Embree Binary: Pathtracer - Model: Asian Dragon Obj OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer - Model: Asian Dragon Obj Linux 6.8 Linux 6.9-rc2 9 18 27 36 45 SE +/- 0.92, N = 15 SE +/- 0.60, N = 15 39.82 36.50 MIN: 26.64 / MAX: 50.91 MIN: 22.66 / MAX: 46.51
Embree Binary: Pathtracer ISPC - Model: Asian Dragon OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Asian Dragon Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.77, N = 13 SE +/- 0.82, N = 15 31.60 32.83 MIN: 20.8 / MAX: 52.83 MIN: 22.56 / MAX: 48.16
Embree Binary: Pathtracer ISPC - Model: Asian Dragon Obj OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Asian Dragon Obj Linux 6.8 Linux 6.9-rc2 9 18 27 36 45 SE +/- 0.94, N = 15 SE +/- 0.88, N = 15 36.44 37.43 MIN: 22.56 / MAX: 49.05 MIN: 27.23 / MAX: 49.01
SVT-AV1 Encoder Mode: Preset 4 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 2.0 Encoder Mode: Preset 4 - Input: Bosphorus 4K Linux 6.8 Linux 6.9-rc2 1.3196 2.6392 3.9588 5.2784 6.598 SE +/- 0.035, N = 3 SE +/- 0.051, N = 15 5.865 5.760 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 8 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 2.0 Encoder Mode: Preset 8 - Input: Bosphorus 4K Linux 6.8 Linux 6.9-rc2 11 22 33 44 55 SE +/- 1.10, N = 12 SE +/- 1.18, N = 15 48.94 48.57 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 12 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 2.0 Encoder Mode: Preset 12 - Input: Bosphorus 4K Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 5.13, N = 15 SE +/- 5.21, N = 15 127.20 126.19 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 13 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 2.0 Encoder Mode: Preset 13 - Input: Bosphorus 4K Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 3.82, N = 12 SE +/- 5.42, N = 15 123.25 123.10 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 4 - Input: Bosphorus 1080p OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 2.0 Encoder Mode: Preset 4 - Input: Bosphorus 1080p Linux 6.8 Linux 6.9-rc2 5 10 15 20 25 SE +/- 0.13, N = 15 SE +/- 0.21, N = 3 18.66 18.93 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
VVenC Video Input: Bosphorus 4K - Video Preset: Fast OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.11 Video Input: Bosphorus 4K - Video Preset: Fast Linux 6.8 Linux 6.9-rc2 1.1106 2.2212 3.3318 4.4424 5.553 SE +/- 0.075, N = 10 SE +/- 0.077, N = 12 4.902 4.936 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
VVenC Video Input: Bosphorus 4K - Video Preset: Faster OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.11 Video Input: Bosphorus 4K - Video Preset: Faster Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.200, N = 12 SE +/- 0.172, N = 15 9.847 9.524 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
VVenC Video Input: Bosphorus 1080p - Video Preset: Fast OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.11 Video Input: Bosphorus 1080p - Video Preset: Fast Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.11, N = 3 SE +/- 0.07, N = 3 15.91 16.15 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
VVenC Video Input: Bosphorus 1080p - Video Preset: Faster OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.11 Video Input: Bosphorus 1080p - Video Preset: Faster Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.29, N = 15 SE +/- 0.36, N = 15 33.99 33.09 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
Intel Open Image Denoise Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.2 Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Only Linux 6.8 Linux 6.9-rc2 0.513 1.026 1.539 2.052 2.565 SE +/- 0.02, N = 15 SE +/- 0.02, N = 15 2.28 2.23
Intel Open Image Denoise Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.2 Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Only Linux 6.8 Linux 6.9-rc2 0.522 1.044 1.566 2.088 2.61 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 2.32 2.22
Intel Open Image Denoise Run: RTLightmap.hdr.4096x4096 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.2 Run: RTLightmap.hdr.4096x4096 - Device: CPU-Only Linux 6.8 Linux 6.9-rc2 0.2385 0.477 0.7155 0.954 1.1925 SE +/- 0.01, N = 7 SE +/- 0.01, N = 15 1.06 1.04
OSPRay Benchmark: particle_volume/ao/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: particle_volume/ao/real_time Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.11, N = 3 SE +/- 0.19, N = 3 29.73 28.88
OSPRay Benchmark: particle_volume/scivis/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: particle_volume/scivis/real_time Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.26, N = 3 SE +/- 0.33, N = 3 29.28 29.07
OSPRay Benchmark: particle_volume/pathtracer/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: particle_volume/pathtracer/real_time Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 1.85, N = 12 SE +/- 1.98, N = 9 102.95 98.87
OSPRay Benchmark: gravity_spheres_volume/dim_512/ao/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: gravity_spheres_volume/dim_512/ao/real_time Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.14, N = 15 SE +/- 0.05, N = 3 10.70 10.51
OSPRay Benchmark: gravity_spheres_volume/dim_512/scivis/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: gravity_spheres_volume/dim_512/scivis/real_time Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.08, N = 15 SE +/- 0.07, N = 3 10.50 10.82
OSPRay Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_time Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.41, N = 12 SE +/- 0.20, N = 3 24.28 25.17
Stockfish Chess Benchmark OpenBenchmarking.org Nodes Per Second, More Is Better Stockfish 16.1 Chess Benchmark Linux 6.8 Linux 6.9-rc2 14M 28M 42M 56M 70M SE +/- 2993041.97, N = 9 SE +/- 2705883.43, N = 9 61999717 64443099 1. (CXX) g++ options: -lgcov -m64 -lpthread -fno-exceptions -std=c++17 -fno-peel-loops -fno-tracer -pedantic -O3 -funroll-loops -msse -msse3 -mpopcnt -mavx2 -mbmi -mavx512f -mavx512bw -mavx512vnni -mavx512dq -mavx512vl -msse4.1 -mssse3 -msse2 -mbmi2 -flto -flto-partition=one -flto=jobserver
Timed FFmpeg Compilation Time To Compile OpenBenchmarking.org Seconds, Fewer Is Better Timed FFmpeg Compilation 6.1 Time To Compile Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.39, N = 15 SE +/- 0.39, N = 15 17.25 17.28
Timed Linux Kernel Compilation Build: defconfig OpenBenchmarking.org Seconds, Fewer Is Better Timed Linux Kernel Compilation 6.8 Build: defconfig Linux 6.8 Linux 6.9-rc2 9 18 27 36 45 SE +/- 0.36, N = 6 SE +/- 0.37, N = 6 38.46 37.70
Timed Linux Kernel Compilation Build: allmodconfig OpenBenchmarking.org Seconds, Fewer Is Better Timed Linux Kernel Compilation 6.8 Build: allmodconfig Linux 6.8 Linux 6.9-rc2 70 140 210 280 350 SE +/- 3.84, N = 3 SE +/- 1.26, N = 3 307.50 311.75
Timed Mesa Compilation Time To Compile OpenBenchmarking.org Seconds, Fewer Is Better Timed Mesa Compilation 24.0 Time To Compile Linux 6.8 Linux 6.9-rc2 5 10 15 20 25 SE +/- 0.51, N = 15 SE +/- 0.39, N = 15 19.51 19.89
Primesieve Length: 1e12 OpenBenchmarking.org Seconds, Fewer Is Better Primesieve 12.1 Length: 1e12 Linux 6.8 Linux 6.9-rc2 0.7551 1.5102 2.2653 3.0204 3.7755 SE +/- 0.024, N = 11 SE +/- 0.044, N = 3 3.356 3.299 1. (CXX) g++ options: -O3
Primesieve Length: 1e13 OpenBenchmarking.org Seconds, Fewer Is Better Primesieve 12.1 Length: 1e13 Linux 6.8 Linux 6.9-rc2 9 18 27 36 45 SE +/- 0.06, N = 3 SE +/- 0.09, N = 3 39.62 39.46 1. (CXX) g++ options: -O3
Y-Cruncher Pi Digits To Calculate: 1B OpenBenchmarking.org Seconds, Fewer Is Better Y-Cruncher 0.8.3 Pi Digits To Calculate: 1B Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.30, N = 15 SE +/- 0.24, N = 15 13.42 13.34
Y-Cruncher Pi Digits To Calculate: 500M OpenBenchmarking.org Seconds, Fewer Is Better Y-Cruncher 0.8.3 Pi Digits To Calculate: 500M Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.143, N = 12 SE +/- 0.083, N = 15 6.483 6.202
oneDNN Harness: IP Shapes 1D - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: IP Shapes 1D - Engine: CPU Linux 6.8 Linux 6.9-rc2 1.0151 2.0302 3.0453 4.0604 5.0755 SE +/- 0.04580, N = 3 SE +/- 0.05180, N = 12 4.50055 4.51166 MIN: 3.52 MIN: 3.37 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
oneDNN Harness: IP Shapes 3D - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: IP Shapes 3D - Engine: CPU Linux 6.8 Linux 6.9-rc2 1.0896 2.1792 3.2688 4.3584 5.448 SE +/- 0.02352, N = 3 SE +/- 0.02357, N = 3 4.79853 4.84245 MIN: 3.98 MIN: 4.01 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
oneDNN Harness: Convolution Batch Shapes Auto - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: Convolution Batch Shapes Auto - Engine: CPU Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.03328, N = 3 SE +/- 0.09863, N = 15 9.00621 8.81043 MIN: 8.12 MIN: 6.25 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
oneDNN Harness: Deconvolution Batch shapes_1d - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: Deconvolution Batch shapes_1d - Engine: CPU Linux 6.8 Linux 6.9-rc2 5 10 15 20 25 SE +/- 0.23, N = 15 SE +/- 0.14, N = 3 18.34 18.03 MIN: 10.09 MIN: 13.29 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
oneDNN Harness: Deconvolution Batch shapes_3d - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: Deconvolution Batch shapes_3d - Engine: CPU Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.10591, N = 15 SE +/- 0.04472, N = 3 6.16576 5.78202 MIN: 3.75 MIN: 3.81 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
oneDNN Harness: Recurrent Neural Network Training - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: Recurrent Neural Network Training - Engine: CPU Linux 6.8 Linux 6.9-rc2 2K 4K 6K 8K 10K SE +/- 22.69, N = 3 SE +/- 498.89, N = 13 4821.32 10195.75 MIN: 4659.99 MIN: 4568.3 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
oneDNN Harness: Recurrent Neural Network Inference - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: Recurrent Neural Network Inference - Engine: CPU Linux 6.8 Linux 6.9-rc2 160 320 480 640 800 SE +/- 7.46, N = 3 SE +/- 6.16, N = 9 748.50 753.36 MIN: 677.68 MIN: 652.83 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
OSPRay Studio Camera: 1 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 400 800 1200 1600 2000 SE +/- 40.26, N = 15 SE +/- 22.91, N = 15 1801 1774
OSPRay Studio Camera: 2 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 400 800 1200 1600 2000 SE +/- 13.93, N = 11 SE +/- 26.21, N = 3 1861 1839
OSPRay Studio Camera: 3 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 500 1000 1500 2000 2500 SE +/- 21.99, N = 4 SE +/- 26.16, N = 13 2108 2164
OSPRay Studio Camera: 1 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 7K 14K 21K 28K 35K SE +/- 294.16, N = 3 SE +/- 454.99, N = 3 34412 34953
OSPRay Studio Camera: 1 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 13K 26K 39K 52K 65K SE +/- 443.96, N = 15 SE +/- 567.54, N = 7 61363 61162
OSPRay Studio Camera: 2 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 8K 16K 24K 32K 40K SE +/- 302.37, N = 15 SE +/- 483.46, N = 3 35317 35422
OSPRay Studio Camera: 2 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 14K 28K 42K 56K 70K SE +/- 591.53, N = 7 SE +/- 885.61, N = 3 63514 63529
OSPRay Studio Camera: 3 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 9K 18K 27K 36K 45K SE +/- 626.71, N = 12 SE +/- 252.89, N = 3 40657 39794
OSPRay Studio Camera: 3 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 16K 32K 48K 64K 80K SE +/- 506.90, N = 3 SE +/- 855.02, N = 15 72172 73367
OSPRay Studio Camera: 1 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 100 200 300 400 500 SE +/- 3.71, N = 3 SE +/- 2.33, N = 3 466 478
OSPRay Studio Camera: 2 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 110 220 330 440 550 SE +/- 5.26, N = 5 SE +/- 2.33, N = 3 501 489
OSPRay Studio Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 120 240 360 480 600 SE +/- 4.36, N = 15 SE +/- 4.85, N = 15 569 570
OSPRay Studio Camera: 1 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 1500 3000 4500 6000 7500 SE +/- 83.68, N = 3 SE +/- 15.17, N = 3 6909 6804
OSPRay Studio Camera: 1 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 3K 6K 9K 12K 15K SE +/- 112.81, N = 9 SE +/- 180.26, N = 3 13867 13807
OSPRay Studio Camera: 2 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 1600 3200 4800 6400 8000 SE +/- 105.66, N = 3 SE +/- 118.69, N = 15 7402 7506
OSPRay Studio Camera: 2 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 3K 6K 9K 12K 15K SE +/- 268.32, N = 12 SE +/- 235.77, N = 15 14836 15415
OSPRay Studio Camera: 3 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 2K 4K 6K 8K 10K SE +/- 70.21, N = 3 SE +/- 26.21, N = 3 8732 8552
OSPRay Studio Camera: 3 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 4K 8K 12K 16K 20K SE +/- 101.18, N = 3 SE +/- 156.23, N = 6 17204 16687
Hackbench Count: 16 - Type: Thread OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 16 - Type: Thread Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.00, N = 3 SE +/- 0.14, N = 15 14.68 14.30 1. (CC) gcc options: -lpthread
Hackbench Count: 16 - Type: Process OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 16 - Type: Process Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.11, N = 3 SE +/- 0.19, N = 3 13.11 13.36 1. (CC) gcc options: -lpthread
Hackbench Count: 32 - Type: Process OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 32 - Type: Process Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.27, N = 3 SE +/- 0.39, N = 3 24.75 27.99 1. (CC) gcc options: -lpthread
OpenSSL Algorithm: SHA256 OpenBenchmarking.org byte/s, More Is Better OpenSSL Algorithm: SHA256 Linux 6.8 Linux 6.9-rc2 20000M 40000M 60000M 80000M 100000M SE +/- 767626108.54, N = 3 SE +/- 717052715.91, N = 12 84678420667 81678284176 1. OpenSSL 3.0.10 1 Aug 2023 (Library: OpenSSL 3.0.10 1 Aug 2023)
OpenSSL Algorithm: SHA512 OpenBenchmarking.org byte/s, More Is Better OpenSSL Algorithm: SHA512 Linux 6.8 Linux 6.9-rc2 6000M 12000M 18000M 24000M 30000M SE +/- 92272355.83, N = 3 SE +/- 252323355.66, N = 3 29940735740 29665499053 1. OpenSSL 3.0.10 1 Aug 2023 (Library: OpenSSL 3.0.10 1 Aug 2023)
OpenSSL Algorithm: RSA4096 OpenBenchmarking.org sign/s, More Is Better OpenSSL Algorithm: RSA4096 Linux 6.8 Linux 6.9-rc2 8K 16K 24K 32K 40K SE +/- 46.27, N = 3 SE +/- 90.23, N = 3 38204.6 38147.4 1. OpenSSL 3.0.10 1 Aug 2023 (Library: OpenSSL 3.0.10 1 Aug 2023)
OpenSSL Algorithm: RSA4096 OpenBenchmarking.org verify/s, More Is Better OpenSSL Algorithm: RSA4096 Linux 6.8 Linux 6.9-rc2 300K 600K 900K 1200K 1500K SE +/- 19064.62, N = 3 SE +/- 833.27, N = 3 1358616.4 1396904.0 1. OpenSSL 3.0.10 1 Aug 2023 (Library: OpenSSL 3.0.10 1 Aug 2023)
GROMACS Implementation: MPI CPU - Input: water_GMX50_bare OpenBenchmarking.org Ns Per Day, More Is Better GROMACS 2024 Implementation: MPI CPU - Input: water_GMX50_bare Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.06, N = 3 SE +/- 0.08, N = 3 12.09 11.76 1. (CXX) g++ options: -O3 -lm
MariaDB mariadb-slap Clients: 512 OpenBenchmarking.org Queries Per Second, More Is Better MariaDB mariadb-slap 11.5 Clients: 512 Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.41, N = 3 SE +/- 0.87, N = 9 86 86 1. (CXX) g++ options: -fPIC -pie -fstack-protector -O3 -shared -lrt -lpthread -lz -ldl -lm -lstdc++
MariaDB mariadb-slap Clients: 1024 OpenBenchmarking.org Queries Per Second, More Is Better MariaDB mariadb-slap 11.5 Clients: 1024 Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 1.23, N = 9 SE +/- 1.41, N = 6 34 36 1. (CXX) g++ options: -fPIC -pie -fstack-protector -O3 -shared -lrt -lpthread -lz -ldl -lm -lstdc++
TensorFlow Device: CPU - Batch Size: 1 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 1 - Model: AlexNet Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.25, N = 12 SE +/- 0.25, N = 3 17.23 18.28
TensorFlow Device: CPU - Batch Size: 16 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 16 - Model: AlexNet Linux 6.8 Linux 6.9-rc2 50 100 150 200 250 SE +/- 1.58, N = 15 SE +/- 1.89, N = 15 213.35 210.71
TensorFlow Device: CPU - Batch Size: 32 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 32 - Model: AlexNet Linux 6.8 Linux 6.9-rc2 80 160 240 320 400 SE +/- 2.86, N = 15 SE +/- 5.16, N = 15 346.10 341.97
TensorFlow Device: CPU - Batch Size: 64 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 64 - Model: AlexNet Linux 6.8 Linux 6.9-rc2 110 220 330 440 550 SE +/- 6.05, N = 3 SE +/- 5.17, N = 15 495.72 471.88
TensorFlow Device: CPU - Batch Size: 1 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 1 - Model: GoogLeNet Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.13, N = 12 SE +/- 0.14, N = 15 6.05 5.98
TensorFlow Device: CPU - Batch Size: 1 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 1 - Model: ResNet-50 Linux 6.8 Linux 6.9-rc2 0.4275 0.855 1.2825 1.71 2.1375 SE +/- 0.03, N = 15 SE +/- 0.03, N = 15 1.88 1.90
TensorFlow Device: CPU - Batch Size: 256 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 256 - Model: AlexNet Linux 6.8 Linux 6.9-rc2 160 320 480 640 800 SE +/- 9.01, N = 3 SE +/- 6.62, N = 7 723.16 711.74
TensorFlow Device: CPU - Batch Size: 16 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 16 - Model: GoogLeNet Linux 6.8 Linux 6.9-rc2 13 26 39 52 65 SE +/- 1.31, N = 12 SE +/- 1.24, N = 12 58.50 58.83
TensorFlow Device: CPU - Batch Size: 16 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 16 - Model: ResNet-50 Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.30, N = 3 SE +/- 0.28, N = 15 24.12 24.80
TensorFlow Device: CPU - Batch Size: 32 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 32 - Model: GoogLeNet Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 1.27, N = 12 SE +/- 1.10, N = 15 96.31 97.53
TensorFlow Device: CPU - Batch Size: 32 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 32 - Model: ResNet-50 Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.48, N = 15 SE +/- 0.42, N = 15 35.30 34.86
TensorFlow Device: CPU - Batch Size: 64 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 64 - Model: GoogLeNet Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 0.94, N = 3 SE +/- 1.70, N = 15 135.18 128.71
TensorFlow Device: CPU - Batch Size: 64 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 64 - Model: ResNet-50 Linux 6.8 Linux 6.9-rc2 11 22 33 44 55 SE +/- 0.54, N = 12 SE +/- 0.42, N = 3 47.03 47.96
TensorFlow Device: CPU - Batch Size: 256 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 256 - Model: GoogLeNet Linux 6.8 Linux 6.9-rc2 40 80 120 160 200 SE +/- 2.20, N = 12 SE +/- 1.95, N = 3 202.37 203.63
TensorFlow Device: CPU - Batch Size: 256 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 256 - Model: ResNet-50 Linux 6.8 Linux 6.9-rc2 16 32 48 64 80 SE +/- 0.38, N = 3 SE +/- 0.47, N = 3 70.10 70.53
Google Draco Model: Lion OpenBenchmarking.org ms, Fewer Is Better Google Draco 1.5.6 Model: Lion Linux 6.8 Linux 6.9-rc2 1300 2600 3900 5200 6500 SE +/- 40.84, N = 3 SE +/- 29.28, N = 3 6129 6054 1. (CXX) g++ options: -O3
Google Draco Model: Church Facade OpenBenchmarking.org ms, Fewer Is Better Google Draco 1.5.6 Model: Church Facade Linux 6.8 Linux 6.9-rc2 1700 3400 5100 6800 8500 SE +/- 0.88, N = 3 SE +/- 9.56, N = 3 7747 7762 1. (CXX) g++ options: -O3
Stress-NG Test: Hash OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Hash Linux 6.8 Linux 6.9-rc2 4M 8M 12M 16M 20M SE +/- 82496.53, N = 3 SE +/- 192774.29, N = 3 17071694.52 17199190.09 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: MMAP OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: MMAP Linux 6.8 Linux 6.9-rc2 1400 2800 4200 5600 7000 SE +/- 70.88, N = 14 SE +/- 66.09, N = 15 6169.36 6427.33 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: NUMA OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: NUMA Linux 6.8 Linux 6.9-rc2 110 220 330 440 550 SE +/- 4.27, N = 3 SE +/- 5.89, N = 4 482.62 487.60 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Pipe OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Pipe Linux 6.8 Linux 6.9-rc2 7M 14M 21M 28M 35M SE +/- 749025.19, N = 15 SE +/- 2106194.07, N = 15 30377557.97 29814480.67 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Poll OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Poll Linux 6.8 Linux 6.9-rc2 2M 4M 6M 8M 10M SE +/- 37074.52, N = 3 SE +/- 30731.99, N = 3 10389186.02 10232390.41 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Zlib OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Zlib Linux 6.8 Linux 6.9-rc2 2K 4K 6K 8K 10K SE +/- 25.13, N = 3 SE +/- 124.93, N = 3 9670.37 9446.07 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: MEMFD OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: MEMFD Linux 6.8 Linux 6.9-rc2 150 300 450 600 750 SE +/- 14.68, N = 15 SE +/- 5.94, N = 8 688.26 688.76 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Mutex OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Mutex Linux 6.8 Linux 6.9-rc2 6M 12M 18M 24M 30M SE +/- 281178.86, N = 3 SE +/- 294297.65, N = 3 28490133.28 28934422.16 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Crypto OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Crypto Linux 6.8 Linux 6.9-rc2 30K 60K 90K 120K 150K SE +/- 1720.03, N = 15 SE +/- 2039.96, N = 15 143967.24 147978.33 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Malloc OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Malloc Linux 6.8 Linux 6.9-rc2 40M 80M 120M 160M 200M SE +/- 201545.00, N = 3 SE +/- 3675496.40, N = 12 163551396.40 161652700.09 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Pthread OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Pthread Linux 6.8 Linux 6.9-rc2 8K 16K 24K 32K 40K SE +/- 49.70, N = 3 SE +/- 450.35, N = 15 39599.26 35919.64 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: AVL Tree OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: AVL Tree Linux 6.8 Linux 6.9-rc2 200 400 600 800 1000 SE +/- 6.53, N = 3 SE +/- 2.13, N = 3 810.19 808.99 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: IO_uring OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: IO_uring Linux 6.8 Linux 6.9-rc2 700K 1400K 2100K 2800K 3500K SE +/- 31595.27, N = 3 SE +/- 43339.28, N = 15 3468957.59 3346567.80 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: SENDFILE OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: SENDFILE Linux 6.8 Linux 6.9-rc2 400K 800K 1200K 1600K 2000K SE +/- 609.53, N = 3 SE +/- 5873.31, N = 3 1798203.12 1900682.08 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: CPU Stress OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: CPU Stress Linux 6.8 Linux 6.9-rc2 40K 80K 120K 160K 200K SE +/- 569.84, N = 3 SE +/- 335.84, N = 3 202073.08 200671.70 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Semaphores OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Semaphores Linux 6.8 Linux 6.9-rc2 30M 60M 90M 120M 150M SE +/- 1498916.25, N = 3 SE +/- 1330349.25, N = 3 154589461.19 150005776.47 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Matrix Math OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Matrix Math Linux 6.8 Linux 6.9-rc2 80K 160K 240K 320K 400K SE +/- 4925.11, N = 15 SE +/- 9278.38, N = 12 384748.54 392750.09 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Floating Point OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Floating Point Linux 6.8 Linux 6.9-rc2 6K 12K 18K 24K 30K SE +/- 632.70, N = 12 SE +/- 214.66, N = 11 29600.02 29139.60 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Matrix 3D Math OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Matrix 3D Math Linux 6.8 Linux 6.9-rc2 7K 14K 21K 28K 35K SE +/- 669.33, N = 15 SE +/- 437.41, N = 15 31442.69 31460.42 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Memory Copying OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Memory Copying Linux 6.8 Linux 6.9-rc2 5K 10K 15K 20K 25K SE +/- 178.25, N = 3 SE +/- 155.03, N = 10 21085.10 20605.39 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Vector Shuffle OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Vector Shuffle Linux 6.8 Linux 6.9-rc2 120K 240K 360K 480K 600K SE +/- 1416.85, N = 3 SE +/- 1897.90, N = 3 576905.92 569593.00 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Mixed Scheduler OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Mixed Scheduler Linux 6.8 Linux 6.9-rc2 15K 30K 45K 60K 75K SE +/- 153.71, N = 3 SE +/- 281.29, N = 3 70407.96 70036.20 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Socket Activity OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Socket Activity Linux 6.8 Linux 6.9-rc2 10K 20K 30K 40K 50K SE +/- 79.93, N = 3 SE +/- 21.40, N = 3 47927.75 46123.32 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Wide Vector Math OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Wide Vector Math Linux 6.8 Linux 6.9-rc2 1.3M 2.6M 3.9M 5.2M 6.5M SE +/- 72950.92, N = 4 SE +/- 55746.31, N = 15 5874353.63 5908316.03 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Context Switching OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Context Switching Linux 6.8 Linux 6.9-rc2 4M 8M 12M 16M 20M SE +/- 178479.09, N = 8 SE +/- 265846.30, N = 3 20343491.52 19390362.90 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Fused Multiply-Add OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Fused Multiply-Add Linux 6.8 Linux 6.9-rc2 40M 80M 120M 160M 200M SE +/- 6187723.80, N = 12 SE +/- 3824188.11, N = 15 204365411.59 200468629.32 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Vector Floating Point OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Vector Floating Point Linux 6.8 Linux 6.9-rc2 50K 100K 150K 200K 250K SE +/- 5616.14, N = 12 SE +/- 8579.53, N = 12 234669.74 230503.34 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Glibc C String Functions OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Glibc C String Functions Linux 6.8 Linux 6.9-rc2 20M 40M 60M 80M 100M SE +/- 2138024.53, N = 13 SE +/- 1504028.14, N = 15 80643523.24 85035507.55 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Glibc Qsort Data Sorting OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Glibc Qsort Data Sorting Linux 6.8 Linux 6.9-rc2 400 800 1200 1600 2000 SE +/- 4.40, N = 3 SE +/- 4.16, N = 3 1866.58 1858.70 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: System V Message Passing OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: System V Message Passing Linux 6.8 Linux 6.9-rc2 4M 8M 12M 16M 20M SE +/- 189741.33, N = 15 SE +/- 35075.39, N = 3 17859304.38 18740350.56 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Blender Blend File: BMW27 - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.1 Blend File: BMW27 - Compute: CPU-Only Linux 6.8 Linux 6.9-rc2 5 10 15 20 25 SE +/- 0.33, N = 14 SE +/- 0.08, N = 3 21.82 22.02
Blender Blend File: Junkshop - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.1 Blend File: Junkshop - Compute: CPU-Only Linux 6.8 Linux 6.9-rc2 11 22 33 44 55 SE +/- 0.47, N = 15 SE +/- 0.40, N = 15 45.60 46.56
Blender Blend File: Classroom - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.1 Blend File: Classroom - Compute: CPU-Only Linux 6.8 Linux 6.9-rc2 12 24 36 48 60 SE +/- 0.10, N = 3 SE +/- 0.09, N = 3 53.98 53.82
Blender Blend File: Fishy Cat - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.1 Blend File: Fishy Cat - Compute: CPU-Only Linux 6.8 Linux 6.9-rc2 9 18 27 36 45 SE +/- 0.38, N = 15 SE +/- 0.44, N = 4 38.09 39.37
Blender Blend File: Barbershop - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.1 Blend File: Barbershop - Compute: CPU-Only Linux 6.8 Linux 6.9-rc2 60 120 180 240 300 SE +/- 6.20, N = 8 SE +/- 6.67, N = 9 271.99 279.74
Blender Blend File: Pabellon Barcelona - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.1 Blend File: Pabellon Barcelona - Compute: CPU-Only Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.83, N = 5 SE +/- 0.61, N = 15 76.20 75.56
OpenVINO Model: Face Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Face Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 0.33, N = 3 SE +/- 0.55, N = 3 119.92 121.93 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Face Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Face Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 70 140 210 280 350 SE +/- 0.84, N = 3 SE +/- 1.30, N = 3 307.74 302.77 MIN: 183.58 / MAX: 728.14 MIN: 181.55 / MAX: 894.73 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Person Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Person Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 100 200 300 400 500 SE +/- 2.33, N = 3 SE +/- 1.20, N = 3 451.13 466.45 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Person Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Person Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.43, N = 3 SE +/- 0.20, N = 3 81.90 79.23 MIN: 49.43 / MAX: 345.35 MIN: 50.1 / MAX: 290.06 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Person Detection FP32 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Person Detection FP32 - Device: CPU Linux 6.8 Linux 6.9-rc2 100 200 300 400 500 SE +/- 3.09, N = 3 SE +/- 3.72, N = 3 445.90 467.79 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Person Detection FP32 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Person Detection FP32 - Device: CPU Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.57, N = 3 SE +/- 0.62, N = 3 82.87 79.00 MIN: 51.69 / MAX: 529.56 MIN: 49.24 / MAX: 319.95 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Vehicle Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Vehicle Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 600 1200 1800 2400 3000 SE +/- 16.72, N = 3 SE +/- 25.96, N = 3 2530.04 2656.42 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Vehicle Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Vehicle Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.10, N = 3 SE +/- 0.14, N = 3 14.58 13.89 MIN: 9.5 / MAX: 109.83 MIN: 9.12 / MAX: 90.57 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Face Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Face Detection FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 70 140 210 280 350 SE +/- 0.75, N = 3 SE +/- 0.82, N = 3 331.23 330.73 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Face Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Face Detection FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 70 140 210 280 350 SE +/- 0.74, N = 3 SE +/- 0.81, N = 3 337.21 337.78 MIN: 261.24 / MAX: 615.09 MIN: 253.21 / MAX: 781.06 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Face Detection Retail FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Face Detection Retail FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 2K 4K 6K 8K 10K SE +/- 23.83, N = 3 SE +/- 26.72, N = 3 11260.98 11295.92 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Face Detection Retail FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Face Detection Retail FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.02, N = 3 SE +/- 0.02, N = 3 9.92 9.89 MIN: 7.45 / MAX: 33.82 MIN: 7.63 / MAX: 36.57 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Road Segmentation ADAS FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Road Segmentation ADAS FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 200 400 600 800 1000 SE +/- 9.82, N = 3 SE +/- 6.69, N = 3 1018.47 1051.30 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Road Segmentation ADAS FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Road Segmentation ADAS FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.35, N = 3 SE +/- 0.23, N = 3 36.28 35.14 MIN: 24.78 / MAX: 205.05 MIN: 25.25 / MAX: 437.27 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Vehicle Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Vehicle Detection FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 1100 2200 3300 4400 5500 SE +/- 9.46, N = 3 SE +/- 18.34, N = 3 5067.40 5052.71 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Vehicle Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Vehicle Detection FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 5 10 15 20 25 SE +/- 0.04, N = 3 SE +/- 0.08, N = 3 22.06 22.12 MIN: 15.51 / MAX: 86.25 MIN: 15.73 / MAX: 71.5 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Weld Porosity Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Weld Porosity Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 4K 8K 12K 16K 20K SE +/- 194.69, N = 3 SE +/- 107.94, N = 3 16147.57 16349.05 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Weld Porosity Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Weld Porosity Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.08, N = 3 SE +/- 0.05, N = 3 6.85 6.76 MIN: 4.47 / MAX: 51.04 MIN: 4.36 / MAX: 70.91 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Face Detection Retail FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Face Detection Retail FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 3K 6K 9K 12K 15K SE +/- 132.85, N = 3 SE +/- 25.09, N = 3 15532.54 15589.59 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Face Detection Retail FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Face Detection Retail FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.07, N = 3 SE +/- 0.00, N = 3 7.17 7.14 MIN: 5.78 / MAX: 54.3 MIN: 5.79 / MAX: 44.66 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Road Segmentation ADAS FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Road Segmentation ADAS FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 300 600 900 1200 1500 SE +/- 1.59, N = 3 SE +/- 5.16, N = 3 1446.42 1442.21 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Road Segmentation ADAS FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Road Segmentation ADAS FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.09, N = 3 SE +/- 0.28, N = 3 77.34 77.57 MIN: 56.48 / MAX: 158.67 MIN: 57.16 / MAX: 148.67 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Machine Translation EN To DE FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Machine Translation EN To DE FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 150 300 450 600 750 SE +/- 5.75, N = 3 SE +/- 3.16, N = 3 671.66 673.46 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Machine Translation EN To DE FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Machine Translation EN To DE FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 12 24 36 48 60 SE +/- 0.47, N = 3 SE +/- 0.25, N = 3 54.98 54.83 MIN: 33.9 / MAX: 423.36 MIN: 34.71 / MAX: 375.04 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Weld Porosity Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Weld Porosity Detection FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 6K 12K 18K 24K 30K SE +/- 285.82, N = 4 SE +/- 155.42, N = 3 26732.88 27790.24 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Weld Porosity Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Weld Porosity Detection FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 0.9113 1.8226 2.7339 3.6452 4.5565 SE +/- 0.05, N = 4 SE +/- 0.01, N = 3 4.05 3.89 MIN: 2.67 / MAX: 52.43 MIN: 2.64 / MAX: 47.09 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Person Vehicle Bike Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Person Vehicle Bike Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 1400 2800 4200 5600 7000 SE +/- 18.88, N = 3 SE +/- 23.77, N = 3 6336.78 6306.42 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Person Vehicle Bike Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Person Vehicle Bike Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.06, N = 3 SE +/- 0.07, N = 3 17.52 17.58 MIN: 11.97 / MAX: 108.98 MIN: 12.03 / MAX: 66.23 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Noise Suppression Poconet-Like FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Noise Suppression Poconet-Like FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 1000 2000 3000 4000 5000 SE +/- 54.03, N = 3 SE +/- 29.05, N = 3 4621.32 4600.95 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Noise Suppression Poconet-Like FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Noise Suppression Poconet-Like FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.17, N = 3 SE +/- 0.20, N = 3 23.68 23.97 MIN: 11.64 / MAX: 175.69 MIN: 11.24 / MAX: 358.1 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Handwritten English Recognition FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Handwritten English Recognition FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 800 1600 2400 3200 4000 SE +/- 3.14, N = 3 SE +/- 10.89, N = 3 3626.41 3608.21 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Handwritten English Recognition FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Handwritten English Recognition FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.03, N = 3 SE +/- 0.09, N = 3 30.85 31.01 MIN: 24.82 / MAX: 82.03 MIN: 24.42 / MAX: 90.86 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Person Re-Identification Retail FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Person Re-Identification Retail FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 2K 4K 6K 8K 10K SE +/- 20.55, N = 3 SE +/- 28.95, N = 3 11638.51 11564.78 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Person Re-Identification Retail FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Person Re-Identification Retail FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.02, N = 3 SE +/- 0.03, N = 3 9.59 9.65 MIN: 7.62 / MAX: 27.98 MIN: 7.62 / MAX: 34.18 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 16K 32K 48K 64K 80K SE +/- 711.41, N = 15 SE +/- 1055.18, N = 3 67789.25 73754.74 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 0.1755 0.351 0.5265 0.702 0.8775 SE +/- 0.01, N = 15 SE +/- 0.03, N = 3 0.78 0.71 MIN: 0.3 / MAX: 138.08 MIN: 0.32 / MAX: 74.71 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Handwritten English Recognition FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Handwritten English Recognition FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 800 1600 2400 3200 4000 SE +/- 10.67, N = 3 SE +/- 8.36, N = 3 3777.56 3704.38 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Handwritten English Recognition FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Handwritten English Recognition FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.08, N = 3 SE +/- 0.07, N = 3 29.61 30.20 MIN: 25.51 / MAX: 92.13 MIN: 22.08 / MAX: 107.87 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 30K 60K 90K 120K 150K SE +/- 1455.69, N = 15 SE +/- 1149.56, N = 15 121702.12 119735.57 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 0.0855 0.171 0.2565 0.342 0.4275 SE +/- 0.01, N = 15 SE +/- 0.00, N = 15 0.38 0.38 MIN: 0.23 / MAX: 46.5 MIN: 0.24 / MAX: 50.65 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
RocksDB Test: Overwrite OpenBenchmarking.org Op/s, More Is Better RocksDB 9.0 Test: Overwrite Linux 6.8 Linux 6.9-rc2 80K 160K 240K 320K 400K SE +/- 881.98, N = 3 SE +/- 3053.81, N = 3 388010 391968 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
RocksDB Test: Random Read OpenBenchmarking.org Op/s, More Is Better RocksDB 9.0 Test: Random Read Linux 6.8 Linux 6.9-rc2 80M 160M 240M 320M 400M SE +/- 1573786.50, N = 3 SE +/- 1447290.34, N = 3 360498518 361586083 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
RocksDB Test: Update Random OpenBenchmarking.org Op/s, More Is Better RocksDB 9.0 Test: Update Random Linux 6.8 Linux 6.9-rc2 80K 160K 240K 320K 400K SE +/- 1470.46, N = 3 SE +/- 1522.72, N = 3 359825 361502 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
RocksDB Test: Read Random Write Random OpenBenchmarking.org Op/s, More Is Better RocksDB 9.0 Test: Read Random Write Random Linux 6.8 Linux 6.9-rc2 500K 1000K 1500K 2000K 2500K SE +/- 17412.72, N = 3 SE +/- 22543.45, N = 5 2120844 2068588 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
Speedb Test: Random Fill OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Random Fill Linux 6.8 Linux 6.9-rc2 70K 140K 210K 280K 350K SE +/- 2148.02, N = 3 SE +/- 3492.28, N = 3 326499 322816 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
Speedb Test: Random Read OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Random Read Linux 6.8 Linux 6.9-rc2 80M 160M 240M 320M 400M SE +/- 1002006.26, N = 3 SE +/- 4146957.10, N = 3 380419389 375837479 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
Speedb Test: Update Random OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Update Random Linux 6.8 Linux 6.9-rc2 60K 120K 180K 240K 300K SE +/- 3488.80, N = 3 SE +/- 3235.16, N = 3 292180 293414 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
Speedb Test: Sequential Fill OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Sequential Fill Linux 6.8 Linux 6.9-rc2 70K 140K 210K 280K 350K SE +/- 1183.38, N = 3 SE +/- 1955.03, N = 3 327637 330167 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
Speedb Test: Random Fill Sync OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Random Fill Sync Linux 6.8 Linux 6.9-rc2 60K 120K 180K 240K 300K SE +/- 1490.88, N = 3 SE +/- 3603.61, N = 3 257446 259603 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
Speedb Test: Read Random Write Random OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Read Random Write Random Linux 6.8 Linux 6.9-rc2 300K 600K 900K 1200K 1500K SE +/- 8674.90, N = 11 SE +/- 13609.90, N = 3 1181758 1175562 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
ONNX Runtime Model: GPT-2 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: GPT-2 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 1.16, N = 3 SE +/- 1.34, N = 15 101.90 105.78 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: GPT-2 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: GPT-2 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.11034, N = 3 SE +/- 0.12308, N = 15 9.80514 9.46496 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: yolov4 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: yolov4 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.07338, N = 3 SE +/- 0.08546, N = 3 7.91321 8.04142 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: yolov4 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: yolov4 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 1.18, N = 3 SE +/- 1.32, N = 3 126.38 124.38 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: yolov4 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: yolov4 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.16273, N = 15 SE +/- 0.10643, N = 13 8.22351 8.90936 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: yolov4 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: yolov4 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 2.31, N = 15 SE +/- 1.33, N = 13 122.24 112.43 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: T5 Encoder - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: T5 Encoder - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 40 80 120 160 200 SE +/- 3.10, N = 12 SE +/- 2.39, N = 14 185.29 196.08 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: T5 Encoder - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: T5 Encoder - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 1.2177 2.4354 3.6531 4.8708 6.0885 SE +/- 0.09595, N = 12 SE +/- 0.06160, N = 14 5.41191 5.10755 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: T5 Encoder - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: T5 Encoder - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 90 180 270 360 450 SE +/- 14.02, N = 15 SE +/- 3.01, N = 3 353.57 402.67 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: T5 Encoder - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: T5 Encoder - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 0.6514 1.3028 1.9542 2.6056 3.257 SE +/- 0.12453, N = 15 SE +/- 0.01855, N = 3 2.89501 2.48234 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: bertsquad-12 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: bertsquad-12 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.04, N = 3 SE +/- 0.05, N = 3 10.78 10.60 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: bertsquad-12 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: bertsquad-12 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.33, N = 3 SE +/- 0.44, N = 3 92.75 94.33 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: bertsquad-12 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: bertsquad-12 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.50, N = 12 SE +/- 0.54, N = 15 14.27 15.41 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: bertsquad-12 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: bertsquad-12 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 16 32 48 64 80 SE +/- 2.24, N = 12 SE +/- 2.28, N = 15 70.96 66.00 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: CaffeNet 12-int8 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: CaffeNet 12-int8 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 60 120 180 240 300 SE +/- 2.79, N = 15 SE +/- 2.09, N = 3 292.17 293.03 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: CaffeNet 12-int8 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: CaffeNet 12-int8 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 0.7703 1.5406 2.3109 3.0812 3.8515 SE +/- 0.03367, N = 15 SE +/- 0.02459, N = 3 3.42357 3.40972 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: CaffeNet 12-int8 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: CaffeNet 12-int8 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 120 240 360 480 600 SE +/- 6.47, N = 15 SE +/- 5.63, N = 3 541.83 571.19 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: CaffeNet 12-int8 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: CaffeNet 12-int8 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 0.4158 0.8316 1.2474 1.6632 2.079 SE +/- 0.02432, N = 15 SE +/- 0.01721, N = 3 1.84807 1.74973 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 0.3382 0.6764 1.0146 1.3528 1.691 SE +/- 0.04472, N = 15 SE +/- 0.03151, N = 12 1.35994 1.50300 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 160 320 480 640 800 SE +/- 25.74, N = 15 SE +/- 14.65, N = 12 747.08 668.70 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: fcn-resnet101-11 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: fcn-resnet101-11 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 1.2178 2.4356 3.6534 4.8712 6.089 SE +/- 0.20811, N = 12 SE +/- 0.05231, N = 3 4.43649 5.41259 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: fcn-resnet101-11 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: fcn-resnet101-11 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 50 100 150 200 250 SE +/- 11.46, N = 12 SE +/- 1.80, N = 3 231.28 184.79 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.23, N = 3 SE +/- 0.17, N = 3 16.78 17.56 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 13 26 39 52 65 SE +/- 0.81, N = 3 SE +/- 0.56, N = 3 59.60 56.95 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.43, N = 12 SE +/- 0.48, N = 15 23.46 24.23 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 10 20 30 40 50 SE +/- 0.81, N = 12 SE +/- 0.86, N = 15 42.78 41.50 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.92, N = 3 SE +/- 1.23, N = 3 94.25 95.14 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.10, N = 3 SE +/- 0.14, N = 3 10.61 10.51 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 2.58, N = 12 SE +/- 3.02, N = 15 111.57 122.60 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.20919, N = 12 SE +/- 0.21896, N = 15 9.01538 8.23121 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: super-resolution-10 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: super-resolution-10 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 0.93, N = 3 SE +/- 1.50, N = 3 152.64 153.56 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: super-resolution-10 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: super-resolution-10 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.04024, N = 3 SE +/- 0.06370, N = 3 6.55011 6.51127 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: super-resolution-10 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: super-resolution-10 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 50 100 150 200 250 SE +/- 2.99, N = 15 SE +/- 0.80, N = 3 205.41 213.74 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: super-resolution-10 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: super-resolution-10 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 1.0984 2.1968 3.2952 4.3936 5.492 SE +/- 0.07270, N = 15 SE +/- 0.01745, N = 3 4.88181 4.67811 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.13, N = 3 SE +/- 0.07, N = 3 23.43 23.17 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 10 20 30 40 50 SE +/- 0.23, N = 3 SE +/- 0.13, N = 3 42.68 43.15 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.40, N = 3 SE +/- 0.25, N = 3 32.45 34.35 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.38, N = 3 SE +/- 0.22, N = 3 30.82 29.11 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
WavPack Audio Encoding WAV To WavPack OpenBenchmarking.org Seconds, Fewer Is Better WavPack Audio Encoding 5.7 WAV To WavPack Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.023, N = 5 SE +/- 0.019, N = 5 6.349 6.340
Llama.cpp Model: llama-2-13b.Q4_0.gguf OpenBenchmarking.org Tokens Per Second, More Is Better Llama.cpp b1808 Model: llama-2-13b.Q4_0.gguf Linux 6.8 Linux 6.9-rc2 0.3443 0.6886 1.0329 1.3772 1.7215 SE +/- 0.01, N = 3 SE +/- 0.02, N = 3 1.48 1.53 1. (CXX) g++ options: -std=c++11 -fPIC -O3 -pthread -march=native -mtune=native -lopenblas
Llama.cpp Model: llama-2-70b-chat.Q5_0.gguf OpenBenchmarking.org Tokens Per Second, More Is Better Llama.cpp b1808 Model: llama-2-70b-chat.Q5_0.gguf Linux 6.8 Linux 6.9-rc2 0.0833 0.1666 0.2499 0.3332 0.4165 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 0.36 0.37 1. (CXX) g++ options: -std=c++11 -fPIC -O3 -pthread -march=native -mtune=native -lopenblas
Llamafile Test: mistral-7b-instruct-v0.2.Q8_0 - Acceleration: CPU OpenBenchmarking.org Tokens Per Second, More Is Better Llamafile 0.6 Test: mistral-7b-instruct-v0.2.Q8_0 - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 1.0305 2.061 3.0915 4.122 5.1525 SE +/- 0.04, N = 3 SE +/- 0.03, N = 3 4.50 4.58
Llamafile Test: wizardcoder-python-34b-v1.0.Q6_K - Acceleration: CPU OpenBenchmarking.org Tokens Per Second, More Is Better Llamafile 0.6 Test: wizardcoder-python-34b-v1.0.Q6_K - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 0.1778 0.3556 0.5334 0.7112 0.889 SE +/- 0.01, N = 4 SE +/- 0.00, N = 3 0.79 0.77
BRL-CAD VGR Performance Metric OpenBenchmarking.org VGR Performance Metric, More Is Better BRL-CAD 7.38.2 VGR Performance Metric Linux 6.8 Linux 6.9-rc2 900K 1800K 2700K 3600K 4500K 3906839 3991802 1. (CXX) g++ options: -std=c++17 -pipe -fvisibility=hidden -fno-strict-aliasing -fno-common -fexceptions -ftemplate-depth-128 -m64 -ggdb3 -O3 -fipa-pta -fstrength-reduce -finline-functions -flto -ltcl8.6 -lnetpbm -lregex_brl -lz_brl -lassimp -ldl -lm -ltk8.6
Chaos Group V-RAY Mode: CPU OpenBenchmarking.org vsamples, More Is Better Chaos Group V-RAY 6.0 Mode: CPU Linux 6.8 Linux 6.9-rc2 13K 26K 39K 52K 65K SE +/- 530.58, N = 15 SE +/- 557.85, N = 3 59852 58995
Phoronix Test Suite v10.8.4