2 x Intel Xeon Max 9480 benchmarks for a future article.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 2404031-NE-XEONMAXLI10 Xeon Max Linux 6.9 Kernel Benchmarks - Phoronix Test Suite Xeon Max Linux 6.9 Kernel Benchmarks 2 x Intel Xeon Max 9480 benchmarks for a future article.
HTML result view exported from: https://openbenchmarking.org/result/2404031-NE-XEONMAXLI10&grs .
Xeon Max Linux 6.9 Kernel Benchmarks Processor Motherboard Chipset Memory Disk Graphics Network OS Kernel Desktop Display Server Compiler File-System Screen Resolution Linux 6.8 Linux 6.9-rc2 2 x Intel Xeon Max 9480 @ 3.50GHz (112 Cores / 224 Threads) Supermicro SYS-221H-TNR X13DEM v1.10 (1.3 BIOS) Intel Device 1bce 512GB 7682GB INTEL SSDPF2KX076TZ ASPEED 2 x Broadcom BCM57508 NetXtreme-E 10Gb/25Gb/40Gb/50Gb/100Gb/200Gb Ubuntu 23.10 6.8.0-060800-generic (x86_64) GNOME Shell 45.0 X Server 1.21.1.7 GCC 13.2.0 ext4 1024x768 6.9.0-060900rc2-generic (x86_64) OpenBenchmarking.org Kernel Details - Transparent Huge Pages: madvise Compiler Details - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details - Scaling Governor: intel_cpufreq performance - CPU Microcode: 0x2c000290 Python Details - Python 3.11.6 Security Details - Linux 6.8: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Not affected - Linux 6.9-rc2: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + reg_file_data_sampling: Not affected + retbleed: Not affected + spec_rstack_overflow: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Not affected
Xeon Max Linux 6.9 Kernel Benchmarks hackbench: 32 - Process stress-ng: Pthread openvino: Age Gender Recognition Retail 0013 FP16 - CPU ffmpeg: libx265 - Platform tensorflow: CPU - 1 - AlexNet onnx: Faster R-CNN R-50-FPN-int8 - CPU - Standard onnx: T5 Encoder - CPU - Parallel stress-ng: SENDFILE onnx: CaffeNet 12-int8 - CPU - Standard tensorflow: CPU - 64 - AlexNet tensorflow: CPU - 64 - GoogLeNet openvino: Vehicle Detection FP16 - CPU openvino: Vehicle Detection FP16 - CPU stress-ng: System V Message Passing webp2: Quality 75, Compression Effort 7 stress-ng: Context Switching openvino: Person Detection FP32 - CPU openvino: Person Detection FP32 - CPU onnx: ArcFace ResNet-100 - CPU - Parallel oidn: RT.ldr_alb_nrm.3840x2160 - CPU-Only stress-ng: MMAP openvino: Weld Porosity Detection FP16-INT8 - CPU jpegxl: JPEG - 90 onnx: super-resolution-10 - CPU - Standard openvino: Weld Porosity Detection FP16-INT8 - CPU stress-ng: Socket Activity onnx: GPT-2 - CPU - Parallel openssl: SHA256 ospray: gravity_spheres_volume/dim_512/pathtracer/real_time stress-ng: IO_uring openvino: Person Detection FP16 - CPU llama-cpp: llama-2-13b.Q4_0.gguf openvino: Person Detection FP16 - CPU blender: Fishy Cat - CPU-Only openvino: Road Segmentation ADAS FP16 - CPU openvino: Road Segmentation ADAS FP16 - CPU ospray-studio: 3 - 1080p - 32 - Path Tracer - CPU stress-ng: Semaphores ospray: gravity_spheres_volume/dim_512/scivis/real_time webp2: Quality 95, Compression Effort 7 ospray: particle_volume/ao/real_time tensorflow: CPU - 16 - ResNet-50 openssl: RSA4096 jpegxl: PNG - 90 stress-ng: Crypto llama-cpp: llama-2-70b-chat.Q5_0.gguf gromacs: MPI CPU - water_GMX50_bare vvenc: Bosphorus 1080p - Faster hackbench: 16 - Thread ospray-studio: 3 - 4K - 1 - Path Tracer - CPU llamafile: wizardcoder-python-34b-v1.0.Q6_K - CPU ospray-studio: 1 - 1080p - 1 - Path Tracer - CPU rocksdb: Read Rand Write Rand ospray-studio: 2 - 1080p - 1 - Path Tracer - CPU stress-ng: Zlib stress-ng: Memory Copying oidn: RT.hdr_alb_nrm.3840x2160 - CPU-Only onednn: Convolution Batch Shapes Auto - CPU brl-cad: VGR Performance Metric ospray-studio: 3 - 4K - 16 - Path Tracer - CPU blender: Junkshop - CPU-Only ospray-studio: 3 - 1080p - 16 - Path Tracer - CPU build-linux-kernel: defconfig xmrig: Monero - 1M openvino: Handwritten English Recognition FP16-INT8 - CPU tensorflow: CPU - 64 - ResNet-50 openvino: Handwritten English Recognition FP16-INT8 - CPU oidn: RTLightmap.hdr.4096x4096 - CPU-Only hackbench: 16 - Process ospray: gravity_spheres_volume/dim_512/ao/real_time svt-av1: Preset 4 - Bosphorus 4K llamafile: mistral-7b-instruct-v0.2.Q8_0 - CPU jpegxl: PNG - 80 primesieve: 1e12 onednn: Deconvolution Batch shapes_1d - CPU onnx: bertsquad-12 - CPU - Parallel openvino: Face Detection FP16 - CPU ospray-studio: 3 - 4K - 32 - Path Tracer - CPU openvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPU openvino: Face Detection FP16 - CPU onnx: yolov4 - CPU - Parallel tensorflow: CPU - 256 - AlexNet ospray-studio: 1 - 4K - 16 - Path Tracer - CPU stress-ng: Mutex ospray-studio: 1 - 1080p - 16 - Path Tracer - CPU stress-ng: Poll vvenc: Bosphorus 1080p - Fast quicksilver: CORAL2 P1 svt-av1: Preset 4 - Bosphorus 1080p v-ray: CPU build-linux-kernel: allmodconfig openvino: Weld Porosity Detection FP16 - CPU compress-lz4: 9 - Decompression Speed stress-ng: Vector Shuffle tensorflow: CPU - 32 - GoogLeNet ffmpeg: libx265 - Video On Demand tensorflow: CPU - 32 - ResNet-50 tensorflow: CPU - 16 - AlexNet openvino: Weld Porosity Detection FP16 - CPU draco: Lion openvino: Noise Suppression Poconet-Like FP16 - CPU speedb: Rand Read tensorflow: CPU - 32 - AlexNet jpegxl-decode: All webp2: Quality 100, Compression Effort 5 ospray-studio: 2 - 4K - 1 - Path Tracer - CPU speedb: Rand Fill onnx: Faster R-CNN R-50-FPN-int8 - CPU - Parallel stress-ng: NUMA rocksdb: Overwrite onnx: ResNet50 v1-12-int8 - CPU - Parallel openssl: SHA512 blender: BMW27 - CPU-Only onednn: IP Shapes 3D - CPU cachebench: Write blender: Pabellon Barcelona - CPU-Only speedb: Rand Fill Sync speedb: Seq Fill compress-lz4: 1 - Decompression Speed stress-ng: Hash ospray: particle_volume/scivis/real_time stress-ng: CPU Stress vvenc: Bosphorus 4K - Fast onednn: Recurrent Neural Network Inference - CPU xmrig: KawPow - 1M openvino: Person Re-Identification Retail FP16 - CPU openvino: Person Re-Identification Retail FP16 - CPU tensorflow: CPU - 256 - GoogLeNet tensorflow: CPU - 256 - ResNet-50 onnx: super-resolution-10 - CPU - Parallel stress-ng: Wide Vector Math stress-ng: Mixed Scheduler speedb: Read Rand Write Rand openvino: Handwritten English Recognition FP16 - CPU openvino: Handwritten English Recognition FP16 - CPU openvino: Person Vehicle Bike Detection FP16 - CPU rocksdb: Update Rand compress-lz4: 3 - Decompression Speed openvino: Noise Suppression Poconet-Like FP16 - CPU ospray-studio: 1 - 1080p - 32 - Path Tracer - CPU stress-ng: Glibc Qsort Data Sorting speedb: Update Rand openvino: Face Detection Retail FP16-INT8 - CPU cachebench: Read / Modify / Write primesieve: 1e13 openvino: Face Detection Retail FP16-INT8 - CPU openvino: Person Vehicle Bike Detection FP16 - CPU ffmpeg: libx264 - Live ospray-studio: 1 - 4K - 32 - Path Tracer - CPU ffmpeg: libx265 - Upload openvino: Face Detection Retail FP16 - CPU openvino: Face Detection Retail FP16 - CPU rocksdb: Rand Read openvino: Road Segmentation ADAS FP16-INT8 - CPU ospray-studio: 2 - 4K - 16 - Path Tracer - CPU blender: Classroom - CPU-Only onnx: CaffeNet 12-int8 - CPU - Parallel openvino: Road Segmentation ADAS FP16-INT8 - CPU openvino: Vehicle Detection FP16-INT8 - CPU openvino: Machine Translation EN To DE FP16 - CPU openvino: Vehicle Detection FP16-INT8 - CPU openvino: Machine Translation EN To DE FP16 - CPU onednn: IP Shapes 1D - CPU xmrig: Wownero - 1M jpegxl: PNG - 100 draco: Church Facade quicksilver: CORAL2 P2 ospray-studio: 3 - 1080p - 1 - Path Tracer - CPU openvino: Face Detection FP16-INT8 - CPU jpegxl: JPEG - 100 compress-lz4: 9 - Compression Speed openvino: Face Detection FP16-INT8 - CPU openssl: RSA4096 stress-ng: AVL Tree encode-wavpack: WAV To WavPack ffmpeg: libx264 - Video On Demand xmrig: GhostRider - 1M xmrig: CryptoNight-Heavy - 1M ffmpeg: libx264 - Upload xmrig: CryptoNight-Femto UPX2 - 1M compress-lz4: 1 - Compression Speed cachebench: Read ospray-studio: 2 - 4K - 32 - Path Tracer - CPU ffmpeg: libx264 - Platform compress-lz4: 3 - Compression Speed openvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPU mysqlslap: 512 webp2: Quality 100, Lossless Compression llamafile: llava-v1.5-7b-q4 - CPU llama-cpp: llama-2-7b.Q4_0.gguf onnx: Faster R-CNN R-50-FPN-int8 - CPU - Standard onnx: Faster R-CNN R-50-FPN-int8 - CPU - Parallel onnx: super-resolution-10 - CPU - Standard onnx: super-resolution-10 - CPU - Parallel onnx: ResNet50 v1-12-int8 - CPU - Standard onnx: ResNet50 v1-12-int8 - CPU - Standard onnx: ResNet50 v1-12-int8 - CPU - Parallel onnx: ArcFace ResNet-100 - CPU - Standard onnx: ArcFace ResNet-100 - CPU - Standard onnx: ArcFace ResNet-100 - CPU - Parallel onnx: fcn-resnet101-11 - CPU - Standard onnx: fcn-resnet101-11 - CPU - Standard onnx: fcn-resnet101-11 - CPU - Parallel onnx: fcn-resnet101-11 - CPU - Parallel onnx: CaffeNet 12-int8 - CPU - Standard onnx: CaffeNet 12-int8 - CPU - Parallel onnx: bertsquad-12 - CPU - Standard onnx: bertsquad-12 - CPU - Standard onnx: bertsquad-12 - CPU - Parallel onnx: T5 Encoder - CPU - Standard onnx: T5 Encoder - CPU - Standard onnx: T5 Encoder - CPU - Parallel onnx: yolov4 - CPU - Standard onnx: yolov4 - CPU - Standard onnx: yolov4 - CPU - Parallel onnx: GPT-2 - CPU - Standard onnx: GPT-2 - CPU - Standard onnx: GPT-2 - CPU - Parallel openvino: Age Gender Recognition Retail 0013 FP16 - CPU blender: Barbershop - CPU-Only stress-ng: Glibc C String Functions stress-ng: Vector Floating Point stress-ng: Fused Multiply-Add stress-ng: Matrix 3D Math stress-ng: Floating Point stress-ng: Matrix Math stress-ng: Malloc stress-ng: MEMFD stress-ng: Pipe tensorflow: CPU - 16 - GoogLeNet tensorflow: CPU - 1 - ResNet-50 tensorflow: CPU - 1 - GoogLeNet mysqlslap: 1024 ospray-studio: 2 - 1080p - 32 - Path Tracer - CPU ospray-studio: 2 - 1080p - 16 - Path Tracer - CPU ospray-studio: 1 - 4K - 1 - Path Tracer - CPU onednn: Recurrent Neural Network Training - CPU onednn: Deconvolution Batch shapes_3d - CPU y-cruncher: 500M y-cruncher: 1B compress-pbzip2: FreeBSD-13.0-RELEASE-amd64-memstick.img Compression build-mesa: Time To Compile build-ffmpeg: Time To Compile stockfish: Chess Benchmark ospray: particle_volume/pathtracer/real_time vvenc: Bosphorus 4K - Faster svt-av1: Preset 13 - Bosphorus 4K svt-av1: Preset 12 - Bosphorus 4K svt-av1: Preset 8 - Bosphorus 4K embree: Pathtracer ISPC - Asian Dragon Obj embree: Pathtracer ISPC - Asian Dragon embree: Pathtracer - Asian Dragon Obj embree: Pathtracer - Asian Dragon embree: Pathtracer ISPC - Crown embree: Pathtracer - Crown srsran: PDSCH Processor Benchmark, Throughput Thread srsran: PDSCH Processor Benchmark, Throughput Total webp2: Default jpegxl: JPEG - 80 ffmpeg: libx265 - Live quicksilver: CTS2 Linux 6.8 Linux 6.9-rc2 24.748 39599.26 67789.25 42.74 17.23 32.4503 185.294 1798203.12 541.827 495.72 135.18 2530.04 14.58 17859304.38 0.64 20343491.52 445.90 82.87 16.7832 2.32 6169.36 4.05 32.562 205.414 26732.88 47927.75 101.901 84678420667 24.2805 3468957.59 451.13 1.48 81.90 38.09 36.28 1018.47 17204 154589461.19 10.50412 0.34 29.7336 24.12 1358616.4 30.969 143967.24 0.36 12.085 33.986 14.679 2108 0.79 466 2120844 501 9670.37 21085.10 2.28 9.00621 3906839 40657 45.60 8732 38.462 32361.2 29.61 47.03 3777.56 1.06 13.112 10.70212 5.865 4.50 32.477 3.356 18.3377 10.7817 119.92 72172 121702.12 307.74 7.91321 723.16 34412 28490133.28 6909 10389186.02 15.913 6814667 18.662 59852 307.502 6.85 3537.0 576905.92 96.31 44.80 35.30 213.35 16147.57 6129 23.68 380419389 346.10 386.232 8.35 1861 326499 23.4276 482.62 388010 94.2487 29940735740 21.82 4.79853 93036.302415 76.20 257446 327637 3601.0 17071694.52 29.2815 202073.08 4.902 748.495 32860.8 11638.51 9.59 202.37 70.10 152.635 5874353.63 70407.96 1181758 30.85 3626.41 6336.78 359825 3364.8 4621.32 13867 1866.58 292180 7.17 99112.699260 39.619 15532.54 17.52 181.51 61363 22.31 11260.98 9.92 360498518 77.34 35317 53.98 292.173 1446.42 5067.40 54.98 22.06 671.66 4.50055 36096.6 25.666 7747 6677333 569 337.21 26.002 32.61 331.23 38204.6 810.19 6.349 42.94 9592.3 32856.3 11.50 33002.2 602.69 13344.699620 63514 42.76 98.43 0.38 86 0.07 1.84 1.99 30.8209 42.6833 4.88181 6.55011 9.01538 111.5706 10.6098 42.7819 23.4613 59.6021 231.284 4.43649 747.084 1.35994 1.84807 3.42357 70.9566 14.2656 92.7467 2.89501 353.573 5.41191 122.236 8.22351 126.384 6.03604 167.343 9.80514 0.78 271.99 80643523.24 234669.74 204365411.59 31442.69 29600.02 384748.54 163551396.40 688.26 30377557.97 58.50 1.88 6.05 34 14836 7402 1801 4821.32 6.16576 6.483 13.416 2.933717 19.511 17.245 61999717 102.9501 9.847 123.246 127.202 48.943 36.4359 31.5957 39.8177 35.4033 38.8350 36.6011 662.4 17006.8 5.90 34.604 112.43 8425444 27.994 35919.64 73754.74 40.20 18.28 34.3512 196.076 1900682.08 571.190 471.88 128.71 2656.42 13.89 18740350.56 0.61 19390362.90 467.79 79.00 17.5623 2.22 6427.33 3.89 33.888 213.737 27790.24 46123.32 105.7827 81678284176 25.1712 3346567.80 466.45 1.53 79.23 39.37 35.14 1051.30 16687 150005776.47 10.8236 0.33 28.8754 24.80 1396904.0 30.128 147978.33 0.37 11.764 33.094 14.298 2164 0.77 478 2068588 489 9446.07 20605.39 2.23 8.81043 3991802 39794 46.56 8552 37.698 33015.5 30.20 47.96 3704.38 1.04 13.357 10.5090 5.760 4.58 33.045 3.299 18.0275 10.6006 121.93 73367 119735.57 302.77 8.04142 711.74 34953 28934422.16 6804 10232390.41 16.152 6916667 18.934 58995 311.745 6.76 3492.0 569593.00 97.53 44.24 34.86 210.71 16349.05 6054 23.97 375837479 341.97 381.657 8.45 1839 322816 23.1727 487.60 391968 95.1441 29665499053 22.02 4.84245 93862.763311 75.56 259603 330167 3628.0 17199190.09 29.0693 200671.70 4.936 753.364 33072.5 11564.78 9.65 203.63 70.53 153.564 5908316.03 70036.20 1175562 31.01 3608.21 6306.42 361502 3349.6 4600.95 13807 1858.70 293414 7.14 98711.076947 39.461 15589.59 17.58 182.12 61162 22.24 11295.92 9.89 361586083 77.57 35422 53.82 293.029 1442.21 5052.71 54.83 22.12 673.46 4.51166 36180.7 25.607 7762 6689667 570 337.78 25.960 32.66 330.73 38147.4 808.99 6.340 42.88 9602.7 32826.1 11.51 32993.8 602.54 13341.467325 63529 42.77 98.41 0.38 86 0.07 1.56 2.49 29.1107 43.1511 4.67811 6.51127 8.23121 122.601 10.5119 41.5048 24.2328 56.9478 184.785 5.41259 668.699 1.50300 1.74973 3.40972 65.9960 15.4110 94.3330 2.48234 402.669 5.10755 112.428 8.90936 124.376 6.57623 154.205 9.46496 0.71 279.74 85035507.55 230503.34 200468629.32 31460.42 29139.60 392750.09 161652700.09 688.76 29814480.67 58.83 1.90 5.98 36 15415 7506 1774 10195.75 5.78202 6.202 13.335 2.926551 19.888 17.283 64443099 98.8690 9.524 123.098 126.187 48.568 37.4335 32.8297 36.5030 35.6507 35.7348 35.0788 650.8 16877.6 5.65 34.187 113.99 8129143 OpenBenchmarking.org
Hackbench Count: 32 - Type: Process OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 32 - Type: Process Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.27, N = 3 SE +/- 0.39, N = 3 24.75 27.99 1. (CC) gcc options: -lpthread
Stress-NG Test: Pthread OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Pthread Linux 6.8 Linux 6.9-rc2 8K 16K 24K 32K 40K SE +/- 49.70, N = 3 SE +/- 450.35, N = 15 39599.26 35919.64 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
OpenVINO Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 16K 32K 48K 64K 80K SE +/- 711.41, N = 15 SE +/- 1055.18, N = 3 67789.25 73754.74 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
FFmpeg Encoder: libx265 - Scenario: Platform OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx265 - Scenario: Platform Linux 6.8 Linux 6.9-rc2 10 20 30 40 50 SE +/- 0.71, N = 12 SE +/- 0.46, N = 12 42.74 40.20 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
TensorFlow Device: CPU - Batch Size: 1 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 1 - Model: AlexNet Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.25, N = 12 SE +/- 0.25, N = 3 17.23 18.28
ONNX Runtime Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.40, N = 3 SE +/- 0.25, N = 3 32.45 34.35 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: T5 Encoder - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: T5 Encoder - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 40 80 120 160 200 SE +/- 3.10, N = 12 SE +/- 2.39, N = 14 185.29 196.08 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
Stress-NG Test: SENDFILE OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: SENDFILE Linux 6.8 Linux 6.9-rc2 400K 800K 1200K 1600K 2000K SE +/- 609.53, N = 3 SE +/- 5873.31, N = 3 1798203.12 1900682.08 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
ONNX Runtime Model: CaffeNet 12-int8 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: CaffeNet 12-int8 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 120 240 360 480 600 SE +/- 6.47, N = 15 SE +/- 5.63, N = 3 541.83 571.19 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
TensorFlow Device: CPU - Batch Size: 64 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 64 - Model: AlexNet Linux 6.8 Linux 6.9-rc2 110 220 330 440 550 SE +/- 6.05, N = 3 SE +/- 5.17, N = 15 495.72 471.88
TensorFlow Device: CPU - Batch Size: 64 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 64 - Model: GoogLeNet Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 0.94, N = 3 SE +/- 1.70, N = 15 135.18 128.71
OpenVINO Model: Vehicle Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Vehicle Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 600 1200 1800 2400 3000 SE +/- 16.72, N = 3 SE +/- 25.96, N = 3 2530.04 2656.42 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Vehicle Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Vehicle Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.10, N = 3 SE +/- 0.14, N = 3 14.58 13.89 MIN: 9.5 / MAX: 109.83 MIN: 9.12 / MAX: 90.57 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
Stress-NG Test: System V Message Passing OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: System V Message Passing Linux 6.8 Linux 6.9-rc2 4M 8M 12M 16M 20M SE +/- 189741.33, N = 15 SE +/- 35075.39, N = 3 17859304.38 18740350.56 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
WebP2 Image Encode Encode Settings: Quality 75, Compression Effort 7 OpenBenchmarking.org MP/s, More Is Better WebP2 Image Encode 20220823 Encode Settings: Quality 75, Compression Effort 7 Linux 6.8 Linux 6.9-rc2 0.144 0.288 0.432 0.576 0.72 SE +/- 0.01, N = 3 SE +/- 0.01, N = 4 0.64 0.61 1. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl
Stress-NG Test: Context Switching OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Context Switching Linux 6.8 Linux 6.9-rc2 4M 8M 12M 16M 20M SE +/- 178479.09, N = 8 SE +/- 265846.30, N = 3 20343491.52 19390362.90 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
OpenVINO Model: Person Detection FP32 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Person Detection FP32 - Device: CPU Linux 6.8 Linux 6.9-rc2 100 200 300 400 500 SE +/- 3.09, N = 3 SE +/- 3.72, N = 3 445.90 467.79 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Person Detection FP32 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Person Detection FP32 - Device: CPU Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.57, N = 3 SE +/- 0.62, N = 3 82.87 79.00 MIN: 51.69 / MAX: 529.56 MIN: 49.24 / MAX: 319.95 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
ONNX Runtime Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.23, N = 3 SE +/- 0.17, N = 3 16.78 17.56 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
Intel Open Image Denoise Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.2 Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Only Linux 6.8 Linux 6.9-rc2 0.522 1.044 1.566 2.088 2.61 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 2.32 2.22
Stress-NG Test: MMAP OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: MMAP Linux 6.8 Linux 6.9-rc2 1400 2800 4200 5600 7000 SE +/- 70.88, N = 14 SE +/- 66.09, N = 15 6169.36 6427.33 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
OpenVINO Model: Weld Porosity Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Weld Porosity Detection FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 0.9113 1.8226 2.7339 3.6452 4.5565 SE +/- 0.05, N = 4 SE +/- 0.01, N = 3 4.05 3.89 MIN: 2.67 / MAX: 52.43 MIN: 2.64 / MAX: 47.09 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
JPEG-XL libjxl Input: JPEG - Quality: 90 OpenBenchmarking.org MP/s, More Is Better JPEG-XL libjxl 0.10.1 Input: JPEG - Quality: 90 Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.37, N = 3 SE +/- 0.38, N = 15 32.56 33.89 1. (CXX) g++ options: -fno-rtti -O3 -fPIE -pie -lm
ONNX Runtime Model: super-resolution-10 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: super-resolution-10 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 50 100 150 200 250 SE +/- 2.99, N = 15 SE +/- 0.80, N = 3 205.41 213.74 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
OpenVINO Model: Weld Porosity Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Weld Porosity Detection FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 6K 12K 18K 24K 30K SE +/- 285.82, N = 4 SE +/- 155.42, N = 3 26732.88 27790.24 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
Stress-NG Test: Socket Activity OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Socket Activity Linux 6.8 Linux 6.9-rc2 10K 20K 30K 40K 50K SE +/- 79.93, N = 3 SE +/- 21.40, N = 3 47927.75 46123.32 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
ONNX Runtime Model: GPT-2 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: GPT-2 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 1.16, N = 3 SE +/- 1.34, N = 15 101.90 105.78 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
OpenSSL Algorithm: SHA256 OpenBenchmarking.org byte/s, More Is Better OpenSSL Algorithm: SHA256 Linux 6.8 Linux 6.9-rc2 20000M 40000M 60000M 80000M 100000M SE +/- 767626108.54, N = 3 SE +/- 717052715.91, N = 12 84678420667 81678284176 1. OpenSSL 3.0.10 1 Aug 2023 (Library: OpenSSL 3.0.10 1 Aug 2023)
OSPRay Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_time Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.41, N = 12 SE +/- 0.20, N = 3 24.28 25.17
Stress-NG Test: IO_uring OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: IO_uring Linux 6.8 Linux 6.9-rc2 700K 1400K 2100K 2800K 3500K SE +/- 31595.27, N = 3 SE +/- 43339.28, N = 15 3468957.59 3346567.80 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
OpenVINO Model: Person Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Person Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 100 200 300 400 500 SE +/- 2.33, N = 3 SE +/- 1.20, N = 3 451.13 466.45 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
Llama.cpp Model: llama-2-13b.Q4_0.gguf OpenBenchmarking.org Tokens Per Second, More Is Better Llama.cpp b1808 Model: llama-2-13b.Q4_0.gguf Linux 6.8 Linux 6.9-rc2 0.3443 0.6886 1.0329 1.3772 1.7215 SE +/- 0.01, N = 3 SE +/- 0.02, N = 3 1.48 1.53 1. (CXX) g++ options: -std=c++11 -fPIC -O3 -pthread -march=native -mtune=native -lopenblas
OpenVINO Model: Person Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Person Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.43, N = 3 SE +/- 0.20, N = 3 81.90 79.23 MIN: 49.43 / MAX: 345.35 MIN: 50.1 / MAX: 290.06 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
Blender Blend File: Fishy Cat - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.1 Blend File: Fishy Cat - Compute: CPU-Only Linux 6.8 Linux 6.9-rc2 9 18 27 36 45 SE +/- 0.38, N = 15 SE +/- 0.44, N = 4 38.09 39.37
OpenVINO Model: Road Segmentation ADAS FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Road Segmentation ADAS FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.35, N = 3 SE +/- 0.23, N = 3 36.28 35.14 MIN: 24.78 / MAX: 205.05 MIN: 25.25 / MAX: 437.27 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Road Segmentation ADAS FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Road Segmentation ADAS FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 200 400 600 800 1000 SE +/- 9.82, N = 3 SE +/- 6.69, N = 3 1018.47 1051.30 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OSPRay Studio Camera: 3 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 4K 8K 12K 16K 20K SE +/- 101.18, N = 3 SE +/- 156.23, N = 6 17204 16687
Stress-NG Test: Semaphores OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Semaphores Linux 6.8 Linux 6.9-rc2 30M 60M 90M 120M 150M SE +/- 1498916.25, N = 3 SE +/- 1330349.25, N = 3 154589461.19 150005776.47 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
OSPRay Benchmark: gravity_spheres_volume/dim_512/scivis/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: gravity_spheres_volume/dim_512/scivis/real_time Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.08, N = 15 SE +/- 0.07, N = 3 10.50 10.82
WebP2 Image Encode Encode Settings: Quality 95, Compression Effort 7 OpenBenchmarking.org MP/s, More Is Better WebP2 Image Encode 20220823 Encode Settings: Quality 95, Compression Effort 7 Linux 6.8 Linux 6.9-rc2 0.0765 0.153 0.2295 0.306 0.3825 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 0.34 0.33 1. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl
OSPRay Benchmark: particle_volume/ao/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: particle_volume/ao/real_time Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.11, N = 3 SE +/- 0.19, N = 3 29.73 28.88
TensorFlow Device: CPU - Batch Size: 16 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 16 - Model: ResNet-50 Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.30, N = 3 SE +/- 0.28, N = 15 24.12 24.80
OpenSSL Algorithm: RSA4096 OpenBenchmarking.org verify/s, More Is Better OpenSSL Algorithm: RSA4096 Linux 6.8 Linux 6.9-rc2 300K 600K 900K 1200K 1500K SE +/- 19064.62, N = 3 SE +/- 833.27, N = 3 1358616.4 1396904.0 1. OpenSSL 3.0.10 1 Aug 2023 (Library: OpenSSL 3.0.10 1 Aug 2023)
JPEG-XL libjxl Input: PNG - Quality: 90 OpenBenchmarking.org MP/s, More Is Better JPEG-XL libjxl 0.10.1 Input: PNG - Quality: 90 Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.28, N = 3 SE +/- 0.18, N = 3 30.97 30.13 1. (CXX) g++ options: -fno-rtti -O3 -fPIE -pie -lm
Stress-NG Test: Crypto OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Crypto Linux 6.8 Linux 6.9-rc2 30K 60K 90K 120K 150K SE +/- 1720.03, N = 15 SE +/- 2039.96, N = 15 143967.24 147978.33 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Llama.cpp Model: llama-2-70b-chat.Q5_0.gguf OpenBenchmarking.org Tokens Per Second, More Is Better Llama.cpp b1808 Model: llama-2-70b-chat.Q5_0.gguf Linux 6.8 Linux 6.9-rc2 0.0833 0.1666 0.2499 0.3332 0.4165 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 0.36 0.37 1. (CXX) g++ options: -std=c++11 -fPIC -O3 -pthread -march=native -mtune=native -lopenblas
GROMACS Implementation: MPI CPU - Input: water_GMX50_bare OpenBenchmarking.org Ns Per Day, More Is Better GROMACS 2024 Implementation: MPI CPU - Input: water_GMX50_bare Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.06, N = 3 SE +/- 0.08, N = 3 12.09 11.76 1. (CXX) g++ options: -O3 -lm
VVenC Video Input: Bosphorus 1080p - Video Preset: Faster OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.11 Video Input: Bosphorus 1080p - Video Preset: Faster Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.29, N = 15 SE +/- 0.36, N = 15 33.99 33.09 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
Hackbench Count: 16 - Type: Thread OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 16 - Type: Thread Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.00, N = 3 SE +/- 0.14, N = 15 14.68 14.30 1. (CC) gcc options: -lpthread
OSPRay Studio Camera: 3 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 500 1000 1500 2000 2500 SE +/- 21.99, N = 4 SE +/- 26.16, N = 13 2108 2164
Llamafile Test: wizardcoder-python-34b-v1.0.Q6_K - Acceleration: CPU OpenBenchmarking.org Tokens Per Second, More Is Better Llamafile 0.6 Test: wizardcoder-python-34b-v1.0.Q6_K - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 0.1778 0.3556 0.5334 0.7112 0.889 SE +/- 0.01, N = 4 SE +/- 0.00, N = 3 0.79 0.77
OSPRay Studio Camera: 1 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 100 200 300 400 500 SE +/- 3.71, N = 3 SE +/- 2.33, N = 3 466 478
RocksDB Test: Read Random Write Random OpenBenchmarking.org Op/s, More Is Better RocksDB 9.0 Test: Read Random Write Random Linux 6.8 Linux 6.9-rc2 500K 1000K 1500K 2000K 2500K SE +/- 17412.72, N = 3 SE +/- 22543.45, N = 5 2120844 2068588 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
OSPRay Studio Camera: 2 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 110 220 330 440 550 SE +/- 5.26, N = 5 SE +/- 2.33, N = 3 501 489
Stress-NG Test: Zlib OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Zlib Linux 6.8 Linux 6.9-rc2 2K 4K 6K 8K 10K SE +/- 25.13, N = 3 SE +/- 124.93, N = 3 9670.37 9446.07 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Memory Copying OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Memory Copying Linux 6.8 Linux 6.9-rc2 5K 10K 15K 20K 25K SE +/- 178.25, N = 3 SE +/- 155.03, N = 10 21085.10 20605.39 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Intel Open Image Denoise Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.2 Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Only Linux 6.8 Linux 6.9-rc2 0.513 1.026 1.539 2.052 2.565 SE +/- 0.02, N = 15 SE +/- 0.02, N = 15 2.28 2.23
oneDNN Harness: Convolution Batch Shapes Auto - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: Convolution Batch Shapes Auto - Engine: CPU Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.03328, N = 3 SE +/- 0.09863, N = 15 9.00621 8.81043 MIN: 8.12 MIN: 6.25 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
BRL-CAD VGR Performance Metric OpenBenchmarking.org VGR Performance Metric, More Is Better BRL-CAD 7.38.2 VGR Performance Metric Linux 6.8 Linux 6.9-rc2 900K 1800K 2700K 3600K 4500K 3906839 3991802 1. (CXX) g++ options: -std=c++17 -pipe -fvisibility=hidden -fno-strict-aliasing -fno-common -fexceptions -ftemplate-depth-128 -m64 -ggdb3 -O3 -fipa-pta -fstrength-reduce -finline-functions -flto -ltcl8.6 -lnetpbm -lregex_brl -lz_brl -lassimp -ldl -lm -ltk8.6
OSPRay Studio Camera: 3 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 9K 18K 27K 36K 45K SE +/- 626.71, N = 12 SE +/- 252.89, N = 3 40657 39794
Blender Blend File: Junkshop - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.1 Blend File: Junkshop - Compute: CPU-Only Linux 6.8 Linux 6.9-rc2 11 22 33 44 55 SE +/- 0.47, N = 15 SE +/- 0.40, N = 15 45.60 46.56
OSPRay Studio Camera: 3 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 2K 4K 6K 8K 10K SE +/- 70.21, N = 3 SE +/- 26.21, N = 3 8732 8552
Timed Linux Kernel Compilation Build: defconfig OpenBenchmarking.org Seconds, Fewer Is Better Timed Linux Kernel Compilation 6.8 Build: defconfig Linux 6.8 Linux 6.9-rc2 9 18 27 36 45 SE +/- 0.36, N = 6 SE +/- 0.37, N = 6 38.46 37.70
Xmrig Variant: Monero - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: Monero - Hash Count: 1M Linux 6.8 Linux 6.9-rc2 7K 14K 21K 28K 35K SE +/- 320.50, N = 13 SE +/- 181.84, N = 3 32361.2 33015.5 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
OpenVINO Model: Handwritten English Recognition FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Handwritten English Recognition FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.08, N = 3 SE +/- 0.07, N = 3 29.61 30.20 MIN: 25.51 / MAX: 92.13 MIN: 22.08 / MAX: 107.87 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
TensorFlow Device: CPU - Batch Size: 64 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 64 - Model: ResNet-50 Linux 6.8 Linux 6.9-rc2 11 22 33 44 55 SE +/- 0.54, N = 12 SE +/- 0.42, N = 3 47.03 47.96
OpenVINO Model: Handwritten English Recognition FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Handwritten English Recognition FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 800 1600 2400 3200 4000 SE +/- 10.67, N = 3 SE +/- 8.36, N = 3 3777.56 3704.38 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
Intel Open Image Denoise Run: RTLightmap.hdr.4096x4096 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.2 Run: RTLightmap.hdr.4096x4096 - Device: CPU-Only Linux 6.8 Linux 6.9-rc2 0.2385 0.477 0.7155 0.954 1.1925 SE +/- 0.01, N = 7 SE +/- 0.01, N = 15 1.06 1.04
Hackbench Count: 16 - Type: Process OpenBenchmarking.org Seconds, Fewer Is Better Hackbench Count: 16 - Type: Process Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.11, N = 3 SE +/- 0.19, N = 3 13.11 13.36 1. (CC) gcc options: -lpthread
OSPRay Benchmark: gravity_spheres_volume/dim_512/ao/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: gravity_spheres_volume/dim_512/ao/real_time Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.14, N = 15 SE +/- 0.05, N = 3 10.70 10.51
SVT-AV1 Encoder Mode: Preset 4 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 2.0 Encoder Mode: Preset 4 - Input: Bosphorus 4K Linux 6.8 Linux 6.9-rc2 1.3196 2.6392 3.9588 5.2784 6.598 SE +/- 0.035, N = 3 SE +/- 0.051, N = 15 5.865 5.760 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
Llamafile Test: mistral-7b-instruct-v0.2.Q8_0 - Acceleration: CPU OpenBenchmarking.org Tokens Per Second, More Is Better Llamafile 0.6 Test: mistral-7b-instruct-v0.2.Q8_0 - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 1.0305 2.061 3.0915 4.122 5.1525 SE +/- 0.04, N = 3 SE +/- 0.03, N = 3 4.50 4.58
JPEG-XL libjxl Input: PNG - Quality: 80 OpenBenchmarking.org MP/s, More Is Better JPEG-XL libjxl 0.10.1 Input: PNG - Quality: 80 Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.38, N = 3 SE +/- 0.36, N = 15 32.48 33.05 1. (CXX) g++ options: -fno-rtti -O3 -fPIE -pie -lm
Primesieve Length: 1e12 OpenBenchmarking.org Seconds, Fewer Is Better Primesieve 12.1 Length: 1e12 Linux 6.8 Linux 6.9-rc2 0.7551 1.5102 2.2653 3.0204 3.7755 SE +/- 0.024, N = 11 SE +/- 0.044, N = 3 3.356 3.299 1. (CXX) g++ options: -O3
oneDNN Harness: Deconvolution Batch shapes_1d - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: Deconvolution Batch shapes_1d - Engine: CPU Linux 6.8 Linux 6.9-rc2 5 10 15 20 25 SE +/- 0.23, N = 15 SE +/- 0.14, N = 3 18.34 18.03 MIN: 10.09 MIN: 13.29 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
ONNX Runtime Model: bertsquad-12 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: bertsquad-12 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.04, N = 3 SE +/- 0.05, N = 3 10.78 10.60 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
OpenVINO Model: Face Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Face Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 0.33, N = 3 SE +/- 0.55, N = 3 119.92 121.93 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OSPRay Studio Camera: 3 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 16K 32K 48K 64K 80K SE +/- 506.90, N = 3 SE +/- 855.02, N = 15 72172 73367
OpenVINO Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 30K 60K 90K 120K 150K SE +/- 1455.69, N = 15 SE +/- 1149.56, N = 15 121702.12 119735.57 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Face Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Face Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 70 140 210 280 350 SE +/- 0.84, N = 3 SE +/- 1.30, N = 3 307.74 302.77 MIN: 183.58 / MAX: 728.14 MIN: 181.55 / MAX: 894.73 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
ONNX Runtime Model: yolov4 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: yolov4 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.07338, N = 3 SE +/- 0.08546, N = 3 7.91321 8.04142 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
TensorFlow Device: CPU - Batch Size: 256 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 256 - Model: AlexNet Linux 6.8 Linux 6.9-rc2 160 320 480 640 800 SE +/- 9.01, N = 3 SE +/- 6.62, N = 7 723.16 711.74
OSPRay Studio Camera: 1 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 7K 14K 21K 28K 35K SE +/- 294.16, N = 3 SE +/- 454.99, N = 3 34412 34953
Stress-NG Test: Mutex OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Mutex Linux 6.8 Linux 6.9-rc2 6M 12M 18M 24M 30M SE +/- 281178.86, N = 3 SE +/- 294297.65, N = 3 28490133.28 28934422.16 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
OSPRay Studio Camera: 1 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 1500 3000 4500 6000 7500 SE +/- 83.68, N = 3 SE +/- 15.17, N = 3 6909 6804
Stress-NG Test: Poll OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Poll Linux 6.8 Linux 6.9-rc2 2M 4M 6M 8M 10M SE +/- 37074.52, N = 3 SE +/- 30731.99, N = 3 10389186.02 10232390.41 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
VVenC Video Input: Bosphorus 1080p - Video Preset: Fast OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.11 Video Input: Bosphorus 1080p - Video Preset: Fast Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.11, N = 3 SE +/- 0.07, N = 3 15.91 16.15 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
Quicksilver Input: CORAL2 P1 OpenBenchmarking.org Figure Of Merit, More Is Better Quicksilver 20230818 Input: CORAL2 P1 Linux 6.8 Linux 6.9-rc2 1.5M 3M 4.5M 6M 7.5M SE +/- 67589.78, N = 6 SE +/- 46491.34, N = 3 6814667 6916667 1. (CXX) g++ options: -fopenmp -O3 -march=native
SVT-AV1 Encoder Mode: Preset 4 - Input: Bosphorus 1080p OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 2.0 Encoder Mode: Preset 4 - Input: Bosphorus 1080p Linux 6.8 Linux 6.9-rc2 5 10 15 20 25 SE +/- 0.13, N = 15 SE +/- 0.21, N = 3 18.66 18.93 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
Chaos Group V-RAY Mode: CPU OpenBenchmarking.org vsamples, More Is Better Chaos Group V-RAY 6.0 Mode: CPU Linux 6.8 Linux 6.9-rc2 13K 26K 39K 52K 65K SE +/- 530.58, N = 15 SE +/- 557.85, N = 3 59852 58995
Timed Linux Kernel Compilation Build: allmodconfig OpenBenchmarking.org Seconds, Fewer Is Better Timed Linux Kernel Compilation 6.8 Build: allmodconfig Linux 6.8 Linux 6.9-rc2 70 140 210 280 350 SE +/- 3.84, N = 3 SE +/- 1.26, N = 3 307.50 311.75
OpenVINO Model: Weld Porosity Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Weld Porosity Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.08, N = 3 SE +/- 0.05, N = 3 6.85 6.76 MIN: 4.47 / MAX: 51.04 MIN: 4.36 / MAX: 70.91 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
LZ4 Compression Compression Level: 9 - Decompression Speed OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.4 Compression Level: 9 - Decompression Speed Linux 6.8 Linux 6.9-rc2 800 1600 2400 3200 4000 SE +/- 16.01, N = 3 SE +/- 57.81, N = 3 3537.0 3492.0 1. (CC) gcc options: -O3
Stress-NG Test: Vector Shuffle OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Vector Shuffle Linux 6.8 Linux 6.9-rc2 120K 240K 360K 480K 600K SE +/- 1416.85, N = 3 SE +/- 1897.90, N = 3 576905.92 569593.00 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
TensorFlow Device: CPU - Batch Size: 32 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 32 - Model: GoogLeNet Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 1.27, N = 12 SE +/- 1.10, N = 15 96.31 97.53
FFmpeg Encoder: libx265 - Scenario: Video On Demand OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx265 - Scenario: Video On Demand Linux 6.8 Linux 6.9-rc2 10 20 30 40 50 SE +/- 0.50, N = 12 SE +/- 0.76, N = 9 44.80 44.24 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
TensorFlow Device: CPU - Batch Size: 32 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 32 - Model: ResNet-50 Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.48, N = 15 SE +/- 0.42, N = 15 35.30 34.86
TensorFlow Device: CPU - Batch Size: 16 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 16 - Model: AlexNet Linux 6.8 Linux 6.9-rc2 50 100 150 200 250 SE +/- 1.58, N = 15 SE +/- 1.89, N = 15 213.35 210.71
OpenVINO Model: Weld Porosity Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Weld Porosity Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 4K 8K 12K 16K 20K SE +/- 194.69, N = 3 SE +/- 107.94, N = 3 16147.57 16349.05 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
Google Draco Model: Lion OpenBenchmarking.org ms, Fewer Is Better Google Draco 1.5.6 Model: Lion Linux 6.8 Linux 6.9-rc2 1300 2600 3900 5200 6500 SE +/- 40.84, N = 3 SE +/- 29.28, N = 3 6129 6054 1. (CXX) g++ options: -O3
OpenVINO Model: Noise Suppression Poconet-Like FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Noise Suppression Poconet-Like FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.17, N = 3 SE +/- 0.20, N = 3 23.68 23.97 MIN: 11.64 / MAX: 175.69 MIN: 11.24 / MAX: 358.1 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
Speedb Test: Random Read OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Random Read Linux 6.8 Linux 6.9-rc2 80M 160M 240M 320M 400M SE +/- 1002006.26, N = 3 SE +/- 4146957.10, N = 3 380419389 375837479 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
TensorFlow Device: CPU - Batch Size: 32 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 32 - Model: AlexNet Linux 6.8 Linux 6.9-rc2 80 160 240 320 400 SE +/- 2.86, N = 15 SE +/- 5.16, N = 15 346.10 341.97
JPEG-XL Decoding libjxl CPU Threads: All OpenBenchmarking.org MP/s, More Is Better JPEG-XL Decoding libjxl 0.10.1 CPU Threads: All Linux 6.8 Linux 6.9-rc2 80 160 240 320 400 SE +/- 4.23, N = 3 SE +/- 2.26, N = 3 386.23 381.66
WebP2 Image Encode Encode Settings: Quality 100, Compression Effort 5 OpenBenchmarking.org MP/s, More Is Better WebP2 Image Encode 20220823 Encode Settings: Quality 100, Compression Effort 5 Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.11, N = 15 SE +/- 0.07, N = 3 8.35 8.45 1. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl
OSPRay Studio Camera: 2 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 400 800 1200 1600 2000 SE +/- 13.93, N = 11 SE +/- 26.21, N = 3 1861 1839
Speedb Test: Random Fill OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Random Fill Linux 6.8 Linux 6.9-rc2 70K 140K 210K 280K 350K SE +/- 2148.02, N = 3 SE +/- 3492.28, N = 3 326499 322816 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
ONNX Runtime Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.13, N = 3 SE +/- 0.07, N = 3 23.43 23.17 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
Stress-NG Test: NUMA OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: NUMA Linux 6.8 Linux 6.9-rc2 110 220 330 440 550 SE +/- 4.27, N = 3 SE +/- 5.89, N = 4 482.62 487.60 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
RocksDB Test: Overwrite OpenBenchmarking.org Op/s, More Is Better RocksDB 9.0 Test: Overwrite Linux 6.8 Linux 6.9-rc2 80K 160K 240K 320K 400K SE +/- 881.98, N = 3 SE +/- 3053.81, N = 3 388010 391968 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
ONNX Runtime Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.92, N = 3 SE +/- 1.23, N = 3 94.25 95.14 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
OpenSSL Algorithm: SHA512 OpenBenchmarking.org byte/s, More Is Better OpenSSL Algorithm: SHA512 Linux 6.8 Linux 6.9-rc2 6000M 12000M 18000M 24000M 30000M SE +/- 92272355.83, N = 3 SE +/- 252323355.66, N = 3 29940735740 29665499053 1. OpenSSL 3.0.10 1 Aug 2023 (Library: OpenSSL 3.0.10 1 Aug 2023)
Blender Blend File: BMW27 - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.1 Blend File: BMW27 - Compute: CPU-Only Linux 6.8 Linux 6.9-rc2 5 10 15 20 25 SE +/- 0.33, N = 14 SE +/- 0.08, N = 3 21.82 22.02
oneDNN Harness: IP Shapes 3D - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: IP Shapes 3D - Engine: CPU Linux 6.8 Linux 6.9-rc2 1.0896 2.1792 3.2688 4.3584 5.448 SE +/- 0.02352, N = 3 SE +/- 0.02357, N = 3 4.79853 4.84245 MIN: 3.98 MIN: 4.01 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
CacheBench Test: Write OpenBenchmarking.org MB/s, More Is Better CacheBench Test: Write Linux 6.8 Linux 6.9-rc2 20K 40K 60K 80K 100K SE +/- 36.95, N = 3 SE +/- 65.10, N = 3 93036.30 93862.76 MIN: 54037.87 / MAX: 103869.49 MIN: 54012.69 / MAX: 104121.12 1. (CC) gcc options: -O3 -lrt
Blender Blend File: Pabellon Barcelona - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.1 Blend File: Pabellon Barcelona - Compute: CPU-Only Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.83, N = 5 SE +/- 0.61, N = 15 76.20 75.56
Speedb Test: Random Fill Sync OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Random Fill Sync Linux 6.8 Linux 6.9-rc2 60K 120K 180K 240K 300K SE +/- 1490.88, N = 3 SE +/- 3603.61, N = 3 257446 259603 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
Speedb Test: Sequential Fill OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Sequential Fill Linux 6.8 Linux 6.9-rc2 70K 140K 210K 280K 350K SE +/- 1183.38, N = 3 SE +/- 1955.03, N = 3 327637 330167 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
LZ4 Compression Compression Level: 1 - Decompression Speed OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.4 Compression Level: 1 - Decompression Speed Linux 6.8 Linux 6.9-rc2 800 1600 2400 3200 4000 SE +/- 44.31, N = 3 SE +/- 18.57, N = 3 3601.0 3628.0 1. (CC) gcc options: -O3
Stress-NG Test: Hash OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Hash Linux 6.8 Linux 6.9-rc2 4M 8M 12M 16M 20M SE +/- 82496.53, N = 3 SE +/- 192774.29, N = 3 17071694.52 17199190.09 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
OSPRay Benchmark: particle_volume/scivis/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: particle_volume/scivis/real_time Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.26, N = 3 SE +/- 0.33, N = 3 29.28 29.07
Stress-NG Test: CPU Stress OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: CPU Stress Linux 6.8 Linux 6.9-rc2 40K 80K 120K 160K 200K SE +/- 569.84, N = 3 SE +/- 335.84, N = 3 202073.08 200671.70 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
VVenC Video Input: Bosphorus 4K - Video Preset: Fast OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.11 Video Input: Bosphorus 4K - Video Preset: Fast Linux 6.8 Linux 6.9-rc2 1.1106 2.2212 3.3318 4.4424 5.553 SE +/- 0.075, N = 10 SE +/- 0.077, N = 12 4.902 4.936 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
oneDNN Harness: Recurrent Neural Network Inference - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: Recurrent Neural Network Inference - Engine: CPU Linux 6.8 Linux 6.9-rc2 160 320 480 640 800 SE +/- 7.46, N = 3 SE +/- 6.16, N = 9 748.50 753.36 MIN: 677.68 MIN: 652.83 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
Xmrig Variant: KawPow - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: KawPow - Hash Count: 1M Linux 6.8 Linux 6.9-rc2 7K 14K 21K 28K 35K SE +/- 68.36, N = 3 SE +/- 137.96, N = 3 32860.8 33072.5 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
OpenVINO Model: Person Re-Identification Retail FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Person Re-Identification Retail FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 2K 4K 6K 8K 10K SE +/- 20.55, N = 3 SE +/- 28.95, N = 3 11638.51 11564.78 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Person Re-Identification Retail FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Person Re-Identification Retail FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.02, N = 3 SE +/- 0.03, N = 3 9.59 9.65 MIN: 7.62 / MAX: 27.98 MIN: 7.62 / MAX: 34.18 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
TensorFlow Device: CPU - Batch Size: 256 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 256 - Model: GoogLeNet Linux 6.8 Linux 6.9-rc2 40 80 120 160 200 SE +/- 2.20, N = 12 SE +/- 1.95, N = 3 202.37 203.63
TensorFlow Device: CPU - Batch Size: 256 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 256 - Model: ResNet-50 Linux 6.8 Linux 6.9-rc2 16 32 48 64 80 SE +/- 0.38, N = 3 SE +/- 0.47, N = 3 70.10 70.53
ONNX Runtime Model: super-resolution-10 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: super-resolution-10 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 0.93, N = 3 SE +/- 1.50, N = 3 152.64 153.56 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
Stress-NG Test: Wide Vector Math OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Wide Vector Math Linux 6.8 Linux 6.9-rc2 1.3M 2.6M 3.9M 5.2M 6.5M SE +/- 72950.92, N = 4 SE +/- 55746.31, N = 15 5874353.63 5908316.03 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Mixed Scheduler OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Mixed Scheduler Linux 6.8 Linux 6.9-rc2 15K 30K 45K 60K 75K SE +/- 153.71, N = 3 SE +/- 281.29, N = 3 70407.96 70036.20 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Speedb Test: Read Random Write Random OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Read Random Write Random Linux 6.8 Linux 6.9-rc2 300K 600K 900K 1200K 1500K SE +/- 8674.90, N = 11 SE +/- 13609.90, N = 3 1181758 1175562 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
OpenVINO Model: Handwritten English Recognition FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Handwritten English Recognition FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.03, N = 3 SE +/- 0.09, N = 3 30.85 31.01 MIN: 24.82 / MAX: 82.03 MIN: 24.42 / MAX: 90.86 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Handwritten English Recognition FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Handwritten English Recognition FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 800 1600 2400 3200 4000 SE +/- 3.14, N = 3 SE +/- 10.89, N = 3 3626.41 3608.21 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Person Vehicle Bike Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Person Vehicle Bike Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 1400 2800 4200 5600 7000 SE +/- 18.88, N = 3 SE +/- 23.77, N = 3 6336.78 6306.42 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
RocksDB Test: Update Random OpenBenchmarking.org Op/s, More Is Better RocksDB 9.0 Test: Update Random Linux 6.8 Linux 6.9-rc2 80K 160K 240K 320K 400K SE +/- 1470.46, N = 3 SE +/- 1522.72, N = 3 359825 361502 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
LZ4 Compression Compression Level: 3 - Decompression Speed OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.4 Compression Level: 3 - Decompression Speed Linux 6.8 Linux 6.9-rc2 700 1400 2100 2800 3500 SE +/- 33.92, N = 3 SE +/- 34.17, N = 3 3364.8 3349.6 1. (CC) gcc options: -O3
OpenVINO Model: Noise Suppression Poconet-Like FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Noise Suppression Poconet-Like FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 1000 2000 3000 4000 5000 SE +/- 54.03, N = 3 SE +/- 29.05, N = 3 4621.32 4600.95 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OSPRay Studio Camera: 1 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 3K 6K 9K 12K 15K SE +/- 112.81, N = 9 SE +/- 180.26, N = 3 13867 13807
Stress-NG Test: Glibc Qsort Data Sorting OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Glibc Qsort Data Sorting Linux 6.8 Linux 6.9-rc2 400 800 1200 1600 2000 SE +/- 4.40, N = 3 SE +/- 4.16, N = 3 1866.58 1858.70 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Speedb Test: Update Random OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Update Random Linux 6.8 Linux 6.9-rc2 60K 120K 180K 240K 300K SE +/- 3488.80, N = 3 SE +/- 3235.16, N = 3 292180 293414 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
OpenVINO Model: Face Detection Retail FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Face Detection Retail FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.07, N = 3 SE +/- 0.00, N = 3 7.17 7.14 MIN: 5.78 / MAX: 54.3 MIN: 5.79 / MAX: 44.66 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
CacheBench Test: Read / Modify / Write OpenBenchmarking.org MB/s, More Is Better CacheBench Test: Read / Modify / Write Linux 6.8 Linux 6.9-rc2 20K 40K 60K 80K 100K SE +/- 2.71, N = 3 SE +/- 71.07, N = 3 99112.70 98711.08 MIN: 88982.6 / MAX: 105154.55 MIN: 85707.72 / MAX: 105115.44 1. (CC) gcc options: -O3 -lrt
Primesieve Length: 1e13 OpenBenchmarking.org Seconds, Fewer Is Better Primesieve 12.1 Length: 1e13 Linux 6.8 Linux 6.9-rc2 9 18 27 36 45 SE +/- 0.06, N = 3 SE +/- 0.09, N = 3 39.62 39.46 1. (CXX) g++ options: -O3
OpenVINO Model: Face Detection Retail FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Face Detection Retail FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 3K 6K 9K 12K 15K SE +/- 132.85, N = 3 SE +/- 25.09, N = 3 15532.54 15589.59 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Person Vehicle Bike Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Person Vehicle Bike Detection FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.06, N = 3 SE +/- 0.07, N = 3 17.52 17.58 MIN: 11.97 / MAX: 108.98 MIN: 12.03 / MAX: 66.23 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
FFmpeg Encoder: libx264 - Scenario: Live OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx264 - Scenario: Live Linux 6.8 Linux 6.9-rc2 40 80 120 160 200 SE +/- 1.29, N = 3 SE +/- 0.70, N = 3 181.51 182.12 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
OSPRay Studio Camera: 1 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 13K 26K 39K 52K 65K SE +/- 443.96, N = 15 SE +/- 567.54, N = 7 61363 61162
FFmpeg Encoder: libx265 - Scenario: Upload OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx265 - Scenario: Upload Linux 6.8 Linux 6.9-rc2 5 10 15 20 25 SE +/- 0.31, N = 3 SE +/- 0.26, N = 4 22.31 22.24 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
OpenVINO Model: Face Detection Retail FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Face Detection Retail FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 2K 4K 6K 8K 10K SE +/- 23.83, N = 3 SE +/- 26.72, N = 3 11260.98 11295.92 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Face Detection Retail FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Face Detection Retail FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.02, N = 3 SE +/- 0.02, N = 3 9.92 9.89 MIN: 7.45 / MAX: 33.82 MIN: 7.63 / MAX: 36.57 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
RocksDB Test: Random Read OpenBenchmarking.org Op/s, More Is Better RocksDB 9.0 Test: Random Read Linux 6.8 Linux 6.9-rc2 80M 160M 240M 320M 400M SE +/- 1573786.50, N = 3 SE +/- 1447290.34, N = 3 360498518 361586083 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
OpenVINO Model: Road Segmentation ADAS FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Road Segmentation ADAS FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.09, N = 3 SE +/- 0.28, N = 3 77.34 77.57 MIN: 56.48 / MAX: 158.67 MIN: 57.16 / MAX: 148.67 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OSPRay Studio Camera: 2 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 8K 16K 24K 32K 40K SE +/- 302.37, N = 15 SE +/- 483.46, N = 3 35317 35422
Blender Blend File: Classroom - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.1 Blend File: Classroom - Compute: CPU-Only Linux 6.8 Linux 6.9-rc2 12 24 36 48 60 SE +/- 0.10, N = 3 SE +/- 0.09, N = 3 53.98 53.82
ONNX Runtime Model: CaffeNet 12-int8 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: CaffeNet 12-int8 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 60 120 180 240 300 SE +/- 2.79, N = 15 SE +/- 2.09, N = 3 292.17 293.03 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
OpenVINO Model: Road Segmentation ADAS FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Road Segmentation ADAS FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 300 600 900 1200 1500 SE +/- 1.59, N = 3 SE +/- 5.16, N = 3 1446.42 1442.21 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Vehicle Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Vehicle Detection FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 1100 2200 3300 4400 5500 SE +/- 9.46, N = 3 SE +/- 18.34, N = 3 5067.40 5052.71 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Machine Translation EN To DE FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Machine Translation EN To DE FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 12 24 36 48 60 SE +/- 0.47, N = 3 SE +/- 0.25, N = 3 54.98 54.83 MIN: 33.9 / MAX: 423.36 MIN: 34.71 / MAX: 375.04 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Vehicle Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Vehicle Detection FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 5 10 15 20 25 SE +/- 0.04, N = 3 SE +/- 0.08, N = 3 22.06 22.12 MIN: 15.51 / MAX: 86.25 MIN: 15.73 / MAX: 71.5 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenVINO Model: Machine Translation EN To DE FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Machine Translation EN To DE FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 150 300 450 600 750 SE +/- 5.75, N = 3 SE +/- 3.16, N = 3 671.66 673.46 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
oneDNN Harness: IP Shapes 1D - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: IP Shapes 1D - Engine: CPU Linux 6.8 Linux 6.9-rc2 1.0151 2.0302 3.0453 4.0604 5.0755 SE +/- 0.04580, N = 3 SE +/- 0.05180, N = 12 4.50055 4.51166 MIN: 3.52 MIN: 3.37 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
Xmrig Variant: Wownero - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: Wownero - Hash Count: 1M Linux 6.8 Linux 6.9-rc2 8K 16K 24K 32K 40K SE +/- 245.88, N = 3 SE +/- 490.33, N = 3 36096.6 36180.7 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
JPEG-XL libjxl Input: PNG - Quality: 100 OpenBenchmarking.org MP/s, More Is Better JPEG-XL libjxl 0.10.1 Input: PNG - Quality: 100 Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.15, N = 3 SE +/- 0.21, N = 3 25.67 25.61 1. (CXX) g++ options: -fno-rtti -O3 -fPIE -pie -lm
Google Draco Model: Church Facade OpenBenchmarking.org ms, Fewer Is Better Google Draco 1.5.6 Model: Church Facade Linux 6.8 Linux 6.9-rc2 1700 3400 5100 6800 8500 SE +/- 0.88, N = 3 SE +/- 9.56, N = 3 7747 7762 1. (CXX) g++ options: -O3
Quicksilver Input: CORAL2 P2 OpenBenchmarking.org Figure Of Merit, More Is Better Quicksilver 20230818 Input: CORAL2 P2 Linux 6.8 Linux 6.9-rc2 1.4M 2.8M 4.2M 5.6M 7M SE +/- 74194.56, N = 9 SE +/- 91964.82, N = 9 6677333 6689667 1. (CXX) g++ options: -fopenmp -O3 -march=native
OSPRay Studio Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 120 240 360 480 600 SE +/- 4.36, N = 15 SE +/- 4.85, N = 15 569 570
OpenVINO Model: Face Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Face Detection FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 70 140 210 280 350 SE +/- 0.74, N = 3 SE +/- 0.81, N = 3 337.21 337.78 MIN: 261.24 / MAX: 615.09 MIN: 253.21 / MAX: 781.06 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
JPEG-XL libjxl Input: JPEG - Quality: 100 OpenBenchmarking.org MP/s, More Is Better JPEG-XL libjxl 0.10.1 Input: JPEG - Quality: 100 Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.11, N = 3 SE +/- 0.08, N = 3 26.00 25.96 1. (CXX) g++ options: -fno-rtti -O3 -fPIE -pie -lm
LZ4 Compression Compression Level: 9 - Compression Speed OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.4 Compression Level: 9 - Compression Speed Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.08, N = 3 SE +/- 0.07, N = 3 32.61 32.66 1. (CC) gcc options: -O3
OpenVINO Model: Face Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2024.0 Model: Face Detection FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 70 140 210 280 350 SE +/- 0.75, N = 3 SE +/- 0.82, N = 3 331.23 330.73 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
OpenSSL Algorithm: RSA4096 OpenBenchmarking.org sign/s, More Is Better OpenSSL Algorithm: RSA4096 Linux 6.8 Linux 6.9-rc2 8K 16K 24K 32K 40K SE +/- 46.27, N = 3 SE +/- 90.23, N = 3 38204.6 38147.4 1. OpenSSL 3.0.10 1 Aug 2023 (Library: OpenSSL 3.0.10 1 Aug 2023)
Stress-NG Test: AVL Tree OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: AVL Tree Linux 6.8 Linux 6.9-rc2 200 400 600 800 1000 SE +/- 6.53, N = 3 SE +/- 2.13, N = 3 810.19 808.99 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
WavPack Audio Encoding WAV To WavPack OpenBenchmarking.org Seconds, Fewer Is Better WavPack Audio Encoding 5.7 WAV To WavPack Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.023, N = 5 SE +/- 0.019, N = 5 6.349 6.340
FFmpeg Encoder: libx264 - Scenario: Video On Demand OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx264 - Scenario: Video On Demand Linux 6.8 Linux 6.9-rc2 10 20 30 40 50 SE +/- 0.06, N = 3 SE +/- 0.10, N = 3 42.94 42.88 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
Xmrig Variant: GhostRider - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: GhostRider - Hash Count: 1M Linux 6.8 Linux 6.9-rc2 2K 4K 6K 8K 10K SE +/- 31.05, N = 3 SE +/- 42.79, N = 3 9592.3 9602.7 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
Xmrig Variant: CryptoNight-Heavy - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: CryptoNight-Heavy - Hash Count: 1M Linux 6.8 Linux 6.9-rc2 7K 14K 21K 28K 35K SE +/- 49.86, N = 3 SE +/- 160.32, N = 3 32856.3 32826.1 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
FFmpeg Encoder: libx264 - Scenario: Upload OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx264 - Scenario: Upload Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 11.50 11.51 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
Xmrig Variant: CryptoNight-Femto UPX2 - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: CryptoNight-Femto UPX2 - Hash Count: 1M Linux 6.8 Linux 6.9-rc2 7K 14K 21K 28K 35K SE +/- 110.64, N = 3 SE +/- 170.37, N = 3 33002.2 32993.8 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
LZ4 Compression Compression Level: 1 - Compression Speed OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.4 Compression Level: 1 - Compression Speed Linux 6.8 Linux 6.9-rc2 130 260 390 520 650 SE +/- 1.15, N = 3 SE +/- 0.16, N = 3 602.69 602.54 1. (CC) gcc options: -O3
CacheBench Test: Read OpenBenchmarking.org MB/s, More Is Better CacheBench Test: Read Linux 6.8 Linux 6.9-rc2 3K 6K 9K 12K 15K SE +/- 0.52, N = 3 SE +/- 0.88, N = 3 13344.70 13341.47 MIN: 13338.06 / MAX: 13346.23 MIN: 13335.31 / MAX: 13343.85 1. (CC) gcc options: -O3 -lrt
OSPRay Studio Camera: 2 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 14K 28K 42K 56K 70K SE +/- 591.53, N = 7 SE +/- 885.61, N = 3 63514 63529
FFmpeg Encoder: libx264 - Scenario: Platform OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx264 - Scenario: Platform Linux 6.8 Linux 6.9-rc2 10 20 30 40 50 SE +/- 0.17, N = 3 SE +/- 0.10, N = 3 42.76 42.77 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
LZ4 Compression Compression Level: 3 - Compression Speed OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.4 Compression Level: 3 - Compression Speed Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.06, N = 3 SE +/- 0.06, N = 3 98.43 98.41 1. (CC) gcc options: -O3
OpenVINO Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU Linux 6.8 Linux 6.9-rc2 0.0855 0.171 0.2565 0.342 0.4275 SE +/- 0.01, N = 15 SE +/- 0.00, N = 15 0.38 0.38 MIN: 0.23 / MAX: 46.5 MIN: 0.24 / MAX: 50.65 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
MariaDB mariadb-slap Clients: 512 OpenBenchmarking.org Queries Per Second, More Is Better MariaDB mariadb-slap 11.5 Clients: 512 Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.41, N = 3 SE +/- 0.87, N = 9 86 86 1. (CXX) g++ options: -fPIC -pie -fstack-protector -O3 -shared -lrt -lpthread -lz -ldl -lm -lstdc++
WebP2 Image Encode Encode Settings: Quality 100, Lossless Compression OpenBenchmarking.org MP/s, More Is Better WebP2 Image Encode 20220823 Encode Settings: Quality 100, Lossless Compression Linux 6.8 Linux 6.9-rc2 0.0158 0.0316 0.0474 0.0632 0.079 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 0.07 0.07 1. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl
Llamafile Test: llava-v1.5-7b-q4 - Acceleration: CPU OpenBenchmarking.org Tokens Per Second, More Is Better Llamafile 0.6 Test: llava-v1.5-7b-q4 - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 0.414 0.828 1.242 1.656 2.07 SE +/- 0.07, N = 12 SE +/- 0.05, N = 15 1.84 1.56
Llama.cpp Model: llama-2-7b.Q4_0.gguf OpenBenchmarking.org Tokens Per Second, More Is Better Llama.cpp b1808 Model: llama-2-7b.Q4_0.gguf Linux 6.8 Linux 6.9-rc2 0.5603 1.1206 1.6809 2.2412 2.8015 SE +/- 0.06, N = 6 SE +/- 0.42, N = 12 1.99 2.49 1. (CXX) g++ options: -std=c++11 -fPIC -O3 -pthread -march=native -mtune=native -lopenblas
ONNX Runtime Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 7 14 21 28 35 SE +/- 0.38, N = 3 SE +/- 0.22, N = 3 30.82 29.11 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 10 20 30 40 50 SE +/- 0.23, N = 3 SE +/- 0.13, N = 3 42.68 43.15 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: super-resolution-10 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: super-resolution-10 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 1.0984 2.1968 3.2952 4.3936 5.492 SE +/- 0.07270, N = 15 SE +/- 0.01745, N = 3 4.88181 4.67811 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: super-resolution-10 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: super-resolution-10 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.04024, N = 3 SE +/- 0.06370, N = 3 6.55011 6.51127 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.20919, N = 12 SE +/- 0.21896, N = 15 9.01538 8.23121 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 2.58, N = 12 SE +/- 3.02, N = 15 111.57 122.60 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.10, N = 3 SE +/- 0.14, N = 3 10.61 10.51 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 10 20 30 40 50 SE +/- 0.81, N = 12 SE +/- 0.86, N = 15 42.78 41.50 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 6 12 18 24 30 SE +/- 0.43, N = 12 SE +/- 0.48, N = 15 23.46 24.23 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 13 26 39 52 65 SE +/- 0.81, N = 3 SE +/- 0.56, N = 3 59.60 56.95 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: fcn-resnet101-11 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: fcn-resnet101-11 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 50 100 150 200 250 SE +/- 11.46, N = 12 SE +/- 1.80, N = 3 231.28 184.79 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: fcn-resnet101-11 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: fcn-resnet101-11 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 1.2178 2.4356 3.6534 4.8712 6.089 SE +/- 0.20811, N = 12 SE +/- 0.05231, N = 3 4.43649 5.41259 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 160 320 480 640 800 SE +/- 25.74, N = 15 SE +/- 14.65, N = 12 747.08 668.70 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 0.3382 0.6764 1.0146 1.3528 1.691 SE +/- 0.04472, N = 15 SE +/- 0.03151, N = 12 1.35994 1.50300 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: CaffeNet 12-int8 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: CaffeNet 12-int8 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 0.4158 0.8316 1.2474 1.6632 2.079 SE +/- 0.02432, N = 15 SE +/- 0.01721, N = 3 1.84807 1.74973 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: CaffeNet 12-int8 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: CaffeNet 12-int8 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 0.7703 1.5406 2.3109 3.0812 3.8515 SE +/- 0.03367, N = 15 SE +/- 0.02459, N = 3 3.42357 3.40972 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: bertsquad-12 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: bertsquad-12 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 16 32 48 64 80 SE +/- 2.24, N = 12 SE +/- 2.28, N = 15 70.96 66.00 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: bertsquad-12 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: bertsquad-12 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.50, N = 12 SE +/- 0.54, N = 15 14.27 15.41 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: bertsquad-12 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: bertsquad-12 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 0.33, N = 3 SE +/- 0.44, N = 3 92.75 94.33 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: T5 Encoder - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: T5 Encoder - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 0.6514 1.3028 1.9542 2.6056 3.257 SE +/- 0.12453, N = 15 SE +/- 0.01855, N = 3 2.89501 2.48234 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: T5 Encoder - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: T5 Encoder - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 90 180 270 360 450 SE +/- 14.02, N = 15 SE +/- 3.01, N = 3 353.57 402.67 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: T5 Encoder - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: T5 Encoder - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 1.2177 2.4354 3.6531 4.8708 6.0885 SE +/- 0.09595, N = 12 SE +/- 0.06160, N = 14 5.41191 5.10755 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: yolov4 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: yolov4 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 2.31, N = 15 SE +/- 1.33, N = 13 122.24 112.43 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: yolov4 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: yolov4 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.16273, N = 15 SE +/- 0.10643, N = 13 8.22351 8.90936 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: yolov4 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: yolov4 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 1.18, N = 3 SE +/- 1.32, N = 3 126.38 124.38 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: GPT-2 - Device: CPU - Executor: Standard OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: GPT-2 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.16304, N = 14 SE +/- 0.19739, N = 13 6.03604 6.57623 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: GPT-2 - Device: CPU - Executor: Standard OpenBenchmarking.org Inferences Per Second, More Is Better ONNX Runtime 1.17 Model: GPT-2 - Device: CPU - Executor: Standard Linux 6.8 Linux 6.9-rc2 40 80 120 160 200 SE +/- 5.43, N = 14 SE +/- 6.41, N = 13 167.34 154.21 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
ONNX Runtime Model: GPT-2 - Device: CPU - Executor: Parallel OpenBenchmarking.org Inference Time Cost (ms), Fewer Is Better ONNX Runtime 1.17 Model: GPT-2 - Device: CPU - Executor: Parallel Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.11034, N = 3 SE +/- 0.12308, N = 15 9.80514 9.46496 1. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt
OpenVINO Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2024.0 Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU Linux 6.8 Linux 6.9-rc2 0.1755 0.351 0.5265 0.702 0.8775 SE +/- 0.01, N = 15 SE +/- 0.03, N = 3 0.78 0.71 MIN: 0.3 / MAX: 138.08 MIN: 0.32 / MAX: 74.71 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl
Blender Blend File: Barbershop - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.1 Blend File: Barbershop - Compute: CPU-Only Linux 6.8 Linux 6.9-rc2 60 120 180 240 300 SE +/- 6.20, N = 8 SE +/- 6.67, N = 9 271.99 279.74
Stress-NG Test: Glibc C String Functions OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Glibc C String Functions Linux 6.8 Linux 6.9-rc2 20M 40M 60M 80M 100M SE +/- 2138024.53, N = 13 SE +/- 1504028.14, N = 15 80643523.24 85035507.55 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Vector Floating Point OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Vector Floating Point Linux 6.8 Linux 6.9-rc2 50K 100K 150K 200K 250K SE +/- 5616.14, N = 12 SE +/- 8579.53, N = 12 234669.74 230503.34 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Fused Multiply-Add OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Fused Multiply-Add Linux 6.8 Linux 6.9-rc2 40M 80M 120M 160M 200M SE +/- 6187723.80, N = 12 SE +/- 3824188.11, N = 15 204365411.59 200468629.32 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Matrix 3D Math OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Matrix 3D Math Linux 6.8 Linux 6.9-rc2 7K 14K 21K 28K 35K SE +/- 669.33, N = 15 SE +/- 437.41, N = 15 31442.69 31460.42 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Floating Point OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Floating Point Linux 6.8 Linux 6.9-rc2 6K 12K 18K 24K 30K SE +/- 632.70, N = 12 SE +/- 214.66, N = 11 29600.02 29139.60 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Matrix Math OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Matrix Math Linux 6.8 Linux 6.9-rc2 80K 160K 240K 320K 400K SE +/- 4925.11, N = 15 SE +/- 9278.38, N = 12 384748.54 392750.09 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Malloc OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Malloc Linux 6.8 Linux 6.9-rc2 40M 80M 120M 160M 200M SE +/- 201545.00, N = 3 SE +/- 3675496.40, N = 12 163551396.40 161652700.09 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: MEMFD OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: MEMFD Linux 6.8 Linux 6.9-rc2 150 300 450 600 750 SE +/- 14.68, N = 15 SE +/- 5.94, N = 8 688.26 688.76 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Pipe OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.16.04 Test: Pipe Linux 6.8 Linux 6.9-rc2 7M 14M 21M 28M 35M SE +/- 749025.19, N = 15 SE +/- 2106194.07, N = 15 30377557.97 29814480.67 1. (CXX) g++ options: -lm -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
TensorFlow Device: CPU - Batch Size: 16 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 16 - Model: GoogLeNet Linux 6.8 Linux 6.9-rc2 13 26 39 52 65 SE +/- 1.31, N = 12 SE +/- 1.24, N = 12 58.50 58.83
TensorFlow Device: CPU - Batch Size: 1 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 1 - Model: ResNet-50 Linux 6.8 Linux 6.9-rc2 0.4275 0.855 1.2825 1.71 2.1375 SE +/- 0.03, N = 15 SE +/- 0.03, N = 15 1.88 1.90
TensorFlow Device: CPU - Batch Size: 1 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.16.1 Device: CPU - Batch Size: 1 - Model: GoogLeNet Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.13, N = 12 SE +/- 0.14, N = 15 6.05 5.98
MariaDB mariadb-slap Clients: 1024 OpenBenchmarking.org Queries Per Second, More Is Better MariaDB mariadb-slap 11.5 Clients: 1024 Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 1.23, N = 9 SE +/- 1.41, N = 6 34 36 1. (CXX) g++ options: -fPIC -pie -fstack-protector -O3 -shared -lrt -lpthread -lz -ldl -lm -lstdc++
OSPRay Studio Camera: 2 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 3K 6K 9K 12K 15K SE +/- 268.32, N = 12 SE +/- 235.77, N = 15 14836 15415
OSPRay Studio Camera: 2 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 1600 3200 4800 6400 8000 SE +/- 105.66, N = 3 SE +/- 118.69, N = 15 7402 7506
OSPRay Studio Camera: 1 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Linux 6.8 Linux 6.9-rc2 400 800 1200 1600 2000 SE +/- 40.26, N = 15 SE +/- 22.91, N = 15 1801 1774
oneDNN Harness: Recurrent Neural Network Training - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: Recurrent Neural Network Training - Engine: CPU Linux 6.8 Linux 6.9-rc2 2K 4K 6K 8K 10K SE +/- 22.69, N = 3 SE +/- 498.89, N = 13 4821.32 10195.75 MIN: 4659.99 MIN: 4568.3 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
oneDNN Harness: Deconvolution Batch shapes_3d - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.4 Harness: Deconvolution Batch shapes_3d - Engine: CPU Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.10591, N = 15 SE +/- 0.04472, N = 3 6.16576 5.78202 MIN: 3.75 MIN: 3.81 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread
Y-Cruncher Pi Digits To Calculate: 500M OpenBenchmarking.org Seconds, Fewer Is Better Y-Cruncher 0.8.3 Pi Digits To Calculate: 500M Linux 6.8 Linux 6.9-rc2 2 4 6 8 10 SE +/- 0.143, N = 12 SE +/- 0.083, N = 15 6.483 6.202
Y-Cruncher Pi Digits To Calculate: 1B OpenBenchmarking.org Seconds, Fewer Is Better Y-Cruncher 0.8.3 Pi Digits To Calculate: 1B Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.30, N = 15 SE +/- 0.24, N = 15 13.42 13.34
Parallel BZIP2 Compression FreeBSD-13.0-RELEASE-amd64-memstick.img Compression OpenBenchmarking.org Seconds, Fewer Is Better Parallel BZIP2 Compression 1.1.13 FreeBSD-13.0-RELEASE-amd64-memstick.img Compression Linux 6.8 Linux 6.9-rc2 0.6601 1.3202 1.9803 2.6404 3.3005 SE +/- 0.081558, N = 15 SE +/- 0.102058, N = 12 2.933717 2.926551 1. (CXX) g++ options: -O2 -pthread -lbz2 -lpthread
Timed Mesa Compilation Time To Compile OpenBenchmarking.org Seconds, Fewer Is Better Timed Mesa Compilation 24.0 Time To Compile Linux 6.8 Linux 6.9-rc2 5 10 15 20 25 SE +/- 0.51, N = 15 SE +/- 0.39, N = 15 19.51 19.89
Timed FFmpeg Compilation Time To Compile OpenBenchmarking.org Seconds, Fewer Is Better Timed FFmpeg Compilation 6.1 Time To Compile Linux 6.8 Linux 6.9-rc2 4 8 12 16 20 SE +/- 0.39, N = 15 SE +/- 0.39, N = 15 17.25 17.28
Stockfish Chess Benchmark OpenBenchmarking.org Nodes Per Second, More Is Better Stockfish 16.1 Chess Benchmark Linux 6.8 Linux 6.9-rc2 14M 28M 42M 56M 70M SE +/- 2993041.97, N = 9 SE +/- 2705883.43, N = 9 61999717 64443099 1. (CXX) g++ options: -lgcov -m64 -lpthread -fno-exceptions -std=c++17 -fno-peel-loops -fno-tracer -pedantic -O3 -funroll-loops -msse -msse3 -mpopcnt -mavx2 -mbmi -mavx512f -mavx512bw -mavx512vnni -mavx512dq -mavx512vl -msse4.1 -mssse3 -msse2 -mbmi2 -flto -flto-partition=one -flto=jobserver
OSPRay Benchmark: particle_volume/pathtracer/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: particle_volume/pathtracer/real_time Linux 6.8 Linux 6.9-rc2 20 40 60 80 100 SE +/- 1.85, N = 12 SE +/- 1.98, N = 9 102.95 98.87
VVenC Video Input: Bosphorus 4K - Video Preset: Faster OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.11 Video Input: Bosphorus 4K - Video Preset: Faster Linux 6.8 Linux 6.9-rc2 3 6 9 12 15 SE +/- 0.200, N = 12 SE +/- 0.172, N = 15 9.847 9.524 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
SVT-AV1 Encoder Mode: Preset 13 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 2.0 Encoder Mode: Preset 13 - Input: Bosphorus 4K Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 3.82, N = 12 SE +/- 5.42, N = 15 123.25 123.10 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 12 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 2.0 Encoder Mode: Preset 12 - Input: Bosphorus 4K Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 5.13, N = 15 SE +/- 5.21, N = 15 127.20 126.19 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 8 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 2.0 Encoder Mode: Preset 8 - Input: Bosphorus 4K Linux 6.8 Linux 6.9-rc2 11 22 33 44 55 SE +/- 1.10, N = 12 SE +/- 1.18, N = 15 48.94 48.57 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
Embree Binary: Pathtracer ISPC - Model: Asian Dragon Obj OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Asian Dragon Obj Linux 6.8 Linux 6.9-rc2 9 18 27 36 45 SE +/- 0.94, N = 15 SE +/- 0.88, N = 15 36.44 37.43 MIN: 22.56 / MAX: 49.05 MIN: 27.23 / MAX: 49.01
Embree Binary: Pathtracer ISPC - Model: Asian Dragon OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Asian Dragon Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.77, N = 13 SE +/- 0.82, N = 15 31.60 32.83 MIN: 20.8 / MAX: 52.83 MIN: 22.56 / MAX: 48.16
Embree Binary: Pathtracer - Model: Asian Dragon Obj OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer - Model: Asian Dragon Obj Linux 6.8 Linux 6.9-rc2 9 18 27 36 45 SE +/- 0.92, N = 15 SE +/- 0.60, N = 15 39.82 36.50 MIN: 26.64 / MAX: 50.91 MIN: 22.66 / MAX: 46.51
Embree Binary: Pathtracer - Model: Asian Dragon OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer - Model: Asian Dragon Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.86, N = 15 SE +/- 0.54, N = 15 35.40 35.65 MIN: 23.5 / MAX: 52.61 MIN: 24.54 / MAX: 51.88
Embree Binary: Pathtracer ISPC - Model: Crown OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Crown Linux 6.8 Linux 6.9-rc2 9 18 27 36 45 SE +/- 0.47, N = 4 SE +/- 1.44, N = 12 38.84 35.73 MIN: 30.82 / MAX: 52.25 MIN: 20.32 / MAX: 57.51
Embree Binary: Pathtracer - Model: Crown OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer - Model: Crown Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.59, N = 15 SE +/- 1.07, N = 15 36.60 35.08 MIN: 25.71 / MAX: 54.49 MIN: 22.8 / MAX: 59.06
srsRAN Project Test: PDSCH Processor Benchmark, Throughput Thread OpenBenchmarking.org Mbps, More Is Better srsRAN Project 23.10.1-20240219 Test: PDSCH Processor Benchmark, Throughput Thread Linux 6.8 Linux 6.9-rc2 140 280 420 560 700 SE +/- 26.44, N = 12 SE +/- 29.77, N = 14 662.4 650.8 1. (CXX) g++ options: -march=native -mavx2 -mavx -msse4.1 -mfma -mavx512f -mavx512cd -mavx512bw -mavx512dq -O3 -fno-trapping-math -fno-math-errno -ldl
srsRAN Project Test: PDSCH Processor Benchmark, Throughput Total OpenBenchmarking.org Mbps, More Is Better srsRAN Project 23.10.1-20240219 Test: PDSCH Processor Benchmark, Throughput Total Linux 6.8 Linux 6.9-rc2 4K 8K 12K 16K 20K SE +/- 263.34, N = 15 SE +/- 377.19, N = 12 17006.8 16877.6 1. (CXX) g++ options: -march=native -mavx2 -mavx -msse4.1 -mfma -mavx512f -mavx512cd -mavx512bw -mavx512dq -O3 -fno-trapping-math -fno-math-errno -ldl
WebP2 Image Encode Encode Settings: Default OpenBenchmarking.org MP/s, More Is Better WebP2 Image Encode 20220823 Encode Settings: Default Linux 6.8 Linux 6.9-rc2 1.3275 2.655 3.9825 5.31 6.6375 SE +/- 0.11, N = 15 SE +/- 0.09, N = 15 5.90 5.65 1. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl
JPEG-XL libjxl Input: JPEG - Quality: 80 OpenBenchmarking.org MP/s, More Is Better JPEG-XL libjxl 0.10.1 Input: JPEG - Quality: 80 Linux 6.8 Linux 6.9-rc2 8 16 24 32 40 SE +/- 0.51, N = 12 SE +/- 0.53, N = 15 34.60 34.19 1. (CXX) g++ options: -fno-rtti -O3 -fPIE -pie -lm
FFmpeg Encoder: libx265 - Scenario: Live OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx265 - Scenario: Live Linux 6.8 Linux 6.9-rc2 30 60 90 120 150 SE +/- 1.36, N = 15 SE +/- 1.80, N = 15 112.43 113.99 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
Quicksilver Input: CTS2 OpenBenchmarking.org Figure Of Merit, More Is Better Quicksilver 20230818 Input: CTS2 Linux 6.8 Linux 6.9-rc2 2M 4M 6M 8M 10M SE +/- 97067.86, N = 9 SE +/- 244177.28, N = 7 8425444 8129143 1. (CXX) g++ options: -fopenmp -O3 -march=native
Phoronix Test Suite v10.8.4