Tests for a future article. AMD EPYC 8324P 32-Core testing with a AMD Cinnabar (RCB1009C BIOS) and ASPEED on Ubuntu 23.10 via the Phoronix Test Suite.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 2401110-NE-NEWTESTS900 new-tests - Phoronix Test Suite new-tests Tests for a future article. AMD EPYC 8324P 32-Core testing with a AMD Cinnabar (RCB1009C BIOS) and ASPEED on Ubuntu 23.10 via the Phoronix Test Suite.
HTML result view exported from: https://openbenchmarking.org/result/2401110-NE-NEWTESTS900&export=pdf&grr&sor&rro .
new-tests Processor Motherboard Chipset Memory Disk Graphics Monitor Network OS Kernel Desktop Display Server OpenGL Compiler File-System Screen Resolution Zen 1 - EPYC 7601 b c 32 32 z 32 c 32 d AMD EPYC 7601 32-Core @ 2.20GHz (32 Cores / 64 Threads) TYAN B8026T70AE24HR (V1.02.B10 BIOS) AMD 17h 128GB 280GB INTEL SSDPE21D280GA + 1000GB INTEL SSDPE2KX010T8 llvmpipe VE228 2 x Broadcom NetXtreme BCM5720 PCIe Ubuntu 23.10 6.6.9-060609-generic (x86_64) GNOME Shell 45.0 X Server 1.21.1.7 4.5 Mesa 23.2.1-1ubuntu3.1 (LLVM 15.0.7 256 bits) GCC 13.2.0 ext4 1920x1080 AMD EPYC 8534PN 64-Core @ 2.00GHz (64 Cores / 128 Threads) AMD Cinnabar (RCB1009C BIOS) AMD Device 14a4 6 x 32 GB DRAM-4800MT/s Samsung M321R4GA0BB0-CQKMG 1000GB INTEL SSDPE2KX010T8 1920x1200 AMD EPYC 8534PN 32-Core @ 2.05GHz (32 Cores / 64 Threads) ASPEED AMD EPYC 8324P 32-Core @ 2.65GHz (32 Cores / 64 Threads) OpenBenchmarking.org Kernel Details - Transparent Huge Pages: madvise Compiler Details - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details - Zen 1 - EPYC 7601: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0x800126e - b: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xaa00212 - c: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xaa00212 - 32: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xaa00212 - 32 z: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xaa00212 - 32 c: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xaa00212 - 32 d: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xaa00212 Security Details - Zen 1 - EPYC 7601: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Mitigation of untrained return thunk; SMT vulnerable + spec_rstack_overflow: Mitigation of Safe RET + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional STIBP: disabled RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected - b: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Mitigation of Safe RET + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected - c: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Mitigation of Safe RET + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected - 32: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Mitigation of Safe RET + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected - 32 z: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Mitigation of Safe RET + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected - 32 c: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Mitigation of Safe RET + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected - 32 d: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Mitigation of Safe RET + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected Java Details - 32, 32 z, 32 c, 32 d: OpenJDK Runtime Environment (build 11.0.21+9-post-Ubuntu-0ubuntu123.10) Python Details - 32, 32 z, 32 c, 32 d: Python 3.11.6
new-tests quicksilver: CTS2 build-linux-kernel: allmodconfig blender: Barbershop - CPU-Only quicksilver: CORAL2 P2 pytorch: CPU - 16 - Efficientnet_v2_l build-gem5: Time To Compile xmrig: GhostRider - 1M quicksilver: CORAL2 P1 ffmpeg: libx265 - Upload ffmpeg: libx265 - Platform ffmpeg: libx265 - Video On Demand ospray-studio: 3 - 4K - 32 - Path Tracer - CPU llama-cpp: llama-2-70b-chat.Q5_0.gguf blender: Pabellon Barcelona - CPU-Only pytorch: CPU - 16 - ResNet-152 ospray-studio: 2 - 4K - 32 - Path Tracer - CPU ospray-studio: 1 - 4K - 32 - Path Tracer - CPU cachebench: Read / Modify / Write cachebench: Write cachebench: Read blender: Classroom - CPU-Only pytorch: CPU - 1 - Efficientnet_v2_l openfoam: drivaerFastback, Small Mesh Size - Execution Time openfoam: drivaerFastback, Small Mesh Size - Mesh Time ospray-studio: 3 - 4K - 16 - Path Tracer - CPU deepsparse: BERT-Large, NLP Question Answering - Asynchronous Multi-Stream deepsparse: BERT-Large, NLP Question Answering - Asynchronous Multi-Stream ospray-studio: 3 - 4K - 1 - Path Tracer - CPU ospray-studio: 2 - 4K - 1 - Path Tracer - CPU ospray-studio: 1 - 4K - 1 - Path Tracer - CPU ospray-studio: 2 - 4K - 16 - Path Tracer - CPU ospray-studio: 1 - 4K - 16 - Path Tracer - CPU tensorflow: CPU - 16 - VGG-16 ffmpeg: libx265 - Live openvino: Face Detection FP16 - CPU openvino: Face Detection FP16 - CPU openvino: Face Detection FP16-INT8 - CPU openvino: Face Detection FP16-INT8 - CPU quantlib: Multi-Threaded pytorch: CPU - 1 - ResNet-152 openvino: Person Detection FP16 - CPU openvino: Person Detection FP16 - CPU openvino: Person Detection FP32 - CPU openvino: Person Detection FP32 - CPU openvino: Machine Translation EN To DE FP16 - CPU openvino: Machine Translation EN To DE FP16 - CPU openvino: Person Vehicle Bike Detection FP16 - CPU openvino: Person Vehicle Bike Detection FP16 - CPU openvino: Road Segmentation ADAS FP16-INT8 - CPU openvino: Road Segmentation ADAS FP16-INT8 - CPU openvino: Road Segmentation ADAS FP16 - CPU openvino: Road Segmentation ADAS FP16 - CPU openvino: Face Detection Retail FP16-INT8 - CPU openvino: Face Detection Retail FP16-INT8 - CPU openvino: Handwritten English Recognition FP16-INT8 - CPU openvino: Handwritten English Recognition FP16-INT8 - CPU openvino: Handwritten English Recognition FP16 - CPU openvino: Handwritten English Recognition FP16 - CPU openvino: Vehicle Detection FP16-INT8 - CPU openvino: Vehicle Detection FP16-INT8 - CPU openvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPU openvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPU openvino: Age Gender Recognition Retail 0013 FP16 - CPU openvino: Age Gender Recognition Retail 0013 FP16 - CPU openvino: Vehicle Detection FP16 - CPU openvino: Vehicle Detection FP16 - CPU openvino: Weld Porosity Detection FP16 - CPU openvino: Weld Porosity Detection FP16 - CPU speedb: Update Rand openvino: Face Detection Retail FP16 - CPU openvino: Face Detection Retail FP16 - CPU openvino: Weld Porosity Detection FP16-INT8 - CPU openvino: Weld Porosity Detection FP16-INT8 - CPU speedb: Read While Writing rocksdb: Update Rand speedb: Read Rand Write Rand speedb: Rand Read rocksdb: Read Rand Write Rand rocksdb: Read While Writing rocksdb: Rand Read dacapobench: Apache Cassandra blender: Fishy Cat - CPU-Only xmrig: Monero - 1M xmrig: CryptoNight-Femto UPX2 - 1M xmrig: KawPow - 1M xmrig: CryptoNight-Heavy - 1M build-linux-kernel: defconfig pytorch: CPU - 16 - ResNet-50 deepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Asynchronous Multi-Stream deepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Asynchronous Multi-Stream dacapobench: Eclipse blender: BMW27 - CPU-Only deepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Stream deepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Stream deepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Stream deepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Stream deepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Stream deepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Stream deepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Stream deepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Stream dacapobench: Apache Lucene Search Index xmrig: Wownero - 1M dacapobench: H2 Database Engine deepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Stream deepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Stream tensorflow: CPU - 16 - ResNet-50 deepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Stream deepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Stream deepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Stream deepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Stream deepsparse: ResNet-50, Baseline - Asynchronous Multi-Stream deepsparse: ResNet-50, Baseline - Asynchronous Multi-Stream deepsparse: ResNet-50, Sparse INT8 - Asynchronous Multi-Stream deepsparse: ResNet-50, Sparse INT8 - Asynchronous Multi-Stream deepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Stream deepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Stream dacapobench: Tradebeans compress-7zip: Decompression Rating compress-7zip: Compression Rating svt-av1: Preset 4 - Bosphorus 4K embree: Pathtracer - Asian Dragon Obj llama-cpp: llama-2-13b.Q4_0.gguf y-cruncher: 1B embree: Pathtracer ISPC - Asian Dragon Obj dacapobench: Tradesoap dacapobench: BioJava Biological Data Framework pytorch: CPU - 1 - ResNet-50 build-ffmpeg: Time To Compile dacapobench: Jython dacapobench: jMonkeyEngine dacapobench: GraphChi embree: Pathtracer - Crown embree: Pathtracer ISPC - Crown dacapobench: H2O In-Memory Platform For Machine Learning llama-cpp: llama-2-7b.Q4_0.gguf dacapobench: Apache Kafka embree: Pathtracer - Asian Dragon tensorflow: CPU - 1 - ResNet-50 dacapobench: Avrora AVR Simulation Framework embree: Pathtracer ISPC - Asian Dragon svt-av1: Preset 8 - Bosphorus 4K y-cruncher: 500M dacapobench: Spring Boot tensorflow: CPU - 16 - GoogLeNet tensorflow: CPU - 1 - VGG-16 dacapobench: Apache Tomcat dacapobench: Apache Lucene Search Engine svt-av1: Preset 13 - Bosphorus 4K svt-av1: Preset 12 - Bosphorus 4K dacapobench: PMD Source Code Analyzer tensorflow: CPU - 16 - AlexNet dacapobench: Batik SVG Toolkit tensorflow: CPU - 1 - GoogLeNet tensorflow: CPU - 1 - AlexNet dacapobench: FOP Print Formatter dacapobench: Apache Xalan XSLT dacapobench: Zxing 1D/2D Barcode Image Processing Zen 1 - EPYC 7601 b c 32 32 z 32 c 32 d 11426667 15013333 12996667 33.923 15.693 16270000 16140000 21180000 10.416 5.202 16260000 16150000 21250000 10.476 5.213 14320000 433.789 410.61 15350000 7.17 254.01 4067.4 18790000 22.28 45.13 45.18 136464 3.42 139.09 15.61 116566 116377 87227.587713 45646.091353 7616.087334 112.03 9.85 72.807288 28.372583 71361 607.935 26.0566 4049 3451 3404 61987 60673 25.15 109.84 929.23 17.17 486.65 32.82 107079.2 19.04 105.48 151.45 105.97 150.8 79.82 199.9 9.12 1741.57 23.95 666.22 27.69 576.18 5.41 5747.65 42.87 745 35.51 898.6 8.07 1960.18 0.48 52441.94 0.65 40123.62 13.36 1190.42 18.69 1704.26 314123 3.91 3921.5 9.56 3300.99 7457600 630575 2231403 179685954 2373654 4284691 176770468 5946 55.65 18845.5 18860.1 18777.2 19004.5 52.133 40.19 19.1107 836.4214 12656 44.73 747.0674 21.2933 41.6314 383.9746 747.314 21.2278 396.2914 40.1688 4613 25814.4 2675 87.4908 182.5085 51.34 129.8749 123.0157 128.8158 123.8817 59.8638 266.8574 7.2332 2208.1537 59.8833 266.8799 8561 212209 241545 5.801 37.284 17.94 11.676 38.9378 5403 7874 52.44 23.557 6703 6914 3536 36.9584 37.2967 3974 29.75 5110 41.5958 8.74 5613 45.9374 48.451 5.656 2444 158.47 9.73 2107 1402 185.665 186.625 1784 272.93 1733 28.99 32.12 751 871 609 14290000 434.187 410.43 15230000 7.11 272.61 4038.6 18760000 22.20 45.05 45.08 136312 3.41 138.6 15.51 116972 115669 87218.210974 45646.816107 7616.334142 112.09 9.82 71.201285 30.75472 71495 608.1326 26.0768 4048 3446 3406 62113 61430 25.2 110.37 927.57 17.18 486.03 32.81 107381.6 18.92 106.44 150.06 106.24 150.37 79.39 201.15 9.16 1735.64 23.95 666.3 27.53 579.41 5.42 5751.58 43.71 730.82 35.59 896.69 8.05 1964.99 0.47 52475.39 0.66 40101.8 13.29 1197.46 18.69 1704.02 314114 3.9 3924.86 9.56 3299.93 7210235 636242 2259344 179434924 2361270 4364996 177167636 5938 55.54 18763.8 18909 18961.3 18936.5 52.012 39.96 19.1338 835.262 12735 44.48 745.1806 21.2711 41.4377 385.6481 746.128 21.289 397.9593 39.9467 4589 25943.7 2655 87.325 182.7643 51.57 129.8035 122.955 128.845 123.785 59.8673 266.9761 7.261 2199.4941 59.6674 267.8417 8600 211584 242399 5.899 36.8586 17.87 11.595 39.107 5168 7858 52.78 23.759 6773 6917 3630 37.2545 37.6791 3868 29.9 5121 41.8198 8.77 5441 46.3088 58.715 5.685 2460 155.77 9.75 2082 1425 184.981 185.562 1820 274.97 1723 28.71 31.92 696 859 599 14430000 453.693 426.3 15180000 7.18 258.307 4136.3 1040000 22.21 45.13 44.95 139685 3.42 148.74 15.32 118980 118221 87238.013197 45645.091133 7615.948086 119.72 10.04 72.384007 30.537591 73024 611.6026 25.8175 4157 3515 3493 63402 62802 24.47 110.02 964.2 16.51 510.79 31.22 98916.2 18.86 106.43 150.07 105.91 150.84 82.18 194.21 9.39 1694.01 25.22 632.92 28.77 554.68 5.78 5416.31 46.17 692.02 37.4 853.38 8.52 1860.99 0.48 52382.31 0.67 39562.87 13.65 1166.56 19.58 1627.93 317758 4.03 3877.91 10.22 3099.2 7746346 633688 2229494 163202721 2327800 4419497 160665305 5955 59.58 18897.5 18887.5 18947.3 18783.9 53.615 40.32 19.5831 816.2785 12826 47.52 753.1229 20.8729 41.5889 384.3164 751.9259 21.0419 411.3435 38.7708 4580 25385.9 2773 88.1952 181.1043 51.56 130.4755 122.3307 129.5421 123.1469 59.9016 266.8428 7.2738 2195.9198 60.0613 266.034 8520 211815 240287 5.829 37.4405 17.87 11.902 39.0046 5366 7904 53.00 24.446 6865 6917 3538 35.9147 36.9967 3979 29.74 5111 41.5696 8.61 5561 45.4648 47.253 5.783 2533 157.6 9.77 2094 1379 183.899 180.955 1966 274.97 1718 27.73 33.14 764 852 569 14280000 452.606 426.37 15100000 7.15 258.934 4095.7 18840000 22.22 44.97 45.10 139445 3.42 148.56 15.35 119783 118802 87854.117672 45643.038713 7615.833145 119.57 10.21 72.305836 30.724194 73329 611.4439 25.7874 4132 3522 3499 62787 63336 24.51 110.29 965.35 16.54 510.9 31.2 98618.7 18.86 105.64 151.25 106.32 150.25 81.87 195.05 9.37 1696.5 25.16 634.5 28.82 553.65 5.78 5423.13 46.28 690.24 37.61 848.62 8.52 1862.24 0.48 52344.6 0.67 39843.05 13.65 1166.83 19.56 1628.91 313683 4.03 3869.7 10.21 3100.95 7105602 630478 2215896 163512432 2351568 4244478 160707812 5927 59.79 18866.1 18818.6 18901.1 18924 53.632 40.31 19.5858 815.9768 12768 47.41 751.2117 21.0932 41.8438 381.7839 750.3997 21.0667 410.3267 38.8343 4602 25396.8 2634 88.2278 181.1155 51.49 130.7937 121.8001 129.8101 122.9312 59.9698 266.5343 7.2896 2189.0655 60.0284 266.2776 8380 211383 241191 5.977 37.4056 18.08 11.975 39.1421 5149 7907 53.30 24.3 6769 6916 3656 36.2812 36.9369 3755 29.85 5114 41.557 8.59 5572 45.6482 58.642 5.751 2452 158.08 9.75 2112 1433 184.099 186.368 1833 276.19 1738 28.79 33.02 758 861 599 OpenBenchmarking.org
Quicksilver Input: CTS2 OpenBenchmarking.org Figure Of Merit, More Is Better Quicksilver 20230818 Input: CTS2 Zen 1 - EPYC 7601 32 d 32 z 32 32 c c b 3M 6M 9M 12M 15M SE +/- 16666.67, N = 3 11426667 14280000 14290000 14320000 14430000 16260000 16270000 1. (CXX) g++ options: -fopenmp -O3 -march=native
Timed Linux Kernel Compilation Build: allmodconfig OpenBenchmarking.org Seconds, Fewer Is Better Timed Linux Kernel Compilation 6.1 Build: allmodconfig 32 c 32 d 32 z 32 100 200 300 400 500 453.69 452.61 434.19 433.79
Blender Blend File: Barbershop - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.0 Blend File: Barbershop - Compute: CPU-Only 32 d 32 c 32 32 z 90 180 270 360 450 426.37 426.30 410.61 410.43
Quicksilver Input: CORAL2 P2 OpenBenchmarking.org Figure Of Merit, More Is Better Quicksilver 20230818 Input: CORAL2 P2 Zen 1 - EPYC 7601 32 d 32 c 32 z 32 b c 3M 6M 9M 12M 15M SE +/- 37118.43, N = 3 15013333 15100000 15180000 15230000 15350000 16140000 16150000 1. (CXX) g++ options: -fopenmp -O3 -march=native
PyTorch Device: CPU - Batch Size: 16 - Model: Efficientnet_v2_l OpenBenchmarking.org batches/sec, More Is Better PyTorch 2.1 Device: CPU - Batch Size: 16 - Model: Efficientnet_v2_l 32 z 32 d 32 32 c 2 4 6 8 10 7.11 7.15 7.17 7.18 MIN: 4.25 / MAX: 7.26 MIN: 4.34 / MAX: 7.3 MIN: 4.45 / MAX: 7.33 MIN: 4.37 / MAX: 7.37
Timed Gem5 Compilation Time To Compile OpenBenchmarking.org Seconds, Fewer Is Better Timed Gem5 Compilation 23.0.1 Time To Compile 32 z 32 d 32 c 32 60 120 180 240 300 272.61 258.93 258.31 254.01
Xmrig Variant: GhostRider - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: GhostRider - Hash Count: 1M 32 z 32 32 d 32 c 900 1800 2700 3600 4500 4038.6 4067.4 4095.7 4136.3 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
Quicksilver Input: CORAL2 P1 OpenBenchmarking.org Figure Of Merit, More Is Better Quicksilver 20230818 Input: CORAL2 P1 32 c Zen 1 - EPYC 7601 32 z 32 32 d b c 5M 10M 15M 20M 25M SE +/- 66916.20, N = 3 1040000 12996667 18760000 18790000 18840000 21180000 21250000 1. (CXX) g++ options: -fopenmp -O3 -march=native
FFmpeg Encoder: libx265 - Scenario: Upload OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx265 - Scenario: Upload 32 z 32 c 32 d 32 5 10 15 20 25 22.20 22.21 22.22 22.28 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
FFmpeg Encoder: libx265 - Scenario: Platform OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx265 - Scenario: Platform 32 d 32 z 32 32 c 10 20 30 40 50 44.97 45.05 45.13 45.13 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
FFmpeg Encoder: libx265 - Scenario: Video On Demand OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx265 - Scenario: Video On Demand 32 c 32 z 32 d 32 10 20 30 40 50 44.95 45.08 45.10 45.18 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
OSPRay Studio Camera: 3 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 0.13 Camera: 3 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU 32 c 32 d 32 32 z 30K 60K 90K 120K 150K 139685 139445 136464 136312
Llama.cpp Model: llama-2-70b-chat.Q5_0.gguf OpenBenchmarking.org Tokens Per Second, More Is Better Llama.cpp b1808 Model: llama-2-70b-chat.Q5_0.gguf 32 z 32 32 c 32 d 0.7695 1.539 2.3085 3.078 3.8475 3.41 3.42 3.42 3.42 1. (CXX) g++ options: -std=c++11 -fPIC -O3 -pthread -march=native -mtune=native -lopenblas
Blender Blend File: Pabellon Barcelona - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.0 Blend File: Pabellon Barcelona - Compute: CPU-Only 32 c 32 d 32 32 z 30 60 90 120 150 148.74 148.56 139.09 138.60
PyTorch Device: CPU - Batch Size: 16 - Model: ResNet-152 OpenBenchmarking.org batches/sec, More Is Better PyTorch 2.1 Device: CPU - Batch Size: 16 - Model: ResNet-152 32 c 32 d 32 z 32 4 8 12 16 20 15.32 15.35 15.51 15.61 MIN: 6.91 / MAX: 15.45 MIN: 8.86 / MAX: 15.52 MIN: 7.3 / MAX: 15.63 MIN: 6.89 / MAX: 15.74
OSPRay Studio Camera: 2 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 0.13 Camera: 2 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU 32 d 32 c 32 z 32 30K 60K 90K 120K 150K 119783 118980 116972 116566
OSPRay Studio Camera: 1 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 0.13 Camera: 1 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU 32 d 32 c 32 32 z 30K 60K 90K 120K 150K 118802 118221 116377 115669
CacheBench Test: Read / Modify / Write OpenBenchmarking.org MB/s, More Is Better CacheBench Test: Read / Modify / Write 32 z 32 32 c 32 d 20K 40K 60K 80K 100K 87218.21 87227.59 87238.01 87854.12 MIN: 65721.62 / MAX: 90703.93 MIN: 65739.52 / MAX: 90694.35 MIN: 65732.92 / MAX: 90706.91 MIN: 72077.93 / MAX: 90708.03 1. (CC) gcc options: -O3 -lrt
CacheBench Test: Write OpenBenchmarking.org MB/s, More Is Better CacheBench Test: Write 32 d 32 c 32 32 z 10K 20K 30K 40K 50K 45643.04 45645.09 45646.09 45646.82 MIN: 45482.26 / MAX: 45696.12 MIN: 45483.02 / MAX: 45696.19 MIN: 45484.29 / MAX: 45698.11 MIN: 45482.27 / MAX: 45698.03 1. (CC) gcc options: -O3 -lrt
CacheBench Test: Read OpenBenchmarking.org MB/s, More Is Better CacheBench Test: Read 32 d 32 c 32 32 z 1600 3200 4800 6400 8000 7615.83 7615.95 7616.09 7616.33 MIN: 7615.4 / MAX: 7616.44 MIN: 7615.46 / MAX: 7616.35 MIN: 7615.65 / MAX: 7616.54 MIN: 7615.95 / MAX: 7616.74 1. (CC) gcc options: -O3 -lrt
Blender Blend File: Classroom - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.0 Blend File: Classroom - Compute: CPU-Only 32 c 32 d 32 z 32 30 60 90 120 150 119.72 119.57 112.09 112.03
PyTorch Device: CPU - Batch Size: 1 - Model: Efficientnet_v2_l OpenBenchmarking.org batches/sec, More Is Better PyTorch 2.1 Device: CPU - Batch Size: 1 - Model: Efficientnet_v2_l 32 z 32 32 c 32 d 3 6 9 12 15 9.82 9.85 10.04 10.21 MIN: 5.63 / MAX: 10.05 MIN: 5.1 / MAX: 9.99 MIN: 5.86 / MAX: 10.23 MIN: 5.69 / MAX: 10.32
OpenFOAM Input: drivaerFastback, Small Mesh Size - Execution Time OpenBenchmarking.org Seconds, Fewer Is Better OpenFOAM 10 Input: drivaerFastback, Small Mesh Size - Execution Time 32 32 c 32 d 32 z 16 32 48 64 80 72.81 72.38 72.31 71.20 1. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm
OpenFOAM Input: drivaerFastback, Small Mesh Size - Mesh Time OpenBenchmarking.org Seconds, Fewer Is Better OpenFOAM 10 Input: drivaerFastback, Small Mesh Size - Mesh Time 32 z 32 d 32 c 32 7 14 21 28 35 30.75 30.72 30.54 28.37 1. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm
OSPRay Studio Camera: 3 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 0.13 Camera: 3 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU 32 d 32 c 32 z 32 16K 32K 48K 64K 80K 73329 73024 71495 71361
Neural Magic DeepSparse Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Stream 32 c 32 d 32 z 32 130 260 390 520 650 611.60 611.44 608.13 607.94
Neural Magic DeepSparse Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Stream 32 d 32 c 32 32 z 6 12 18 24 30 25.79 25.82 26.06 26.08
OSPRay Studio Camera: 3 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 0.13 Camera: 3 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU 32 c 32 d 32 32 z 900 1800 2700 3600 4500 4157 4132 4049 4048
OSPRay Studio Camera: 2 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 0.13 Camera: 2 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU 32 d 32 c 32 32 z 800 1600 2400 3200 4000 3522 3515 3451 3446
OSPRay Studio Camera: 1 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 0.13 Camera: 1 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU 32 d 32 c 32 z 32 800 1600 2400 3200 4000 3499 3493 3406 3404
OSPRay Studio Camera: 2 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 0.13 Camera: 2 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU 32 c 32 d 32 z 32 14K 28K 42K 56K 70K 63402 62787 62113 61987
OSPRay Studio Camera: 1 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 0.13 Camera: 1 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU 32 d 32 c 32 z 32 14K 28K 42K 56K 70K 63336 62802 61430 60673
TensorFlow Device: CPU - Batch Size: 16 - Model: VGG-16 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 16 - Model: VGG-16 32 c 32 d 32 32 z 6 12 18 24 30 24.47 24.51 25.15 25.20
FFmpeg Encoder: libx265 - Scenario: Live OpenBenchmarking.org FPS, More Is Better FFmpeg 6.1 Encoder: libx265 - Scenario: Live 32 32 c 32 d 32 z 20 40 60 80 100 109.84 110.02 110.29 110.37 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
OpenVINO Model: Face Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Face Detection FP16 - Device: CPU 32 d 32 c 32 32 z 200 400 600 800 1000 965.35 964.20 929.23 927.57 MIN: 922.7 / MAX: 1047.5 MIN: 905.78 / MAX: 1053.38 MIN: 907.01 / MAX: 1013.02 MIN: 895.6 / MAX: 1019.94 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Face Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Face Detection FP16 - Device: CPU 32 c 32 d 32 32 z 4 8 12 16 20 16.51 16.54 17.17 17.18 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Face Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Face Detection FP16-INT8 - Device: CPU 32 d 32 c 32 32 z 110 220 330 440 550 510.90 510.79 486.65 486.03 MIN: 470.7 / MAX: 595.97 MIN: 473.86 / MAX: 584.54 MIN: 465.68 / MAX: 570.73 MIN: 454.31 / MAX: 580.9 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Face Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Face Detection FP16-INT8 - Device: CPU 32 d 32 c 32 z 32 8 16 24 32 40 31.20 31.22 32.81 32.82 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
QuantLib Configuration: Multi-Threaded OpenBenchmarking.org MFLOPS, More Is Better QuantLib 1.32 Configuration: Multi-Threaded 32 d 32 c 32 32 z 20K 40K 60K 80K 100K 98618.7 98916.2 107079.2 107381.6 1. (CXX) g++ options: -O3 -march=native -fPIE -pie
PyTorch Device: CPU - Batch Size: 1 - Model: ResNet-152 OpenBenchmarking.org batches/sec, More Is Better PyTorch 2.1 Device: CPU - Batch Size: 1 - Model: ResNet-152 32 c 32 d 32 z 32 5 10 15 20 25 18.86 18.86 18.92 19.04 MIN: 10.78 / MAX: 19.02 MIN: 7.91 / MAX: 19.03 MIN: 7.59 / MAX: 19.04 MIN: 6.89 / MAX: 19.18
OpenVINO Model: Person Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Person Detection FP16 - Device: CPU 32 z 32 c 32 d 32 20 40 60 80 100 106.44 106.43 105.64 105.48 MIN: 81.71 / MAX: 196.1 MIN: 80.87 / MAX: 199.77 MIN: 54.2 / MAX: 154.42 MIN: 82.05 / MAX: 167.92 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Person Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Person Detection FP16 - Device: CPU 32 z 32 c 32 d 32 30 60 90 120 150 150.06 150.07 151.25 151.45 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Person Detection FP32 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Person Detection FP32 - Device: CPU 32 d 32 z 32 32 c 20 40 60 80 100 106.32 106.24 105.97 105.91 MIN: 81.37 / MAX: 177.41 MIN: 81.06 / MAX: 185.99 MIN: 81.88 / MAX: 218.45 MIN: 82.12 / MAX: 188.16 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Person Detection FP32 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Person Detection FP32 - Device: CPU 32 d 32 z 32 32 c 30 60 90 120 150 150.25 150.37 150.80 150.84 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Machine Translation EN To DE FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Machine Translation EN To DE FP16 - Device: CPU 32 c 32 d 32 32 z 20 40 60 80 100 82.18 81.87 79.82 79.39 MIN: 58.39 / MAX: 175.7 MIN: 52.13 / MAX: 175.84 MIN: 42.02 / MAX: 179.47 MIN: 43.97 / MAX: 186.13 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Machine Translation EN To DE FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Machine Translation EN To DE FP16 - Device: CPU 32 c 32 d 32 32 z 40 80 120 160 200 194.21 195.05 199.90 201.15 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Person Vehicle Bike Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Person Vehicle Bike Detection FP16 - Device: CPU 32 c 32 d 32 z 32 3 6 9 12 15 9.39 9.37 9.16 9.12 MIN: 5.95 / MAX: 68.66 MIN: 6.07 / MAX: 71.06 MIN: 5.99 / MAX: 67.91 MIN: 6.22 / MAX: 56.95 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Person Vehicle Bike Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Person Vehicle Bike Detection FP16 - Device: CPU 32 c 32 d 32 z 32 400 800 1200 1600 2000 1694.01 1696.50 1735.64 1741.57 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Road Segmentation ADAS FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Road Segmentation ADAS FP16-INT8 - Device: CPU 32 c 32 d 32 z 32 6 12 18 24 30 25.22 25.16 23.95 23.95 MIN: 21.61 / MAX: 89.16 MIN: 19.24 / MAX: 86.7 MIN: 15.19 / MAX: 90.71 MIN: 13.94 / MAX: 114.01 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Road Segmentation ADAS FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Road Segmentation ADAS FP16-INT8 - Device: CPU 32 c 32 d 32 32 z 140 280 420 560 700 632.92 634.50 666.22 666.30 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Road Segmentation ADAS FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Road Segmentation ADAS FP16 - Device: CPU 32 d 32 c 32 32 z 7 14 21 28 35 28.82 28.77 27.69 27.53 MIN: 19.39 / MAX: 99.16 MIN: 17.12 / MAX: 135.79 MIN: 18.56 / MAX: 147.54 MIN: 18.86 / MAX: 82.58 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Road Segmentation ADAS FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Road Segmentation ADAS FP16 - Device: CPU 32 d 32 c 32 32 z 130 260 390 520 650 553.65 554.68 576.18 579.41 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Face Detection Retail FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Face Detection Retail FP16-INT8 - Device: CPU 32 d 32 c 32 z 32 1.3005 2.601 3.9015 5.202 6.5025 5.78 5.78 5.42 5.41 MIN: 3.37 / MAX: 65.27 MIN: 3.21 / MAX: 58.78 MIN: 3.15 / MAX: 67.23 MIN: 3.17 / MAX: 57.08 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Face Detection Retail FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Face Detection Retail FP16-INT8 - Device: CPU 32 c 32 d 32 32 z 1200 2400 3600 4800 6000 5416.31 5423.13 5747.65 5751.58 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Handwritten English Recognition FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Handwritten English Recognition FP16-INT8 - Device: CPU 32 d 32 c 32 z 32 10 20 30 40 50 46.28 46.17 43.71 42.87 MIN: 30.15 / MAX: 108.49 MIN: 39.81 / MAX: 161.92 MIN: 35.06 / MAX: 153.84 MIN: 35.14 / MAX: 107.5 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Handwritten English Recognition FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Handwritten English Recognition FP16-INT8 - Device: CPU 32 d 32 c 32 z 32 160 320 480 640 800 690.24 692.02 730.82 745.00 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Handwritten English Recognition FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Handwritten English Recognition FP16 - Device: CPU 32 d 32 c 32 z 32 9 18 27 36 45 37.61 37.40 35.59 35.51 MIN: 24.11 / MAX: 127.49 MIN: 27.33 / MAX: 92.33 MIN: 24.72 / MAX: 147.24 MIN: 22.8 / MAX: 100.53 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Handwritten English Recognition FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Handwritten English Recognition FP16 - Device: CPU 32 d 32 c 32 z 32 200 400 600 800 1000 848.62 853.38 896.69 898.60 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Vehicle Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Vehicle Detection FP16-INT8 - Device: CPU 32 d 32 c 32 32 z 2 4 6 8 10 8.52 8.52 8.07 8.05 MIN: 4.8 / MAX: 75.53 MIN: 4.97 / MAX: 67.6 MIN: 4.55 / MAX: 69.24 MIN: 4.56 / MAX: 76.48 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Vehicle Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Vehicle Detection FP16-INT8 - Device: CPU 32 c 32 d 32 32 z 400 800 1200 1600 2000 1860.99 1862.24 1960.18 1964.99 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU 32 d 32 c 32 32 z 0.108 0.216 0.324 0.432 0.54 0.48 0.48 0.48 0.47 MIN: 0.27 / MAX: 65.55 MIN: 0.27 / MAX: 50.17 MIN: 0.27 / MAX: 50.11 MIN: 0.27 / MAX: 64.47 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU 32 d 32 c 32 32 z 11K 22K 33K 44K 55K 52344.60 52382.31 52441.94 52475.39 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU 32 d 32 c 32 z 32 0.1508 0.3016 0.4524 0.6032 0.754 0.67 0.67 0.66 0.65 MIN: 0.36 / MAX: 50.74 MIN: 0.36 / MAX: 62.87 MIN: 0.36 / MAX: 65.79 MIN: 0.36 / MAX: 51.48 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU 32 c 32 d 32 z 32 9K 18K 27K 36K 45K 39562.87 39843.05 40101.80 40123.62 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Vehicle Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Vehicle Detection FP16 - Device: CPU 32 d 32 c 32 32 z 4 8 12 16 20 13.65 13.65 13.36 13.29 MIN: 6.73 / MAX: 75.18 MIN: 9.08 / MAX: 67.03 MIN: 7.26 / MAX: 78.85 MIN: 8.3 / MAX: 73.59 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Vehicle Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Vehicle Detection FP16 - Device: CPU 32 c 32 d 32 32 z 300 600 900 1200 1500 1166.56 1166.83 1190.42 1197.46 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Weld Porosity Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Weld Porosity Detection FP16 - Device: CPU 32 c 32 d 32 z 32 5 10 15 20 25 19.58 19.56 18.69 18.69 MIN: 10.24 / MAX: 83.63 MIN: 13.73 / MAX: 73.6 MIN: 9.78 / MAX: 86.93 MIN: 9.97 / MAX: 81.33 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Weld Porosity Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Weld Porosity Detection FP16 - Device: CPU 32 c 32 d 32 z 32 400 800 1200 1600 2000 1627.93 1628.91 1704.02 1704.26 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
Speedb Test: Update Random OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Update Random 32 d 32 z 32 32 c 70K 140K 210K 280K 350K 313683 314114 314123 317758 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
OpenVINO Model: Face Detection Retail FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Face Detection Retail FP16 - Device: CPU 32 d 32 c 32 32 z 0.9068 1.8136 2.7204 3.6272 4.534 4.03 4.03 3.91 3.90 MIN: 2.23 / MAX: 62.26 MIN: 2.23 / MAX: 54.09 MIN: 2.2 / MAX: 72.73 MIN: 2.18 / MAX: 64.81 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Face Detection Retail FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Face Detection Retail FP16 - Device: CPU 32 d 32 c 32 32 z 800 1600 2400 3200 4000 3869.70 3877.91 3921.50 3924.86 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Weld Porosity Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Weld Porosity Detection FP16-INT8 - Device: CPU 32 c 32 d 32 z 32 3 6 9 12 15 10.22 10.21 9.56 9.56 MIN: 5.48 / MAX: 68.07 MIN: 5.17 / MAX: 61.15 MIN: 5.09 / MAX: 75.37 MIN: 5.1 / MAX: 77.12 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
OpenVINO Model: Weld Porosity Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Weld Porosity Detection FP16-INT8 - Device: CPU 32 c 32 d 32 z 32 700 1400 2100 2800 3500 3099.20 3100.95 3299.93 3300.99 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie
Speedb Test: Read While Writing OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Read While Writing 32 d 32 z 32 32 c 1.7M 3.4M 5.1M 6.8M 8.5M 7105602 7210235 7457600 7746346 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
RocksDB Test: Update Random OpenBenchmarking.org Op/s, More Is Better RocksDB 8.0 Test: Update Random 32 d 32 32 c 32 z 140K 280K 420K 560K 700K 630478 630575 633688 636242 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
Speedb Test: Read Random Write Random OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Read Random Write Random 32 d 32 c 32 32 z 500K 1000K 1500K 2000K 2500K 2215896 2229494 2231403 2259344 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
Speedb Test: Random Read OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Random Read 32 c 32 d 32 z 32 40M 80M 120M 160M 200M 163202721 163512432 179434924 179685954 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
RocksDB Test: Read Random Write Random OpenBenchmarking.org Op/s, More Is Better RocksDB 8.0 Test: Read Random Write Random 32 c 32 d 32 z 32 500K 1000K 1500K 2000K 2500K 2327800 2351568 2361270 2373654 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
RocksDB Test: Read While Writing OpenBenchmarking.org Op/s, More Is Better RocksDB 8.0 Test: Read While Writing 32 d 32 32 z 32 c 900K 1800K 2700K 3600K 4500K 4244478 4284691 4364996 4419497 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
RocksDB Test: Random Read OpenBenchmarking.org Op/s, More Is Better RocksDB 8.0 Test: Random Read 32 c 32 d 32 32 z 40M 80M 120M 160M 200M 160665305 160707812 176770468 177167636 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
DaCapo Benchmark Java Test: Apache Cassandra OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Apache Cassandra 32 c 32 32 z 32 d 1300 2600 3900 5200 6500 5955 5946 5938 5927
Blender Blend File: Fishy Cat - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.0 Blend File: Fishy Cat - Compute: CPU-Only 32 d 32 c 32 32 z 13 26 39 52 65 59.79 59.58 55.65 55.54
Xmrig Variant: Monero - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: Monero - Hash Count: 1M 32 z 32 32 d 32 c 4K 8K 12K 16K 20K 18763.8 18845.5 18866.1 18897.5 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
Xmrig Variant: CryptoNight-Femto UPX2 - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: CryptoNight-Femto UPX2 - Hash Count: 1M 32 d 32 32 c 32 z 4K 8K 12K 16K 20K 18818.6 18860.1 18887.5 18909.0 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
Xmrig Variant: KawPow - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: KawPow - Hash Count: 1M 32 32 d 32 c 32 z 4K 8K 12K 16K 20K 18777.2 18901.1 18947.3 18961.3 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
Xmrig Variant: CryptoNight-Heavy - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: CryptoNight-Heavy - Hash Count: 1M 32 c 32 d 32 z 32 4K 8K 12K 16K 20K 18783.9 18924.0 18936.5 19004.5 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
Timed Linux Kernel Compilation Build: defconfig OpenBenchmarking.org Seconds, Fewer Is Better Timed Linux Kernel Compilation 6.1 Build: defconfig 32 d 32 c 32 32 z 12 24 36 48 60 53.63 53.62 52.13 52.01
PyTorch Device: CPU - Batch Size: 16 - Model: ResNet-50 OpenBenchmarking.org batches/sec, More Is Better PyTorch 2.1 Device: CPU - Batch Size: 16 - Model: ResNet-50 32 z 32 32 d 32 c 9 18 27 36 45 39.96 40.19 40.31 40.32 MIN: 15.13 / MAX: 40.53 MIN: 15.55 / MAX: 40.67 MIN: 15.27 / MAX: 40.73 MIN: 15.51 / MAX: 40.87
Neural Magic DeepSparse Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream 32 d 32 c 32 z 32 5 10 15 20 25 19.59 19.58 19.13 19.11
Neural Magic DeepSparse Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream 32 d 32 c 32 z 32 200 400 600 800 1000 815.98 816.28 835.26 836.42
DaCapo Benchmark Java Test: Eclipse OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Eclipse 32 c 32 d 32 z 32 3K 6K 9K 12K 15K 12826 12768 12735 12656
Blender Blend File: BMW27 - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 4.0 Blend File: BMW27 - Compute: CPU-Only 32 c 32 d 32 32 z 11 22 33 44 55 47.52 47.41 44.73 44.48
Neural Magic DeepSparse Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream 32 c 32 d 32 32 z 160 320 480 640 800 753.12 751.21 747.07 745.18
Neural Magic DeepSparse Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream 32 c 32 d 32 z 32 5 10 15 20 25 20.87 21.09 21.27 21.29
Neural Magic DeepSparse Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream 32 d 32 32 c 32 z 10 20 30 40 50 41.84 41.63 41.59 41.44
Neural Magic DeepSparse Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream 32 d 32 32 c 32 z 80 160 240 320 400 381.78 383.97 384.32 385.65
Neural Magic DeepSparse Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream 32 c 32 d 32 32 z 160 320 480 640 800 751.93 750.40 747.31 746.13
Neural Magic DeepSparse Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream 32 c 32 d 32 32 z 5 10 15 20 25 21.04 21.07 21.23 21.29
Neural Magic DeepSparse Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream 32 c 32 d 32 z 32 90 180 270 360 450 411.34 410.33 397.96 396.29
Neural Magic DeepSparse Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream 32 c 32 d 32 z 32 9 18 27 36 45 38.77 38.83 39.95 40.17
DaCapo Benchmark Java Test: Apache Lucene Search Index OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Apache Lucene Search Index 32 32 d 32 z 32 c 1000 2000 3000 4000 5000 4613 4602 4589 4580
Xmrig Variant: Wownero - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.21 Variant: Wownero - Hash Count: 1M 32 c 32 d 32 32 z 6K 12K 18K 24K 30K 25385.9 25396.8 25814.4 25943.7 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
DaCapo Benchmark Java Test: H2 Database Engine OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: H2 Database Engine 32 c 32 32 z 32 d 600 1200 1800 2400 3000 2773 2675 2655 2634
Neural Magic DeepSparse Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream 32 d 32 c 32 32 z 20 40 60 80 100 88.23 88.20 87.49 87.33
Neural Magic DeepSparse Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream 32 c 32 d 32 32 z 40 80 120 160 200 181.10 181.12 182.51 182.76
TensorFlow Device: CPU - Batch Size: 16 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 16 - Model: ResNet-50 32 32 d 32 c 32 z 12 24 36 48 60 51.34 51.49 51.56 51.57
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream 32 d 32 c 32 32 z 30 60 90 120 150 130.79 130.48 129.87 129.80
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream 32 d 32 c 32 z 32 30 60 90 120 150 121.80 122.33 122.96 123.02
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream 32 d 32 c 32 z 32 30 60 90 120 150 129.81 129.54 128.85 128.82
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream 32 d 32 c 32 z 32 30 60 90 120 150 122.93 123.15 123.79 123.88
Neural Magic DeepSparse Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream 32 d 32 c 32 z 32 13 26 39 52 65 59.97 59.90 59.87 59.86
Neural Magic DeepSparse Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream 32 d 32 c 32 32 z 60 120 180 240 300 266.53 266.84 266.86 266.98
Neural Magic DeepSparse Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream 32 d 32 c 32 z 32 2 4 6 8 10 7.2896 7.2738 7.2610 7.2332
Neural Magic DeepSparse Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream 32 d 32 c 32 z 32 500 1000 1500 2000 2500 2189.07 2195.92 2199.49 2208.15
Neural Magic DeepSparse Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream 32 c 32 d 32 32 z 13 26 39 52 65 60.06 60.03 59.88 59.67
Neural Magic DeepSparse Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream 32 c 32 d 32 32 z 60 120 180 240 300 266.03 266.28 266.88 267.84
DaCapo Benchmark Java Test: Tradebeans OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Tradebeans 32 z 32 32 c 32 d 2K 4K 6K 8K 10K 8600 8561 8520 8380
7-Zip Compression Test: Decompression Rating OpenBenchmarking.org MIPS, More Is Better 7-Zip Compression 22.01 Test: Decompression Rating 32 d 32 z 32 c 32 50K 100K 150K 200K 250K 211383 211584 211815 212209 1. (CXX) g++ options: -lpthread -ldl -O2 -fPIC
7-Zip Compression Test: Compression Rating OpenBenchmarking.org MIPS, More Is Better 7-Zip Compression 22.01 Test: Compression Rating 32 c 32 d 32 32 z 50K 100K 150K 200K 250K 240287 241191 241545 242399 1. (CXX) g++ options: -lpthread -ldl -O2 -fPIC
SVT-AV1 Encoder Mode: Preset 4 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.8 Encoder Mode: Preset 4 - Input: Bosphorus 4K 32 32 c 32 z 32 d 1.3448 2.6896 4.0344 5.3792 6.724 5.801 5.829 5.899 5.977 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
Embree Binary: Pathtracer - Model: Asian Dragon Obj OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer - Model: Asian Dragon Obj 32 z 32 32 d 32 c 9 18 27 36 45 36.86 37.28 37.41 37.44 MIN: 36.67 / MAX: 37.11 MIN: 37.09 / MAX: 37.7 MIN: 37.22 / MAX: 37.69 MIN: 37.24 / MAX: 37.71
Llama.cpp Model: llama-2-13b.Q4_0.gguf OpenBenchmarking.org Tokens Per Second, More Is Better Llama.cpp b1808 Model: llama-2-13b.Q4_0.gguf 32 z 32 c 32 32 d 4 8 12 16 20 17.87 17.87 17.94 18.08 1. (CXX) g++ options: -std=c++11 -fPIC -O3 -pthread -march=native -mtune=native -lopenblas
Y-Cruncher Pi Digits To Calculate: 1B OpenBenchmarking.org Seconds, Fewer Is Better Y-Cruncher 0.8.3 Pi Digits To Calculate: 1B Zen 1 - EPYC 7601 32 d 32 c 32 32 z c b 8 16 24 32 40 SE +/- 0.09, N = 3 33.92 11.98 11.90 11.68 11.60 10.48 10.42
Embree Binary: Pathtracer ISPC - Model: Asian Dragon Obj OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Asian Dragon Obj 32 32 c 32 z 32 d 9 18 27 36 45 38.94 39.00 39.11 39.14 MIN: 38.69 / MAX: 39.29 MIN: 38.78 / MAX: 39.64 MIN: 38.88 / MAX: 39.43 MIN: 38.92 / MAX: 39.84
DaCapo Benchmark Java Test: Tradesoap OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Tradesoap 32 32 c 32 z 32 d 1200 2400 3600 4800 6000 5403 5366 5168 5149
DaCapo Benchmark Java Test: BioJava Biological Data Framework OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: BioJava Biological Data Framework 32 d 32 c 32 32 z 2K 4K 6K 8K 10K 7907 7904 7874 7858
PyTorch Device: CPU - Batch Size: 1 - Model: ResNet-50 OpenBenchmarking.org batches/sec, More Is Better PyTorch 2.1 Device: CPU - Batch Size: 1 - Model: ResNet-50 32 32 z 32 c 32 d 12 24 36 48 60 52.44 52.78 53.00 53.30 MIN: 15.02 / MAX: 53.14 MIN: 17.43 / MAX: 53.32 MIN: 50.62 / MAX: 53.51 MIN: 50.97 / MAX: 53.84
Timed FFmpeg Compilation Time To Compile OpenBenchmarking.org Seconds, Fewer Is Better Timed FFmpeg Compilation 6.1 Time To Compile 32 c 32 d 32 z 32 6 12 18 24 30 24.45 24.30 23.76 23.56
DaCapo Benchmark Java Test: Jython OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Jython 32 c 32 z 32 d 32 1500 3000 4500 6000 7500 6865 6773 6769 6703
DaCapo Benchmark Java Test: jMonkeyEngine OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: jMonkeyEngine 32 c 32 z 32 d 32 1500 3000 4500 6000 7500 6917 6917 6916 6914
DaCapo Benchmark Java Test: GraphChi OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: GraphChi 32 d 32 z 32 c 32 800 1600 2400 3200 4000 3656 3630 3538 3536
Embree Binary: Pathtracer - Model: Crown OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer - Model: Crown 32 c 32 d 32 32 z 9 18 27 36 45 35.91 36.28 36.96 37.25 MIN: 35.53 / MAX: 37.08 MIN: 35.88 / MAX: 37.13 MIN: 36.61 / MAX: 37.43 MIN: 36.89 / MAX: 37.75
Embree Binary: Pathtracer ISPC - Model: Crown OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Crown 32 d 32 c 32 32 z 9 18 27 36 45 36.94 37.00 37.30 37.68 MIN: 36.46 / MAX: 37.76 MIN: 36.53 / MAX: 38.11 MIN: 36.86 / MAX: 38.04 MIN: 37.25 / MAX: 38.37
DaCapo Benchmark Java Test: H2O In-Memory Platform For Machine Learning OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: H2O In-Memory Platform For Machine Learning 32 c 32 32 z 32 d 900 1800 2700 3600 4500 3979 3974 3868 3755
Llama.cpp Model: llama-2-7b.Q4_0.gguf OpenBenchmarking.org Tokens Per Second, More Is Better Llama.cpp b1808 Model: llama-2-7b.Q4_0.gguf 32 c 32 32 d 32 z 7 14 21 28 35 29.74 29.75 29.85 29.90 1. (CXX) g++ options: -std=c++11 -fPIC -O3 -pthread -march=native -mtune=native -lopenblas
DaCapo Benchmark Java Test: Apache Kafka OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Apache Kafka 32 z 32 d 32 c 32 1100 2200 3300 4400 5500 5121 5114 5111 5110
Embree Binary: Pathtracer - Model: Asian Dragon OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer - Model: Asian Dragon 32 d 32 c 32 32 z 10 20 30 40 50 41.56 41.57 41.60 41.82 MIN: 41.33 / MAX: 41.84 MIN: 41.37 / MAX: 41.9 MIN: 41.36 / MAX: 41.86 MIN: 41.6 / MAX: 42.16
TensorFlow Device: CPU - Batch Size: 1 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 1 - Model: ResNet-50 32 d 32 c 32 32 z 2 4 6 8 10 8.59 8.61 8.74 8.77
DaCapo Benchmark Java Test: Avrora AVR Simulation Framework OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Avrora AVR Simulation Framework 32 32 d 32 c 32 z 1200 2400 3600 4800 6000 5613 5572 5561 5441
Embree Binary: Pathtracer ISPC - Model: Asian Dragon OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Asian Dragon 32 c 32 d 32 32 z 11 22 33 44 55 45.46 45.65 45.94 46.31 MIN: 45.22 / MAX: 46.6 MIN: 45.37 / MAX: 46.89 MIN: 45.66 / MAX: 46.38 MIN: 46.05 / MAX: 46.74
SVT-AV1 Encoder Mode: Preset 8 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.8 Encoder Mode: Preset 8 - Input: Bosphorus 4K 32 c 32 32 d 32 z 13 26 39 52 65 47.25 48.45 58.64 58.72 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
Y-Cruncher Pi Digits To Calculate: 500M OpenBenchmarking.org Seconds, Fewer Is Better Y-Cruncher 0.8.3 Pi Digits To Calculate: 500M Zen 1 - EPYC 7601 32 c 32 d 32 z 32 c b 4 8 12 16 20 SE +/- 0.118, N = 3 15.693 5.783 5.751 5.685 5.656 5.213 5.202
DaCapo Benchmark Java Test: Spring Boot OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Spring Boot 32 c 32 z 32 d 32 500 1000 1500 2000 2500 2533 2460 2452 2444
TensorFlow Device: CPU - Batch Size: 16 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 16 - Model: GoogLeNet 32 z 32 c 32 d 32 40 80 120 160 200 155.77 157.60 158.08 158.47
TensorFlow Device: CPU - Batch Size: 1 - Model: VGG-16 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 1 - Model: VGG-16 32 32 z 32 d 32 c 3 6 9 12 15 9.73 9.75 9.75 9.77
DaCapo Benchmark Java Test: Apache Tomcat OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Apache Tomcat 32 d 32 32 c 32 z 500 1000 1500 2000 2500 2112 2107 2094 2082
DaCapo Benchmark Java Test: Apache Lucene Search Engine OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Apache Lucene Search Engine 32 d 32 z 32 32 c 300 600 900 1200 1500 1433 1425 1402 1379
SVT-AV1 Encoder Mode: Preset 13 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.8 Encoder Mode: Preset 13 - Input: Bosphorus 4K 32 c 32 d 32 z 32 40 80 120 160 200 183.90 184.10 184.98 185.67 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 12 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.8 Encoder Mode: Preset 12 - Input: Bosphorus 4K 32 c 32 z 32 d 32 40 80 120 160 200 180.96 185.56 186.37 186.63 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
DaCapo Benchmark Java Test: PMD Source Code Analyzer OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: PMD Source Code Analyzer 32 c 32 d 32 z 32 400 800 1200 1600 2000 1966 1833 1820 1784
TensorFlow Device: CPU - Batch Size: 16 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 16 - Model: AlexNet 32 32 z 32 c 32 d 60 120 180 240 300 272.93 274.97 274.97 276.19
DaCapo Benchmark Java Test: Batik SVG Toolkit OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Batik SVG Toolkit 32 d 32 32 z 32 c 400 800 1200 1600 2000 1738 1733 1723 1718
TensorFlow Device: CPU - Batch Size: 1 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 1 - Model: GoogLeNet 32 c 32 z 32 d 32 7 14 21 28 35 27.73 28.71 28.79 28.99
TensorFlow Device: CPU - Batch Size: 1 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 1 - Model: AlexNet 32 z 32 32 d 32 c 8 16 24 32 40 31.92 32.12 33.02 33.14
DaCapo Benchmark Java Test: FOP Print Formatter OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: FOP Print Formatter 32 c 32 d 32 32 z 160 320 480 640 800 764 758 751 696
DaCapo Benchmark Java Test: Apache Xalan XSLT OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Apache Xalan XSLT 32 32 d 32 z 32 c 200 400 600 800 1000 871 861 859 852
DaCapo Benchmark Java Test: Zxing 1D/2D Barcode Image Processing OpenBenchmarking.org msec, Fewer Is Better DaCapo Benchmark 23.11 Java Test: Zxing 1D/2D Barcode Image Processing 32 32 d 32 z 32 c 130 260 390 520 650 609 599 599 569
CPU Power Consumption Monitor Phoronix Test Suite System Monitoring OpenBenchmarking.org Watts CPU Power Consumption Monitor Phoronix Test Suite System Monitoring Zen 1 - EPYC 7601 130 260 390 520 650 Min: 242.58 / Avg: 585.92 / Max: 718
Meta Performance Per Watts Performance Per Watts OpenBenchmarking.org Performance Per Watts, More Is Better Meta Performance Per Watts Performance Per Watts Zen 1 - EPYC 7601 3M 6M 9M 12M 15M 13064001.66
Y-Cruncher CPU Power Consumption Monitor Min Avg Max Zen 1 - EPYC 7601 263 602 718 OpenBenchmarking.org Watts, Fewer Is Better Y-Cruncher 0.8.3 CPU Power Consumption Monitor 200 400 600 800 1000
Y-Cruncher CPU Power Consumption Monitor Min Avg Max Zen 1 - EPYC 7601 262 543 712 OpenBenchmarking.org Watts, Fewer Is Better Y-Cruncher 0.8.3 CPU Power Consumption Monitor 200 400 600 800 1000
Quicksilver CPU Power Consumption Monitor OpenBenchmarking.org Watts, Fewer Is Better Quicksilver 20230818 CPU Power Consumption Monitor Zen 1 - EPYC 7601 120 240 360 480 600 Min: 258.88 / Avg: 624.15 / Max: 662.04
Quicksilver Input: CTS2 OpenBenchmarking.org Figure Of Merit Per Watt, More Is Better Quicksilver 20230818 Input: CTS2 Zen 1 - EPYC 7601 4K 8K 12K 16K 20K 18307.66
Quicksilver CPU Power Consumption Monitor Min Avg Max Zen 1 - EPYC 7601 255.2 553.7 594.9 OpenBenchmarking.org Watts, Fewer Is Better Quicksilver 20230818 CPU Power Consumption Monitor 160 320 480 640 800
Quicksilver Input: CORAL2 P2 OpenBenchmarking.org Figure Of Merit Per Watt, More Is Better Quicksilver 20230818 Input: CORAL2 P2 Zen 1 - EPYC 7601 6K 12K 18K 24K 30K 27116.87
Quicksilver CPU Power Consumption Monitor Min Avg Max Zen 1 - EPYC 7601 243 584 648 OpenBenchmarking.org Watts, Fewer Is Better Quicksilver 20230818 CPU Power Consumption Monitor 200 400 600 800 1000
Quicksilver Input: CORAL2 P1 OpenBenchmarking.org Figure Of Merit Per Watt, More Is Better Quicksilver 20230818 Input: CORAL2 P1 Zen 1 - EPYC 7601 5K 10K 15K 20K 25K 22248.55
Phoronix Test Suite v10.8.4