AVX-512 Analysis

AMD Ryzen 7 7840U testing with a PHX Ray_PEU (V1.04 BIOS) and AMD Phoenix1 512MB on Ubuntu 23.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2307083-NE-AVX512ANA67
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results

Limit displaying results to tests within:

CPU Massive 4 Tests
Creator Workloads 6 Tests
Cryptography 2 Tests
HPC - High Performance Computing 5 Tests
Machine Learning 4 Tests
Multi-Core 7 Tests
Intel oneAPI 6 Tests
Python Tests 2 Tests
Raytracing 2 Tests
Renderers 2 Tests
Server CPU Tests 2 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Disable Color Branding
Prefer Vertical Bar Graphs
No Box Plots
On Line Graphs With Missing Data, Connect The Line Gaps

Multi-Way Comparison

Condense Comparison
Transpose Comparison

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
Ryzen 7 7840U: AVX512 Off
July 08 2023
  9 Hours, 49 Minutes
Ryzen 7 7840U: AVX512 On
July 08 2023
  8 Hours, 19 Minutes
Invert Hiding All Results Option
  9 Hours, 4 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


AVX-512 AnalysisOpenBenchmarking.orgPhoronix Test SuiteAMD Ryzen 7 7840U @ 3.30GHz (8 Cores / 16 Threads)PHX Ray_PEU (V1.04 BIOS)AMD Device 14e816GB1024GB Micron_3400_MTFDKBA1T0TFHAMD Phoenix1 512MB (2700/400MHz)AMD Rembrandt Radeon HD AudioMEDIATEK MT7922 802.11ax PCIUbuntu 23.046.2.0-24-generic (x86_64)KDE Plasma 5.27.4X Server 1.21.1.74.6 Mesa 23.0.2 (LLVM 15.0.7 DRM 3.49)GCC 12.2.0ext43200x2000ProcessorMotherboardChipsetMemoryDiskGraphicsAudioNetworkOSKernelDesktopDisplay ServerOpenGLCompilerFile-SystemScreen ResolutionAVX-512 Analysis PerformanceSystem Logs- Transparent Huge Pages: madvise- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0xa704101 - Python 3.11.2- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected

openvino: Face Detection FP16-INT8 - CPUopenvino: Face Detection FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16 - CPUopenvino: Age Gender Recognition Retail 0013 FP16 - CPUopenvino: Machine Translation EN To DE FP16 - CPUopenvino: Machine Translation EN To DE FP16 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUcpuminer-opt: LBC, LBRY Creditstensorflow: CPU - 16 - GoogLeNettensorflow: CPU - 64 - ResNet-50openvino: Person Detection FP16 - CPUopenvino: Person Detection FP16 - CPUopenvino: Weld Porosity Detection FP16 - CPUopenvino: Person Detection FP32 - CPUopenvino: Weld Porosity Detection FP16 - CPUopenvino: Person Detection FP32 - CPUtensorflow: CPU - 64 - GoogLeNetcpuminer-opt: Myriad-Groestlonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUtensorflow: CPU - 16 - ResNet-50openvino: Weld Porosity Detection FP16-INT8 - CPUopenvino: Weld Porosity Detection FP16-INT8 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUtensorflow: CPU - 64 - AlexNetospray: gravity_spheres_volume/dim_512/scivis/real_timecpuminer-opt: Blake-2 Stensorflow: CPU - 16 - AlexNetospray: gravity_spheres_volume/dim_512/ao/real_timeopenvino: Person Vehicle Bike Detection FP16 - CPUopenvino: Person Vehicle Bike Detection FP16 - CPUopenvino: Vehicle Detection FP16-INT8 - CPUopenvino: Vehicle Detection FP16-INT8 - CPUcpuminer-opt: Quad SHA-256, Pyriteopenvino: Vehicle Detection FP16 - CPUopenvino: Vehicle Detection FP16 - CPUcpuminer-opt: Garlicoincpuminer-opt: scryptminibude: OpenMP - BM1minibude: OpenMP - BM1ospray: gravity_spheres_volume/dim_512/pathtracer/real_timeopenvkl: vklBenchmark ISPCembree: Pathtracer ISPC - Asian Dragoncpuminer-opt: Skeincoinembree: Pathtracer ISPC - Asian Dragon Objospray-studio: 3 - 4K - 1 - Path Tracerospray-studio: 3 - 4K - 32 - Path Traceropenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUospray-studio: 3 - 1080p - 32 - Path Traceronednn: Recurrent Neural Network Inference - u8s8f32 - CPUembree: Pathtracer ISPC - Crownospray-studio: 3 - 1080p - 1 - Path Tracersmhasher: FarmHash32 x86_64 AVXospray-studio: 1 - 4K - 1 - Path Tracerospray-studio: 1 - 4K - 32 - Path Tracerospray: particle_volume/pathtracer/real_timelibxsmm: 64cpuminer-opt: x25xospray-studio: 1 - 1080p - 32 - Path Tracerospray-studio: 1 - 1080p - 1 - Path Tracercpuminer-opt: Magismhasher: FarmHash32 x86_64 AVXopenvino: Face Detection FP16 - CPUopenvino: Face Detection FP16 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUlczero: EigenRyzen 7 7840U AVX512 Off AVX512 On3.181252.114781.191.6517.13233.173.369982203025.788.544390.070.9123.484387.55170.200.9124.87139102.869458.8423.20344.821.6805864.791.1372431592546.211.1948718.83212.1516.84237.455853330.53131.132009.38128.05290.69711.6281.95764928.0917639077.2329224797395751.495336.561878532748.206.8556570736442.761808859792393.5182155.3348.691477644456365.6126.4382570.311.565003.234636.91578.3510275.450.7736.46109.681.616334390351.1916.742242.451.7812.052252.82331.311.7748.05261601.5270716.4412.49639.050.944432113.581.9781453147777.651.9887011.88336.1311.10359.978762721.53185.722752.27166.57372.10014.8842.469561149.9584783008.5595190646287671.276240.041612582371.627.9186506040077.5916567548903101.808168.0374.951385534192351.2126.3901143.203.504354.17655OpenBenchmarking.org

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Face Detection FP16-INT8 - Device: CPUAVX512 OffAVX512 On246810SE +/- 0.02, N = 3SE +/- 0.01, N = 33.186.911. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Face Detection FP16-INT8 - Device: CPUAVX512 OffAVX512 On3691215Min: 3.14 / Avg: 3.18 / Max: 3.22Min: 6.89 / Avg: 6.91 / Max: 6.931. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Face Detection FP16-INT8 - Device: CPUAVX512 OffAVX512 On30060090012001500SE +/- 9.28, N = 3SE +/- 0.78, N = 31252.11578.35MIN: 1052.1 / MAX: 1331.96MIN: 526.21 / MAX: 640.831. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Face Detection FP16-INT8 - Device: CPUAVX512 OffAVX512 On2004006008001000Min: 1237.65 / Avg: 1252.11 / Max: 1269.41Min: 576.85 / Avg: 578.35 / Max: 579.461. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Age Gender Recognition Retail 0013 FP16 - Device: CPUAVX512 OffAVX512 On2K4K6K8K10KSE +/- 49.53, N = 5SE +/- 23.21, N = 34781.1910275.451. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Age Gender Recognition Retail 0013 FP16 - Device: CPUAVX512 OffAVX512 On2K4K6K8K10KMin: 4684.8 / Avg: 4781.19 / Max: 4967.08Min: 10245.86 / Avg: 10275.45 / Max: 10321.221. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Age Gender Recognition Retail 0013 FP16 - Device: CPUAVX512 OffAVX512 On0.37130.74261.11391.48521.8565SE +/- 0.02, N = 5SE +/- 0.00, N = 31.650.77MIN: 1.04 / MAX: 65.91MIN: 0.44 / MAX: 31.761. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Age Gender Recognition Retail 0013 FP16 - Device: CPUAVX512 OffAVX512 On246810Min: 1.58 / Avg: 1.65 / Max: 1.68Min: 0.77 / Avg: 0.77 / Max: 0.771. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Machine Translation EN To DE FP16 - Device: CPUAVX512 OffAVX512 On816243240SE +/- 0.10, N = 3SE +/- 0.33, N = 317.1336.461. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Machine Translation EN To DE FP16 - Device: CPUAVX512 OffAVX512 On816243240Min: 16.94 / Avg: 17.13 / Max: 17.26Min: 35.81 / Avg: 36.46 / Max: 36.911. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Machine Translation EN To DE FP16 - Device: CPUAVX512 OffAVX512 On50100150200250SE +/- 1.31, N = 3SE +/- 0.99, N = 3233.17109.68MIN: 159.4 / MAX: 539.64MIN: 84.64 / MAX: 182.051. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Machine Translation EN To DE FP16 - Device: CPUAVX512 OffAVX512 On4080120160200Min: 231.55 / Avg: 233.17 / Max: 235.76Min: 108.35 / Avg: 109.68 / Max: 111.621. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUAVX512 OffAVX512 On0.75821.51642.27463.03283.791SE +/- 0.03292, N = 15SE +/- 0.01465, N = 153.369981.61633MIN: 2.69MIN: 1.181. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUAVX512 OffAVX512 On246810Min: 3.23 / Avg: 3.37 / Max: 3.67Min: 1.54 / Avg: 1.62 / Max: 1.741. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: LBC, LBRY CreditsAVX512 OffAVX512 On9K18K27K36K45KSE +/- 83.27, N = 3SE +/- 327.43, N = 322030439031. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: LBC, LBRY CreditsAVX512 OffAVX512 On8K16K24K32K40KMin: 21910 / Avg: 22030 / Max: 22190Min: 43310 / Avg: 43903.33 / Max: 444401. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries if desired for complementary metrics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 16 - Model: GoogLeNetAVX512 OffAVX512 On1224364860SE +/- 0.19, N = 11SE +/- 0.32, N = 325.7851.19
OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 16 - Model: GoogLeNetAVX512 OffAVX512 On1020304050Min: 25.41 / Avg: 25.78 / Max: 27.56Min: 50.55 / Avg: 51.19 / Max: 51.53

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 64 - Model: ResNet-50AVX512 OffAVX512 On48121620SE +/- 0.01, N = 3SE +/- 0.20, N = 38.5416.74
OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 64 - Model: ResNet-50AVX512 OffAVX512 On48121620Min: 8.53 / Avg: 8.54 / Max: 8.55Min: 16.42 / Avg: 16.74 / Max: 17.11

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Person Detection FP16 - Device: CPUAVX512 OffAVX512 On9001800270036004500SE +/- 43.35, N = 15SE +/- 36.55, N = 124390.072242.45MIN: 3908.72 / MAX: 4679.85MIN: 1833.67 / MAX: 2481.741. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Person Detection FP16 - Device: CPUAVX512 OffAVX512 On8001600240032004000Min: 3992.51 / Avg: 4390.07 / Max: 4539.49Min: 2000.36 / Avg: 2242.45 / Max: 2353.131. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Person Detection FP16 - Device: CPUAVX512 OffAVX512 On0.40050.8011.20151.6022.0025SE +/- 0.01, N = 15SE +/- 0.03, N = 120.911.781. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Person Detection FP16 - Device: CPUAVX512 OffAVX512 On246810Min: 0.88 / Avg: 0.91 / Max: 1Min: 1.69 / Avg: 1.78 / Max: 1.991. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Weld Porosity Detection FP16 - Device: CPUAVX512 OffAVX512 On612182430SE +/- 0.22, N = 3SE +/- 0.12, N = 323.4812.05MIN: 17.58 / MAX: 65.4MIN: 8.18 / MAX: 20.361. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Weld Porosity Detection FP16 - Device: CPUAVX512 OffAVX512 On510152025Min: 23.05 / Avg: 23.48 / Max: 23.76Min: 11.9 / Avg: 12.05 / Max: 12.281. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Person Detection FP32 - Device: CPUAVX512 OffAVX512 On9001800270036004500SE +/- 16.00, N = 3SE +/- 21.58, N = 34387.552252.82MIN: 4147.08 / MAX: 4543.03MIN: 2082.67 / MAX: 23671. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Person Detection FP32 - Device: CPUAVX512 OffAVX512 On8001600240032004000Min: 4355.58 / Avg: 4387.55 / Max: 4404.76Min: 2210.26 / Avg: 2252.82 / Max: 2280.281. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Weld Porosity Detection FP16 - Device: CPUAVX512 OffAVX512 On70140210280350SE +/- 1.57, N = 3SE +/- 3.20, N = 3170.20331.311. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Weld Porosity Detection FP16 - Device: CPUAVX512 OffAVX512 On60120180240300Min: 168.16 / Avg: 170.2 / Max: 173.28Min: 325.01 / Avg: 331.31 / Max: 335.441. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Person Detection FP32 - Device: CPUAVX512 OffAVX512 On0.39830.79661.19491.59321.9915SE +/- 0.00, N = 3SE +/- 0.02, N = 30.911.771. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Person Detection FP32 - Device: CPUAVX512 OffAVX512 On246810Min: 0.91 / Avg: 0.91 / Max: 0.92Min: 1.75 / Avg: 1.77 / Max: 1.81. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries if desired for complementary metrics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 64 - Model: GoogLeNetAVX512 OffAVX512 On1122334455SE +/- 0.08, N = 3SE +/- 0.53, N = 424.8748.05
OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 64 - Model: GoogLeNetAVX512 OffAVX512 On1020304050Min: 24.78 / Avg: 24.87 / Max: 25.02Min: 47.33 / Avg: 48.05 / Max: 49.61

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Myriad-GroestlAVX512 OffAVX512 On6K12K18K24K30KSE +/- 62.45, N = 3SE +/- 73.71, N = 313910261601. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Myriad-GroestlAVX512 OffAVX512 On5K10K15K20K25KMin: 13820 / Avg: 13910 / Max: 14030Min: 26050 / Avg: 26160 / Max: 263001. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUAVX512 OffAVX512 On0.64561.29121.93682.58243.228SE +/- 0.03232, N = 3SE +/- 0.01623, N = 32.869451.52707MIN: 2.23MIN: 1.21. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUAVX512 OffAVX512 On246810Min: 2.83 / Avg: 2.87 / Max: 2.93Min: 1.51 / Avg: 1.53 / Max: 1.561. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries if desired for complementary metrics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 16 - Model: ResNet-50AVX512 OffAVX512 On48121620SE +/- 0.00, N = 3SE +/- 0.08, N = 38.8416.44
OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 16 - Model: ResNet-50AVX512 OffAVX512 On48121620Min: 8.83 / Avg: 8.84 / Max: 8.84Min: 16.34 / Avg: 16.44 / Max: 16.59

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Weld Porosity Detection FP16-INT8 - Device: CPUAVX512 OffAVX512 On612182430SE +/- 0.18, N = 15SE +/- 0.08, N = 323.2012.49MIN: 17.03 / MAX: 63.15MIN: 8.3 / MAX: 21.541. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Weld Porosity Detection FP16-INT8 - Device: CPUAVX512 OffAVX512 On510152025Min: 22.59 / Avg: 23.2 / Max: 25.12Min: 12.39 / Avg: 12.49 / Max: 12.641. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Weld Porosity Detection FP16-INT8 - Device: CPUAVX512 OffAVX512 On140280420560700SE +/- 2.55, N = 15SE +/- 3.92, N = 3344.82639.051. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Weld Porosity Detection FP16-INT8 - Device: CPUAVX512 OffAVX512 On110220330440550Min: 318.22 / Avg: 344.82 / Max: 353.91Min: 631.38 / Avg: 639.05 / Max: 644.271. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUAVX512 OffAVX512 On0.37810.75621.13431.51241.8905SE +/- 0.015865, N = 4SE +/- 0.005008, N = 41.6805800.944432MIN: 1.41MIN: 0.81. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUAVX512 OffAVX512 On246810Min: 1.64 / Avg: 1.68 / Max: 1.71Min: 0.93 / Avg: 0.94 / Max: 0.961. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries if desired for complementary metrics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 64 - Model: AlexNetAVX512 OffAVX512 On306090120150SE +/- 0.71, N = 5SE +/- 0.62, N = 364.79113.58
OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 64 - Model: AlexNetAVX512 OffAVX512 On20406080100Min: 63.46 / Avg: 64.79 / Max: 67.38Min: 112.94 / Avg: 113.58 / Max: 114.82

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: gravity_spheres_volume/dim_512/scivis/real_timeAVX512 OffAVX512 On0.44510.89021.33531.78042.2255SE +/- 0.00398, N = 3SE +/- 0.01375, N = 31.137241.97814
OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: gravity_spheres_volume/dim_512/scivis/real_timeAVX512 OffAVX512 On246810Min: 1.13 / Avg: 1.14 / Max: 1.14Min: 1.95 / Avg: 1.98 / Max: 2

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Blake-2 SAVX512 OffAVX512 On110K220K330K440K550KSE +/- 2454.54, N = 15SE +/- 4566.14, N = 33159255314771. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Blake-2 SAVX512 OffAVX512 On90K180K270K360K450KMin: 299430 / Avg: 315924.67 / Max: 331560Min: 523180 / Avg: 531476.67 / Max: 5389301. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries if desired for complementary metrics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 16 - Model: AlexNetAVX512 OffAVX512 On20406080100SE +/- 0.03, N = 3SE +/- 0.18, N = 346.2177.65
OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 16 - Model: AlexNetAVX512 OffAVX512 On1530456075Min: 46.15 / Avg: 46.21 / Max: 46.26Min: 77.31 / Avg: 77.65 / Max: 77.9

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: gravity_spheres_volume/dim_512/ao/real_timeAVX512 OffAVX512 On0.44750.8951.34251.792.2375SE +/- 0.00129, N = 3SE +/- 0.01322, N = 31.194871.98870
OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: gravity_spheres_volume/dim_512/ao/real_timeAVX512 OffAVX512 On246810Min: 1.19 / Avg: 1.19 / Max: 1.2Min: 1.97 / Avg: 1.99 / Max: 2.01

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Person Vehicle Bike Detection FP16 - Device: CPUAVX512 OffAVX512 On510152025SE +/- 0.07, N = 3SE +/- 0.11, N = 318.8311.88MIN: 12.5 / MAX: 37.96MIN: 7.65 / MAX: 48.461. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Person Vehicle Bike Detection FP16 - Device: CPUAVX512 OffAVX512 On510152025Min: 18.75 / Avg: 18.83 / Max: 18.96Min: 11.65 / Avg: 11.88 / Max: 121. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Person Vehicle Bike Detection FP16 - Device: CPUAVX512 OffAVX512 On70140210280350SE +/- 0.75, N = 3SE +/- 3.29, N = 3212.15336.131. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Person Vehicle Bike Detection FP16 - Device: CPUAVX512 OffAVX512 On60120180240300Min: 210.65 / Avg: 212.15 / Max: 212.98Min: 332.55 / Avg: 336.13 / Max: 342.71. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Vehicle Detection FP16-INT8 - Device: CPUAVX512 OffAVX512 On48121620SE +/- 0.19, N = 3SE +/- 0.09, N = 1516.8411.10MIN: 12.24 / MAX: 32.26MIN: 6.52 / MAX: 49.241. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Vehicle Detection FP16-INT8 - Device: CPUAVX512 OffAVX512 On48121620Min: 16.52 / Avg: 16.84 / Max: 17.18Min: 10.13 / Avg: 11.1 / Max: 11.341. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Vehicle Detection FP16-INT8 - Device: CPUAVX512 OffAVX512 On80160240320400SE +/- 2.68, N = 3SE +/- 3.22, N = 15237.45359.971. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Vehicle Detection FP16-INT8 - Device: CPUAVX512 OffAVX512 On60120180240300Min: 232.65 / Avg: 237.45 / Max: 241.93Min: 352.02 / Avg: 359.97 / Max: 394.051. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Quad SHA-256, PyriteAVX512 OffAVX512 On20K40K60K80K100KSE +/- 69.60, N = 3SE +/- 217.36, N = 358533876271. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Quad SHA-256, PyriteAVX512 OffAVX512 On15K30K45K60K75KMin: 58420 / Avg: 58533.33 / Max: 58660Min: 87380 / Avg: 87626.67 / Max: 880601. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Vehicle Detection FP16 - Device: CPUAVX512 OffAVX512 On714212835SE +/- 0.33, N = 15SE +/- 0.29, N = 330.5321.53MIN: 20.33 / MAX: 68.23MIN: 12.47 / MAX: 43.641. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Vehicle Detection FP16 - Device: CPUAVX512 OffAVX512 On714212835Min: 27.42 / Avg: 30.53 / Max: 32.02Min: 20.94 / Avg: 21.53 / Max: 21.871. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Vehicle Detection FP16 - Device: CPUAVX512 OffAVX512 On4080120160200SE +/- 1.48, N = 15SE +/- 2.58, N = 3131.13185.721. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Vehicle Detection FP16 - Device: CPUAVX512 OffAVX512 On306090120150Min: 124.82 / Avg: 131.13 / Max: 145.72Min: 182.7 / Avg: 185.72 / Max: 190.861. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: GarlicoinAVX512 OffAVX512 On6001200180024003000SE +/- 17.03, N = 3SE +/- 15.25, N = 32009.382752.271. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: GarlicoinAVX512 OffAVX512 On5001000150020002500Min: 1982.19 / Avg: 2009.38 / Max: 2040.75Min: 2724.09 / Avg: 2752.27 / Max: 2776.471. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: scryptAVX512 OffAVX512 On4080120160200SE +/- 0.43, N = 3SE +/- 0.83, N = 3128.05166.571. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: scryptAVX512 OffAVX512 On306090120150Min: 127.36 / Avg: 128.05 / Max: 128.83Min: 165.23 / Avg: 166.57 / Max: 168.11. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

miniBUDE

MiniBUDE is a mini application for the the core computation of the Bristol University Docking Engine (BUDE). This test profile currently makes use of the OpenMP implementation of miniBUDE for CPU benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFInst/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM1AVX512 OffAVX512 On80160240320400SE +/- 3.42, N = 3SE +/- 3.72, N = 3290.70372.101. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm
OpenBenchmarking.orgGFInst/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM1AVX512 OffAVX512 On70140210280350Min: 284.1 / Avg: 290.7 / Max: 295.55Min: 365.06 / Avg: 372.1 / Max: 377.71. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

OpenBenchmarking.orgBillion Interactions/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM1AVX512 OffAVX512 On48121620SE +/- 0.14, N = 3SE +/- 0.15, N = 311.6314.881. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm
OpenBenchmarking.orgBillion Interactions/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM1AVX512 OffAVX512 On48121620Min: 11.36 / Avg: 11.63 / Max: 11.82Min: 14.6 / Avg: 14.88 / Max: 15.111. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_timeAVX512 OffAVX512 On0.55571.11141.66712.22282.7785SE +/- 0.00675, N = 3SE +/- 0.01551, N = 31.957642.46956
OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_timeAVX512 OffAVX512 On246810Min: 1.94 / Avg: 1.96 / Max: 1.97Min: 2.45 / Avg: 2.47 / Max: 2.5

OpenVKL

OpenVKL is the Intel Open Volume Kernel Library that offers high-performance volume computation kernels and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 1.3.1Benchmark: vklBenchmark ISPCAVX512 OffAVX512 On306090120150SE +/- 1.30, N = 9SE +/- 1.61, N = 992114MIN: 10 / MAX: 1437MIN: 13 / MAX: 1839
OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 1.3.1Benchmark: vklBenchmark ISPCAVX512 OffAVX512 On20406080100Min: 89 / Avg: 92.11 / Max: 102Min: 110 / Avg: 113.78 / Max: 124

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer ISPC - Model: Asian DragonAVX512 OffAVX512 On3691215SE +/- 0.0474, N = 3SE +/- 0.0930, N = 38.09179.9584MIN: 7.92 / MAX: 8.46MIN: 9.63 / MAX: 10.55
OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer ISPC - Model: Asian DragonAVX512 OffAVX512 On3691215Min: 8.02 / Avg: 8.09 / Max: 8.18Min: 9.81 / Avg: 9.96 / Max: 10.13

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: SkeincoinAVX512 OffAVX512 On20K40K60K80K100KSE +/- 336.52, N = 3SE +/- 587.57, N = 363907783001. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: SkeincoinAVX512 OffAVX512 On14K28K42K56K70KMin: 63240 / Avg: 63906.67 / Max: 64320Min: 77130 / Avg: 78300 / Max: 789801. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer ISPC - Model: Asian Dragon ObjAVX512 OffAVX512 On246810SE +/- 0.0412, N = 3SE +/- 0.1142, N = 37.23298.5595MIN: 7.04 / MAX: 7.55MIN: 8.31 / MAX: 8.97
OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer ISPC - Model: Asian Dragon ObjAVX512 OffAVX512 On3691215Min: 7.15 / Avg: 7.23 / Max: 7.28Min: 8.4 / Avg: 8.56 / Max: 8.78

OSPRay Studio

Intel OSPRay Studio is an open-source, interactive visualization and ray-tracing software package. OSPRay Studio makes use of Intel OSPRay, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 3 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path TracerAVX512 OffAVX512 On5K10K15K20K25KSE +/- 44.38, N = 3SE +/- 180.02, N = 322479190641. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 3 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path TracerAVX512 OffAVX512 On4K8K12K16K20KMin: 22391 / Avg: 22479 / Max: 22533Min: 18776 / Avg: 19063.67 / Max: 193951. (CXX) g++ options: -O3 -lm -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 3 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path TracerAVX512 OffAVX512 On160K320K480K640K800KSE +/- 6114.98, N = 3SE +/- 4373.49, N = 37395756287671. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 3 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path TracerAVX512 OffAVX512 On130K260K390K520K650KMin: 727821 / Avg: 739575 / Max: 748378Min: 622415 / Avg: 628767.33 / Max: 6371511. (CXX) g++ options: -O3 -lm -ldl

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUAVX512 OffAVX512 On0.33530.67061.00591.34121.6765SE +/- 0.02, N = 3SE +/- 0.02, N = 31.491.27MIN: 0.87 / MAX: 5.84MIN: 0.8 / MAX: 27.281. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUAVX512 OffAVX512 On246810Min: 1.47 / Avg: 1.49 / Max: 1.52Min: 1.25 / Avg: 1.27 / Max: 1.31. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUAVX512 OffAVX512 On13002600390052006500SE +/- 61.26, N = 3SE +/- 89.90, N = 35336.566240.041. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUAVX512 OffAVX512 On11002200330044005500Min: 5215.02 / Avg: 5336.56 / Max: 5410.67Min: 6066.06 / Avg: 6240.04 / Max: 6366.321. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OSPRay Studio

Intel OSPRay Studio is an open-source, interactive visualization and ray-tracing software package. OSPRay Studio makes use of Intel OSPRay, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 3 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path TracerAVX512 OffAVX512 On40K80K120K160K200KSE +/- 2250.38, N = 3SE +/- 466.38, N = 31878531612581. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 3 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path TracerAVX512 OffAVX512 On30K60K90K120K150KMin: 183725 / Avg: 187853.33 / Max: 191470Min: 160336 / Avg: 161258.33 / Max: 1618401. (CXX) g++ options: -O3 -lm -ldl

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUAVX512 OffAVX512 On6001200180024003000SE +/- 10.85, N = 3SE +/- 20.46, N = 32748.202371.62MIN: 2675.29MIN: 2280.951. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUAVX512 OffAVX512 On5001000150020002500Min: 2731.97 / Avg: 2748.2 / Max: 2768.79Min: 2334.89 / Avg: 2371.62 / Max: 2405.621. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer ISPC - Model: CrownAVX512 OffAVX512 On246810SE +/- 0.0389, N = 3SE +/- 0.0891, N = 36.85567.9186MIN: 6.66 / MAX: 7.15MIN: 7.65 / MAX: 8.35
OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer ISPC - Model: CrownAVX512 OffAVX512 On3691215Min: 6.78 / Avg: 6.86 / Max: 6.91Min: 7.75 / Avg: 7.92 / Max: 8.05

OSPRay Studio

Intel OSPRay Studio is an open-source, interactive visualization and ray-tracing software package. OSPRay Studio makes use of Intel OSPRay, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path TracerAVX512 OffAVX512 On12002400360048006000SE +/- 29.78, N = 3SE +/- 57.01, N = 3570750601. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path TracerAVX512 OffAVX512 On10002000300040005000Min: 5664 / Avg: 5706.67 / Max: 5764Min: 4955 / Avg: 5060 / Max: 51511. (CXX) g++ options: -O3 -lm -ldl

SMHasher

SMHasher is a hash function tester supporting various algorithms and able to make use of AVX and other modern CPU instruction set extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: FarmHash32 x86_64 AVXAVX512 OffAVX512 On9K18K27K36K45KSE +/- 89.33, N = 6SE +/- 5.74, N = 636442.7640077.591. (CXX) g++ options: -march=native -O3 -flto=auto -fno-fat-lto-objects
OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: FarmHash32 x86_64 AVXAVX512 OffAVX512 On7K14K21K28K35KMin: 36190.75 / Avg: 36442.76 / Max: 36681.47Min: 40049.01 / Avg: 40077.59 / Max: 40085.281. (CXX) g++ options: -march=native -O3 -flto=auto -fno-fat-lto-objects

OSPRay Studio

Intel OSPRay Studio is an open-source, interactive visualization and ray-tracing software package. OSPRay Studio makes use of Intel OSPRay, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 1 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path TracerAVX512 OffAVX512 On4K8K12K16K20KSE +/- 66.23, N = 3SE +/- 185.50, N = 418088165671. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 1 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path TracerAVX512 OffAVX512 On3K6K9K12K15KMin: 17981 / Avg: 18087.67 / Max: 18209Min: 16013 / Avg: 16566.5 / Max: 168001. (CXX) g++ options: -O3 -lm -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 1 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path TracerAVX512 OffAVX512 On130K260K390K520K650KSE +/- 5872.37, N = 3SE +/- 3927.76, N = 35979235489031. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 1 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path TracerAVX512 OffAVX512 On100K200K300K400K500KMin: 591078 / Avg: 597923.33 / Max: 609611Min: 541062 / Avg: 548903 / Max: 5532371. (CXX) g++ options: -O3 -lm -ldl

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: particle_volume/pathtracer/real_timeAVX512 OffAVX512 On20406080100SE +/- 0.14, N = 3SE +/- 0.25, N = 393.52101.81
OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: particle_volume/pathtracer/real_timeAVX512 OffAVX512 On20406080100Min: 93.34 / Avg: 93.52 / Max: 93.79Min: 101.36 / Avg: 101.81 / Max: 102.21

libxsmm

Libxsmm is an open-source library for specialized dense and sparse matrix operations and deep learning primitives. Libxsmm supports making use of Intel AMX, AVX-512, and other modern CPU instruction set capabilities. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFLOPS/s, More Is Betterlibxsmm 2-1.17-3645M N K: 64AVX512 OffAVX512 On4080120160200SE +/- 1.56, N = 3SE +/- 1.19, N = 3155.3168.01. (CXX) g++ options: -dynamic -Bstatic -static-libgcc -lgomp -lm -lrt -ldl -lquadmath -lstdc++ -pthread -fPIC -std=c++14 -pedantic -O2 -fopenmp -fopenmp-simd -funroll-loops -ftree-vectorize -fdata-sections -ffunction-sections -fvisibility=hidden -march=core-avx2
OpenBenchmarking.orgGFLOPS/s, More Is Betterlibxsmm 2-1.17-3645M N K: 64AVX512 OffAVX512 On306090120150Min: 152.6 / Avg: 155.33 / Max: 158Min: 165.7 / Avg: 167.97 / Max: 169.71. (CXX) g++ options: -dynamic -Bstatic -static-libgcc -lgomp -lm -lrt -ldl -lquadmath -lstdc++ -pthread -fPIC -std=c++14 -pedantic -O2 -fopenmp -fopenmp-simd -funroll-loops -ftree-vectorize -fdata-sections -ffunction-sections -fvisibility=hidden -march=core-avx2

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: x25xAVX512 OffAVX512 On80160240320400SE +/- 0.87, N = 3SE +/- 2.44, N = 3348.69374.951. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: x25xAVX512 OffAVX512 On70140210280350Min: 347.64 / Avg: 348.69 / Max: 350.41Min: 371.85 / Avg: 374.95 / Max: 379.761. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OSPRay Studio

Intel OSPRay Studio is an open-source, interactive visualization and ray-tracing software package. OSPRay Studio makes use of Intel OSPRay, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 1 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path TracerAVX512 OffAVX512 On30K60K90K120K150KSE +/- 922.16, N = 3SE +/- 1719.79, N = 41477641385531. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 1 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path TracerAVX512 OffAVX512 On30K60K90K120K150KMin: 146036 / Avg: 147764.33 / Max: 149186Min: 133843 / Avg: 138553 / Max: 1414521. (CXX) g++ options: -O3 -lm -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 1 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path TracerAVX512 OffAVX512 On10002000300040005000SE +/- 46.49, N = 3SE +/- 36.74, N = 12445641921. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 1 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path TracerAVX512 OffAVX512 On8001600240032004000Min: 4394 / Avg: 4456 / Max: 4547Min: 3849 / Avg: 4192.17 / Max: 42881. (CXX) g++ options: -O3 -lm -ldl

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: MagiAVX512 OffAVX512 On80160240320400SE +/- 1.15, N = 3SE +/- 1.18, N = 3365.61351.211. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: MagiAVX512 OffAVX512 On70140210280350Min: 363.48 / Avg: 365.61 / Max: 367.43Min: 349.05 / Avg: 351.21 / Max: 353.11. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

System Temperature Monitor

OpenBenchmarking.orgCelsiusSystem Temperature MonitorPhoronix Test Suite System MonitoringAVX512 OffAVX512 On20406080100Min: 38 / Avg: 67.22 / Max: 92Min: 31 / Avg: 66.6 / Max: 92

OpenVINO

OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.3System Temperature MonitorAVX512 OffAVX512 On1632486480Min: 54 / Avg: 67.41 / Max: 83Min: 53 / Avg: 69.05 / Max: 81

oneDNN

OpenBenchmarking.orgCelsius, Fewer Is BetteroneDNN 3.1System Temperature MonitorAVX512 OffAVX512 On1632486480Min: 52 / Avg: 70.71 / Max: 85Min: 53 / Avg: 69.82 / Max: 83

LeelaChessZero

OpenBenchmarking.orgCelsius, Fewer Is BetterLeelaChessZero 0.28System Temperature MonitorAVX512 OffAVX512 On20406080100Min: 39 / Avg: 67.13 / Max: 89Min: 31 / Avg: 67.44 / Max: 85

OpenBenchmarking.orgNodes Per Second Per Watt, More Is BetterLeelaChessZero 0.28Backend: EigenAVX512 OffAVX512 On91827364528.5038.12

CPU Temperature Monitor

OpenBenchmarking.orgCelsiusCPU Temperature MonitorPhoronix Test Suite System MonitoringAVX512 OffAVX512 On20406080100Min: 38.88 / Avg: 67.66 / Max: 92Min: 31.25 / Avg: 67.05 / Max: 92.13

CPU Power Consumption Monitor

OpenBenchmarking.orgWattsCPU Power Consumption MonitorPhoronix Test Suite System MonitoringAVX512 OffAVX512 On714212835Min: 1.3 / Avg: 16.39 / Max: 30.21Min: 1.27 / Avg: 15.88 / Max: 30.82

CPU Peak Freq (Highest CPU Core Frequency) Monitor

OpenBenchmarking.orgMegahertzCPU Peak Freq (Highest CPU Core Frequency) MonitorPhoronix Test Suite System MonitoringAVX512 OffAVX512 On9001800270036004500Min: 1114 / Avg: 2621.67 / Max: 5115Min: 1397 / Avg: 2761.55 / Max: 5115

OpenVINO

OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.3CPU Temperature MonitorAVX512 OffAVX512 On1632486480Min: 54.13 / Avg: 67.88 / Max: 85.25Min: 54.13 / Avg: 69.52 / Max: 82

OpenBenchmarking.orgWatts, Fewer Is BetterOpenVINO 2022.3CPU Power Consumption MonitorAVX512 OffAVX512 On612182430Min: 2.49 / Avg: 16.5 / Max: 28.33Min: 2.44 / Avg: 17.42 / Max: 26.03

OpenBenchmarking.orgMegahertz, More Is BetterOpenVINO 2022.3CPU Peak Freq (Highest CPU Core Frequency) MonitorAVX512 OffAVX512 On9001800270036004500Min: 1397 / Avg: 2336.56 / Max: 5069Min: 1600 / Avg: 2546.19 / Max: 5109

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Face Detection FP16 - Device: CPUAVX512 OffAVX512 On6001200180024003000SE +/- 44.91, N = 15SE +/- 15.69, N = 122570.311143.20MIN: 2049 / MAX: 2793.24MIN: 992.31 / MAX: 1244.551. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Face Detection FP16 - Device: CPUAVX512 OffAVX512 On400800120016002000Min: 2161.2 / Avg: 2570.31 / Max: 2734.63Min: 1024.85 / Avg: 1143.2 / Max: 1194.141. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Face Detection FP16 - Device: CPUAVX512 OffAVX512 On0.78751.5752.36253.153.9375SE +/- 0.03, N = 15SE +/- 0.05, N = 121.563.501. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Face Detection FP16 - Device: CPUAVX512 OffAVX512 On246810Min: 1.46 / Avg: 1.56 / Max: 1.84Min: 3.35 / Avg: 3.5 / Max: 3.91. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

oneDNN

OpenBenchmarking.orgCelsius, Fewer Is BetteroneDNN 3.1CPU Temperature MonitorAVX512 OffAVX512 On1632486480Min: 52.5 / Avg: 71.17 / Max: 85.13Min: 53 / Avg: 70.24 / Max: 84.25

OpenBenchmarking.orgWatts, Fewer Is BetteroneDNN 3.1CPU Power Consumption MonitorAVX512 OffAVX512 On714212835Min: 2.34 / Avg: 17.93 / Max: 30.03Min: 2.3 / Avg: 17.63 / Max: 30.03

OpenBenchmarking.orgMegahertz, More Is BetteroneDNN 3.1CPU Peak Freq (Highest CPU Core Frequency) MonitorAVX512 OffAVX512 On9001800270036004500Min: 1397 / Avg: 2543.71 / Max: 5065Min: 1397 / Avg: 2672.77 / Max: 5042

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUAVX512 OffAVX512 On11002200330044005500SE +/- 91.64, N = 15SE +/- 84.98, N = 125003.234354.17MIN: 4002.46MIN: 3672.421. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUAVX512 OffAVX512 On9001800270036004500Min: 4134.6 / Avg: 5003.23 / Max: 5296.23Min: 3765.03 / Avg: 4354.17 / Max: 4756.161. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

LeelaChessZero

OpenBenchmarking.orgCelsius, Fewer Is BetterLeelaChessZero 0.28CPU Temperature MonitorAVX512 OffAVX512 On20406080100Min: 39.5 / Avg: 67.57 / Max: 89.38Min: 31.25 / Avg: 67.9 / Max: 85.88

OpenBenchmarking.orgWatts, Fewer Is BetterLeelaChessZero 0.28CPU Power Consumption MonitorAVX512 OffAVX512 On714212835Min: 2.08 / Avg: 16.25 / Max: 30.12Min: 1.86 / Avg: 17.18 / Max: 30.82

OpenBenchmarking.orgMegahertz, More Is BetterLeelaChessZero 0.28CPU Peak Freq (Highest CPU Core Frequency) MonitorAVX512 OffAVX512 On9001800270036004500Min: 1114 / Avg: 2776.7 / Max: 5092Min: 1397 / Avg: 2875.81 / Max: 5096

OpenBenchmarking.orgNodes Per Second, More Is BetterLeelaChessZero 0.28Backend: EigenAVX512 OffAVX512 On140280420560700SE +/- 11.93, N = 9SE +/- 12.23, N = 94636551. (CXX) g++ options: -flto -pthread
OpenBenchmarking.orgNodes Per Second, More Is BetterLeelaChessZero 0.28Backend: EigenAVX512 OffAVX512 On120240360480600Min: 436 / Avg: 463.44 / Max: 549Min: 611 / Avg: 654.67 / Max: 7101. (CXX) g++ options: -flto -pthread

82 Results Shown

OpenVINO:
  Face Detection FP16-INT8 - CPU:
    FPS
    ms
  Age Gender Recognition Retail 0013 FP16 - CPU:
    FPS
    ms
  Machine Translation EN To DE FP16 - CPU:
    FPS
    ms
oneDNN
Cpuminer-Opt
TensorFlow:
  CPU - 16 - GoogLeNet
  CPU - 64 - ResNet-50
OpenVINO:
  Person Detection FP16 - CPU:
    ms
    FPS
  Weld Porosity Detection FP16 - CPU:
    ms
  Person Detection FP32 - CPU:
    ms
  Weld Porosity Detection FP16 - CPU:
    FPS
  Person Detection FP32 - CPU:
    FPS
TensorFlow
Cpuminer-Opt
oneDNN
TensorFlow
OpenVINO:
  Weld Porosity Detection FP16-INT8 - CPU:
    ms
    FPS
oneDNN
TensorFlow
OSPRay
Cpuminer-Opt
TensorFlow
OSPRay
OpenVINO:
  Person Vehicle Bike Detection FP16 - CPU:
    ms
    FPS
  Vehicle Detection FP16-INT8 - CPU:
    ms
    FPS
Cpuminer-Opt
OpenVINO:
  Vehicle Detection FP16 - CPU:
    ms
    FPS
Cpuminer-Opt:
  Garlicoin
  scrypt
miniBUDE:
  OpenMP - BM1:
    GFInst/s
    Billion Interactions/s
OSPRay
OpenVKL
Embree
Cpuminer-Opt
Embree
OSPRay Studio:
  3 - 4K - 1 - Path Tracer
  3 - 4K - 32 - Path Tracer
OpenVINO:
  Age Gender Recognition Retail 0013 FP16-INT8 - CPU:
    ms
    FPS
OSPRay Studio
oneDNN
Embree
OSPRay Studio
SMHasher
OSPRay Studio:
  1 - 4K - 1 - Path Tracer
  1 - 4K - 32 - Path Tracer
OSPRay
libxsmm
Cpuminer-Opt
OSPRay Studio:
  1 - 1080p - 32 - Path Tracer
  1 - 1080p - 1 - Path Tracer
Cpuminer-Opt
System Temperature Monitor:
  Phoronix Test Suite System Monitoring
  System Temp Monitor
  System Temp Monitor
  System Temp Monitor
  Eigen
  Phoronix Test Suite System Monitoring
  Phoronix Test Suite System Monitoring
  Phoronix Test Suite System Monitoring
  CPU Temp Monitor
  CPU Power Consumption Monitor
  CPU Peak Freq (Highest CPU Core Frequency) Monitor
OpenVINO:
  Face Detection FP16 - CPU:
    ms
    FPS
oneDNN:
  CPU Temp Monitor
  CPU Power Consumption Monitor
  CPU Peak Freq (Highest CPU Core Frequency) Monitor
oneDNN
LeelaChessZero:
  CPU Temp Monitor
  CPU Power Consumption Monitor
  CPU Peak Freq (Highest CPU Core Frequency) Monitor
LeelaChessZero