sss

AMD Ryzen 9 5950X 16-Core testing with a ASUS ROG CROSSHAIR VIII HERO (WI-FI) (4006 BIOS) and llvmpipe on Ubuntu 20.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2210145-NE-SSS86725756
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results

Limit displaying results to tests within:

CPU Massive 2 Tests
Creator Workloads 2 Tests
HPC - High Performance Computing 5 Tests
Machine Learning 4 Tests
Multi-Core 2 Tests
Python Tests 3 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
A
October 13 2022
  1 Hour, 20 Minutes
B
October 13 2022
  4 Hours, 14 Minutes
C
October 14 2022
  4 Hours, 53 Minutes
Invert Hiding All Results Option
  3 Hours, 29 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


sssOpenBenchmarking.orgPhoronix Test SuiteAMD Ryzen 9 5950X 16-Core @ 3.40GHz (16 Cores / 32 Threads)ASUS ROG CROSSHAIR VIII HERO (WI-FI) (4006 BIOS)AMD Starship/Matisse32GB500GB Western Digital WDS500G3X0C-00SJG0llvmpipe (2450MHz)Intel Device 4f92ASUS MG28URealtek RTL8125 2.5GbE + Intel I211 + Intel Wi-Fi 6 AX200Ubuntu 20.046.0.0-060000rc5daily20220915-generic (x86_64)GNOME Shell 3.36.9X Server 1.20.134.5 Mesa 21.2.6 (LLVM 12.0.0 256 bits)OpenCL 3.01.1.182GCC 9.4.0ext43840x2160ProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerOpenGLOpenCLVulkanCompilerFile-SystemScreen ResolutionSss PerformanceSystem Logs- i915.force_probe=56a5 - Transparent Huge Pages: madvise- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-9-Av3uEd/gcc-9-9.4.0/debian/tmp-nvptx/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: acpi-cpufreq ondemand (Boost: Enabled) - CPU Microcode: 0xa201016 - Python 2.7.18 + Python 3.8.10- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected

ABCResult OverviewPhoronix Test Suite100%100%101%101%SMHasheroneDNNOpenRadiossNeural Magic DeepSparsespaCyTensorFlowY-CruncherQuadRay

sssopenradioss: INIVOL and Fluid Structure Interaction Drop Containersmhasher: SHA3-256smhasher: SHA3-256tensorflow: CPU - 32 - ResNet-50openradioss: Bird Strike on Windshieldonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUtensorflow: CPU - 16 - ResNet-50tensorflow: CPU - 32 - GoogLeNetopenradioss: Bumper Beamonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUopenradioss: Rubber O-Ring Seal Installationopenradioss: Cell Phone Drop Testonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUspacy: en_core_web_trfspacy: en_core_web_lgtensorflow: CPU - 16 - GoogLeNetdeepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Streamdeepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Streamdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Streamdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Streamdeepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Synchronous Single-Streamdeepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Synchronous Single-Streamtensorflow: CPU - 32 - AlexNetdeepsparse: NLP Text Classification, BERT base uncased SST2 - Synchronous Single-Streamdeepsparse: NLP Text Classification, BERT base uncased SST2 - Synchronous Single-Streamdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Synchronous Single-Streamdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Synchronous Single-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Synchronous Single-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Synchronous Single-Streamy-cruncher: 1Bdeepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Streamdeepsparse: CV Detection,YOLOv5s COCO - Asynchronous Multi-Streamdeepsparse: CV Detection,YOLOv5s COCO - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, DistilBERT mnli - Synchronous Single-Streamdeepsparse: NLP Text Classification, DistilBERT mnli - Synchronous Single-Streamdeepsparse: CV Detection,YOLOv5s COCO - Synchronous Single-Streamdeepsparse: CV Detection,YOLOv5s COCO - Synchronous Single-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Streamonednn: Deconvolution Batch shapes_1d - f32 - CPUdeepsparse: CV Classification, ResNet-50 ImageNet - Synchronous Single-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Synchronous Single-Streamtensorflow: CPU - 16 - AlexNetonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUquadray: 5 - 4Kquadray: 2 - 4Kquadray: 3 - 4Kquadray: 1 - 4Kquadray: 5 - 1080pquadray: 3 - 1080pquadray: 2 - 1080pquadray: 1 - 1080py-cruncher: 500Monednn: IP Shapes 1D - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUsmhasher: MeowHash x86_64 AES-NIsmhasher: MeowHash x86_64 AES-NIsmhasher: FarmHash128smhasher: FarmHash128smhasher: t1ha0_aes_avx2 x86_64smhasher: t1ha0_aes_avx2 x86_64onednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUsmhasher: Spooky32smhasher: Spooky32smhasher: FarmHash32 x86_64 AVXsmhasher: FarmHash32 x86_64 AVXsmhasher: t1ha2_atoncesmhasher: t1ha2_atoncesmhasher: fasthash32smhasher: fasthash32onednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUsmhasher: wyhashsmhasher: wyhashonednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: IP Shapes 1D - bf16bf16bf16 - CPUABC512.251852.225209.8511.5225.541637.8711.7932.36110.062602.441641.1592.0887.552594.592608.121647.7310741600034.05192.355741.582622.838412.7346626.728612.7051144.473355.295531.326531.914880.0124.180541.346488.011411.361487.487511.429538.8266.9923119.3666117.845767.805912.41880.496117.775856.227850.1021159.5434.407528.8777112.557256.171.071590.893.412.8711.13.5411.4413.4446.0318.0583.866250.7761190.7862430.68193753.59146922.8352.11420617.6822.09783813.938.180950.46766929.94419879.9328.72533990.6922.87818113.8523.518321.9717.033218.422414.96828600.243.578051.65982511.381898.738204.9111.49226.961679.6411.8332.48110.292624.271655.4793.0587.662590.102597.671666.7910761602134.16192.825141.4807625.546512.7366626.008712.6875144.701155.242031.360031.880879.6724.212741.291888.095911.350688.431111.307538.81967.0315119.2849116.965568.327412.443880.327717.751256.304450.4242158.59034.360028.8852112.463256.671.072450.893.392.9011.113.5511.4413.3345.9418.1263.923860.7775070.7954730.68274753.46246726.1151.83220202.8121.95082904.447.957400.45990930.81519280.7228.52334077.2023.11118237.8224.6907974.2116.812218.231316.31527320.493.634451.68678518.291889.905206.2911.47227.521703.1111.7832.53107.762670.401664.4494.3088.812627.852598.171744.7810691601634.20197.096340.5820628.285612.6662631.115412.6218144.773455.223231.406031.834579.5424.248641.230788.276011.327488.132911.346038.71067.2081118.9598118.405767.491112.488080.043217.876855.909650.5237158.24964.395048.8930112.364556.511.076020.893.392.9011.113.5711.4413.3645.7518.1183.963220.7750550.7899100.68222853.36147376.7351.63020364.5022.09781425.199.085800.47311331.19819028.9728.56033941.3322.88018384.2724.3168089.0217.121318.928315.42027857.673.603701.67046OpenBenchmarking.org

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: INIVOL and Fluid Structure Interaction Drop ContainerABC110220330440550SE +/- 0.30, N = 3SE +/- 4.41, N = 8512.25511.38518.29
OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: INIVOL and Fluid Structure Interaction Drop ContainerABC90180270360450Min: 510.78 / Avg: 511.38 / Max: 511.69Min: 511.29 / Avg: 518.29 / Max: 546.31

SMHasher

SMHasher is a hash function tester supporting various algorithms and able to make use of AVX and other modern CPU instruction set extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: SHA3-256ABC400800120016002000SE +/- 14.44, N = 3SE +/- 3.87, N = 31852.231898.741889.911. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: SHA3-256ABC30060090012001500Min: 1870.64 / Avg: 1898.74 / Max: 1918.58Min: 1884.12 / Avg: 1889.91 / Max: 1897.251. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: SHA3-256ABC50100150200250SE +/- 1.57, N = 3SE +/- 0.30, N = 3209.85204.91206.291. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: SHA3-256ABC4080120160200Min: 202.59 / Avg: 204.91 / Max: 207.91Min: 205.99 / Avg: 206.29 / Max: 206.891. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 32 - Model: ResNet-50ABC3691215SE +/- 0.01, N = 3SE +/- 0.02, N = 311.5011.4911.47
OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 32 - Model: ResNet-50ABC3691215Min: 11.48 / Avg: 11.49 / Max: 11.51Min: 11.44 / Avg: 11.47 / Max: 11.51

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: Bird Strike on WindshieldABC50100150200250SE +/- 0.53, N = 3SE +/- 0.78, N = 3225.54226.96227.52
OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: Bird Strike on WindshieldABC4080120160200Min: 226.33 / Avg: 226.96 / Max: 228.02Min: 226.58 / Avg: 227.52 / Max: 229.06

oneDNN

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUABC400800120016002000SE +/- 12.64, N = 11SE +/- 15.79, N = 61637.871679.641703.11MIN: 1622.93MIN: 1619.79MIN: 1634.781. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUABC30060090012001500Min: 1635.47 / Avg: 1679.64 / Max: 1768.94Min: 1644.08 / Avg: 1703.11 / Max: 1742.611. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 16 - Model: ResNet-50ABC3691215SE +/- 0.01, N = 3SE +/- 0.03, N = 311.7911.8311.78
OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 16 - Model: ResNet-50ABC3691215Min: 11.81 / Avg: 11.83 / Max: 11.84Min: 11.72 / Avg: 11.78 / Max: 11.82

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 32 - Model: GoogLeNetABC816243240SE +/- 0.02, N = 3SE +/- 0.02, N = 332.3632.4832.53
OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 32 - Model: GoogLeNetABC714212835Min: 32.44 / Avg: 32.48 / Max: 32.51Min: 32.51 / Avg: 32.53 / Max: 32.57

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: Bumper BeamABC20406080100SE +/- 0.53, N = 3SE +/- 0.29, N = 3110.06110.29107.76
OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: Bumper BeamABC20406080100Min: 109.39 / Avg: 110.29 / Max: 111.24Min: 107.22 / Avg: 107.76 / Max: 108.19

oneDNN

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUABC6001200180024003000SE +/- 11.17, N = 3SE +/- 28.26, N = 52602.442624.272670.40MIN: 2590.62MIN: 2596.85MIN: 2578.831. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUABC5001000150020002500Min: 2610.05 / Avg: 2624.27 / Max: 2646.3Min: 2590.12 / Avg: 2670.4 / Max: 2734.131. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUABC400800120016002000SE +/- 15.95, N = 3SE +/- 17.63, N = 51641.151655.471664.44MIN: 1622.41MIN: 1622.09MIN: 1618.121. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUABC30060090012001500Min: 1637.35 / Avg: 1655.47 / Max: 1687.28Min: 1629.66 / Avg: 1664.44 / Max: 1707.681. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: Rubber O-Ring Seal InstallationABC20406080100SE +/- 0.43, N = 3SE +/- 0.41, N = 392.0893.0594.30
OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: Rubber O-Ring Seal InstallationABC20406080100Min: 92.3 / Avg: 93.05 / Max: 93.8Min: 93.5 / Avg: 94.3 / Max: 94.84

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: Cell Phone Drop TestABC20406080100SE +/- 0.05, N = 3SE +/- 0.52, N = 387.5587.6688.81
OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: Cell Phone Drop TestABC20406080100Min: 87.56 / Avg: 87.66 / Max: 87.75Min: 87.99 / Avg: 88.81 / Max: 89.77

oneDNN

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUABC6001200180024003000SE +/- 6.69, N = 3SE +/- 31.84, N = 32594.592590.102627.85MIN: 2584.09MIN: 2567.75MIN: 2582.81. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUABC5001000150020002500Min: 2576.72 / Avg: 2590.1 / Max: 2597.11Min: 2595.22 / Avg: 2627.85 / Max: 2691.531. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUABC6001200180024003000SE +/- 8.68, N = 3SE +/- 7.36, N = 32608.122597.672598.17MIN: 2597.2MIN: 2576.33MIN: 2573.51. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUABC5001000150020002500Min: 2586.38 / Avg: 2597.67 / Max: 2614.73Min: 2586.37 / Avg: 2598.17 / Max: 2611.71. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUABC400800120016002000SE +/- 19.69, N = 3SE +/- 19.41, N = 31647.731666.791744.78MIN: 1635.21MIN: 1620.68MIN: 1706.541. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUABC30060090012001500Min: 1628.71 / Avg: 1666.79 / Max: 1694.53Min: 1720.62 / Avg: 1744.78 / Max: 1783.181. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

spaCy

The spaCy library is an open-source solution for advanced neural language processing (NLP). The spaCy library leverages Python and is a leading neural language processing solution. This test profile times the spaCy CPU performance with various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgtokens/sec, More Is BetterspaCy 3.4.1Model: en_core_web_trfABC2004006008001000SE +/- 3.48, N = 3SE +/- 4.93, N = 3107410761069
OpenBenchmarking.orgtokens/sec, More Is BetterspaCy 3.4.1Model: en_core_web_trfABC2004006008001000Min: 1070 / Avg: 1076.33 / Max: 1082Min: 1061 / Avg: 1069 / Max: 1078

OpenBenchmarking.orgtokens/sec, More Is BetterspaCy 3.4.1Model: en_core_web_lgABC3K6K9K12K15KSE +/- 18.22, N = 3SE +/- 4.58, N = 3160001602116016
OpenBenchmarking.orgtokens/sec, More Is BetterspaCy 3.4.1Model: en_core_web_lgABC3K6K9K12K15KMin: 15988 / Avg: 16020.67 / Max: 16051Min: 16010 / Avg: 16016 / Max: 16025

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 16 - Model: GoogLeNetABC816243240SE +/- 0.02, N = 3SE +/- 0.02, N = 334.0534.1634.20
OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 16 - Model: GoogLeNetABC714212835Min: 34.12 / Avg: 34.16 / Max: 34.2Min: 34.18 / Avg: 34.2 / Max: 34.23

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-StreamABC4080120160200SE +/- 0.11, N = 3SE +/- 0.14, N = 3192.36192.83197.10
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-StreamABC4080120160200Min: 192.64 / Avg: 192.83 / Max: 193.02Min: 196.93 / Avg: 197.1 / Max: 197.37

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-StreamABC918273645SE +/- 0.02, N = 3SE +/- 0.03, N = 341.5841.4840.58
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-StreamABC918273645Min: 41.44 / Avg: 41.48 / Max: 41.52Min: 40.53 / Avg: 40.58 / Max: 40.62

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-StreamABC140280420560700SE +/- 2.50, N = 3SE +/- 2.78, N = 3622.84625.55628.29
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-StreamABC110220330440550Min: 620.65 / Avg: 625.55 / Max: 628.89Min: 623.81 / Avg: 628.29 / Max: 633.37

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-StreamABC3691215SE +/- 0.07, N = 3SE +/- 0.04, N = 312.7312.7412.67
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-StreamABC48121620Min: 12.65 / Avg: 12.74 / Max: 12.88Min: 12.6 / Avg: 12.67 / Max: 12.72

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-StreamABC140280420560700SE +/- 1.55, N = 3SE +/- 1.66, N = 3626.73626.01631.12
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-StreamABC110220330440550Min: 623.58 / Avg: 626.01 / Max: 628.9Min: 628.14 / Avg: 631.12 / Max: 633.88

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-StreamABC3691215SE +/- 0.02, N = 3SE +/- 0.03, N = 312.7112.6912.62
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-StreamABC48121620Min: 12.64 / Avg: 12.69 / Max: 12.72Min: 12.58 / Avg: 12.62 / Max: 12.68

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-StreamABC306090120150SE +/- 0.23, N = 3SE +/- 0.12, N = 3144.47144.70144.77
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-StreamABC306090120150Min: 144.44 / Avg: 144.7 / Max: 145.17Min: 144.63 / Avg: 144.77 / Max: 145.02

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-StreamABC1224364860SE +/- 0.08, N = 3SE +/- 0.05, N = 355.3055.2455.22
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-StreamABC1122334455Min: 55.09 / Avg: 55.24 / Max: 55.35Min: 55.12 / Avg: 55.22 / Max: 55.28

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Synchronous Single-StreamABC714212835SE +/- 0.06, N = 3SE +/- 0.10, N = 331.3331.3631.41
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Synchronous Single-StreamABC714212835Min: 31.25 / Avg: 31.36 / Max: 31.45Min: 31.25 / Avg: 31.41 / Max: 31.58

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Synchronous Single-StreamABC714212835SE +/- 0.06, N = 3SE +/- 0.10, N = 331.9131.8831.83
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Synchronous Single-StreamABC714212835Min: 31.79 / Avg: 31.88 / Max: 31.99Min: 31.65 / Avg: 31.83 / Max: 31.99

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 32 - Model: AlexNetABC20406080100SE +/- 0.15, N = 3SE +/- 0.12, N = 380.0179.6779.54
OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 32 - Model: AlexNetABC1530456075Min: 79.37 / Avg: 79.67 / Max: 79.86Min: 79.33 / Avg: 79.54 / Max: 79.75

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Synchronous Single-StreamABC612182430SE +/- 0.01, N = 3SE +/- 0.03, N = 324.1824.2124.25
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Synchronous Single-StreamABC612182430Min: 24.19 / Avg: 24.21 / Max: 24.23Min: 24.21 / Avg: 24.25 / Max: 24.3

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Synchronous Single-StreamABC918273645SE +/- 0.02, N = 3SE +/- 0.05, N = 341.3541.2941.23
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Synchronous Single-StreamABC918273645Min: 41.27 / Avg: 41.29 / Max: 41.32Min: 41.15 / Avg: 41.23 / Max: 41.3

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-StreamABC20406080100SE +/- 0.13, N = 3SE +/- 0.06, N = 388.0188.1088.28
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-StreamABC20406080100Min: 87.83 / Avg: 88.1 / Max: 88.23Min: 88.15 / Avg: 88.28 / Max: 88.35

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-StreamABC3691215SE +/- 0.02, N = 3SE +/- 0.01, N = 311.3611.3511.33
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-StreamABC3691215Min: 11.33 / Avg: 11.35 / Max: 11.38Min: 11.32 / Avg: 11.33 / Max: 11.34

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-StreamABC20406080100SE +/- 0.09, N = 3SE +/- 0.22, N = 387.4988.4388.13
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-StreamABC20406080100Min: 88.27 / Avg: 88.43 / Max: 88.57Min: 87.69 / Avg: 88.13 / Max: 88.4

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-StreamABC3691215SE +/- 0.01, N = 3SE +/- 0.03, N = 311.4311.3111.35
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-StreamABC3691215Min: 11.29 / Avg: 11.31 / Max: 11.33Min: 11.31 / Avg: 11.35 / Max: 11.4

Y-Cruncher

Y-Cruncher is a multi-threaded Pi benchmark capable of computing Pi to trillions of digits. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterY-Cruncher 0.7.10.9513Pi Digits To Calculate: 1BABC918273645SE +/- 0.13, N = 3SE +/- 0.09, N = 338.8238.8238.71
OpenBenchmarking.orgSeconds, Fewer Is BetterY-Cruncher 0.7.10.9513Pi Digits To Calculate: 1BABC816243240Min: 38.6 / Avg: 38.82 / Max: 39.06Min: 38.54 / Avg: 38.71 / Max: 38.85

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-StreamABC1530456075SE +/- 0.03, N = 3SE +/- 0.02, N = 366.9967.0367.21
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-StreamABC1326395265Min: 67 / Avg: 67.03 / Max: 67.1Min: 67.16 / Avg: 67.21 / Max: 67.24

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-StreamABC306090120150SE +/- 0.08, N = 3SE +/- 0.06, N = 3119.37119.28118.96
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-StreamABC20406080100Min: 119.12 / Avg: 119.28 / Max: 119.37Min: 118.87 / Avg: 118.96 / Max: 119.07

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: CV Detection,YOLOv5s COCO - Scenario: Asynchronous Multi-StreamABC306090120150SE +/- 1.56, N = 3SE +/- 0.21, N = 3117.85116.97118.41
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: CV Detection,YOLOv5s COCO - Scenario: Asynchronous Multi-StreamABC20406080100Min: 114.22 / Avg: 116.97 / Max: 119.61Min: 118.19 / Avg: 118.41 / Max: 118.83

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: CV Detection,YOLOv5s COCO - Scenario: Asynchronous Multi-StreamABC1530456075SE +/- 0.91, N = 3SE +/- 0.13, N = 367.8168.3367.49
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: CV Detection,YOLOv5s COCO - Scenario: Asynchronous Multi-StreamABC1326395265Min: 66.82 / Avg: 68.33 / Max: 69.95Min: 67.24 / Avg: 67.49 / Max: 67.64

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-StreamABC3691215SE +/- 0.01, N = 3SE +/- 0.03, N = 312.4212.4412.49
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-StreamABC48121620Min: 12.42 / Avg: 12.44 / Max: 12.46Min: 12.44 / Avg: 12.49 / Max: 12.54

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-StreamABC20406080100SE +/- 0.08, N = 3SE +/- 0.18, N = 380.5080.3380.04
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-StreamABC1530456075Min: 80.21 / Avg: 80.33 / Max: 80.47Min: 79.72 / Avg: 80.04 / Max: 80.35

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: CV Detection,YOLOv5s COCO - Scenario: Synchronous Single-StreamABC48121620SE +/- 0.01, N = 3SE +/- 0.01, N = 317.7817.7517.88
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: CV Detection,YOLOv5s COCO - Scenario: Synchronous Single-StreamABC510152025Min: 17.73 / Avg: 17.75 / Max: 17.78Min: 17.86 / Avg: 17.88 / Max: 17.89

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: CV Detection,YOLOv5s COCO - Scenario: Synchronous Single-StreamABC1326395265SE +/- 0.05, N = 3SE +/- 0.03, N = 356.2356.3055.91
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: CV Detection,YOLOv5s COCO - Scenario: Synchronous Single-StreamABC1122334455Min: 56.21 / Avg: 56.3 / Max: 56.36Min: 55.87 / Avg: 55.91 / Max: 55.96

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-StreamABC1122334455SE +/- 0.23, N = 3SE +/- 0.18, N = 350.1050.4250.52
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-StreamABC1020304050Min: 49.96 / Avg: 50.42 / Max: 50.7Min: 50.17 / Avg: 50.52 / Max: 50.79

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-StreamABC4080120160200SE +/- 0.71, N = 3SE +/- 0.56, N = 3159.54158.59158.25
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-StreamABC306090120150Min: 157.74 / Avg: 158.59 / Max: 160Min: 157.47 / Avg: 158.25 / Max: 159.34

oneDNN

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUABC0.99171.98342.97513.96684.9585SE +/- 0.03778, N = 8SE +/- 0.01829, N = 34.407524.360024.39504MIN: 3.5MIN: 3.52MIN: 3.541. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUABC246810Min: 4.25 / Avg: 4.36 / Max: 4.56Min: 4.37 / Avg: 4.4 / Max: 4.431. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-StreamABC246810SE +/- 0.0017, N = 3SE +/- 0.0051, N = 38.87778.88528.8930
OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.1Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-StreamABC3691215Min: 8.88 / Avg: 8.89 / Max: 8.89Min: 8.89 / Avg: 8.89 / Max: 8.9

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-StreamABC306090120150SE +/- 0.02, N = 3SE +/- 0.06, N = 3112.56112.46112.36
OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.1Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-StreamABC20406080100Min: 112.43 / Avg: 112.46 / Max: 112.49Min: 112.24 / Avg: 112.36 / Max: 112.45

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 16 - Model: AlexNetABC1326395265SE +/- 0.14, N = 3SE +/- 0.15, N = 356.1756.6756.51
OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 16 - Model: AlexNetABC1122334455Min: 56.51 / Avg: 56.67 / Max: 56.95Min: 56.22 / Avg: 56.51 / Max: 56.69

oneDNN

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUABC0.24210.48420.72630.96841.2105SE +/- 0.00134, N = 3SE +/- 0.00183, N = 31.071591.072451.07602MIN: 0.97MIN: 0.97MIN: 0.971. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUABC246810Min: 1.07 / Avg: 1.07 / Max: 1.08Min: 1.07 / Avg: 1.08 / Max: 1.081. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

QuadRay

VectorChief's QuadRay is a real-time ray-tracing engine written to support SIMD across ARM, MIPS, PPC, and x86/x86_64 processors. QuadRay supports SSE/SSE2/SSE4 and AVX/AVX2/AVX-512 usage on Intel/AMD CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 5 - Resolution: 4KABC0.20030.40060.60090.80121.0015SE +/- 0.00, N = 3SE +/- 0.00, N = 30.890.890.891. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread
OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 5 - Resolution: 4KABC246810Min: 0.89 / Avg: 0.89 / Max: 0.89Min: 0.89 / Avg: 0.89 / Max: 0.891. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 2 - Resolution: 4KABC0.76731.53462.30193.06923.8365SE +/- 0.01, N = 3SE +/- 0.01, N = 33.413.393.391. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread
OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 2 - Resolution: 4KABC246810Min: 3.37 / Avg: 3.39 / Max: 3.4Min: 3.38 / Avg: 3.39 / Max: 3.411. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 3 - Resolution: 4KABC0.65251.3051.95752.613.2625SE +/- 0.01, N = 3SE +/- 0.01, N = 32.872.902.901. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread
OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 3 - Resolution: 4KABC246810Min: 2.89 / Avg: 2.9 / Max: 2.91Min: 2.89 / Avg: 2.9 / Max: 2.911. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 1 - Resolution: 4KABC3691215SE +/- 0.03, N = 3SE +/- 0.02, N = 311.1011.1111.111. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread
OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 1 - Resolution: 4KABC3691215Min: 11.05 / Avg: 11.11 / Max: 11.15Min: 11.08 / Avg: 11.11 / Max: 11.161. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 5 - Resolution: 1080pABC0.80331.60662.40993.21324.0165SE +/- 0.01, N = 3SE +/- 0.00, N = 33.543.553.571. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread
OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 5 - Resolution: 1080pABC246810Min: 3.53 / Avg: 3.55 / Max: 3.56Min: 3.56 / Avg: 3.57 / Max: 3.571. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 3 - Resolution: 1080pABC3691215SE +/- 0.03, N = 3SE +/- 0.01, N = 311.4411.4411.441. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread
OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 3 - Resolution: 1080pABC3691215Min: 11.41 / Avg: 11.44 / Max: 11.5Min: 11.43 / Avg: 11.44 / Max: 11.461. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 2 - Resolution: 1080pABC3691215SE +/- 0.02, N = 3SE +/- 0.04, N = 313.4413.3313.361. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread
OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 2 - Resolution: 1080pABC48121620Min: 13.29 / Avg: 13.33 / Max: 13.36Min: 13.29 / Avg: 13.36 / Max: 13.411. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 1 - Resolution: 1080pABC1020304050SE +/- 0.09, N = 3SE +/- 0.04, N = 346.0345.9445.751. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread
OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 1 - Resolution: 1080pABC918273645Min: 45.77 / Avg: 45.94 / Max: 46.06Min: 45.67 / Avg: 45.75 / Max: 45.821. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

Y-Cruncher

Y-Cruncher is a multi-threaded Pi benchmark capable of computing Pi to trillions of digits. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterY-Cruncher 0.7.10.9513Pi Digits To Calculate: 500MABC48121620SE +/- 0.02, N = 3SE +/- 0.01, N = 318.0618.1318.12
OpenBenchmarking.orgSeconds, Fewer Is BetterY-Cruncher 0.7.10.9513Pi Digits To Calculate: 500MABC510152025Min: 18.09 / Avg: 18.13 / Max: 18.16Min: 18.1 / Avg: 18.12 / Max: 18.14

oneDNN

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUABC0.89171.78342.67513.56684.4585SE +/- 0.03093, N = 3SE +/- 0.00111, N = 33.866253.923863.96322MIN: 3.66MIN: 3.68MIN: 3.731. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUABC246810Min: 3.87 / Avg: 3.92 / Max: 3.98Min: 3.96 / Avg: 3.96 / Max: 3.961. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUABC0.17490.34980.52470.69960.8745SE +/- 0.000399, N = 3SE +/- 0.001575, N = 30.7761190.7775070.775055MIN: 0.7MIN: 0.71MIN: 0.71. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUABC246810Min: 0.78 / Avg: 0.78 / Max: 0.78Min: 0.77 / Avg: 0.78 / Max: 0.781. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUABC0.1790.3580.5370.7160.895SE +/- 0.008536, N = 3SE +/- 0.002635, N = 30.7862430.7954730.789910MIN: 0.7MIN: 0.7MIN: 0.71. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUABC246810Min: 0.78 / Avg: 0.8 / Max: 0.81Min: 0.79 / Avg: 0.79 / Max: 0.791. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUABC0.15360.30720.46080.61440.768SE +/- 0.002859, N = 3SE +/- 0.002683, N = 30.6819370.6827470.682228MIN: 0.59MIN: 0.58MIN: 0.581. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUABC246810Min: 0.68 / Avg: 0.68 / Max: 0.69Min: 0.68 / Avg: 0.68 / Max: 0.691. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

SMHasher

SMHasher is a hash function tester supporting various algorithms and able to make use of AVX and other modern CPU instruction set extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: MeowHash x86_64 AES-NIABC1224364860SE +/- 0.18, N = 3SE +/- 0.30, N = 353.5953.4653.361. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: MeowHash x86_64 AES-NIABC1122334455Min: 53.1 / Avg: 53.46 / Max: 53.65Min: 53 / Avg: 53.36 / Max: 53.951. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: MeowHash x86_64 AES-NIABC10K20K30K40K50KSE +/- 324.08, N = 3SE +/- 226.78, N = 346922.8346726.1147376.731. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: MeowHash x86_64 AES-NIABC8K16K24K32K40KMin: 46108.23 / Avg: 46726.11 / Max: 47204.61Min: 46948.21 / Avg: 47376.73 / Max: 47719.71. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: FarmHash128ABC1224364860SE +/- 0.44, N = 3SE +/- 0.45, N = 352.1151.8351.631. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: FarmHash128ABC1020304050Min: 50.96 / Avg: 51.83 / Max: 52.42Min: 50.73 / Avg: 51.63 / Max: 52.171. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: FarmHash128ABC4K8K12K16K20KSE +/- 147.41, N = 3SE +/- 92.08, N = 320617.6820202.8120364.501. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: FarmHash128ABC4K8K12K16K20KMin: 20018.31 / Avg: 20202.81 / Max: 20494.21Min: 20186.52 / Avg: 20364.5 / Max: 20494.441. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: t1ha0_aes_avx2 x86_64ABC510152025SE +/- 0.17, N = 9SE +/- 0.00, N = 322.1021.9522.101. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: t1ha0_aes_avx2 x86_64ABC510152025Min: 21.1 / Avg: 21.95 / Max: 22.57Min: 22.1 / Avg: 22.1 / Max: 22.11. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: t1ha0_aes_avx2 x86_64ABC20K40K60K80K100KSE +/- 679.73, N = 9SE +/- 940.93, N = 383813.9382904.4481425.191. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: t1ha0_aes_avx2 x86_64ABC15K30K45K60K75KMin: 79048.39 / Avg: 82904.44 / Max: 85305.34Min: 80253.43 / Avg: 81425.19 / Max: 83286.341. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

oneDNN

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUABC3691215SE +/- 0.05903, N = 3SE +/- 0.00928, N = 38.180957.957409.08580MIN: 7.86MIN: 7.47MIN: 8.891. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUABC3691215Min: 7.85 / Avg: 7.96 / Max: 8.05Min: 9.07 / Avg: 9.09 / Max: 9.11. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUABC0.10650.2130.31950.4260.5325SE +/- 0.004656, N = 3SE +/- 0.001374, N = 30.4676690.4599090.473113MIN: 0.43MIN: 0.41MIN: 0.441. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUABC12345Min: 0.45 / Avg: 0.46 / Max: 0.47Min: 0.47 / Avg: 0.47 / Max: 0.481. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

SMHasher

SMHasher is a hash function tester supporting various algorithms and able to make use of AVX and other modern CPU instruction set extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: Spooky32ABC714212835SE +/- 0.09, N = 3SE +/- 0.11, N = 329.9430.8231.201. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: Spooky32ABC714212835Min: 30.71 / Avg: 30.81 / Max: 30.99Min: 30.97 / Avg: 31.2 / Max: 31.341. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: Spooky32ABC4K8K12K16K20KSE +/- 10.48, N = 3SE +/- 43.76, N = 319879.9319280.7219028.971. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: Spooky32ABC3K6K9K12K15KMin: 19260.01 / Avg: 19280.72 / Max: 19293.83Min: 18964.14 / Avg: 19028.97 / Max: 19112.31. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: FarmHash32 x86_64 AVXABC714212835SE +/- 0.01, N = 3SE +/- 0.02, N = 328.7328.5228.561. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: FarmHash32 x86_64 AVXABC612182430Min: 28.52 / Avg: 28.52 / Max: 28.53Min: 28.54 / Avg: 28.56 / Max: 28.591. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: FarmHash32 x86_64 AVXABC7K14K21K28K35KSE +/- 102.23, N = 3SE +/- 52.01, N = 333990.6934077.2033941.331. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: FarmHash32 x86_64 AVXABC6K12K18K24K30KMin: 33874.95 / Avg: 34077.2 / Max: 34204.29Min: 33851.74 / Avg: 33941.33 / Max: 34031.891. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: t1ha2_atonceABC612182430SE +/- 0.27, N = 3SE +/- 0.43, N = 422.8823.1122.881. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: t1ha2_atonceABC510152025Min: 22.84 / Avg: 23.11 / Max: 23.66Min: 21.84 / Avg: 22.88 / Max: 23.941. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: t1ha2_atonceABC4K8K12K16K20KSE +/- 55.92, N = 3SE +/- 219.35, N = 418113.8518237.8218384.271. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: t1ha2_atonceABC3K6K9K12K15KMin: 18152.9 / Avg: 18237.82 / Max: 18343.31Min: 18117.32 / Avg: 18384.27 / Max: 19034.061. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: fasthash32ABC612182430SE +/- 0.10, N = 3SE +/- 0.16, N = 323.5124.6924.321. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: fasthash32ABC612182430Min: 24.5 / Avg: 24.69 / Max: 24.79Min: 24.02 / Avg: 24.32 / Max: 24.581. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: fasthash32ABC2K4K6K8K10KSE +/- 27.99, N = 3SE +/- 59.27, N = 38321.977974.218089.021. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: fasthash32ABC14002800420056007000Min: 7946.18 / Avg: 7974.21 / Max: 8030.18Min: 7990.29 / Avg: 8089.02 / Max: 8195.21. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

oneDNN

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUABC48121620SE +/- 0.04, N = 3SE +/- 0.03, N = 317.0316.8117.12MIN: 16.72MIN: 16.43MIN: 16.811. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUABC48121620Min: 16.73 / Avg: 16.81 / Max: 16.86Min: 17.09 / Avg: 17.12 / Max: 17.171. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUABC510152025SE +/- 0.06, N = 3SE +/- 0.05, N = 318.4218.2318.93MIN: 18.06MIN: 17.9MIN: 18.481. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUABC510152025Min: 18.12 / Avg: 18.23 / Max: 18.33Min: 18.85 / Avg: 18.93 / Max: 19.011. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

SMHasher

SMHasher is a hash function tester supporting various algorithms and able to make use of AVX and other modern CPU instruction set extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: wyhashABC48121620SE +/- 0.35, N = 3SE +/- 0.16, N = 314.9716.3215.421. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgcycles/hash, Fewer Is BetterSMHasher 2022-08-22Hash: wyhashABC48121620Min: 15.97 / Avg: 16.31 / Max: 17.01Min: 15.11 / Avg: 15.42 / Max: 15.631. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: wyhashABC6K12K18K24K30KSE +/- 37.17, N = 3SE +/- 110.01, N = 328600.2427320.4927857.671. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread
OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: wyhashABC5K10K15K20K25KMin: 27247.6 / Avg: 27320.49 / Max: 27369.58Min: 27739.5 / Avg: 27857.67 / Max: 28077.491. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects -lpthread

oneDNN

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUABC0.81781.63562.45343.27124.089SE +/- 0.00327, N = 3SE +/- 0.01247, N = 33.578053.634453.60370MIN: 3.41MIN: 3.41MIN: 3.41. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUABC246810Min: 3.63 / Avg: 3.63 / Max: 3.64Min: 3.58 / Avg: 3.6 / Max: 3.621. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUABC0.37950.7591.13851.5181.8975SE +/- 0.01458, N = 3SE +/- 0.00419, N = 31.659821.686781.67046MIN: 1.53MIN: 1.52MIN: 1.541. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUABC246810Min: 1.67 / Avg: 1.69 / Max: 1.72Min: 1.67 / Avg: 1.67 / Max: 1.681. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result.

C: The test run did not produce a result.

Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result.

C: The test run did not produce a result.

Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result.

C: The test run did not produce a result.

Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result.

C: The test run did not produce a result.

Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result.

C: The test run did not produce a result.

Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result.

C: The test run did not produce a result.

87 Results Shown

OpenRadioss
SMHasher:
  SHA3-256:
    cycles/hash
    MiB/sec
TensorFlow
OpenRadioss
oneDNN
TensorFlow:
  CPU - 16 - ResNet-50
  CPU - 32 - GoogLeNet
OpenRadioss
oneDNN:
  Recurrent Neural Network Training - f32 - CPU
  Recurrent Neural Network Inference - u8s8f32 - CPU
OpenRadioss:
  Rubber O-Ring Seal Installation
  Cell Phone Drop Test
oneDNN:
  Recurrent Neural Network Training - u8s8f32 - CPU
  Recurrent Neural Network Training - bf16bf16bf16 - CPU
  Recurrent Neural Network Inference - f32 - CPU
spaCy:
  en_core_web_trf
  en_core_web_lg
TensorFlow
Neural Magic DeepSparse:
  NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Synchronous Single-Stream:
    ms/batch
    items/sec
TensorFlow
Neural Magic DeepSparse:
  NLP Text Classification, BERT base uncased SST2 - Synchronous Single-Stream:
    ms/batch
    items/sec
  NLP Document Classification, oBERT base uncased on IMDB - Synchronous Single-Stream:
    ms/batch
    items/sec
  NLP Token Classification, BERT base uncased conll2003 - Synchronous Single-Stream:
    ms/batch
    items/sec
Y-Cruncher
Neural Magic DeepSparse:
  NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  CV Detection,YOLOv5s COCO - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  NLP Text Classification, DistilBERT mnli - Synchronous Single-Stream:
    ms/batch
    items/sec
  CV Detection,YOLOv5s COCO - Synchronous Single-Stream:
    ms/batch
    items/sec
  CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Stream:
    ms/batch
    items/sec
oneDNN
Neural Magic DeepSparse:
  CV Classification, ResNet-50 ImageNet - Synchronous Single-Stream:
    ms/batch
    items/sec
TensorFlow
oneDNN
QuadRay:
  5 - 4K
  2 - 4K
  3 - 4K
  1 - 4K
  5 - 1080p
  3 - 1080p
  2 - 1080p
  1 - 1080p
Y-Cruncher
oneDNN:
  IP Shapes 1D - f32 - CPU
  IP Shapes 1D - u8s8f32 - CPU
  Matrix Multiply Batch Shapes Transformer - f32 - CPU
  Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU
SMHasher:
  MeowHash x86_64 AES-NI:
    cycles/hash
    MiB/sec
  FarmHash128:
    cycles/hash
    MiB/sec
  t1ha0_aes_avx2 x86_64:
    cycles/hash
    MiB/sec
oneDNN:
  IP Shapes 3D - f32 - CPU
  IP Shapes 3D - u8s8f32 - CPU
SMHasher:
  Spooky32:
    cycles/hash
    MiB/sec
  FarmHash32 x86_64 AVX:
    cycles/hash
    MiB/sec
  t1ha2_atonce:
    cycles/hash
    MiB/sec
  fasthash32:
    cycles/hash
    MiB/sec
oneDNN:
  Convolution Batch Shapes Auto - f32 - CPU
  Convolution Batch Shapes Auto - u8s8f32 - CPU
SMHasher:
  wyhash:
    cycles/hash
    MiB/sec
oneDNN:
  Deconvolution Batch shapes_3d - f32 - CPU
  Deconvolution Batch shapes_3d - u8s8f32 - CPU