ss4

Intel Core i9-7980XE testing with a ASRock X299E-ITX/ac (P1.60 BIOS) and llvmpipe on Ubuntu 20.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2107058-IB-SS431414552
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results

Limit displaying results to tests within:

C/C++ Compiler Tests 2 Tests
CPU Massive 5 Tests
Creator Workloads 2 Tests
HPC - High Performance Computing 3 Tests
Multi-Core 4 Tests
OpenMPI Tests 2 Tests
Server CPU Tests 4 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
sysbench-i9Ph10
July 05 2021
  7 Minutes
graphics-magick-i9Ph10
July 05 2021
  24 Minutes
ipc-benchmark-i9Ph10
July 05 2021
  4 Minutes
amg-i9Ph10
July 05 2021
  2 Minutes
ramspeed-i9Ph10
July 05 2021
  9 Minutes
ramspeed-i9Ph10-2
July 05 2021
  20 Minutes
npb-i9Ph10
July 05 2021
  5 Minutes
onednn-i9Ph10
July 05 2021
  39 Minutes
scimark-i9Ph10
July 05 2021
  2 Minutes
Invert Hiding All Results Option
  13 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


ss4OpenBenchmarking.orgPhoronix Test SuiteIntel Core i9-7980XE @ 4.40GHz (18 Cores / 36 Threads)ASRock X299E-ITX/ac (P1.60 BIOS)Intel Sky Lake-E DMI3 Registers32GB512GB Western Digital CL SN520 SDAPNUW-512G-1022llvmpipeRealtek ALC1220Intel I219-V + Intel I211 + 2 x Intel 10-Gigabit X540-AT2 + Intel 8265 / 8275Ubuntu 20.045.8.0-55-generic (x86_64)GNOME Shell 3.36.7X Server 1.20.9NVIDIA4.5 Mesa 20.2.6 (LLVM 11.0.0 256 bits)GCC 9.3.0ext43840x1080ProcessorMotherboardChipsetMemoryDiskGraphicsAudioNetworkOSKernelDesktopDisplay ServerDisplay DriverOpenGLCompilerFile-SystemScreen ResolutionSs4 BenchmarksSystem Logs- Transparent Huge Pages: madvise- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-9-HskZEa/gcc-9-9.3.0/debian/tmp-nvptx/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: intel_cpufreq ondemand - CPU Microcode: 0x2006b06- itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Mitigation of PTE Inversion; VMX: conditional cache flushes SMT vulnerable + mds: Mitigation of Clear buffers; SMT vulnerable + meltdown: Mitigation of PTI + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + srbds: Not affected + tsx_async_abort: Mitigation of Clear buffers; SMT vulnerable

ss4sysbench: RAM / Memorysysbench: CPUgraphics-magick: Swirlgraphics-magick: Rotategraphics-magick: Sharpengraphics-magick: Enhancedgraphics-magick: Resizinggraphics-magick: Noise-Gaussiangraphics-magick: HWB Color Spaceipc-benchmark: TCP Socket - 128ipc-benchmark: TCP Socket - 1024ipc-benchmark: Unnamed Pipe - 128ipc-benchmark: Unnamed Pipe - 1024ipc-benchmark: FIFO Named Pipe - 128ipc-benchmark: FIFO Named Pipe - 1024ipc-benchmark: Unnamed Unix Domain Socket - 128ipc-benchmark: Unnamed Unix Domain Socket - 1024amg: ramspeed: Add - Integerramspeed: Scale - Integerramspeed: Average - Integerramspeed: Add - Floating Pointramspeed: Scale - Floating Pointramspeed: Average - Floating Pointnpb: EP.Cnpb: EP.Donednn: IP Shapes 1D - f32 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUonednn: IP Shapes 1D - bf16bf16bf16 - CPUonednn: IP Shapes 3D - bf16bf16bf16 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - bf16bf16bf16 - CPUonednn: Deconvolution Batch shapes_1d - bf16bf16bf16 - CPUonednn: Deconvolution Batch shapes_3d - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - bf16bf16bf16 - CPUscimark2: Compositescimark2: Monte Carloscimark2: Fast Fourier Transformscimark2: Sparse Matrix Multiplyscimark2: Dense LU Matrix Factorizationscimark2: Jacobi Successive Over-Relaxationsysbench-i9Ph10graphics-magick-i9Ph10ipc-benchmark-i9Ph10amg-i9Ph10ramspeed-i9Ph10ramspeed-i9Ph10-2npb-i9Ph10onednn-i9Ph10scimark-i9Ph1012623.7412520.7024475877125581155711161872012314281736829152904618001081547106110114298201342166376729880.8428949.5129409.0829791.0728949.6028998.3529206.7328492.5928905.771506.402072.0429.894360.18063.1982229.512415.335722.07399.3850410.26986.224218.692303.327603.756184016.912528.574016.0722.183231.225526.74282530.552.542614015.842529.351.858195.43817667.44145.14306.37759.82938.771187.12OpenBenchmarking.org

Sysbench

This is a benchmark of Sysbench with the built-in CPU and memory sub-tests. Sysbench is a scriptable multi-threaded benchmark tool based on LuaJIT. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/sec, More Is BetterSysbench 1.0.20Test: RAM / Memorysysbench-i9Ph103K6K9K12K15KSE +/- 43.24, N = 312623.741. (CC) gcc options: -pthread -O2 -funroll-loops -rdynamic -ldl -laio -lm

OpenBenchmarking.orgEvents Per Second, More Is BetterSysbench 1.0.20Test: CPUsysbench-i9Ph103K6K9K12K15KSE +/- 3.16, N = 312520.701. (CC) gcc options: -pthread -O2 -funroll-loops -rdynamic -ldl -laio -lm

GraphicsMagick

This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.33Operation: Swirlgraphics-magick-i9Ph10501001502002502441. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lz -lm -lpthread

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.33Operation: Rotategraphics-magick-i9Ph10160320480640800SE +/- 8.25, N = 57581. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lz -lm -lpthread

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.33Operation: Sharpengraphics-magick-i9Ph1020406080100SE +/- 0.33, N = 3771. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lz -lm -lpthread

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.33Operation: Enhancedgraphics-magick-i9Ph103060901201501251. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lz -lm -lpthread

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.33Operation: Resizinggraphics-magick-i9Ph10130260390520650SE +/- 0.58, N = 35811. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lz -lm -lpthread

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.33Operation: Noise-Gaussiangraphics-magick-i9Ph103060901201501551. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lz -lm -lpthread

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.33Operation: HWB Color Spacegraphics-magick-i9Ph10150300450600750SE +/- 1.20, N = 37111. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lz -lm -lpthread

IPC_benchmark

IPC_benchmark is a Linux inter-process communication benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMessages Per Second, More Is BetterIPC_benchmarkType: TCP Socket - Message Bytes: 128ipc-benchmark-i9Ph10300K600K900K1200K1500KSE +/- 1126.58, N = 31618720

OpenBenchmarking.orgMessages Per Second, More Is BetterIPC_benchmarkType: TCP Socket - Message Bytes: 1024ipc-benchmark-i9Ph10300K600K900K1200K1500KSE +/- 7215.07, N = 31231428

OpenBenchmarking.orgMessages Per Second, More Is BetterIPC_benchmarkType: Unnamed Pipe - Message Bytes: 128ipc-benchmark-i9Ph10400K800K1200K1600K2000KSE +/- 29706.06, N = 151736829

OpenBenchmarking.orgMessages Per Second, More Is BetterIPC_benchmarkType: Unnamed Pipe - Message Bytes: 1024ipc-benchmark-i9Ph10300K600K900K1200K1500KSE +/- 21356.75, N = 31529046

OpenBenchmarking.orgMessages Per Second, More Is BetterIPC_benchmarkType: FIFO Named Pipe - Message Bytes: 128ipc-benchmark-i9Ph10400K800K1200K1600K2000KSE +/- 25575.15, N = 31800108

OpenBenchmarking.orgMessages Per Second, More Is BetterIPC_benchmarkType: FIFO Named Pipe - Message Bytes: 1024ipc-benchmark-i9Ph10300K600K900K1200K1500KSE +/- 5801.12, N = 31547106

OpenBenchmarking.orgMessages Per Second, More Is BetterIPC_benchmarkType: Unnamed Unix Domain Socket - Message Bytes: 128ipc-benchmark-i9Ph10200K400K600K800K1000KSE +/- 1288.51, N = 31101142

OpenBenchmarking.orgMessages Per Second, More Is BetterIPC_benchmarkType: Unnamed Unix Domain Socket - Message Bytes: 1024ipc-benchmark-i9Ph10200K400K600K800K1000KSE +/- 660.09, N = 3982013

Algebraic Multi-Grid Benchmark

AMG is a parallel algebraic multigrid solver for linear systems arising from problems on unstructured grids. The driver provided with AMG builds linear systems for various 3-dimensional problems. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFigure Of Merit, More Is BetterAlgebraic Multi-Grid Benchmark 1.2amg-i9Ph1090M180M270M360M450MSE +/- 226558.75, N = 34216637671. (CC) gcc options: -lparcsr_ls -lparcsr_mv -lseq_mv -lIJ_mv -lkrylov -lHYPRE_utilities -lm -fopenmp -pthread -lmpi

RAMspeed SMP

This benchmark tests the system memory (RAM) performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterRAMspeed SMP 3.5.0Type: Add - Benchmark: Integerramspeed-i9Ph10ramspeed-i9Ph10-26K12K18K24K30KSE +/- 10.07, N = 3SE +/- 33.36, N = 329880.8429791.071. (CC) gcc options: -O3 -march=native
OpenBenchmarking.orgMB/s, More Is BetterRAMspeed SMP 3.5.0Type: Add - Benchmark: Integerramspeed-i9Ph10ramspeed-i9Ph10-25K10K15K20K25KMin: 29869.58 / Avg: 29880.84 / Max: 29900.92Min: 29726.13 / Avg: 29791.07 / Max: 29836.781. (CC) gcc options: -O3 -march=native

OpenBenchmarking.orgMB/s, More Is BetterRAMspeed SMP 3.5.0Type: Scale - Benchmark: Integerramspeed-i9Ph10ramspeed-i9Ph10-26K12K18K24K30KSE +/- 19.97, N = 3SE +/- 11.65, N = 328949.5128949.601. (CC) gcc options: -O3 -march=native
OpenBenchmarking.orgMB/s, More Is BetterRAMspeed SMP 3.5.0Type: Scale - Benchmark: Integerramspeed-i9Ph10ramspeed-i9Ph10-25K10K15K20K25KMin: 28920.24 / Avg: 28949.51 / Max: 28987.67Min: 28926.83 / Avg: 28949.6 / Max: 28965.221. (CC) gcc options: -O3 -march=native

OpenBenchmarking.orgMB/s, More Is BetterRAMspeed SMP 3.5.0Type: Average - Benchmark: Integerramspeed-i9Ph10ramspeed-i9Ph10-26K12K18K24K30KSE +/- 3.06, N = 3SE +/- 14.58, N = 329409.0828998.351. (CC) gcc options: -O3 -march=native
OpenBenchmarking.orgMB/s, More Is BetterRAMspeed SMP 3.5.0Type: Average - Benchmark: Integerramspeed-i9Ph10ramspeed-i9Ph10-25K10K15K20K25KMin: 29402.97 / Avg: 29409.08 / Max: 29412.51Min: 28973.6 / Avg: 28998.35 / Max: 29024.071. (CC) gcc options: -O3 -march=native

OpenBenchmarking.orgMB/s, More Is BetterRAMspeed SMP 3.5.0Type: Add - Benchmark: Floating Pointramspeed-i9Ph10-26K12K18K24K30KSE +/- 156.20, N = 329206.731. (CC) gcc options: -O3 -march=native

OpenBenchmarking.orgMB/s, More Is BetterRAMspeed SMP 3.5.0Type: Scale - Benchmark: Floating Pointramspeed-i9Ph10-26K12K18K24K30KSE +/- 42.13, N = 328492.591. (CC) gcc options: -O3 -march=native

OpenBenchmarking.orgMB/s, More Is BetterRAMspeed SMP 3.5.0Type: Average - Benchmark: Floating Pointramspeed-i9Ph10-26K12K18K24K30KSE +/- 20.30, N = 328905.771. (CC) gcc options: -O3 -march=native

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.Cnpb-i9Ph1030060090012001500SE +/- 14.95, N = 151506.401. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi2. Open MPI 4.0.3

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.Dnpb-i9Ph10400800120016002000SE +/- 11.23, N = 32072.041. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi2. Open MPI 4.0.3

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUonednn-i9Ph10714212835SE +/- 0.31, N = 329.89MIN: 17.051. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUonednn-i9Ph101326395265SE +/- 0.57, N = 360.18MIN: 27.051. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUonednn-i9Ph100.71961.43922.15882.87843.598SE +/- 0.00253, N = 33.19822MIN: 2.791. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUonednn-i9Ph10714212835SE +/- 0.24, N = 1529.51MIN: 4.141. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPUonednn-i9Ph1048121620SE +/- 0.01, N = 315.34MIN: 14.731. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPUonednn-i9Ph10510152025SE +/- 0.20, N = 1522.07MIN: 6.981. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUonednn-i9Ph103691215SE +/- 0.00136, N = 39.38504MIN: 9.241. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUonednn-i9Ph103691215SE +/- 0.04, N = 310.27MIN: 7.431. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUonednn-i9Ph10246810SE +/- 0.00670, N = 36.22421MIN: 6.11. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUonednn-i9Ph10246810SE +/- 0.00069, N = 38.69230MIN: 8.51. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUonednn-i9Ph100.74871.49742.24612.99483.7435SE +/- 0.00103, N = 33.32760MIN: 2.811. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUonednn-i9Ph100.84511.69022.53533.38044.2255SE +/- 0.02427, N = 33.75618MIN: 3.641. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUonednn-i9Ph109001800270036004500SE +/- 0.94, N = 34016.91MIN: 3981.581. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUonednn-i9Ph105001000150020002500SE +/- 1.65, N = 32528.57MIN: 2493.371. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUonednn-i9Ph109001800270036004500SE +/- 1.46, N = 34016.07MIN: 3984.541. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPUonednn-i9Ph10510152025SE +/- 0.01, N = 322.18MIN: 21.951. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPUonednn-i9Ph10714212835SE +/- 0.01, N = 331.23MIN: 30.351. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPUonednn-i9Ph10612182430SE +/- 0.03, N = 326.74MIN: 26.271. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUonednn-i9Ph105001000150020002500SE +/- 0.81, N = 32530.55MIN: 2497.651. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUonednn-i9Ph100.57211.14421.71632.28842.8605SE +/- 0.00332, N = 32.54261MIN: 2.151. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUonednn-i9Ph109001800270036004500SE +/- 1.16, N = 34015.84MIN: 3980.891. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUonednn-i9Ph105001000150020002500SE +/- 1.04, N = 32529.35MIN: 2494.551. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUonednn-i9Ph100.41810.83621.25431.67242.0905SE +/- 0.00144, N = 31.85819MIN: 1.491. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPUonednn-i9Ph101.22362.44723.67084.89446.118SE +/- 0.00321, N = 35.43817MIN: 4.91. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

SciMark

This test runs the ANSI C version of SciMark 2.0, which is a benchmark for scientific and numerical computing developed by programmers at the National Institute of Standards and Technology. This test is made up of Fast Foruier Transform, Jacobi Successive Over-relaxation, Monte Carlo, Sparse Matrix Multiply, and dense LU matrix factorization benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMflops, More Is BetterSciMark 2.0Computational Test: Compositescimark-i9Ph10140280420560700SE +/- 0.38, N = 3667.441. (CC) gcc options: -lm

OpenBenchmarking.orgMflops, More Is BetterSciMark 2.0Computational Test: Monte Carloscimark-i9Ph10306090120150SE +/- 0.02, N = 3145.141. (CC) gcc options: -lm

OpenBenchmarking.orgMflops, More Is BetterSciMark 2.0Computational Test: Fast Fourier Transformscimark-i9Ph1070140210280350SE +/- 0.93, N = 3306.371. (CC) gcc options: -lm

OpenBenchmarking.orgMflops, More Is BetterSciMark 2.0Computational Test: Sparse Matrix Multiplyscimark-i9Ph10160320480640800SE +/- 0.38, N = 3759.821. (CC) gcc options: -lm

OpenBenchmarking.orgMflops, More Is BetterSciMark 2.0Computational Test: Dense LU Matrix Factorizationscimark-i9Ph102004006008001000SE +/- 0.62, N = 3938.771. (CC) gcc options: -lm

OpenBenchmarking.orgMflops, More Is BetterSciMark 2.0Computational Test: Jacobi Successive Over-Relaxationscimark-i9Ph1030060090012001500SE +/- 0.14, N = 31187.121. (CC) gcc options: -lm

56 Results Shown

Sysbench:
  RAM / Memory
  CPU
GraphicsMagick:
  Swirl
  Rotate
  Sharpen
  Enhanced
  Resizing
  Noise-Gaussian
  HWB Color Space
IPC_benchmark:
  TCP Socket - 128
  TCP Socket - 1024
  Unnamed Pipe - 128
  Unnamed Pipe - 1024
  FIFO Named Pipe - 128
  FIFO Named Pipe - 1024
  Unnamed Unix Domain Socket - 128
  Unnamed Unix Domain Socket - 1024
Algebraic Multi-Grid Benchmark
RAMspeed SMP:
  Add - Integer
  Scale - Integer
  Average - Integer
  Add - Floating Point
  Scale - Floating Point
  Average - Floating Point
NAS Parallel Benchmarks:
  EP.C
  EP.D
oneDNN:
  IP Shapes 1D - f32 - CPU
  IP Shapes 3D - f32 - CPU
  IP Shapes 1D - u8s8f32 - CPU
  IP Shapes 3D - u8s8f32 - CPU
  IP Shapes 1D - bf16bf16bf16 - CPU
  IP Shapes 3D - bf16bf16bf16 - CPU
  Convolution Batch Shapes Auto - f32 - CPU
  Deconvolution Batch shapes_1d - f32 - CPU
  Deconvolution Batch shapes_3d - f32 - CPU
  Convolution Batch Shapes Auto - u8s8f32 - CPU
  Deconvolution Batch shapes_1d - u8s8f32 - CPU
  Deconvolution Batch shapes_3d - u8s8f32 - CPU
  Recurrent Neural Network Training - f32 - CPU
  Recurrent Neural Network Inference - f32 - CPU
  Recurrent Neural Network Training - u8s8f32 - CPU
  Convolution Batch Shapes Auto - bf16bf16bf16 - CPU
  Deconvolution Batch shapes_1d - bf16bf16bf16 - CPU
  Deconvolution Batch shapes_3d - bf16bf16bf16 - CPU
  Recurrent Neural Network Inference - u8s8f32 - CPU
  Matrix Multiply Batch Shapes Transformer - f32 - CPU
  Recurrent Neural Network Training - bf16bf16bf16 - CPU
  Recurrent Neural Network Inference - bf16bf16bf16 - CPU
  Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU
  Matrix Multiply Batch Shapes Transformer - bf16bf16bf16 - CPU
SciMark:
  Composite
  Monte Carlo
  Fast Fourier Transform
  Sparse Matrix Multiply
  Dense LU Matrix Factorization
  Jacobi Successive Over-Relaxation