10900K sysbench onednn

Intel Core i9-10900K testing with a Gigabyte Z490 AORUS MASTER (F3 BIOS) and Gigabyte AMD Radeon RX 5500/5500M / Pro 5500M 8GB on Ubuntu 20.10 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2103134-PTS-10900KSY85
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts
Allow Limiting Results To Certain Suite(s)

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Toggle/Hide
Result
Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
1
March 13 2021
  40 Minutes
2
March 13 2021
  40 Minutes
3
March 13 2021
  40 Minutes
4
March 13 2021
  40 Minutes
Invert Behavior (Only Show Selected Data)
  40 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


10900K sysbench onednnProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilerFile-SystemScreen Resolution1234Intel Core i9-10900K @ 5.30GHz (10 Cores / 20 Threads)Gigabyte Z490 AORUS MASTER (F3 BIOS)Intel Comet Lake PCH16GBSamsung SSD 970 EVO 250GBGigabyte AMD Radeon RX 5500/5500M / Pro 5500M 8GB (1900/875MHz)Realtek ALC1220ASUS MG28UIntel + Intel Wi-Fi 6 AX201Ubuntu 20.105.11.0-051100rc2daily20210106-generic (x86_64) 20210105GNOME Shell 3.38.1X Server 1.20.94.6 Mesa 20.2.1 (LLVM 11.0.0)1.2.131GCC 10.2.0ext43840x2160OpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- Scaling Governor: intel_pstate powersave - CPU Microcode: 0xe0 - Thermald 2.3Security Details- itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected

1234Result OverviewPhoronix Test Suite100%101%102%103%104%oneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNSysbenchoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNSysbenchD.B.s - f32 - CPUR.N.N.I - u8s8f32 - CPUR.N.N.I - f32 - CPUIP Shapes 3D - u8s8f32 - CPUC.B.S.A - u8s8f32 - CPUR.N.N.I - bf16bf16bf16 - CPUM.M.B.S.T - f32 - CPUR.N.N.T - f32 - CPUM.M.B.S.T - u8s8f32 - CPUR.N.N.T - u8s8f32 - CPUD.B.s - u8s8f32 - CPURAM / MemoryD.B.s - f32 - CPUIP Shapes 3D - f32 - CPUD.B.s - u8s8f32 - CPUIP Shapes 1D - u8s8f32 - CPUR.N.N.T - bf16bf16bf16 - CPUC.B.S.A - f32 - CPUIP Shapes 1D - f32 - CPUCPU

10900K sysbench onednnonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUsysbench: RAM / Memoryonednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: IP Shapes 1D - f32 - CPUsysbench: CPUonednn: Deconvolution Batch shapes_1d - f32 - CPU12341779.521745.362.4500317.92801748.033.948072843.831.911152847.383.7651734129.114.8227012.21501.491381.172482834.4921.36213.3045626017.896.677241732.401729.052.4276917.65311742.263.924552837.761.932762837.573.7615034243.574.8135312.17851.494451.174652836.9221.33393.3054526029.636.854651754.381742.272.4720617.68991725.153.898272854.281.913192818.433.7745334315.674.8244112.16381.491961.172582836.9021.32773.3022826028.926.622391744.701764.342.4546817.77911749.053.897012818.991.933512844.563.7497634146.514.8345812.17121.492791.174242832.2521.32983.3005226019.116.59741OpenBenchmarking.org

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU4321400800120016002000SE +/- 6.63, N = 3SE +/- 19.63, N = 3SE +/- 13.35, N = 3SE +/- 8.77, N = 31744.701754.381732.401779.52MIN: 1681.33MIN: 1691.38MIN: 1680.02MIN: 1736.711. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU4321400800120016002000SE +/- 6.63, N = 3SE +/- 21.30, N = 3SE +/- 21.18, N = 3SE +/- 15.68, N = 31764.341742.271729.051745.36MIN: 1689.05MIN: 1683.83MIN: 1683.97MIN: 1705.831. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU43210.55621.11241.66862.22482.781SE +/- 0.01102, N = 3SE +/- 0.01612, N = 3SE +/- 0.01028, N = 3SE +/- 0.03980, N = 32.454682.472062.427692.45003MIN: 2.31MIN: 2.33MIN: 2.31MIN: 2.341. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU432148121620SE +/- 0.08, N = 3SE +/- 0.13, N = 3SE +/- 0.19, N = 3SE +/- 0.07, N = 317.7817.6917.6517.93MIN: 17.16MIN: 17.18MIN: 17.18MIN: 17.281. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU4321400800120016002000SE +/- 27.03, N = 3SE +/- 2.77, N = 3SE +/- 26.98, N = 3SE +/- 14.77, N = 31749.051725.151742.261748.03MIN: 1667.38MIN: 1683.55MIN: 1671.61MIN: 1687.971. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU43210.88831.77662.66493.55324.4415SE +/- 0.00827, N = 3SE +/- 0.00147, N = 3SE +/- 0.00768, N = 3SE +/- 0.01417, N = 33.897013.898273.924553.94807MIN: 3.72MIN: 3.76MIN: 3.77MIN: 3.791. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU43216001200180024003000SE +/- 16.37, N = 3SE +/- 23.35, N = 3SE +/- 5.58, N = 3SE +/- 10.02, N = 32818.992854.282837.762843.83MIN: 2765.73MIN: 2775.53MIN: 2798.07MIN: 2768.241. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU43210.4350.871.3051.742.175SE +/- 0.01806, N = 3SE +/- 0.00284, N = 3SE +/- 0.01672, N = 3SE +/- 0.00175, N = 31.933511.913191.932761.91115MIN: 1.88MIN: 1.88MIN: 1.88MIN: 1.881. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU43216001200180024003000SE +/- 5.94, N = 3SE +/- 6.65, N = 3SE +/- 15.45, N = 3SE +/- 3.12, N = 32844.562818.432837.572847.38MIN: 2768.09MIN: 2766.26MIN: 2767.4MIN: 2765.941. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU43210.84931.69862.54793.39724.2465SE +/- 0.00653, N = 3SE +/- 0.00173, N = 3SE +/- 0.00158, N = 3SE +/- 0.00296, N = 33.749763.774533.761503.76517MIN: 3.72MIN: 3.75MIN: 3.74MIN: 3.741. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Sysbench

This is a benchmark of Sysbench with the built-in CPU and memory sub-tests. Sysbench is a scriptable multi-threaded benchmark tool based on LuaJIT. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/sec, More Is BetterSysbench 1.0.20Test: RAM / Memory43217K14K21K28K35KSE +/- 148.70, N = 3SE +/- 108.17, N = 3SE +/- 99.86, N = 3SE +/- 68.06, N = 334146.5134315.6734243.5734129.111. (CC) gcc options: -pthread -O2 -funroll-loops -rdynamic -ldl -laio -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU43211.08782.17563.26344.35125.439SE +/- 0.00473, N = 3SE +/- 0.00815, N = 3SE +/- 0.00575, N = 3SE +/- 0.00884, N = 34.834584.824414.813534.82270MIN: 4.73MIN: 4.75MIN: 4.71MIN: 4.791. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU43213691215SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 312.1712.1612.1812.22MIN: 12.06MIN: 12.05MIN: 12.08MIN: 12.11. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU43210.33630.67261.00891.34521.6815SE +/- 0.00086, N = 3SE +/- 0.00095, N = 3SE +/- 0.00187, N = 3SE +/- 0.00072, N = 31.492791.491961.494451.49138MIN: 1.48MIN: 1.48MIN: 1.48MIN: 1.481. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU43210.26430.52860.79291.05721.3215SE +/- 0.00107, N = 3SE +/- 0.00090, N = 3SE +/- 0.00194, N = 3SE +/- 0.00221, N = 31.174241.172581.174651.17248MIN: 1.16MIN: 1.16MIN: 1.16MIN: 1.161. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU43216001200180024003000SE +/- 11.43, N = 3SE +/- 6.62, N = 3SE +/- 0.84, N = 3SE +/- 12.24, N = 32832.252836.902836.922834.49MIN: 2770.76MIN: 2764.69MIN: 2770.47MIN: 2770.11. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU4321510152025SE +/- 0.02, N = 3SE +/- 0.01, N = 3SE +/- 0.02, N = 3SE +/- 0.01, N = 321.3321.3321.3321.36MIN: 21.24MIN: 21.25MIN: 21.24MIN: 21.291. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU43210.74371.48742.23112.97483.7185SE +/- 0.01635, N = 3SE +/- 0.00348, N = 3SE +/- 0.00232, N = 3SE +/- 0.00272, N = 33.300523.302283.305453.30456MIN: 3.03MIN: 3.02MIN: 3.03MIN: 3.031. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Sysbench

This is a benchmark of Sysbench with the built-in CPU and memory sub-tests. Sysbench is a scriptable multi-threaded benchmark tool based on LuaJIT. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgEvents Per Second, More Is BetterSysbench 1.0.20Test: CPU43216K12K18K24K30KSE +/- 1.33, N = 3SE +/- 1.44, N = 3SE +/- 5.92, N = 3SE +/- 2.59, N = 326019.1126028.9226029.6326017.891. (CC) gcc options: -pthread -O2 -funroll-loops -rdynamic -ldl -laio -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU4321246810SE +/- 0.11275, N = 15SE +/- 0.09531, N = 15SE +/- 0.13332, N = 15SE +/- 0.09219, N = 156.597416.622396.854656.67724MIN: 3.47MIN: 3.47MIN: 3.45MIN: 3.451. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl