3990X sysbench onednn

AMD Ryzen Threadripper 3990X 64-Core testing with a System76 Thelio Major (F4c Z5 BIOS) and AMD Radeon RX 5600 OEM/5600 XT / 5700/5700 8GB on Pop 20.10 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2103132-PTS-3990XSYS94
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts
Allow Limiting Results To Certain Suite(s)

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Toggle/Hide
Result
Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
1
March 13 2021
  37 Minutes
2
March 13 2021
  37 Minutes
3
March 13 2021
  37 Minutes
4
March 13 2021
  37 Minutes
Invert Behavior (Only Show Selected Data)
  37 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


3990X sysbench onednnProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilerFile-SystemScreen Resolution1234AMD Ryzen Threadripper 3990X 64-Core @ 2.90GHz (64 Cores / 128 Threads)System76 Thelio Major (F4c Z5 BIOS)AMD Starship/Matisse126GBSamsung SSD 970 EVO Plus 500GBAMD Radeon RX 5600 OEM/5600 XT / 5700/5700 8GB (1750/875MHz)AMD Navi 10 HDMI AudioDELL P2415QIntel I211 + Intel Wi-Fi 6 AX200Pop 20.105.8.0-7630-generic (x86_64)GNOME Shell 3.38.2X Server 1.20.84.6 Mesa 21.1.0-devel (git-96d7555 2021-01-22 groovy-oibaf-ppa) (LLVM 11.0.1)1.2.145GCC 10.2.0 + Clang 11.0.1-1~oibaf~gext43840x2160OpenBenchmarking.orgKernel Details- snd_usb_audio.ignore_ctl_error=1 - Transparent Huge Pages: madviseCompiler Details- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- Scaling Governor: acpi-cpufreq ondemand (Boost: Enabled) - CPU Microcode: 0x8301025Security Details- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional STIBP: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected

1234Result OverviewPhoronix Test Suite100%115%130%145%160%oneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNoneDNNSysbenchoneDNNSysbenchoneDNNIP Shapes 3D - f32 - CPUC.B.S.A - f32 - CPUIP Shapes 3D - u8s8f32 - CPUIP Shapes 1D - u8s8f32 - CPUC.B.S.A - u8s8f32 - CPUD.B.s - f32 - CPUIP Shapes 1D - f32 - CPUR.N.N.T - u8s8f32 - CPUM.M.B.S.T - f32 - CPUR.N.N.T - f32 - CPUD.B.s - f32 - CPUD.B.s - u8s8f32 - CPUD.B.s - u8s8f32 - CPUR.N.N.T - bf16bf16bf16 - CPUR.N.N.I - u8s8f32 - CPUM.M.B.S.T - u8s8f32 - CPUCPUR.N.N.I - bf16bf16bf16 - CPURAM / MemoryR.N.N.I - f32 - CPU

3990X sysbench onednnonednn: IP Shapes 3D - f32 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: IP Shapes 1D - f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUsysbench: CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUsysbench: RAM / Memoryonednn: Recurrent Neural Network Inference - f32 - CPU12345.372210.9461251.0021511.472436.527837.566501.235083780.640.4119153741.532.021310.8909760.9479413750.35743.1970.667705124312.60742.7887515.51743.9728.442701.025491.026891.489456.636257.624991.257243750.450.4154203770.992.035850.8967070.9547593759.67747.6330.668714123807.73744.7877520.15744.4718.576651.081291.032521.501436.588487.591921.249533760.170.4119183747.702.043020.8954180.9538203747.38745.8170.667192123768.96743.5817511.92746.4508.119461.082801.035481.462916.656087.706341.238323721.680.4171223730.172.026470.8894030.9540163733.74747.8140.670735123660.73746.7027484.70744.352OpenBenchmarking.org

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU1234246810SE +/- 0.01315, N = 3SE +/- 0.11123, N = 3SE +/- 0.01501, N = 3SE +/- 0.02779, N = 35.372218.442708.576658.11946MIN: 5.17MIN: 8.09MIN: 8.33MIN: 7.821. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU12340.24360.48720.73080.97441.218SE +/- 0.006847, N = 3SE +/- 0.016517, N = 3SE +/- 0.002000, N = 3SE +/- 0.009734, N = 30.9461251.0254901.0812901.082800MIN: 0.85MIN: 0.92MIN: 0.98MIN: 0.981. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU12340.2330.4660.6990.9321.165SE +/- 0.001528, N = 3SE +/- 0.002666, N = 3SE +/- 0.002913, N = 3SE +/- 0.002595, N = 31.0021511.0268901.0325201.035480MIN: 0.95MIN: 0.98MIN: 0.98MIN: 0.981. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU12340.33780.67561.01341.35121.689SE +/- 0.00777, N = 3SE +/- 0.00450, N = 3SE +/- 0.00467, N = 3SE +/- 0.00627, N = 31.472431.489451.501431.46291MIN: 1.26MIN: 1.28MIN: 1.28MIN: 1.241. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU1234246810SE +/- 0.01403, N = 3SE +/- 0.02161, N = 3SE +/- 0.00793, N = 3SE +/- 0.01940, N = 36.527836.636256.588486.65608MIN: 6.36MIN: 6.5MIN: 6.47MIN: 6.511. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU1234246810SE +/- 0.02931, N = 3SE +/- 0.02617, N = 3SE +/- 0.00967, N = 3SE +/- 0.01670, N = 37.566507.624997.591927.70634MIN: 6.16MIN: 6.17MIN: 6.56MIN: 6.611. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU12340.28290.56580.84871.13161.4145SE +/- 0.00292, N = 3SE +/- 0.00760, N = 3SE +/- 0.00826, N = 3SE +/- 0.00828, N = 31.235081.257241.249531.23832MIN: 1.19MIN: 1.2MIN: 1.2MIN: 1.191. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU12348001600240032004000SE +/- 10.06, N = 3SE +/- 8.77, N = 3SE +/- 6.98, N = 3SE +/- 13.27, N = 33780.643750.453760.173721.68MIN: 3744.41MIN: 3724.77MIN: 3731.98MIN: 36821. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU12340.09390.18780.28170.37560.4695SE +/- 0.000765, N = 3SE +/- 0.000908, N = 3SE +/- 0.002080, N = 3SE +/- 0.005514, N = 30.4119150.4154200.4119180.417122MIN: 0.4MIN: 0.4MIN: 0.4MIN: 0.41. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU12348001600240032004000SE +/- 3.42, N = 3SE +/- 3.71, N = 3SE +/- 5.93, N = 3SE +/- 10.20, N = 33741.533770.993747.703730.17MIN: 3720.38MIN: 3740.38MIN: 3723.93MIN: 3690.551. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU12340.45970.91941.37911.83882.2985SE +/- 0.00410, N = 3SE +/- 0.00592, N = 3SE +/- 0.00417, N = 3SE +/- 0.00241, N = 32.021312.035852.043022.02647MIN: 1.93MIN: 1.96MIN: 1.97MIN: 1.961. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU12340.20180.40360.60540.80721.009SE +/- 0.001294, N = 3SE +/- 0.002296, N = 3SE +/- 0.002330, N = 3SE +/- 0.000457, N = 30.8909760.8967070.8954180.889403MIN: 0.84MIN: 0.84MIN: 0.84MIN: 0.841. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU12340.21480.42960.64440.85921.074SE +/- 0.002129, N = 3SE +/- 0.001253, N = 3SE +/- 0.000086, N = 3SE +/- 0.000529, N = 30.9479410.9547590.9538200.954016MIN: 0.9MIN: 0.9MIN: 0.89MIN: 0.91. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU12348001600240032004000SE +/- 3.01, N = 3SE +/- 4.29, N = 3SE +/- 2.18, N = 3SE +/- 13.03, N = 33750.353759.673747.383733.74MIN: 3727.41MIN: 3726.65MIN: 3726.51MIN: 3691.911. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU1234160320480640800SE +/- 1.10, N = 3SE +/- 1.00, N = 3SE +/- 0.40, N = 3SE +/- 1.95, N = 3743.20747.63745.82747.81MIN: 734.28MIN: 738.57MIN: 737.38MIN: 737.341. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU12340.15090.30180.45270.60360.7545SE +/- 0.001823, N = 3SE +/- 0.000941, N = 3SE +/- 0.000661, N = 3SE +/- 0.001611, N = 30.6677050.6687140.6671920.670735MIN: 0.64MIN: 0.64MIN: 0.64MIN: 0.641. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Sysbench

This is a benchmark of Sysbench with the built-in CPU and memory sub-tests. Sysbench is a scriptable multi-threaded benchmark tool based on LuaJIT. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgEvents Per Second, More Is BetterSysbench 1.0.20Test: CPU123430K60K90K120K150KSE +/- 312.36, N = 3SE +/- 228.83, N = 3SE +/- 281.49, N = 3SE +/- 209.17, N = 3124312.60123807.73123768.96123660.731. (CC) gcc options: -pthread -O2 -funroll-loops -rdynamic -ldl -laio -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU1234160320480640800SE +/- 0.53, N = 3SE +/- 0.96, N = 3SE +/- 1.75, N = 3SE +/- 0.55, N = 3742.79744.79743.58746.70MIN: 734.92MIN: 736.3MIN: 733.02MIN: 738.991. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Sysbench

This is a benchmark of Sysbench with the built-in CPU and memory sub-tests. Sysbench is a scriptable multi-threaded benchmark tool based on LuaJIT. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/sec, More Is BetterSysbench 1.0.20Test: RAM / Memory123416003200480064008000SE +/- 11.15, N = 3SE +/- 9.49, N = 3SE +/- 12.72, N = 3SE +/- 8.95, N = 37515.517520.157511.927484.701. (CC) gcc options: -pthread -O2 -funroll-loops -rdynamic -ldl -laio -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU1234160320480640800SE +/- 2.22, N = 3SE +/- 0.50, N = 3SE +/- 1.44, N = 3SE +/- 1.34, N = 3743.97744.47746.45744.35MIN: 733.98MIN: 735.64MIN: 736.4MIN: 734.631. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl