5900X sysbench onednn

AMD Ryzen 9 5900X 12-Core testing with a ASUS ROG CROSSHAIR VIII HERO (3202 BIOS) and Sapphire AMD Radeon RX 5600 OEM/5600 XT / 5700/5700 6GB on Ubuntu 20.10 via the Phoronix Test Suite.

HTML result view exported from: https://openbenchmarking.org/result/2103138-SYST-5900XSY95.

5900X sysbench onednnProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilerFile-SystemScreen Resolution123AMD Ryzen 9 5900X 12-Core @ 3.70GHz (12 Cores / 24 Threads)ASUS ROG CROSSHAIR VIII HERO (3202 BIOS)AMD Starship/Matisse16GB1000GB Sabrent Rocket 4.0 PlusSapphire AMD Radeon RX 5600 OEM/5600 XT / 5700/5700 6GB (1780/875MHz)AMD Navi 10 HDMI AudioASUS VP28URealtek RTL8125 2.5GbE + Intel I211Ubuntu 20.105.12.0-051200rc2-generic (x86_64) 20210306GNOME Shell 3.38.1X Server 1.20.94.6 Mesa 20.2.1 (LLVM 11.0.0)1.2.131GCC 10.2.0ext43840x2160OpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0xa201009Security Details- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional IBRS_FW STIBP: always-on RSB filling + srbds: Not affected + tsx_async_abort: Not affected

5900X sysbench onednnonednn: IP Shapes 1D - f32 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUsysbench: RAM / Memorysysbench: CPU1233.752678.627171.069660.52449916.19405.538454.2792817.35411.316292.016132890.591672.052904.051707.040.7601812897.911697.371.7408313950.1068475.653.771077.533911.069810.86221615.72015.320474.2662416.46571.309402.013202892.531689.612901.371698.310.7588232873.621701.821.7407313926.3168421.773.770327.533641.070250.51752015.69905.286684.2840616.49451.314212.009252905.721679.322924.951701.380.7603942875.281693.141.7404513917.9868422.23OpenBenchmarking.org

oneDNN

Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU1230.84851.6972.54553.3944.2425SE +/- 0.00813, N = 3SE +/- 0.00424, N = 3SE +/- 0.02135, N = 33.752673.771073.77032MIN: 3.45MIN: 3.45MIN: 3.461. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU123246810SE +/- 0.01236, N = 3SE +/- 0.00401, N = 3SE +/- 0.01646, N = 38.627177.533917.53364MIN: 8.47MIN: 7.33MIN: 7.321. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU1230.24080.48160.72240.96321.204SE +/- 0.00321, N = 3SE +/- 0.00243, N = 3SE +/- 0.00399, N = 31.069661.069811.07025MIN: 1MIN: 1MIN: 0.991. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU1230.1940.3880.5820.7760.97SE +/- 0.002916, N = 3SE +/- 0.348175, N = 15SE +/- 0.003242, N = 30.5244990.8622160.517520MIN: 0.47MIN: 0.46MIN: 0.471. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU12348121620SE +/- 0.03, N = 3SE +/- 0.01, N = 3SE +/- 0.02, N = 316.1915.7215.70MIN: 15.68MIN: 15.38MIN: 15.311. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU1231.24622.49243.73864.98486.231SE +/- 0.21782, N = 12SE +/- 0.19863, N = 15SE +/- 0.19023, N = 155.538455.320475.28668MIN: 3.21MIN: 3.23MIN: 3.241. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU1230.96391.92782.89173.85564.8195SE +/- 0.00766, N = 3SE +/- 0.00980, N = 3SE +/- 0.00550, N = 34.279284.266244.28406MIN: 4.15MIN: 4.15MIN: 4.171. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU12348121620SE +/- 0.01, N = 3SE +/- 0.08, N = 3SE +/- 0.09, N = 317.3516.4716.49MIN: 16.94MIN: 15.93MIN: 15.811. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU1230.29620.59240.88861.18481.481SE +/- 0.00569, N = 3SE +/- 0.00533, N = 3SE +/- 0.00800, N = 31.316291.309401.31421MIN: 1.22MIN: 1.22MIN: 1.221. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU1230.45360.90721.36081.81442.268SE +/- 0.00253, N = 3SE +/- 0.00271, N = 3SE +/- 0.00223, N = 32.016132.013202.00925MIN: 1.94MIN: 1.91MIN: 1.781. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU1236001200180024003000SE +/- 14.19, N = 3SE +/- 10.81, N = 3SE +/- 12.12, N = 32890.592892.532905.72MIN: 2861.19MIN: 2864.07MIN: 2873.261. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU123400800120016002000SE +/- 3.61, N = 3SE +/- 6.12, N = 3SE +/- 9.45, N = 31672.051689.611679.32MIN: 1639.18MIN: 1668.94MIN: 1646.621. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU1236001200180024003000SE +/- 14.19, N = 3SE +/- 9.96, N = 3SE +/- 17.20, N = 32904.052901.372924.95MIN: 2876.45MIN: 2874.88MIN: 2881.731. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU123400800120016002000SE +/- 2.15, N = 3SE +/- 3.80, N = 3SE +/- 10.80, N = 31707.041698.311701.38MIN: 1692.22MIN: 1678.43MIN: 1662.081. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU1230.17110.34220.51330.68440.8555SE +/- 0.001551, N = 3SE +/- 0.001642, N = 3SE +/- 0.004859, N = 30.7601810.7588230.760394MIN: 0.71MIN: 0.72MIN: 0.721. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU1236001200180024003000SE +/- 4.58, N = 3SE +/- 15.32, N = 3SE +/- 9.27, N = 32897.912873.622875.28MIN: 2882.62MIN: 2840.96MIN: 2845.111. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU123400800120016002000SE +/- 20.36, N = 3SE +/- 1.80, N = 3SE +/- 4.28, N = 31697.371701.821693.14MIN: 1650.11MIN: 1684.14MIN: 1661.051. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU1230.39170.78341.17511.56681.9585SE +/- 0.00034, N = 3SE +/- 0.00092, N = 3SE +/- 0.00070, N = 31.740831.740731.74045MIN: 1.63MIN: 1.64MIN: 1.631. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Sysbench

Test: RAM / Memory

OpenBenchmarking.orgMiB/sec, More Is BetterSysbench 1.0.20Test: RAM / Memory1233K6K9K12K15KSE +/- 13.75, N = 3SE +/- 13.11, N = 3SE +/- 16.45, N = 313950.1013926.3113917.981. (CC) gcc options: -pthread -O2 -funroll-loops -rdynamic -ldl -laio -lm

Sysbench

Test: CPU

OpenBenchmarking.orgEvents Per Second, More Is BetterSysbench 1.0.20Test: CPU12315K30K45K60K75KSE +/- 66.39, N = 3SE +/- 41.87, N = 3SE +/- 46.63, N = 368475.6568421.7768422.231. (CC) gcc options: -pthread -O2 -funroll-loops -rdynamic -ldl -laio -lm


Phoronix Test Suite v10.8.4