Ryzen 5 5600X AMD Ryzen 5 5600X 6-Core testing with a ASUS TUF GAMING B550M-PLUS (WI-FI) (1216 BIOS) and NVIDIA NV166 6GB on Ubuntu 20.10 via the Phoronix Test Suite.
HTML result view exported from: https://openbenchmarking.org/result/2012091-HA-RYZEN556051&grw&rdt .
Ryzen 5 5600X Processor Motherboard Chipset Memory Disk Graphics Audio Monitor Network OS Kernel Desktop Display Server Display Driver OpenGL Compiler File-System Screen Resolution 1 2 3 AMD Ryzen 5 5600X 6-Core @ 3.70GHz (6 Cores / 12 Threads) ASUS TUF GAMING B550M-PLUS (WI-FI) (1216 BIOS) AMD Starship/Matisse 16GB 1000GB Samsung SSD 980 PRO 1TB NVIDIA NV166 6GB NVIDIA TU106 HD Audio G237HL Realtek RTL8125 2.5GbE + Intel Wi-Fi 6 AX200 Ubuntu 20.10 5.9.10-050910-generic (x86_64) GNOME Shell 3.38.1 X Server 1.20.9 modesetting 1.20.9 4.3 Mesa 20.2.1 GCC 10.2.0 ext4 1920x1080 OpenBenchmarking.org Compiler Details - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details - Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0xa201009 Security Details - itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional IBRS_FW STIBP: always-on RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Ryzen 5 5600X onednn: IP Shapes 1D - f32 - CPU onednn: IP Shapes 3D - f32 - CPU onednn: IP Shapes 1D - u8s8f32 - CPU onednn: IP Shapes 3D - u8s8f32 - CPU onednn: Convolution Batch Shapes Auto - f32 - CPU onednn: Deconvolution Batch shapes_1d - f32 - CPU onednn: Deconvolution Batch shapes_3d - f32 - CPU onednn: Convolution Batch Shapes Auto - u8s8f32 - CPU onednn: Deconvolution Batch shapes_1d - u8s8f32 - CPU onednn: Deconvolution Batch shapes_3d - u8s8f32 - CPU onednn: Recurrent Neural Network Training - f32 - CPU onednn: Recurrent Neural Network Inference - f32 - CPU onednn: Recurrent Neural Network Training - u8s8f32 - CPU onednn: Recurrent Neural Network Inference - u8s8f32 - CPU onednn: Matrix Multiply Batch Shapes Transformer - f32 - CPU onednn: Recurrent Neural Network Training - bf16bf16bf16 - CPU onednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPU onednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU 1 2 3 4.41650 10.3754 1.90790 1.70309 20.9535 5.60692 7.62028 15.9555 5.21935 3.40352 3922.55 2231.84 3932.25 2236.45 2.54205 3934.05 2241.50 3.47200 4.42871 10.3136 1.91613 1.71015 21.0805 5.65211 7.62532 16.2839 5.24175 3.41444 3933.03 2235.83 3935.58 2239.18 2.53107 3931.97 2233.39 3.46802 4.40609 10.3063 1.89870 1.71271 21.0348 5.65539 7.62753 16.1671 5.23835 3.41929 3924.44 2232.21 3930.91 2231.20 2.51502 3936.22 2237.51 3.46915 OpenBenchmarking.org
oneDNN Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU 1 2 3 0.9965 1.993 2.9895 3.986 4.9825 SE +/- 0.01727, N = 3 SE +/- 0.00512, N = 3 SE +/- 0.00992, N = 3 4.41650 4.42871 4.40609 MIN: 4 MIN: 4.09 MIN: 3.97 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU 1 2 3 3 6 9 12 15 SE +/- 0.04, N = 3 SE +/- 0.06, N = 3 SE +/- 0.02, N = 3 10.38 10.31 10.31 MIN: 10.15 MIN: 10.1 MIN: 10.09 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU 1 2 3 0.4311 0.8622 1.2933 1.7244 2.1555 SE +/- 0.00615, N = 3 SE +/- 0.00969, N = 3 SE +/- 0.00568, N = 3 1.90790 1.91613 1.89870 MIN: 1.77 MIN: 1.79 MIN: 1.77 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU 1 2 3 0.3854 0.7708 1.1562 1.5416 1.927 SE +/- 0.00829, N = 3 SE +/- 0.02447, N = 3 SE +/- 0.00859, N = 3 1.70309 1.71015 1.71271 MIN: 1.56 MIN: 1.57 MIN: 1.64 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU 1 2 3 5 10 15 20 25 SE +/- 0.03, N = 3 SE +/- 0.02, N = 3 SE +/- 0.09, N = 3 20.95 21.08 21.03 MIN: 19.47 MIN: 19.47 MIN: 19.29 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU 1 2 3 1.2725 2.545 3.8175 5.09 6.3625 SE +/- 0.01589, N = 3 SE +/- 0.03636, N = 3 SE +/- 0.00895, N = 3 5.60692 5.65211 5.65539 MIN: 5.38 MIN: 5.39 MIN: 5.41 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU 1 2 3 2 4 6 8 10 SE +/- 0.02300, N = 3 SE +/- 0.02048, N = 3 SE +/- 0.01570, N = 3 7.62028 7.62532 7.62753 MIN: 7.36 MIN: 7.33 MIN: 7.3 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU 1 2 3 4 8 12 16 20 SE +/- 0.17, N = 3 SE +/- 0.09, N = 3 SE +/- 0.24, N = 15 15.96 16.28 16.17 MIN: 15.28 MIN: 15.5 MIN: 15.09 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU 1 2 3 1.1794 2.3588 3.5382 4.7176 5.897 SE +/- 0.01757, N = 3 SE +/- 0.01439, N = 3 SE +/- 0.00626, N = 3 5.21935 5.24175 5.23835 MIN: 4.97 MIN: 4.96 MIN: 4.98 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU 1 2 3 0.7693 1.5386 2.3079 3.0772 3.8465 SE +/- 0.00990, N = 3 SE +/- 0.01180, N = 3 SE +/- 0.01474, N = 3 3.40352 3.41444 3.41929 MIN: 3.18 MIN: 3.19 MIN: 3.23 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU 1 2 3 800 1600 2400 3200 4000 SE +/- 2.28, N = 3 SE +/- 2.23, N = 3 SE +/- 5.60, N = 3 3922.55 3933.03 3924.44 MIN: 3910.11 MIN: 3923.42 MIN: 3908.49 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU 1 2 3 500 1000 1500 2000 2500 SE +/- 2.14, N = 3 SE +/- 4.41, N = 3 SE +/- 2.05, N = 3 2231.84 2235.83 2232.21 MIN: 2221.91 MIN: 2222.52 MIN: 2222.49 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU 1 2 3 800 1600 2400 3200 4000 SE +/- 1.26, N = 3 SE +/- 0.57, N = 3 SE +/- 2.37, N = 3 3932.25 3935.58 3930.91 MIN: 3923.02 MIN: 3929.47 MIN: 3915.64 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU 1 2 3 500 1000 1500 2000 2500 SE +/- 5.43, N = 3 SE +/- 4.33, N = 3 SE +/- 2.83, N = 3 2236.45 2239.18 2231.20 MIN: 2221.99 MIN: 2226.62 MIN: 2221.77 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU 1 2 3 0.572 1.144 1.716 2.288 2.86 SE +/- 0.01010, N = 3 SE +/- 0.00699, N = 3 SE +/- 0.00472, N = 3 2.54205 2.53107 2.51502 MIN: 2.45 MIN: 2.42 MIN: 2.4 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU 1 2 3 800 1600 2400 3200 4000 SE +/- 0.95, N = 3 SE +/- 1.41, N = 3 SE +/- 1.56, N = 3 3934.05 3931.97 3936.22 MIN: 3925.31 MIN: 3922.96 MIN: 3927.76 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU 1 2 3 500 1000 1500 2000 2500 SE +/- 3.89, N = 3 SE +/- 2.09, N = 3 SE +/- 4.43, N = 3 2241.50 2233.39 2237.51 MIN: 2228.95 MIN: 2223.36 MIN: 2225.32 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU 1 2 3 0.7812 1.5624 2.3436 3.1248 3.906 SE +/- 0.00857, N = 3 SE +/- 0.01356, N = 3 SE +/- 0.00870, N = 3 3.47200 3.46802 3.46915 MIN: 3.32 MIN: 3.32 MIN: 3.31 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
Phoronix Test Suite v10.8.5