Ryzen 7 2700 oneDNN 2.0 AMD Ryzen 7 2700 Eight-Core testing with a Gigabyte AB350N-Gaming WIFI-CF (F20 BIOS) and HIS AMD Radeon HD 6450/7450/8450 / R5 230 OEM 1GB on Ubuntu 19.10 via the Phoronix Test Suite.
HTML result view exported from: https://openbenchmarking.org/result/2012101-HA-RYZEN727034 .
Ryzen 7 2700 oneDNN 2.0 Processor Motherboard Chipset Memory Disk Graphics Audio Monitor Network OS Kernel Desktop Display Server Display Driver OpenGL Compiler File-System Screen Resolution 1 2 3 AMD Ryzen 7 2700 Eight-Core @ 3.20GHz (8 Cores / 16 Threads) Gigabyte AB350N-Gaming WIFI-CF (F20 BIOS) AMD 17h 16GB 120GB ADATA SU700 HIS AMD Radeon HD 6450/7450/8450 / R5 230 OEM 1GB AMD Caicos HDMI Audio DELL S2409W Realtek RTL8111/8168/8411 + Intel 3165 Ubuntu 19.10 5.9.0-050900rc7daily20201004-generic (x86_64) 20201003 GNOME Shell 3.34.1 X Server 1.20.5 modesetting 1.20.5 3.3 Mesa 19.2.8 (LLVM 9.0.0) GCC 9.2.1 20191008 ext4 1920x1080 OpenBenchmarking.org Compiler Details - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-offload-targets=nvptx-none,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details - Scaling Governor: acpi-cpufreq ondemand - CPU Microcode: 0x800820b Security Details - itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional STIBP: disabled RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Ryzen 7 2700 oneDNN 2.0 onednn: IP Shapes 1D - f32 - CPU onednn: IP Shapes 3D - f32 - CPU onednn: IP Shapes 1D - u8s8f32 - CPU onednn: IP Shapes 3D - u8s8f32 - CPU onednn: Convolution Batch Shapes Auto - f32 - CPU onednn: Deconvolution Batch shapes_1d - f32 - CPU onednn: Deconvolution Batch shapes_3d - f32 - CPU onednn: Convolution Batch Shapes Auto - u8s8f32 - CPU onednn: Deconvolution Batch shapes_1d - u8s8f32 - CPU onednn: Deconvolution Batch shapes_3d - u8s8f32 - CPU onednn: Recurrent Neural Network Training - f32 - CPU onednn: Recurrent Neural Network Inference - f32 - CPU onednn: Recurrent Neural Network Training - u8s8f32 - CPU onednn: Recurrent Neural Network Inference - u8s8f32 - CPU onednn: Matrix Multiply Batch Shapes Transformer - f32 - CPU onednn: Recurrent Neural Network Training - bf16bf16bf16 - CPU onednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPU onednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU 1 2 3 8.30216 12.0536 6.41140 2.61603 23.7473 10.2586 14.3959 24.2433 12.8285 12.6217 8137.11 4324.22 8213.83 4329.68 5.16128 8201.12 4336.81 6.04440 8.38407 11.9905 6.47135 2.56768 23.7574 10.5255 14.4047 24.1902 14.0855 12.6613 8199.88 4339.60 8222.43 4340.57 5.12423 8212.31 4335.51 6.04323 8.31945 12.0197 6.44652 2.56705 23.7334 10.3230 14.3789 24.1868 12.2584 12.6210 8159.58 4321.11 8188.93 4336.59 5.12143 8165.85 4338.29 6.08775 OpenBenchmarking.org
oneDNN Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU 1 2 3 2 4 6 8 10 SE +/- 0.01659, N = 3 SE +/- 0.00291, N = 3 SE +/- 0.02622, N = 3 8.30216 8.38407 8.31945 MIN: 8.09 MIN: 8.22 MIN: 8.14 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU 1 2 3 3 6 9 12 15 SE +/- 0.03, N = 3 SE +/- 0.01, N = 3 SE +/- 0.03, N = 3 12.05 11.99 12.02 MIN: 11.7 MIN: 11.59 MIN: 11.56 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU 1 2 3 2 4 6 8 10 SE +/- 0.00691, N = 3 SE +/- 0.00778, N = 3 SE +/- 0.01020, N = 3 6.41140 6.47135 6.44652 MIN: 6.24 MIN: 6.34 MIN: 6.26 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU 1 2 3 0.5886 1.1772 1.7658 2.3544 2.943 SE +/- 0.00622, N = 3 SE +/- 0.00219, N = 3 SE +/- 0.00426, N = 3 2.61603 2.56768 2.56705 MIN: 2.43 MIN: 2.38 MIN: 2.38 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU 1 2 3 6 12 18 24 30 SE +/- 0.01, N = 3 SE +/- 0.05, N = 3 SE +/- 0.01, N = 3 23.75 23.76 23.73 MIN: 23.07 MIN: 23.17 MIN: 23.31 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU 1 2 3 3 6 9 12 15 SE +/- 0.03, N = 3 SE +/- 0.18, N = 13 SE +/- 0.03, N = 3 10.26 10.53 10.32 MIN: 10.09 MIN: 10 MIN: 10.12 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU 1 2 3 4 8 12 16 20 SE +/- 0.04, N = 3 SE +/- 0.03, N = 3 SE +/- 0.01, N = 3 14.40 14.40 14.38 MIN: 13.97 MIN: 13.99 MIN: 14 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU 1 2 3 6 12 18 24 30 SE +/- 0.03, N = 3 SE +/- 0.02, N = 3 SE +/- 0.01, N = 3 24.24 24.19 24.19 MIN: 23.55 MIN: 23.15 MIN: 23.43 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU 1 2 3 4 8 12 16 20 SE +/- 0.24, N = 15 SE +/- 0.34, N = 15 SE +/- 0.03, N = 3 12.83 14.09 12.26 MIN: 11.99 MIN: 12 MIN: 12.03 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU 1 2 3 3 6 9 12 15 SE +/- 0.01, N = 3 SE +/- 0.02, N = 3 SE +/- 0.02, N = 3 12.62 12.66 12.62 MIN: 12.38 MIN: 12.45 MIN: 12.46 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU 1 2 3 2K 4K 6K 8K 10K SE +/- 13.50, N = 3 SE +/- 27.53, N = 3 SE +/- 28.96, N = 3 8137.11 8199.88 8159.58 MIN: 8105.8 MIN: 8153.31 MIN: 8099.23 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU 1 2 3 900 1800 2700 3600 4500 SE +/- 5.49, N = 3 SE +/- 6.18, N = 3 SE +/- 17.01, N = 3 4324.22 4339.60 4321.11 MIN: 4308.63 MIN: 4322.49 MIN: 4291.26 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU 1 2 3 2K 4K 6K 8K 10K SE +/- 16.68, N = 3 SE +/- 19.08, N = 3 SE +/- 33.99, N = 3 8213.83 8222.43 8188.93 MIN: 8172.42 MIN: 8176.89 MIN: 8126.7 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU 1 2 3 900 1800 2700 3600 4500 SE +/- 15.08, N = 3 SE +/- 9.76, N = 3 SE +/- 18.03, N = 3 4329.68 4340.57 4336.59 MIN: 4300.39 MIN: 4313.07 MIN: 4309.62 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU 1 2 3 1.1613 2.3226 3.4839 4.6452 5.8065 SE +/- 0.00286, N = 3 SE +/- 0.00240, N = 3 SE +/- 0.00382, N = 3 5.16128 5.12423 5.12143 MIN: 5.06 MIN: 5.01 MIN: 4.9 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU 1 2 3 2K 4K 6K 8K 10K SE +/- 15.02, N = 3 SE +/- 24.03, N = 3 SE +/- 15.62, N = 3 8201.12 8212.31 8165.85 MIN: 8176.44 MIN: 8161.35 MIN: 8128.51 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU 1 2 3 900 1800 2700 3600 4500 SE +/- 11.38, N = 3 SE +/- 15.49, N = 3 SE +/- 9.76, N = 3 4336.81 4335.51 4338.29 MIN: 4303.78 MIN: 4295.9 MIN: 4311.48 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
oneDNN Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU 1 2 3 2 4 6 8 10 SE +/- 0.01033, N = 3 SE +/- 0.01029, N = 3 SE +/- 0.02684, N = 3 6.04440 6.04323 6.08775 MIN: 5.77 MIN: 5.79 MIN: 5.77 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
Phoronix Test Suite v10.8.4