onednn 5500U

AMD Ryzen 5 5500U testing with a LENOVO LNVNB161216 (GLCN22WW BIOS) and AMD Lucienne 2GB on Ubuntu 21.10 via the Phoronix Test Suite.

HTML result view exported from: https://openbenchmarking.org/result/2203306-PTS-ONEDNN5567&grt.

onednn 5500UProcessorMotherboardChipsetMemoryDiskGraphicsAudioNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilerFile-SystemScreen ResolutionABCAMD Ryzen 5 5500U @ 4.06GHz (6 Cores / 12 Threads)LENOVO LNVNB161216 (GLCN22WW BIOS)AMD Renoir/Cezanne6GB256GB SAMSUNG MZALQ256HBJD-00BL2AMD Lucienne 2GB (1800/400MHz)AMD Renoir Radeon HD AudioQualcomm Atheros QCA6174 802.11acUbuntu 21.105.17.0-051700-generic (x86_64)GNOME Shell 40.5X Server 1.20.13 + Wayland4.6 Mesa 22.1.0-devel (git-729f95a 2022-03-24 impish-oibaf-ppa) (LLVM 13.0.1 DRM 3.44)1.3.207GCC 11.2.0ext41920x1080OpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- Scaling Governor: amd-pstate schedutil (Boost: Enabled) - CPU Microcode: 0x8608102Security Details- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected

onednn 5500Uonednn: IP Shapes 1D - f32 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - bf16bf16bf16 - CPUABC12.105414.50033.729994.5530434.452613.622211.755737.77115.628428.032157127.164688.217099.284694.727.906327073.684687.065.1179012.087611.32583.796903.7546934.124413.641711.718533.73655.619118.028147087.034690.557098.274701.507.890747067.164683.435.1131112.086411.38643.794453.7632734.158613.325311.720533.72105.706478.031247043.604690.507066.774676.627.852767094.514662.025.11709OpenBenchmarking.org

oneDNN

Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUABC3691215SE +/- 0.03, N = 3SE +/- 0.02, N = 3SE +/- 0.02, N = 312.1112.0912.09MIN: 11.64MIN: 11.57MIN: 11.631. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUABC48121620SE +/- 0.00, N = 3SE +/- 0.06, N = 3SE +/- 0.04, N = 314.5011.3311.39MIN: 14.37MIN: 11.07MIN: 11.161. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUABC0.85431.70862.56293.41724.2715SE +/- 0.03657, N = 3SE +/- 0.02821, N = 3SE +/- 0.03002, N = 33.729993.796903.79445MIN: 3.5MIN: 3.57MIN: 3.41. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUABC1.02442.04883.07324.09765.122SE +/- 0.00367, N = 3SE +/- 0.04023, N = 5SE +/- 0.05002, N = 34.553043.754693.76327MIN: 4.49MIN: 3.55MIN: 3.581. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUABC816243240SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.03, N = 334.4534.1234.16MIN: 34.09MIN: 33.66MIN: 33.591. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUABC48121620SE +/- 0.14, N = 15SE +/- 0.13, N = 15SE +/- 0.17, N = 1513.6213.6413.33MIN: 8.96MIN: 9MIN: 9.031. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUABC3691215SE +/- 0.01, N = 3SE +/- 0.02, N = 3SE +/- 0.01, N = 311.7611.7211.72MIN: 11.4MIN: 11.49MIN: 11.511. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUABC918273645SE +/- 2.09, N = 12SE +/- 0.04, N = 3SE +/- 0.01, N = 337.7733.7433.72MIN: 33.4MIN: 33.42MIN: 33.441. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUABC1.2842.5683.8525.1366.42SE +/- 0.05563, N = 6SE +/- 0.06253, N = 5SE +/- 0.05850, N = 125.628425.619115.70647MIN: 5.07MIN: 4.89MIN: 4.881. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUABC246810SE +/- 0.00982, N = 3SE +/- 0.00293, N = 3SE +/- 0.01365, N = 38.032158.028148.03124MIN: 7.71MIN: 7.79MIN: 7.551. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUABC15003000450060007500SE +/- 8.00, N = 3SE +/- 9.99, N = 3SE +/- 14.12, N = 37127.167087.037043.60MIN: 7078.19MIN: 7040.89MIN: 6995.51. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUABC10002000300040005000SE +/- 13.87, N = 3SE +/- 11.37, N = 3SE +/- 23.16, N = 34688.214690.554690.50MIN: 4643MIN: 4647.66MIN: 4623.541. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUABC15003000450060007500SE +/- 16.66, N = 3SE +/- 10.05, N = 3SE +/- 7.85, N = 37099.287098.277066.77MIN: 7031.56MIN: 7058.13MIN: 7019.151. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUABC10002000300040005000SE +/- 12.56, N = 3SE +/- 20.21, N = 3SE +/- 15.17, N = 34694.724701.504676.62MIN: 4650.38MIN: 4642.31MIN: 4628.391. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUABC246810SE +/- 0.01256, N = 3SE +/- 0.01478, N = 3SE +/- 0.00587, N = 37.906327.890747.85276MIN: 7.78MIN: 7.75MIN: 7.731. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUABC15003000450060007500SE +/- 5.15, N = 3SE +/- 12.42, N = 3SE +/- 12.08, N = 37073.687067.167094.51MIN: 7037.09MIN: 7017.84MIN: 7042.11. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUABC10002000300040005000SE +/- 8.81, N = 3SE +/- 9.44, N = 3SE +/- 5.70, N = 34687.064683.434662.02MIN: 4647.19MIN: 4644.98MIN: 4631.351. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUABC1.15152.3033.45454.6065.7575SE +/- 0.00093, N = 3SE +/- 0.00391, N = 3SE +/- 0.00500, N = 35.117905.113115.11709MIN: 4.83MIN: 4.9MIN: 4.921. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread


Phoronix Test Suite v10.8.5