onednn 5500U

AMD Ryzen 5 5500U testing with a LENOVO LNVNB161216 (GLCN22WW BIOS) and AMD Lucienne 2GB on Ubuntu 21.10 via the Phoronix Test Suite.

HTML result view exported from: https://openbenchmarking.org/result/2203306-PTS-ONEDNN5567&grs.

onednn 5500UProcessorMotherboardChipsetMemoryDiskGraphicsAudioNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilerFile-SystemScreen ResolutionABCAMD Ryzen 5 5500U @ 4.06GHz (6 Cores / 12 Threads)LENOVO LNVNB161216 (GLCN22WW BIOS)AMD Renoir/Cezanne6GB256GB SAMSUNG MZALQ256HBJD-00BL2AMD Lucienne 2GB (1800/400MHz)AMD Renoir Radeon HD AudioQualcomm Atheros QCA6174 802.11acUbuntu 21.105.17.0-051700-generic (x86_64)GNOME Shell 40.5X Server 1.20.13 + Wayland4.6 Mesa 22.1.0-devel (git-729f95a 2022-03-24 impish-oibaf-ppa) (LLVM 13.0.1 DRM 3.44)1.3.207GCC 11.2.0ext41920x1080OpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-ZPT0kp/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- Scaling Governor: amd-pstate schedutil (Boost: Enabled) - CPU Microcode: 0x8608102Security Details- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected

onednn 5500Uonednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUonednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: IP Shapes 1D - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: IP Shapes 1D - bf16bf16bf16 - CPUABC14.50034.5530413.62223.729995.628427127.1634.45267.906324687.064694.727099.287073.6811.755712.10545.117908.032154688.2137.771111.32583.7546913.64173.796905.619117087.0334.12447.890744683.434701.507098.277067.1611.718512.08765.113118.028144690.5533.736511.38643.7632713.32533.794455.706477043.6034.15867.852764662.024676.627066.777094.5111.720512.08645.117098.031244690.5033.7210OpenBenchmarking.org

oneDNN

Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUABC48121620SE +/- 0.00, N = 3SE +/- 0.06, N = 3SE +/- 0.04, N = 314.5011.3311.39MIN: 14.37MIN: 11.07MIN: 11.161. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUABC1.02442.04883.07324.09765.122SE +/- 0.00367, N = 3SE +/- 0.04023, N = 5SE +/- 0.05002, N = 34.553043.754693.76327MIN: 4.49MIN: 3.55MIN: 3.581. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUABC48121620SE +/- 0.14, N = 15SE +/- 0.13, N = 15SE +/- 0.17, N = 1513.6213.6413.33MIN: 8.96MIN: 9MIN: 9.031. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUABC0.85431.70862.56293.41724.2715SE +/- 0.03657, N = 3SE +/- 0.02821, N = 3SE +/- 0.03002, N = 33.729993.796903.79445MIN: 3.5MIN: 3.57MIN: 3.41. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUABC1.2842.5683.8525.1366.42SE +/- 0.05563, N = 6SE +/- 0.06253, N = 5SE +/- 0.05850, N = 125.628425.619115.70647MIN: 5.07MIN: 4.89MIN: 4.881. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUABC15003000450060007500SE +/- 8.00, N = 3SE +/- 9.99, N = 3SE +/- 14.12, N = 37127.167087.037043.60MIN: 7078.19MIN: 7040.89MIN: 6995.51. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUABC816243240SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.03, N = 334.4534.1234.16MIN: 34.09MIN: 33.66MIN: 33.591. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUABC246810SE +/- 0.01256, N = 3SE +/- 0.01478, N = 3SE +/- 0.00587, N = 37.906327.890747.85276MIN: 7.78MIN: 7.75MIN: 7.731. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUABC10002000300040005000SE +/- 8.81, N = 3SE +/- 9.44, N = 3SE +/- 5.70, N = 34687.064683.434662.02MIN: 4647.19MIN: 4644.98MIN: 4631.351. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUABC10002000300040005000SE +/- 12.56, N = 3SE +/- 20.21, N = 3SE +/- 15.17, N = 34694.724701.504676.62MIN: 4650.38MIN: 4642.31MIN: 4628.391. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUABC15003000450060007500SE +/- 16.66, N = 3SE +/- 10.05, N = 3SE +/- 7.85, N = 37099.287098.277066.77MIN: 7031.56MIN: 7058.13MIN: 7019.151. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUABC15003000450060007500SE +/- 5.15, N = 3SE +/- 12.42, N = 3SE +/- 12.08, N = 37073.687067.167094.51MIN: 7037.09MIN: 7017.84MIN: 7042.11. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUABC3691215SE +/- 0.01, N = 3SE +/- 0.02, N = 3SE +/- 0.01, N = 311.7611.7211.72MIN: 11.4MIN: 11.49MIN: 11.511. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUABC3691215SE +/- 0.03, N = 3SE +/- 0.02, N = 3SE +/- 0.02, N = 312.1112.0912.09MIN: 11.64MIN: 11.57MIN: 11.631. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUABC1.15152.3033.45454.6065.7575SE +/- 0.00093, N = 3SE +/- 0.00391, N = 3SE +/- 0.00500, N = 35.117905.113115.11709MIN: 4.83MIN: 4.9MIN: 4.921. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUABC246810SE +/- 0.00982, N = 3SE +/- 0.00293, N = 3SE +/- 0.01365, N = 38.032158.028148.03124MIN: 7.71MIN: 7.79MIN: 7.551. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUABC10002000300040005000SE +/- 13.87, N = 3SE +/- 11.37, N = 3SE +/- 23.16, N = 34688.214690.554690.50MIN: 4643MIN: 4647.66MIN: 4623.541. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUABC918273645SE +/- 2.09, N = 12SE +/- 0.04, N = 3SE +/- 0.01, N = 337.7733.7433.72MIN: 33.4MIN: 33.42MIN: 33.441. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread


Phoronix Test Suite v10.8.5