atomicrulez_ubuntu_2310_mkl-dnn

Intel Core i9-10850K testing with a ASUS ROG MAXIMUS XII APEX (2701 BIOS) and NVIDIA GeForce RTX 3090 Ti 24GB on Ubuntu 23.10 via the Phoronix Test Suite.

HTML result view exported from: https://openbenchmarking.org/result/2310017-MICH-ATOMICR47.

atomicrulez_ubuntu_2310_mkl-dnnProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerDisplay DriverOpenGLOpenCLCompilerFile-SystemScreen Resolutioni9-10850KIntel Core i9-10850K @ 5.20GHz (10 Cores / 20 Threads)ASUS ROG MAXIMUS XII APEX (2701 BIOS)Intel Comet Lake PCH64GB280GB INTEL SSDPE21D280GANVIDIA GeForce RTX 3090 Ti 24GBRealtek ALC1220ROG PG259QNIntel I225-V + Intel Comet Lake PCH CNVi WiFiUbuntu 23.106.5.0-5-generic (x86_64)GNOME Shell 45.0X Server 1.21.1.7NVIDIA 535.104.054.6.0OpenCL 3.0 CUDA 12.2.138GCC 13.2.0 + CUDA 12.0ext41920x1080OpenBenchmarking.org- Transparent Huge Pages: madvise- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: intel_pstate powersave (EPP: performance) - CPU Microcode: 0xf8 - Thermald 2.5.4 - gather_data_sampling: Mitigation of Microcode + itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Mitigation of Clear buffers; SMT vulnerable + retbleed: Mitigation of Enhanced IBRS + spec_rstack_overflow: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Mitigation of Microcode + tsx_async_abort: Not affected

atomicrulez_ubuntu_2310_mkl-dnnmkl-dnn: IP Batch 1D - f32mkl-dnn: IP Batch All - f32mkl-dnn: IP Batch 1D - u8s8f32mkl-dnn: IP Batch All - u8s8f32mkl-dnn: Deconvolution Batch deconv_1d - f32mkl-dnn: Deconvolution Batch deconv_3d - f32mkl-dnn: Deconvolution Batch deconv_1d - u8s8f32mkl-dnn: Deconvolution Batch deconv_3d - u8s8f32mkl-dnn: Recurrent Neural Network Training - f32mkl-dnn: Recurrent Neural Network Inference - f32mkl-dnn: Recurrent Neural Network Inference - bf16bf16bf16i9-10850K2.9350443.85591.2391016.84293.052294.75091124.5242.18324167.67425.7463OpenBenchmarking.org

oneDNN MKL-DNN

Harness: IP Batch 1D - Data Type: f32

OpenBenchmarking.orgms, Fewer Is BetteroneDNN MKL-DNN 1.3Harness: IP Batch 1D - Data Type: f32i9-10850K0.66041.32081.98122.64163.302SE +/- 0.03799, N = 32.93504MIN: 2.531. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -ldl

oneDNN MKL-DNN

Harness: IP Batch All - Data Type: f32

OpenBenchmarking.orgms, Fewer Is BetteroneDNN MKL-DNN 1.3Harness: IP Batch All - Data Type: f32i9-10850K1020304050SE +/- 0.01, N = 343.86MIN: 43.361. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -ldl

oneDNN MKL-DNN

Harness: IP Batch 1D - Data Type: u8s8f32

OpenBenchmarking.orgms, Fewer Is BetteroneDNN MKL-DNN 1.3Harness: IP Batch 1D - Data Type: u8s8f32i9-10850K0.27880.55760.83641.11521.394SE +/- 0.01401, N = 41.23910MIN: 1.161. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -ldl

oneDNN MKL-DNN

Harness: IP Batch All - Data Type: u8s8f32

OpenBenchmarking.orgms, Fewer Is BetteroneDNN MKL-DNN 1.3Harness: IP Batch All - Data Type: u8s8f32i9-10850K48121620SE +/- 0.03, N = 316.84MIN: 16.41. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -ldl

oneDNN MKL-DNN

Harness: Deconvolution Batch deconv_1d - Data Type: f32

OpenBenchmarking.orgms, Fewer Is BetteroneDNN MKL-DNN 1.3Harness: Deconvolution Batch deconv_1d - Data Type: f32i9-10850K0.68681.37362.06042.74723.434SE +/- 0.00699, N = 33.05229MIN: 2.951. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -ldl

oneDNN MKL-DNN

Harness: Deconvolution Batch deconv_3d - Data Type: f32

OpenBenchmarking.orgms, Fewer Is BetteroneDNN MKL-DNN 1.3Harness: Deconvolution Batch deconv_3d - Data Type: f32i9-10850K1.0692.1383.2074.2765.345SE +/- 0.00615, N = 34.75091MIN: 4.631. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -ldl

oneDNN MKL-DNN

Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32

OpenBenchmarking.orgms, Fewer Is BetteroneDNN MKL-DNN 1.3Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32i9-10850K306090120150SE +/- 0.49, N = 3124.52MIN: 117.91. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -ldl

oneDNN MKL-DNN

Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32

OpenBenchmarking.orgms, Fewer Is BetteroneDNN MKL-DNN 1.3Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32i9-10850K0.49120.98241.47361.96482.456SE +/- 0.01126, N = 32.18324MIN: 2.121. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -ldl

oneDNN MKL-DNN

Harness: Recurrent Neural Network Training - Data Type: f32

OpenBenchmarking.orgms, Fewer Is BetteroneDNN MKL-DNN 1.3Harness: Recurrent Neural Network Training - Data Type: f32i9-10850K4080120160200SE +/- 0.41, N = 3167.67MIN: 160.271. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -ldl

oneDNN MKL-DNN

Harness: Recurrent Neural Network Inference - Data Type: f32

OpenBenchmarking.orgms, Fewer Is BetteroneDNN MKL-DNN 1.3Harness: Recurrent Neural Network Inference - Data Type: f32i9-10850K612182430SE +/- 0.08, N = 325.75MIN: 24.551. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -ldl


Phoronix Test Suite v10.8.4