2-x-intel-xeon-gold-5220r-onednn 2 x Intel Xeon Gold 5220R testing with a Supermicro X11DPG-QT (3.3 BIOS) and ASPEED 32GB on ManjaroLinux 20.1 via the Phoronix Test Suite.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 2009072-NE-2XINTELXE31 manjaro-anaconda Processor: 2 x Intel Xeon Gold 5220R @ 4.00GHz (48 Cores / 96 Threads), Motherboard: Supermicro X11DPG-QT (3.3 BIOS), Chipset: Intel Sky Lake-E DMI3 Registers, Memory: 252GB, Disk: 2000GB INTEL SSDPELKX020T8 + 8002GB HUS728T8TAL5204, Graphics: ASPEED 32GB, Audio: Realtek ALC888-VD, Monitor: PHL 276E9Q, Network: 2 x Intel 10G X550T
OS: ManjaroLinux 20.1, Kernel: 5.4.60-2-MANJARO (x86_64), Display Server: X Server 1.20.8, Compiler: GCC 10.2.0, File-System: ext4, Screen Resolution: 1920x1080
Compiler Notes: --disable-libssp --disable-libstdcxx-pch --disable-libunwind-exceptions --disable-werror --enable-__cxa_atexit --enable-cet=auto --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-default-ssp --enable-gnu-indirect-function --enable-gnu-unique-object --enable-install-libiberty --enable-languages=c,c++,ada,fortran,go,lto,objc,obj-c++,d --enable-lto --enable-multilib --enable-plugin --enable-shared --enable-threads=posix --mandir=/usr/share/man --with-isl --with-linker-hash-style=gnuProcessor Notes: Scaling Governor: intel_pstate performance - CPU Microcode: 0x5002f01Security Notes: itlb_multihit: KVM: Mitigation of Split huge pages + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Mitigation of TSX disabled
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch 1D - Data Type: f32 - Engine: CPU manjaro-anaconda 0.3604 0.7208 1.0812 1.4416 1.802 SE +/- 0.00195, N = 3 1.60190 MIN: 1.51 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch All - Data Type: f32 - Engine: CPU manjaro-anaconda 7 14 21 28 35 SE +/- 0.08, N = 3 31.81 MIN: 30.11 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch 1D - Data Type: u8s8f32 - Engine: CPU manjaro-anaconda 0.5181 1.0362 1.5543 2.0724 2.5905 SE +/- 0.01439, N = 3 2.30253 MIN: 2.06 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch All - Data Type: u8s8f32 - Engine: CPU manjaro-anaconda 2 4 6 8 10 SE +/- 0.00196, N = 3 6.20202 MIN: 5.54 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch 1D - Data Type: bf16bf16bf16 - Engine: CPU manjaro-anaconda 1.2431 2.4862 3.7293 4.9724 6.2155 SE +/- 0.00499, N = 3 5.52492 MIN: 5.24 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch All - Data Type: bf16bf16bf16 - Engine: CPU manjaro-anaconda 10 20 30 40 50 SE +/- 0.08, N = 3 42.12 MIN: 40.36 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU manjaro-anaconda 3 6 9 12 15 SE +/- 0.03, N = 3 12.21 MIN: 11.83 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_1d - Data Type: f32 - Engine: CPU manjaro-anaconda 0.4213 0.8426 1.2639 1.6852 2.1065 SE +/- 0.00479, N = 3 1.87255 MIN: 1.71 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_3d - Data Type: f32 - Engine: CPU manjaro-anaconda 0.4057 0.8114 1.2171 1.6228 2.0285 SE +/- 0.00266, N = 3 1.80290 MIN: 1.74 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU manjaro-anaconda 3 6 9 12 15 SE +/- 0.00, N = 3 11.52 MIN: 11.23 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32 - Engine: CPU manjaro-anaconda 0.1175 0.235 0.3525 0.47 0.5875 SE +/- 0.004497, N = 3 0.522418 MIN: 0.47 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32 - Engine: CPU manjaro-anaconda 0.1078 0.2156 0.3234 0.4312 0.539 SE +/- 0.000386, N = 3 0.479023 MIN: 0.45 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU manjaro-anaconda 70 140 210 280 350 SE +/- 4.28, N = 3 330.63 MIN: 312.27 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU manjaro-anaconda 20 40 60 80 100 SE +/- 1.34, N = 3 89.27 MIN: 83.67 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU manjaro-anaconda 1.0827 2.1654 3.2481 4.3308 5.4135 SE +/- 0.00955, N = 3 4.81206 MIN: 4.67 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_1d - Data Type: bf16bf16bf16 - Engine: CPU manjaro-anaconda 1.2798 2.5596 3.8394 5.1192 6.399 SE +/- 0.00510, N = 3 5.68810 MIN: 5.46 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_3d - Data Type: bf16bf16bf16 - Engine: CPU manjaro-anaconda 2 4 6 8 10 SE +/- 0.00539, N = 3 6.24086 MIN: 6.11 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU manjaro-anaconda 0.096 0.192 0.288 0.384 0.48 SE +/- 0.000220, N = 3 0.426886 MIN: 0.39 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU manjaro-anaconda 0.0713 0.1426 0.2139 0.2852 0.3565 SE +/- 0.001171, N = 3 0.316890 MIN: 0.26 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU manjaro-anaconda 0.2543 0.5086 0.7629 1.0172 1.2715 SE +/- 0.00275, N = 3 1.13003 MIN: 1.07 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
manjaro-anaconda Processor: 2 x Intel Xeon Gold 5220R @ 4.00GHz (48 Cores / 96 Threads), Motherboard: Supermicro X11DPG-QT (3.3 BIOS), Chipset: Intel Sky Lake-E DMI3 Registers, Memory: 252GB, Disk: 2000GB INTEL SSDPELKX020T8 + 8002GB HUS728T8TAL5204, Graphics: ASPEED 32GB, Audio: Realtek ALC888-VD, Monitor: PHL 276E9Q, Network: 2 x Intel 10G X550T
OS: ManjaroLinux 20.1, Kernel: 5.4.60-2-MANJARO (x86_64), Display Server: X Server 1.20.8, Compiler: GCC 10.2.0, File-System: ext4, Screen Resolution: 1920x1080
Compiler Notes: --disable-libssp --disable-libstdcxx-pch --disable-libunwind-exceptions --disable-werror --enable-__cxa_atexit --enable-cet=auto --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-default-ssp --enable-gnu-indirect-function --enable-gnu-unique-object --enable-install-libiberty --enable-languages=c,c++,ada,fortran,go,lto,objc,obj-c++,d --enable-lto --enable-multilib --enable-plugin --enable-shared --enable-threads=posix --mandir=/usr/share/man --with-isl --with-linker-hash-style=gnuProcessor Notes: Scaling Governor: intel_pstate performance - CPU Microcode: 0x5002f01Security Notes: itlb_multihit: KVM: Mitigation of Split huge pages + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Mitigation of TSX disabled
Testing initiated at 7 September 2020 09:38 by user mwojc.