sle.hpc-wk1-ML-29aug2020

VMware testing on SUSE Linux Enterprise High Performance Computing 15 SP2 15.2 via the Phoronix Test Suite.

HTML result view exported from: https://openbenchmarking.org/result/2008293-NI-SLEHPCWK139.

sle.hpc-wk1-ML-29aug2020ProcessorMotherboardChipsetMemoryDiskGraphicsAudioNetworkOSKernelDesktopDisplay ServerOpenGLCompilerFile-SystemScreen ResolutionSystem Layersle.hpc-wk1-ML-29aug20208 x AMD Ryzen Threadripper 3960X 24-Core (16 Cores)Intel 440BX (6.00 BIOS)Intel 440BX/ZX/DX16GB129GB VMware Virtual SSVGA3D; build: RELEASE; LLVM;Ensoniq ES1371/ES1373Intel 82545EM + 3 x AMD 79c970SUSE Linux Enterprise High Performance Computing 15 SP2 15.25.3.18-24.9-default (x86_64)GNOME Shell 3.34.4X Server2.1 Mesa 19.3.4GCC 7.5.0btrfs800x600VMwareOpenBenchmarking.org- --build=x86_64-suse-linux --disable-libcc1 --disable-libssp --disable-libstdcxx-pch --disable-libvtv --disable-plugin --disable-werror --enable-checking=release --enable-gnu-indirect-function --enable-languages=c,c++,objc,fortran,obj-c++,ada,go --enable-libstdcxx-allocator=new --enable-linux-futex --enable-multilib --enable-offload-targets=hsa,nvptx-none=/usr/nvptx-none, --enable-ssp --enable-version-specific-runtime-libs --host=x86_64-suse-linux --mandir=/usr/share/man --with-arch-32=x86-64 --with-gcc-major-version-only --with-slibdir=/lib64 --with-tune=generic --without-cuda-driver --without-system-libunwind - CPU Microcode: 0x8301039- Python 2.7.17 + Python 3.6.10- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline IBPB: conditional STIBP: disabled RSB filling + srbds: Not affected + tsx_async_abort: Not affected

sle.hpc-wk1-ML-29aug2020onednn: IP Batch 1D - f32 - CPUonednn: IP Batch All - f32 - CPUonednn: IP Batch 1D - u8s8f32 - CPUonednn: IP Batch All - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Deconvolution Batch deconv_1d - f32 - CPUonednn: Deconvolution Batch deconv_3d - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Deconvolution Batch deconv_1d - u8s8f32 - CPUonednn: Deconvolution Batch deconv_3d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUnumpy: deepspeech: tensorflow-lite: SqueezeNettensorflow-lite: Inception V4tensorflow-lite: NASNet Mobiletensorflow-lite: Mobilenet Floattensorflow-lite: Mobilenet Quanttensorflow-lite: Inception ResNet V2plaidml: No - Inference - VGG16 - CPUplaidml: No - Inference - ResNet 50 - CPUnumenta-nab: EXPoSEnumenta-nab: Relative Entropynumenta-nab: Windowed Gaussiannumenta-nab: Earthgecko Skylinenumenta-nab: Bayesian Changepointsle.hpc-wk1-ML-29aug20204.1920860.41832.8684436.132911.97604.165136.4774513.59746.820405.60745353.99372.17591.473872.70409300.6861.730981591402189237155238108052117076204052018.627.65792.21023.19112.642116.22344.493OpenBenchmarking.org

oneDNN

Harness: IP Batch 1D - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch 1D - Data Type: f32 - Engine: CPUsle.hpc-wk1-ML-29aug20200.94321.88642.82963.77284.716SE +/- 0.07009, N = 34.19208MIN: 3.271. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: IP Batch All - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch All - Data Type: f32 - Engine: CPUsle.hpc-wk1-ML-29aug20201428425670SE +/- 0.18, N = 360.42MIN: 54.961. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: IP Batch 1D - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch 1D - Data Type: u8s8f32 - Engine: CPUsle.hpc-wk1-ML-29aug20200.64541.29081.93622.58163.227SE +/- 0.07072, N = 152.86844MIN: 1.421. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: IP Batch All - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch All - Data Type: u8s8f32 - Engine: CPUsle.hpc-wk1-ML-29aug2020816243240SE +/- 0.24, N = 336.13MIN: 28.431. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUsle.hpc-wk1-ML-29aug20203691215SE +/- 0.11, N = 1011.98MIN: 9.931. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch deconv_1d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_1d - Data Type: f32 - Engine: CPUsle.hpc-wk1-ML-29aug20200.93721.87442.81163.74884.686SE +/- 0.06948, N = 34.16513MIN: 2.721. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch deconv_3d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_3d - Data Type: f32 - Engine: CPUsle.hpc-wk1-ML-29aug2020246810SE +/- 0.10618, N = 36.47745MIN: 4.321. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUsle.hpc-wk1-ML-29aug20203691215SE +/- 0.07, N = 313.60MIN: 12.111. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32 - Engine: CPUsle.hpc-wk1-ML-29aug2020246810SE +/- 0.10143, N = 46.82040MIN: 5.351. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32 - Engine: CPUsle.hpc-wk1-ML-29aug20201.26172.52343.78515.04686.3085SE +/- 0.06814, N = 155.60745MIN: 3.241. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUsle.hpc-wk1-ML-29aug202080160240320400SE +/- 21.78, N = 15353.99MIN: 220.111. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUsle.hpc-wk1-ML-29aug20201632486480SE +/- 2.73, N = 1272.18MIN: 44.581. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUsle.hpc-wk1-ML-29aug20200.33160.66320.99481.32641.658SE +/- 0.04378, N = 151.47387MIN: 0.731. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUsle.hpc-wk1-ML-29aug20200.60841.21681.82522.43363.042SE +/- 0.03671, N = 32.70409MIN: 1.951. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Numpy Benchmark

OpenBenchmarking.orgScore, More Is BetterNumpy Benchmarksle.hpc-wk1-ML-29aug202070140210280350SE +/- 8.83, N = 9300.68

DeepSpeech

OpenBenchmarking.orgSeconds, Fewer Is BetterDeepSpeech 0.6sle.hpc-wk1-ML-29aug20201428425670SE +/- 0.33, N = 361.73

TensorFlow Lite

Model: SqueezeNet

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: SqueezeNetsle.hpc-wk1-ML-29aug202030K60K90K120K150KSE +/- 1671.97, N = 3159140

TensorFlow Lite

Model: Inception V4

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception V4sle.hpc-wk1-ML-29aug2020500K1000K1500K2000K2500KSE +/- 4045.79, N = 32189237

TensorFlow Lite

Model: NASNet Mobile

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: NASNet Mobilesle.hpc-wk1-ML-29aug202030K60K90K120K150KSE +/- 845.46, N = 3155238

TensorFlow Lite

Model: Mobilenet Float

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet Floatsle.hpc-wk1-ML-29aug202020K40K60K80K100KSE +/- 1797.81, N = 3108052

TensorFlow Lite

Model: Mobilenet Quant

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet Quantsle.hpc-wk1-ML-29aug202030K60K90K120K150KSE +/- 639.67, N = 3117076

TensorFlow Lite

Model: Inception ResNet V2

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception ResNet V2sle.hpc-wk1-ML-29aug2020400K800K1200K1600K2000KSE +/- 11541.24, N = 32040520

PlaidML

FP16: No - Mode: Inference - Network: VGG16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: VGG16 - Device: CPUsle.hpc-wk1-ML-29aug2020510152025SE +/- 0.26, N = 1218.62

PlaidML

FP16: No - Mode: Inference - Network: ResNet 50 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: ResNet 50 - Device: CPUsle.hpc-wk1-ML-29aug2020246810SE +/- 0.01, N = 37.65

Numenta Anomaly Benchmark

Detector: EXPoSE

OpenBenchmarking.orgSeconds, Fewer Is BetterNumenta Anomaly Benchmark 1.1Detector: EXPoSEsle.hpc-wk1-ML-29aug20202004006008001000SE +/- 1.82, N = 3792.21

Numenta Anomaly Benchmark

Detector: Relative Entropy

OpenBenchmarking.orgSeconds, Fewer Is BetterNumenta Anomaly Benchmark 1.1Detector: Relative Entropysle.hpc-wk1-ML-29aug2020612182430SE +/- 0.18, N = 323.19

Numenta Anomaly Benchmark

Detector: Windowed Gaussian

OpenBenchmarking.orgSeconds, Fewer Is BetterNumenta Anomaly Benchmark 1.1Detector: Windowed Gaussiansle.hpc-wk1-ML-29aug20203691215SE +/- 0.04, N = 312.64

Numenta Anomaly Benchmark

Detector: Earthgecko Skyline

OpenBenchmarking.orgSeconds, Fewer Is BetterNumenta Anomaly Benchmark 1.1Detector: Earthgecko Skylinesle.hpc-wk1-ML-29aug2020306090120150SE +/- 1.40, N = 3116.22

Numenta Anomaly Benchmark

Detector: Bayesian Changepoint

OpenBenchmarking.orgSeconds, Fewer Is BetterNumenta Anomaly Benchmark 1.1Detector: Bayesian Changepointsle.hpc-wk1-ML-29aug20201020304050SE +/- 0.24, N = 344.49


Phoronix Test Suite v10.8.4