5218-oneDNN

2 x Intel Xeon Gold 5218 testing with a Intel S2600WFT (SE5C620.86B.02.01.0008.031920191559 BIOS) and ASPEED on CentOS Linux 7 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2009213-NE-5218ONEDN12
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
2 x Intel Xeon Gold 5218
September 21 2020
  22 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


5218-oneDNNOpenBenchmarking.orgPhoronix Test Suite2 x Intel Xeon Gold 5218 @ 3.90GHz (32 Cores / 64 Threads)Intel S2600WFT (SE5C620.86B.02.01.0008.031920191559 BIOS)16384 MB + 129408 MB + 16384 MB + 129408 MB + 16384 MB + 129408 MB + 16384 MB + 129408 MB + 16384 MB + 129408 MB + 16384 MB + 129408 MB + 16384 MB + 129408 MB + 16384 MB + 129408 MB + 16384 MB + 129408 MB + 16384 MB + 129408 MB + 16384 MB + 129408 MB + 16384 MB + 129408 MB DDR4-2666MT/s7 x 1600GB INTEL SSDPE2KE016T8ASPEED2 x Mellanox MT27710CentOS Linux 73.10.0-1127.el7.x86_64 (x86_64)GCC 4.8.5 20150623xfs1024x768ProcessorMotherboardMemoryDiskGraphicsNetworkOSKernelCompilerFile-SystemScreen Resolution5218-oneDNN BenchmarksSystem Logs- --build=x86_64-redhat-linux --disable-libgcj --disable-libunwind-exceptions --enable-__cxa_atexit --enable-bootstrap --enable-checking=release --enable-gnu-indirect-function --enable-gnu-unique-object --enable-initfini-array --enable-languages=c,c++,objc,obj-c++,java,fortran,ada,go,lto --enable-plugin --enable-shared --enable-threads=posix --mandir=/usr/share/man --with-arch_32=x86-64 --with-linker-hash-style=gnu --with-tune=generic - Scaling Governor: intel_pstate powersave - CPU Microcode: 0x500002c- SELinux + itlb_multihit: vulnerable + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of Load fences usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full retpoline IBPB + tsx_async_abort: Mitigation of Clear buffers; SMT vulnerable

5218-oneDNNonednn: IP Batch 1D - f32 - CPUonednn: IP Batch All - f32 - CPUonednn: IP Batch 1D - u8s8f32 - CPUonednn: IP Batch All - u8s8f32 - CPUonednn: IP Batch 1D - bf16bf16bf16 - CPUonednn: IP Batch All - bf16bf16bf16 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Deconvolution Batch deconv_1d - f32 - CPUonednn: Deconvolution Batch deconv_3d - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Deconvolution Batch deconv_1d - u8s8f32 - CPUonednn: Deconvolution Batch deconv_3d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Convolution Batch Shapes Auto - bf16bf16bf16 - CPUonednn: Deconvolution Batch deconv_1d - bf16bf16bf16 - CPUonednn: Deconvolution Batch deconv_3d - bf16bf16bf16 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - bf16bf16bf16 - CPU2 x Intel Xeon Gold 52182.0685229.46400.6769097.950064.7190863.61305.582142.481783.018325.258710.6990110.768639286.34874.84668.500779.3529611.15090.6583100.4118401.90325OpenBenchmarking.org

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch 1D - Data Type: f32 - Engine: CPU2 x Intel Xeon Gold 52180.46540.93081.39621.86162.327SE +/- 0.00619, N = 32.06852MIN: 1.911. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch All - Data Type: f32 - Engine: CPU2 x Intel Xeon Gold 5218714212835SE +/- 0.05, N = 329.46MIN: 27.921. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch 1D - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon Gold 52180.15230.30460.45690.60920.7615SE +/- 0.003884, N = 30.676909MIN: 0.611. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch All - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon Gold 5218246810SE +/- 0.00984, N = 37.95006MIN: 7.561. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch 1D - Data Type: bf16bf16bf16 - Engine: CPU2 x Intel Xeon Gold 52181.06182.12363.18544.24725.309SE +/- 0.00336, N = 34.71908MIN: 4.511. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch All - Data Type: bf16bf16bf16 - Engine: CPU2 x Intel Xeon Gold 52181428425670SE +/- 0.04, N = 363.61MIN: 62.641. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU2 x Intel Xeon Gold 52181.2562.5123.7685.0246.28SE +/- 0.03121, N = 35.58214MIN: 4.81. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_1d - Data Type: f32 - Engine: CPU2 x Intel Xeon Gold 52180.55841.11681.67522.23362.792SE +/- 0.00251, N = 32.48178MIN: 2.341. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_3d - Data Type: f32 - Engine: CPU2 x Intel Xeon Gold 52180.67911.35822.03732.71643.3955SE +/- 0.01722, N = 33.01832MIN: 2.941. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon Gold 52181.18322.36643.54964.73285.916SE +/- 0.01078, N = 35.25871MIN: 5.111. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon Gold 52180.15730.31460.47190.62920.7865SE +/- 0.001093, N = 30.699011MIN: 0.631. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon Gold 52180.17290.34580.51870.69160.8645SE +/- 0.002329, N = 30.768639MIN: 0.741. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU2 x Intel Xeon Gold 521860120180240300SE +/- 1.50, N = 3286.35MIN: 274.771. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU2 x Intel Xeon Gold 521820406080100SE +/- 0.10, N = 374.85MIN: 71.61. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU2 x Intel Xeon Gold 5218246810SE +/- 0.00738, N = 38.50077MIN: 8.361. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_1d - Data Type: bf16bf16bf16 - Engine: CPU2 x Intel Xeon Gold 52183691215SE +/- 0.00637, N = 39.35296MIN: 9.021. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_3d - Data Type: bf16bf16bf16 - Engine: CPU2 x Intel Xeon Gold 52183691215SE +/- 0.01, N = 311.15MIN: 10.991. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU2 x Intel Xeon Gold 52180.14810.29620.44430.59240.7405SE +/- 0.001104, N = 30.658310MIN: 0.611. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon Gold 52180.09270.18540.27810.37080.4635SE +/- 0.002937, N = 30.411840MIN: 0.351. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU2 x Intel Xeon Gold 52180.42820.85641.28461.71282.141SE +/- 0.00146, N = 31.90325MIN: 1.831. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -rdynamic -lpthread -ldl