2023-01-18-1402

2 x Intel Xeon E5-2680 v2 testing with a Supermicro X9DR3-F v0123456789 (3.4 BIOS) and Matrox MGA G200eW WPCM450 on Arch Linux via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2301186-NE-20230118120
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs
No Box Plots
On Line Graphs With Missing Data, Connect The Line Gaps

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs
Condense Test Profiles With Multiple Version Results Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
2 x Intel Xeon E5-2680 v2
January 18 2023
  47 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


2023-01-18-1402OpenBenchmarking.orgPhoronix Test Suite2 x Intel Xeon E5-2680 v2 @ 3.60GHz (20 Cores / 40 Threads)Supermicro X9DR3-F v0123456789 (3.4 BIOS)Intel Xeon E7 v2/Xeon128GB120GB KINGSTON SA400S3Matrox MGA G200eW WPCM4502 x Intel I350Arch Linux6.1.6-arch1-1 (x86_64)GCC 12.2.0xfs1280x1024ProcessorMotherboardChipsetMemoryDiskGraphicsNetworkOSKernelCompilerFile-SystemScreen Resolution2023-01-18-1402 BenchmarksSystem Logs- Transparent Huge Pages: always- --disable-libssp --disable-libstdcxx-pch --disable-werror --enable-__cxa_atexit --enable-bootstrap --enable-cet=auto --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-default-ssp --enable-gnu-indirect-function --enable-gnu-unique-object --enable-languages=c,c++,ada,fortran,go,lto,objc,obj-c++,d --enable-libstdcxx-backtrace --enable-link-serialization=1 --enable-lto --enable-multilib --enable-plugin --enable-shared --enable-threads=posix --mandir=/usr/share/man --with-build-config=bootstrap-lto --with-linker-hash-style=gnu - Scaling Governor: intel_cpufreq performance - CPU Microcode: 0x42e- itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Mitigation of PTE Inversion; VMX: conditional cache flushes SMT vulnerable + mds: Mitigation of Clear buffers; SMT vulnerable + meltdown: Mitigation of PTI + mmio_stale_data: Unknown: No mitigations + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: conditional RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected

2023-01-18-1402onednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: IP Shapes 1D - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPU2 x Intel Xeon E5-2680 v213.613924.352410.002854.0095111.93175.552974.976584.659068.037462.233292.306255.900426163.966167.956260.783364.453393.933351.30OpenBenchmarking.org

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU2 x Intel Xeon E5-2680 v23691215SE +/- 0.28, N = 1513.61MIN: 7.461. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon E5-2680 v2612182430SE +/- 0.25, N = 1524.35MIN: 16.921. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU

2 x Intel Xeon E5-2680 v2: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU2 x Intel Xeon E5-2680 v23691215SE +/- 0.12, N = 310.00MIN: 9.591. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon E5-2680 v20.90211.80422.70633.60844.5105SE +/- 0.03154, N = 154.00951MIN: 3.761. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU

2 x Intel Xeon E5-2680 v2: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU2 x Intel Xeon E5-2680 v23691215SE +/- 0.03, N = 911.93MIN: 11.841. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon E5-2680 v21.24942.49883.74824.99766.247SE +/- 0.00785, N = 95.55297MIN: 5.511. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU

2 x Intel Xeon E5-2680 v2: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU2 x Intel Xeon E5-2680 v21.11972.23943.35914.47885.5985SE +/- 0.02322, N = 44.97658MIN: 4.851. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon E5-2680 v21.04832.09663.14494.19325.2415SE +/- 0.03286, N = 44.65906MIN: 4.31. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU

2 x Intel Xeon E5-2680 v2: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU2 x Intel Xeon E5-2680 v2246810SE +/- 0.01391, N = 58.03746MIN: 7.91. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon E5-2680 v20.50251.0051.50752.012.5125SE +/- 0.01188, N = 52.23329MIN: 1.861. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU

2 x Intel Xeon E5-2680 v2: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU2 x Intel Xeon E5-2680 v20.51891.03781.55672.07562.5945SE +/- 0.02065, N = 42.30625MIN: 2.051. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon E5-2680 v21.32762.65523.98285.31046.638SE +/- 0.78217, N = 155.90042MIN: 2.251. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU

2 x Intel Xeon E5-2680 v2: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU2 x Intel Xeon E5-2680 v213002600390052006500SE +/- 61.01, N = 36163.96MIN: 6045.161. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon E5-2680 v213002600390052006500SE +/- 59.21, N = 36167.95MIN: 6057.961. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU2 x Intel Xeon E5-2680 v213002600390052006500SE +/- 82.86, N = 36260.78MIN: 6085.331. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU2 x Intel Xeon E5-2680 v27001400210028003500SE +/- 25.88, N = 33364.45MIN: 3296.181. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU2 x Intel Xeon E5-2680 v27001400210028003500SE +/- 31.48, N = 33393.93MIN: 3323.371. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU2 x Intel Xeon E5-2680 v27001400210028003500SE +/- 24.00, N = 33351.30MIN: 3292.81. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

CPU Temperature Monitor

MinAvgMax2 x Intel Xeon E5-2680 v230.041.455.0OpenBenchmarking.orgCelsiusCPU Temperature MonitorPhoronix Test Suite System Monitoring1530456075

System Temperature Monitor

MinAvgMax2 x Intel Xeon E5-2680 v240.048.360.0OpenBenchmarking.orgCelsiusSystem Temperature MonitorPhoronix Test Suite System Monitoring1632486480