3700X More march

AMD Ryzen 7 3700X 8-Core testing with a Gigabyte A320M-S2H-CF (F52a BIOS) and HIS AMD Radeon HD 7750/8740 / R7 250E 1GB on Ubuntu 20.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2103170-IB-3700XMORE69
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts

Limit displaying results to tests within:

CPU Massive 4 Tests
Creator Workloads 3 Tests
Encoding 2 Tests
HPC - High Performance Computing 2 Tests
Multi-Core 5 Tests
Server CPU Tests 4 Tests
Video Encoding 2 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
1
March 17 2021
  1 Hour, 3 Minutes
2
March 17 2021
  1 Hour, 8 Minutes
3
March 17 2021
  43 Minutes
Invert Hiding All Results Option
  58 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


3700X More marchProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerOpenGLCompilerFile-SystemScreen Resolution123AMD Ryzen 7 3700X 8-Core @ 3.60GHz (8 Cores / 16 Threads)Gigabyte A320M-S2H-CF (F52a BIOS)AMD Starship/Matisse8GB240GB TOSHIBA RC100HIS AMD Radeon HD 7750/8740 / R7 250E 1GBAMD Oland/Hainan/CapeDELL S2409WRealtek RTL8111/8168/8411Ubuntu 20.045.8.1-050801-generic (x86_64)GNOME Shell 3.36.4X Server 1.20.94.5 Mesa 20.0.8 (LLVM 10.0.0)GCC 9.3.0ext41920x1080OpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-9-HskZEa/gcc-9-9.3.0/debian/tmp-nvptx/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- Scaling Governor: acpi-cpufreq ondemand (Boost: Enabled) - CPU Microcode: 0x8701021 Python Details- Python 3.8.5Security Details- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional STIBP: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected

123Result OverviewPhoronix Test Suite100%104%108%111%oneDNNTimed Mesa CompilationSVT-VP9Xcompact3d Incompact3dSVT-HEVC

3700X More marchonednn: IP Shapes 3D - u8s8f32 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUincompact3d: input.i3d 129 Cells Per Directiononednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: IP Shapes 1D - f32 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUbuild-mesa: Time To Compileincompact3d: input.i3d 192 Cells Per Directionsvt-hevc: 1 - Bosphorus 1080ponednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUsvt-hevc: 7 - Bosphorus 1080psvt-vp9: Visual Quality Optimized - Bosphorus 1080psysbench: CPUsvt-vp9: PSNR/SSIM Optimized - Bosphorus 1080ponednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUsysbench: RAM / Memorysvt-hevc: 10 - Bosphorus 1080ponednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUsvt-vp9: VMAF Optimized - Bosphorus 1080p1232.3862310.23333899.222746.852758.754.8277921.05842743.143873.423858.1140.34667716.697245.5671523.038055.883318.2113547.634.622873.05660106.19114.3917383.33143.243.6397610289.34209.258.992452.63018136.602.6882710.74003936.822819.512830.694.9447721.5222801.033938.803919.2040.47938796.748625.5320423.094355.559316.5289417.594.643393.06767105.83114.4517341.94143.513.6447810276.90209.218.705082.62524134.182.4609610.95203811.1321.244740.71856696.735585.5343822.953855.758317.0066227.614.63664105.92114.17143.343.64264209.458.7141111.59989135.52OpenBenchmarking.org

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU3210.60491.20981.81472.41963.0245SE +/- 0.00446, N = 3SE +/- 0.00780, N = 3SE +/- 0.01088, N = 32.460962.688272.38623MIN: 2.4MIN: 2.61MIN: 2.31. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU3213691215SE +/- 0.01, N = 3SE +/- 0.02, N = 3SE +/- 0.03, N = 310.9510.7410.23MIN: 10.66MIN: 10.43MIN: 9.781. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU3218001600240032004000SE +/- 4.52, N = 3SE +/- 19.48, N = 3SE +/- 2.05, N = 33811.133936.823899.22MIN: 3793.22MIN: 3902.02MIN: 3886.771. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU216001200180024003000SE +/- 7.65, N = 3SE +/- 11.92, N = 32819.512746.85MIN: 2784.96MIN: 2702.151. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU216001200180024003000SE +/- 7.42, N = 3SE +/- 12.62, N = 32830.692758.75MIN: 2793.58MIN: 2706.381. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU211.11262.22523.33784.45045.563SE +/- 0.01312, N = 3SE +/- 0.01300, N = 34.944774.82779MIN: 4.77MIN: 4.661. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU321510152025SE +/- 0.03, N = 3SE +/- 0.02, N = 3SE +/- 0.04, N = 321.2421.5221.06MIN: 20.94MIN: 21.31MIN: 20.681. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU216001200180024003000SE +/- 8.10, N = 3SE +/- 16.15, N = 32801.032743.14MIN: 2771.6MIN: 2706.621. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU218001600240032004000SE +/- 6.78, N = 3SE +/- 16.07, N = 33938.803873.42MIN: 3920.91MIN: 3842.11. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU218001600240032004000SE +/- 2.50, N = 3SE +/- 13.76, N = 33919.203858.11MIN: 3903.26MIN: 3834.741. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Xcompact3d Incompact3d

Xcompact3d Incompact3d is a Fortran-MPI based, finite difference high-performance code for solving the incompressible Navier-Stokes equation and as many as you need scalar transport equations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterXcompact3d Incompact3d 2021-03-11Input: input.i3d 129 Cells Per Direction321918273645SE +/- 0.22, N = 3SE +/- 0.05, N = 3SE +/- 0.04, N = 340.7240.4840.351. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU321246810SE +/- 0.00795, N = 3SE +/- 0.01474, N = 3SE +/- 0.00606, N = 36.735586.748626.69724MIN: 6.65MIN: 6.64MIN: 6.621. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU3211.25262.50523.75785.01046.263SE +/- 0.00906, N = 3SE +/- 0.01225, N = 3SE +/- 0.02182, N = 35.534385.532045.56715MIN: 5.4MIN: 5.37MIN: 5.41. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU321612182430SE +/- 0.01, N = 3SE +/- 0.03, N = 3SE +/- 0.00, N = 322.9523.0923.04MIN: 22.74MIN: 22.81MIN: 22.721. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Timed Mesa Compilation

This test profile times how long it takes to compile Mesa with Meson/Ninja. For minimizing build dependencies and avoid versioning conflicts, test this is just the core Mesa build without LLVM or the extra Gallium3D/Mesa drivers enabled. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Mesa Compilation 21.0Time To Compile3211326395265SE +/- 0.06, N = 3SE +/- 0.10, N = 3SE +/- 0.01, N = 355.7655.5655.88

Xcompact3d Incompact3d

Xcompact3d Incompact3d is a Fortran-MPI based, finite difference high-performance code for solving the incompressible Navier-Stokes equation and as many as you need scalar transport equations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterXcompact3d Incompact3d 2021-03-11Input: input.i3d 192 Cells Per Direction32170140210280350SE +/- 0.35, N = 3SE +/- 0.22, N = 3SE +/- 1.26, N = 3317.01316.53318.211. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 1 - Input: Bosphorus 1080p321246810SE +/- 0.03, N = 3SE +/- 0.01, N = 3SE +/- 0.02, N = 37.617.597.631. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU3211.04482.08963.13444.17925.224SE +/- 0.00935, N = 3SE +/- 0.00974, N = 3SE +/- 0.00500, N = 34.636644.643394.62287MIN: 4.45MIN: 4.47MIN: 4.431. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU210.69021.38042.07062.76083.451SE +/- 0.00076, N = 3SE +/- 0.00195, N = 33.067673.05660MIN: 3MIN: 2.981. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 7 - Input: Bosphorus 1080p32120406080100SE +/- 0.22, N = 3SE +/- 0.29, N = 3SE +/- 0.04, N = 3105.92105.83106.191. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

SVT-VP9

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample YUV input video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-VP9 0.3Tuning: Visual Quality Optimized - Input: Bosphorus 1080p321306090120150SE +/- 0.46, N = 3SE +/- 0.10, N = 3SE +/- 0.20, N = 3114.17114.45114.391. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm

Sysbench

This is a benchmark of Sysbench with the built-in CPU and memory sub-tests. Sysbench is a scriptable multi-threaded benchmark tool based on LuaJIT. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgEvents Per Second, More Is BetterSysbench 1.0.20Test: CPU214K8K12K16K20KSE +/- 33.37, N = 3SE +/- 5.91, N = 317341.9417383.331. (CC) gcc options: -pthread -O2 -funroll-loops -rdynamic -ldl -laio -lm

SVT-VP9

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample YUV input video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-VP9 0.3Tuning: PSNR/SSIM Optimized - Input: Bosphorus 1080p321306090120150SE +/- 0.23, N = 3SE +/- 0.39, N = 3SE +/- 0.20, N = 3143.34143.51143.241. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU3210.82011.64022.46033.28044.1005SE +/- 0.01166, N = 3SE +/- 0.00448, N = 3SE +/- 0.00803, N = 33.642643.644783.63976MIN: 3.47MIN: 3.47MIN: 3.471. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Sysbench

This is a benchmark of Sysbench with the built-in CPU and memory sub-tests. Sysbench is a scriptable multi-threaded benchmark tool based on LuaJIT. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/sec, More Is BetterSysbench 1.0.20Test: RAM / Memory212K4K6K8K10KSE +/- 23.22, N = 3SE +/- 4.79, N = 310276.9010289.341. (CC) gcc options: -pthread -O2 -funroll-loops -rdynamic -ldl -laio -lm

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 10 - Input: Bosphorus 1080p32150100150200250SE +/- 0.49, N = 3SE +/- 0.26, N = 3SE +/- 0.31, N = 3209.45209.21209.251. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU3213691215SE +/- 0.21403, N = 15SE +/- 0.16196, N = 15SE +/- 0.01522, N = 38.714118.705088.99245MIN: 5.16MIN: 5.13MIN: 5.171. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU3213691215SE +/- 5.08322, N = 12SE +/- 0.00360, N = 3SE +/- 0.00406, N = 311.599892.625242.63018MIN: 2.57MIN: 2.58MIN: 2.581. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

SVT-VP9

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample YUV input video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-VP9 0.3Tuning: VMAF Optimized - Input: Bosphorus 1080p321306090120150SE +/- 3.07, N = 12SE +/- 4.27, N = 12SE +/- 1.75, N = 3135.52134.18136.601. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm