dl560gen10-1tb-klasuite

4 x Intel Xeon Platinum 8160 testing with a HPE ProLiant DL560 Gen10 (U34 BIOS) and llvmpipe on CentOS 7.6.1810 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2303274-NE-DL560GEN147
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
1TB-Run1
March 24 2023
  3 Days, 5 Hours, 39 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


dl560gen10-1tb-klasuiteOpenBenchmarking.orgPhoronix Test Suite4 x Intel Xeon Platinum 8160 (96 Cores / 192 Threads)HPE ProLiant DL560 Gen10 (U34 BIOS)Intel Sky Lake-E DMI3 Registers32 x GB DDR4-2666MT/s HPE429GB P824i-p MR Gen10 + 107GB P824i-p MR Gen10 + 1100GB P824i-p MR Gen10 + 36757GB P824i-p MR Gen10llvmpipeSmart Cable4 x Intel I350 + 2 x Intel X710 for 10GbE SFP+ + 2 x Intel 10G X550TCentOS 7.6.18103.10.0-957.21.3.el7.x86_64 (x86_64)GNOME Shell 3.28.3X Server 1.20.12.1 Mesa 18.0.5 (LLVM 6.0 256 bits)GCC 4.8.5 20150623ext41024x768ProcessorMotherboardChipsetMemoryDiskGraphicsMonitorNetworkOSKernelDesktopDisplay ServerOpenGLCompilerFile-SystemScreen ResolutionDl560gen10-1tb-klasuite BenchmarksSystem Logs- Transparent Huge Pages: always- --build=x86_64-redhat-linux --disable-libgcj --disable-libunwind-exceptions --enable-__cxa_atexit --enable-bootstrap --enable-checking=release --enable-gnu-indirect-function --enable-gnu-unique-object --enable-initfini-array --enable-languages=c,c++,objc,obj-c++,java,fortran,ada,go,lto --enable-plugin --enable-shared --enable-threads=posix --mandir=/usr/share/man --with-arch_32=x86-64 --with-linker-hash-style=gnu --with-tune=generic - CPU Microcode: 0x2006e05- OpenJDK Runtime Environment (build 1.8.0_222-b10)- Python 2.7.5 + Python 3.6.8- l1tf: Mitigation of PTE Inversion + mds: Mitigation of Clear buffers; SMT vulnerable + meltdown: Mitigation of PTI + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of Load fences __user pointer sanitization + spectre_v2: Mitigation of IBRS (kernel) IBPB

dl560gen10-1tb-klasuiteonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUincompact3d: X3D-benchmarking input.i3dopencv: Image Processingcompress-zstd: 8 - Decompression Speedcompress-zstd: 8 - Compression Speedcompress-zstd: 8, Long Mode - Decompression Speedcompress-zstd: 8, Long Mode - Compression Speedcompress-zstd: 12 - Decompression Speedcompress-zstd: 12 - Compression Speedjohn-the-ripper: MD5john-the-ripper: HMAC-SHA512opencv: Stitchingcloverleaf: Lagrangian-Eulerian Hydrodynamicslammps: 20k Atomsopencv: Features 2Dopencv: Corestress-ng: Atomicstress-ng: Pthreadstress-ng: Mutexstress-ng: Context Switchingstress-ng: Zlibjohn-the-ripper: WPA PSKstress-ng: Glibc Qsort Data Sortingjohn-the-ripper: bcryptstress-ng: System V Message Passingstress-ng: Function Callstress-ng: SENDFILEstress-ng: Glibc C String Functionsstress-ng: CPU Cachestress-ng: Socket Activitystress-ng: MMAPstress-ng: Forkingstress-ng: x86_64 RdRandcompress-zstd: 3, Long Mode - Decompression Speedcompress-zstd: 3, Long Mode - Compression Speedonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUstress-ng: NUMAncnn: CPU - FastestDetncnn: CPU - vision_transformerncnn: CPU - regnety_400mncnn: CPU - squeezenet_ssdncnn: CPU - yolov4-tinyncnn: CPU - resnet50ncnn: CPU - alexnetncnn: CPU - resnet18ncnn: CPU - vgg16ncnn: CPU - googlenetncnn: CPU - blazefacencnn: CPU - efficientnet-b0ncnn: CPU - mnasnetncnn: CPU - shufflenet-v2ncnn: CPU-v3-v3 - mobilenet-v3ncnn: CPU-v2-v2 - mobilenet-v2ncnn: CPU - mobilenetstress-ng: Memory Copyingstress-ng: Vector Mathstress-ng: Hashstress-ng: Futexopencv: DNN - Deep Neural Networkonednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUminibude: OpenMP - BM2minibude: OpenMP - BM2numpy: compress-zstd: 19, Long Mode - Decompression Speedcompress-zstd: 19, Long Mode - Compression Speedonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUsvt-av1: Preset 4 - Bosphorus 4Ksvt-hevc: 1 - Bosphorus 4Kcompress-zstd: 19 - Decompression Speedcompress-zstd: 19 - Compression Speedcompress-zstd: 3 - Decompression Speedcompress-zstd: 3 - Compression Speeddacapobench: H2onednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 3D - bf16bf16bf16 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUsvt-av1: Preset 13 - Bosphorus 4Ksvt-hevc: 10 - Bosphorus 4Knamd: ATPase Simulation - 327,506 Atomssvt-av1: Preset 8 - Bosphorus 4Kpybench: Total For Average Test Timesonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - bf16bf16bf16 - CPUamg: phpbench: PHP Benchmark Suitejohn-the-ripper: Blowfishstress-ng: Semaphoresstress-ng: Pollstress-ng: Matrix Mathstress-ng: Mallocstress-ng: CPU Stressstress-ng: Cryptoonednn: Deconvolution Batch shapes_1d - bf16bf16bf16 - CPUsvt-av1: Preset 4 - Bosphorus 1080pcython-bench: N-Queensincompact3d: input.i3d 193 Cells Per Directiontjbench: Decompression Throughputdacapobench: Eclipsedolfyn: Computational Fluid Dynamicssvt-hevc: 1 - Bosphorus 1080pminibude: OpenMP - BM1minibude: OpenMP - BM1onednn: Deconvolution Batch shapes_3d - bf16bf16bf16 - CPUonednn: IP Shapes 1D - f32 - CPUonednn: IP Shapes 1D - bf16bf16bf16 - CPUlulesh: opencv: Object Detectiononednn: Matrix Multiply Batch Shapes Transformer - bf16bf16bf16 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUsvt-av1: Preset 12 - Bosphorus 4Kincompact3d: input.i3d 129 Cells Per Directionsvt-hevc: 7 - Bosphorus 4Ksvt-av1: Preset 8 - Bosphorus 1080pdacapobench: Jythonsvt-hevc: 7 - Bosphorus 1080psvt-av1: Preset 12 - Bosphorus 1080ponednn: Deconvolution Batch shapes_3d - f32 - CPUsvt-hevc: 10 - Bosphorus 1080ponednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUsvt-av1: Preset 13 - Bosphorus 1080plammps: Rhodopsin Proteinopencv: Graph API1TB-Run1597.540596.686436.610931370799840.7610.4850.8559.3777.8168.661460507613378033444259.4027.28322288422140666.0471318.00933053.6529527790.454415.01206487807.081108694711689.93390289.591363842.774959677.9835.073142.2510864.3340359.971051693.30850.6676.61187.9958.5139.51232.60171.5263.4857.3374.4617.2330.61153.0256.3716.9842.4930.0935.0630.5034.4049.301300.94166779.7412668067.86248.707706418.71781.05454116.337408.423259.15668.67.871183.021199.89597.3584.261082.4189.22670.014.3820.61979.971206.7805831.18592.337771.25277101.531110.720.3264841.83515733.826142.285792.42126191511380019605511061116591808.115837581.44332280.3191315767.44120051.2783163.044.856996.13028.50614.5665817163.8144542061622.47930.3813.360334.0074.195861.970244.0083534264.225458476.028120.50225194.0983.2422503692.4596.2804151215.31312.8811.18163360.290.668686337.64231.953OpenBenchmarking.org

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU1TB-Run1130260390520650SE +/- 0.87, N = 100597.54MIN: 555.831. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU1TB-Run1130260390520650SE +/- 0.92, N = 100596.69MIN: 555.791. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Xcompact3d Incompact3d

Xcompact3d Incompact3d is a Fortran-MPI based, finite difference high-performance code for solving the incompressible Navier-Stokes equation and as many as you need scalar transport equations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterXcompact3d Incompact3d 2021-03-11Input: X3D-benchmarking input.i3d1TB-Run190180270360450SE +/- 25.18, N = 18436.611. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -pthread -m64 -lmpi_usempi -lmpi_mpifh -lmpi

OpenCV

This is a benchmark of the OpenCV (Computer Vision) library's built-in performance tests. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenCV 4.7Test: Image Processing1TB-Run180K160K240K320K400KSE +/- 2962.47, N = 183707991. (CXX) g++ options: -fPIC -fsigned-char -pthread -fomit-frame-pointer -ffunction-sections -fdata-sections -msse -msse2 -msse3 -fvisibility=hidden -O3 -shared

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8 - Decompression Speed1TB-Run12004006008001000SE +/- 0.47, N = 100840.71. (CC) gcc options: -O3 -pthread -lz -llzma

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8 - Compression Speed1TB-Run1130260390520650SE +/- 2.72, N = 100610.41. (CC) gcc options: -O3 -pthread -lz -llzma

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8, Long Mode - Decompression Speed1TB-Run12004006008001000SE +/- 0.60, N = 100850.81. (CC) gcc options: -O3 -pthread -lz -llzma

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8, Long Mode - Compression Speed1TB-Run1120240360480600SE +/- 1.69, N = 100559.31. (CC) gcc options: -O3 -pthread -lz -llzma

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 12 - Decompression Speed1TB-Run12004006008001000SE +/- 0.76, N = 100777.81. (CC) gcc options: -O3 -pthread -lz -llzma

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 12 - Compression Speed1TB-Run14080120160200SE +/- 0.54, N = 100168.61. (CC) gcc options: -O3 -pthread -lz -llzma

John The Ripper

This is a benchmark of John The Ripper, which is a password cracker. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgReal C/S, More Is BetterJohn The Ripper 2023.03.14Test: MD51TB-Run11.3M2.6M3.9M5.2M6.5MSE +/- 42604.32, N = 10061460501. (CC) gcc options: -m64 -lssl -lcrypto -fopenmp -lgmp -pthread -lm -lrt -lz -ldl -lcrypt -lbz2

OpenBenchmarking.orgReal C/S, More Is BetterJohn The Ripper 2023.03.14Test: HMAC-SHA5121TB-Run116M32M48M64M80MSE +/- 392784.50, N = 100761337801. (CC) gcc options: -m64 -lssl -lcrypto -fopenmp -lgmp -pthread -lm -lrt -lz -ldl -lcrypt -lbz2

OpenCV

This is a benchmark of the OpenCV (Computer Vision) library's built-in performance tests. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenCV 4.7Test: Stitching1TB-Run170K140K210K280K350KSE +/- 2269.46, N = 183344421. (CXX) g++ options: -fPIC -fsigned-char -pthread -fomit-frame-pointer -ffunction-sections -fdata-sections -msse -msse2 -msse3 -fvisibility=hidden -O3 -shared

CloverLeaf

CloverLeaf is a Lagrangian-Eulerian hydrodynamics benchmark. This test profile currently makes use of CloverLeaf's OpenMP version and benchmarked with the clover_bm.in input file (Problem 5). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterCloverLeafLagrangian-Eulerian Hydrodynamics1TB-Run11326395265SE +/- 0.83, N = 10059.401. (F9X) gfortran options: -O3 -march=native -funroll-loops -fopenmp

LAMMPS Molecular Dynamics Simulator

LAMMPS is a classical molecular dynamics code, and an acronym for Large-scale Atomic/Molecular Massively Parallel Simulator. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgns/day, More Is BetterLAMMPS Molecular Dynamics Simulator 23Jun2022Model: 20k Atoms1TB-Run1612182430SE +/- 0.21, N = 1827.281. (CXX) g++ options: -O3 -pthread -ldl

OpenCV

This is a benchmark of the OpenCV (Computer Vision) library's built-in performance tests. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenCV 4.7Test: Features 2D1TB-Run150K100K150K200K250KSE +/- 11211.58, N = 242228841. (CXX) g++ options: -fPIC -fsigned-char -pthread -fomit-frame-pointer -ffunction-sections -fdata-sections -msse -msse2 -msse3 -fvisibility=hidden -O3 -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenCV 4.7Test: Core1TB-Run150K100K150K200K250KSE +/- 3547.19, N = 182214061. (CXX) g++ options: -fPIC -fsigned-char -pthread -fomit-frame-pointer -ffunction-sections -fdata-sections -msse -msse2 -msse3 -fvisibility=hidden -O3 -shared

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Atomic1TB-Run11530456075SE +/- 0.55, N = 10066.041. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Pthread1TB-Run115K30K45K60K75KSE +/- 3937.61, N = 10071318.001. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Mutex1TB-Run1200K400K600K800K1000KSE +/- 74916.71, N = 100933053.651. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Context Switching1TB-Run16M12M18M24M30MSE +/- 879721.20, N = 10029527790.451. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Zlib1TB-Run19001800270036004500SE +/- 79.90, N = 1004415.011. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

John The Ripper

This is a benchmark of John The Ripper, which is a password cracker. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgReal C/S, More Is BetterJohn The Ripper 2023.03.14Test: WPA PSK1TB-Run140K80K120K160K200KSE +/- 957.18, N = 1002064871. (CC) gcc options: -m64 -lssl -lcrypto -fopenmp -lgmp -pthread -lm -lrt -lz -ldl -lcrypt -lbz2

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Glibc Qsort Data Sorting1TB-Run12004006008001000SE +/- 14.74, N = 100807.081. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

John The Ripper

This is a benchmark of John The Ripper, which is a password cracker. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgReal C/S, More Is BetterJohn The Ripper 2023.03.14Test: bcrypt1TB-Run120K40K60K80K100KSE +/- 258.97, N = 1001108691. (CC) gcc options: -m64 -lssl -lcrypto -fopenmp -lgmp -pthread -lm -lrt -lz -ldl -lcrypt -lbz2

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: System V Message Passing1TB-Run11000K2000K3000K4000K5000KSE +/- 268866.31, N = 1004711689.931. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Function Call1TB-Run180K160K240K320K400KSE +/- 4589.77, N = 100390289.591. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: SENDFILE1TB-Run1300K600K900K1200K1500KSE +/- 15306.66, N = 1001363842.771. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Glibc C String Functions1TB-Run11.1M2.2M3.3M4.4M5.5MSE +/- 99455.84, N = 1004959677.981. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: CPU Cache1TB-Run1816243240SE +/- 7.41, N = 10035.071. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Socket Activity1TB-Run17001400210028003500SE +/- 465.54, N = 1003142.251. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: MMAP1TB-Run12K4K6K8K10KSE +/- 310.71, N = 10010864.331. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Forking1TB-Run19K18K27K36K45KSE +/- 307.02, N = 10040359.971. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: x86_64 RdRand1TB-Run1200K400K600K800K1000KSE +/- 26362.36, N = 961051693.301. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3, Long Mode - Decompression Speed1TB-Run12004006008001000SE +/- 1.69, N = 8850.61. (CC) gcc options: -O3 -pthread -lz -llzma

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3, Long Mode - Compression Speed1TB-Run1150300450600750SE +/- 2.50, N = 45676.61. (CC) gcc options: -O3 -pthread -lz -llzma

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU1TB-Run130060090012001500SE +/- 5.07, N = 341187.99MIN: 1074.421. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: NUMA1TB-Run11326395265SE +/- 1.31, N = 8058.511. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: FastestDet1TB-Run1918273645SE +/- 0.32, N = 639.51MIN: 37.77 / MAX: 261.211. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: vision_transformer1TB-Run150100150200250SE +/- 10.15, N = 6232.60MIN: 191.01 / MAX: 4140.651. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: regnety_400m1TB-Run14080120160200SE +/- 5.81, N = 6171.52MIN: 156.86 / MAX: 4341.861. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: squeezenet_ssd1TB-Run11428425670SE +/- 1.05, N = 663.48MIN: 54.4 / MAX: 732.881. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: yolov4-tiny1TB-Run11326395265SE +/- 2.91, N = 657.33MIN: 48.4 / MAX: 3482.321. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: resnet501TB-Run120406080100SE +/- 8.44, N = 674.46MIN: 50.79 / MAX: 8860.41. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: alexnet1TB-Run148121620SE +/- 0.45, N = 617.23MIN: 14.52 / MAX: 56.931. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: resnet181TB-Run1714212835SE +/- 1.33, N = 630.61MIN: 23.87 / MAX: 483.051. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: vgg161TB-Run1306090120150SE +/- 30.15, N = 6153.02MIN: 41.73 / MAX: 25291. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: googlenet1TB-Run11326395265SE +/- 0.74, N = 656.37MIN: 49.86 / MAX: 528.261. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: blazeface1TB-Run148121620SE +/- 0.10, N = 616.98MIN: 16.28 / MAX: 49.91. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: efficientnet-b01TB-Run11020304050SE +/- 0.51, N = 642.49MIN: 40.21 / MAX: 398.391. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: mnasnet1TB-Run1714212835SE +/- 0.14, N = 630.09MIN: 27.93 / MAX: 182.231. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: shufflenet-v21TB-Run1816243240SE +/- 0.19, N = 635.06MIN: 33.54 / MAX: 86.851. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU-v3-v3 - Model: mobilenet-v31TB-Run1714212835SE +/- 0.20, N = 630.50MIN: 28.91 / MAX: 343.391. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU-v2-v2 - Model: mobilenet-v21TB-Run1816243240SE +/- 2.77, N = 634.40MIN: 28.61 / MAX: 4378.561. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: mobilenet1TB-Run11122334455SE +/- 0.43, N = 649.30MIN: 42.57 / MAX: 703.971. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Memory Copying1TB-Run130060090012001500SE +/- 290.96, N = 801300.941. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Vector Math1TB-Run140K80K120K160K200KSE +/- 2261.24, N = 80166779.741. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Hash1TB-Run13M6M9M12M15MSE +/- 72150.37, N = 8012668067.861. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Futex1TB-Run150100150200250SE +/- 76.38, N = 80248.701. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenCV

This is a benchmark of the OpenCV (Computer Vision) library's built-in performance tests. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenCV 4.7Test: DNN - Deep Neural Network1TB-Run117K34K51K68K85KSE +/- 1179.12, N = 30770641. (CXX) g++ options: -fPIC -fsigned-char -pthread -fomit-frame-pointer -ffunction-sections -fdata-sections -msse -msse2 -msse3 -fvisibility=hidden -O3 -shared

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU1TB-Run1510152025SE +/- 0.09, N = 10018.72MIN: 14.061. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU1TB-Run10.23730.47460.71190.94921.1865SE +/- 0.016991, N = 1001.054541MIN: 0.711. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

miniBUDE

MiniBUDE is a mini application for the the core computation of the Bristol University Docking Engine (BUDE). This test profile currently makes use of the OpenMP implementation of miniBUDE for CPU benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBillion Interactions/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM21TB-Run148121620SE +/- 0.02, N = 616.341. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

OpenBenchmarking.orgGFInst/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM21TB-Run190180270360450SE +/- 0.59, N = 6408.421. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

Numpy Benchmark

This is a test to obtain the general Numpy performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgScore, More Is BetterNumpy Benchmark1TB-Run160120180240300SE +/- 0.66, N = 6259.15

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19, Long Mode - Decompression Speed1TB-Run1140280420560700SE +/- 1.22, N = 20668.61. (CC) gcc options: -O3 -pthread -lz -llzma

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19, Long Mode - Compression Speed1TB-Run1246810SE +/- 0.02, N = 207.871. (CC) gcc options: -O3 -pthread -lz -llzma

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU1TB-Run130060090012001500SE +/- 6.04, N = 201183.02MIN: 1073.91. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU1TB-Run130060090012001500SE +/- 6.02, N = 201199.89MIN: 1105.111. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU1TB-Run1130260390520650SE +/- 1.43, N = 20597.36MIN: 558.681. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU1TB-Run10.95871.91742.87613.83484.7935SE +/- 0.01688, N = 1004.26108MIN: 3.411. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 4 - Input: Bosphorus 4K1TB-Run10.54411.08821.63232.17642.7205SE +/- 0.008, N = 202.4181. (CXX) g++ options: -march=native -mno-avx

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 1 - Input: Bosphorus 4K1TB-Run13691215SE +/- 0.02, N = 209.221. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19 - Decompression Speed1TB-Run1140280420560700SE +/- 1.68, N = 20670.01. (CC) gcc options: -O3 -pthread -lz -llzma

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19 - Compression Speed1TB-Run148121620SE +/- 0.07, N = 2014.31. (CC) gcc options: -O3 -pthread -lz -llzma

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3 - Decompression Speed1TB-Run12004006008001000SE +/- 0.68, N = 16820.61. (CC) gcc options: -O3 -pthread -lz -llzma

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3 - Compression Speed1TB-Run1400800120016002000SE +/- 10.65, N = 201979.91. (CC) gcc options: -O3 -pthread -lz -llzma

DaCapo Benchmark

This test runs the DaCapo Benchmarks written in Java and intended to test system/CPU performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgmsec, Fewer Is BetterDaCapo Benchmark 9.12-MR1Java Test: H21TB-Run115003000450060007500SE +/- 28.93, N = 1007120

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU1TB-Run1246810SE +/- 0.01507, N = 1006.78058MIN: 5.81. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU1TB-Run1714212835SE +/- 0.14, N = 10031.19MIN: 12.071. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU1TB-Run10.5261.0521.5782.1042.63SE +/- 0.00635, N = 1002.33777MIN: 2.041. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU1TB-Run10.28190.56380.84571.12761.4095SE +/- 0.01966, N = 1001.25277MIN: 0.931. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 13 - Input: Bosphorus 4K1TB-Run120406080100SE +/- 0.34, N = 100101.531. (CXX) g++ options: -march=native -mno-avx

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 10 - Input: Bosphorus 4K1TB-Run120406080100SE +/- 0.48, N = 100110.721. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

NAMD

NAMD is a parallel molecular dynamics code designed for high-performance simulation of large biomolecular systems. NAMD was developed by the Theoretical and Computational Biophysics Group in the Beckman Institute for Advanced Science and Technology at the University of Illinois at Urbana-Champaign. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgdays/ns, Fewer Is BetterNAMD 2.14ATPase Simulation - 327,506 Atoms1TB-Run10.07350.1470.22050.2940.3675SE +/- 0.00162, N = 250.32648

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 8 - Input: Bosphorus 4K1TB-Run11020304050SE +/- 0.18, N = 3541.841. (CXX) g++ options: -march=native -mno-avx

PyBench

This test profile reports the total time of the different average timed test results from PyBench. PyBench reports average test times for different functions such as BuiltinFunctionCalls and NestedForLoops, with this total result providing a rough estimate as to Python's average performance on a given system. This test profile runs PyBench each time for 20 rounds. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyBench 2018-02-16Total For Average Test Times1TB-Run130060090012001500SE +/- 1.36, N = 201573

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU1TB-Run10.86091.72182.58273.44364.3045SE +/- 0.01989, N = 1003.82614MIN: 3.281. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU1TB-Run10.51431.02861.54292.05722.5715SE +/- 0.03967, N = 1002.28579MIN: 1.471. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU1TB-Run10.54481.08961.63442.17922.724SE +/- 0.00373, N = 1002.42126MIN: 2.161. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Algebraic Multi-Grid Benchmark

AMG is a parallel algebraic multigrid solver for linear systems arising from problems on unstructured grids. The driver provided with AMG builds linear systems for various 3-dimensional problems. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFigure Of Merit, More Is BetterAlgebraic Multi-Grid Benchmark 1.21TB-Run1400M800M1200M1600M2000MSE +/- 833796.71, N = 2019151138001. (CC) gcc options: -lparcsr_ls -lparcsr_mv -lseq_mv -lIJ_mv -lkrylov -lHYPRE_utilities -lm -fopenmp -pthread -m64 -lmpi

PHPBench

PHPBench is a benchmark suite for PHP. It performs a large number of simple tests in order to bench various aspects of the PHP interpreter. PHPBench can be used to compare hardware, operating systems, PHP versions, PHP accelerators and caches, compiler options, etc. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgScore, More Is BetterPHPBench 0.8.1PHP Benchmark Suite1TB-Run140K80K120K160K200KSE +/- 455.27, N = 6196055

John The Ripper

This is a benchmark of John The Ripper, which is a password cracker. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgReal C/S, More Is BetterJohn The Ripper 2023.03.14Test: Blowfish1TB-Run120K40K60K80K100KSE +/- 289.73, N = 201106111. (CC) gcc options: -m64 -lssl -lcrypto -fopenmp -lgmp -pthread -lm -lrt -lz -ldl -lcrypt -lbz2

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Semaphores1TB-Run14M8M12M16M20MSE +/- 7878.32, N = 2016591808.111. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Poll1TB-Run11.3M2.6M3.9M5.2M6.5MSE +/- 1356.27, N = 205837581.441. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Matrix Math1TB-Run170K140K210K280K350KSE +/- 86.84, N = 20332280.311. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Malloc1TB-Run120M40M60M80M100MSE +/- 74227.52, N = 2091315767.441. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: CPU Stress1TB-Run130K60K90K120K150KSE +/- 30.91, N = 20120051.271. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Crypto1TB-Run120K40K60K80K100KSE +/- 26.26, N = 2083163.041. (CC) gcc options: -std=gnu99 -O2 -lm -laio -lc -lcrypt -ldl -ljpeg -lrt -lz -pthread

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU1TB-Run11.09282.18563.27844.37125.464SE +/- 0.02263, N = 284.85699MIN: 4.441. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 4 - Input: Bosphorus 1080p1TB-Run1246810SE +/- 0.020, N = 206.1301. (CXX) g++ options: -march=native -mno-avx

Cython Benchmark

Cython provides a superset of Python that is geared to deliver C-like levels of performance. This test profile makes use of Cython's bundled benchmark tests and runs an N-Queens sample test as a simple benchmark to the system's Cython performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterCython Benchmark 0.29.21Test: N-Queens1TB-Run1714212835SE +/- 0.07, N = 2028.51

Xcompact3d Incompact3d

Xcompact3d Incompact3d is a Fortran-MPI based, finite difference high-performance code for solving the incompressible Navier-Stokes equation and as many as you need scalar transport equations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterXcompact3d Incompact3d 2021-03-11Input: input.i3d 193 Cells Per Direction1TB-Run148121620SE +/- 0.08, N = 3014.571. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -pthread -m64 -lmpi_usempi -lmpi_mpifh -lmpi

libjpeg-turbo tjbench

tjbench is a JPEG decompression/compression benchmark that is part of libjpeg-turbo, a JPEG image codec library optimized for SIMD instructions on modern CPU architectures. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMegapixels/sec, More Is Betterlibjpeg-turbo tjbench 2.1.0Test: Decompression Throughput1TB-Run14080120160200SE +/- 0.33, N = 20163.811. (CC) gcc options: -O3 -rdynamic

DaCapo Benchmark

This test runs the DaCapo Benchmarks written in Java and intended to test system/CPU performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgmsec, Fewer Is BetterDaCapo Benchmark 9.12-MR1Java Test: Eclipse1TB-Run14K8K12K16K20KSE +/- 78.57, N = 2020616

Dolfyn

Dolfyn is a Computational Fluid Dynamics (CFD) code of modern numerical simulation techniques. The Dolfyn test profile measures the execution time of the bundled computational fluid dynamics demos that are bundled with Dolfyn. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterDolfyn 0.527Computational Fluid Dynamics1TB-Run1510152025SE +/- 0.03, N = 2022.48

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 1 - Input: Bosphorus 1080p1TB-Run1714212835SE +/- 0.07, N = 2030.381. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

miniBUDE

MiniBUDE is a mini application for the the core computation of the Bristol University Docking Engine (BUDE). This test profile currently makes use of the OpenMP implementation of miniBUDE for CPU benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBillion Interactions/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM11TB-Run13691215SE +/- 0.03, N = 613.361. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

OpenBenchmarking.orgGFInst/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM11TB-Run170140210280350SE +/- 0.80, N = 6334.011. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU1TB-Run10.94411.88822.83233.77644.7205SE +/- 0.00586, N = 1004.19586MIN: 4.031. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU1TB-Run10.44330.88661.32991.77322.2165SE +/- 0.00379, N = 201.97024MIN: 1.461. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU1TB-Run10.90191.80382.70573.60764.5095SE +/- 0.00608, N = 204.00835MIN: 3.611. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

LULESH

LULESH is the Livermore Unstructured Lagrangian Explicit Shock Hydrodynamics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgz/s, More Is BetterLULESH 2.0.31TB-Run17K14K21K28K35KSE +/- 13.03, N = 2034264.231. (CXX) g++ options: -O3 -fopenmp -lm -pthread -m64 -lmpi_cxx -lmpi

OpenCV

This is a benchmark of the OpenCV (Computer Vision) library's built-in performance tests. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenCV 4.7Test: Object Detection1TB-Run110K20K30K40K50KSE +/- 467.86, N = 6458471. (CXX) g++ options: -fPIC -fsigned-char -pthread -fomit-frame-pointer -ffunction-sections -fdata-sections -msse -msse2 -msse3 -fvisibility=hidden -O3 -shared

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU1TB-Run1246810SE +/- 0.02799, N = 206.02812MIN: 5.321. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU1TB-Run10.1130.2260.3390.4520.565SE +/- 0.001016, N = 200.502251MIN: 0.361. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 12 - Input: Bosphorus 4K1TB-Run120406080100SE +/- 0.46, N = 2094.101. (CXX) g++ options: -march=native -mno-avx

Xcompact3d Incompact3d

Xcompact3d Incompact3d is a Fortran-MPI based, finite difference high-performance code for solving the incompressible Navier-Stokes equation and as many as you need scalar transport equations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterXcompact3d Incompact3d 2021-03-11Input: input.i3d 129 Cells Per Direction1TB-Run10.72951.4592.18852.9183.6475SE +/- 0.03923848, N = 303.242250361. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -pthread -m64 -lmpi_usempi -lmpi_mpifh -lmpi

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 7 - Input: Bosphorus 4K1TB-Run120406080100SE +/- 0.33, N = 2092.451. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 8 - Input: Bosphorus 1080p1TB-Run120406080100SE +/- 0.25, N = 2096.281. (CXX) g++ options: -march=native -mno-avx

DaCapo Benchmark

This test runs the DaCapo Benchmarks written in Java and intended to test system/CPU performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgmsec, Fewer Is BetterDaCapo Benchmark 9.12-MR1Java Test: Jython1TB-Run19001800270036004500SE +/- 18.95, N = 204151

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 7 - Input: Bosphorus 1080p1TB-Run150100150200250SE +/- 0.38, N = 20215.311. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 12 - Input: Bosphorus 1080p1TB-Run170140210280350SE +/- 1.04, N = 20312.881. (CXX) g++ options: -march=native -mno-avx

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU1TB-Run10.26590.53180.79771.06361.3295SE +/- 0.00289, N = 201.18163MIN: 1.091. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 10 - Input: Bosphorus 1080p1TB-Run180160240320400SE +/- 0.99, N = 20360.291. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU1TB-Run10.15050.3010.45150.6020.7525SE +/- 0.001146, N = 200.668686MIN: 0.621. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 13 - Input: Bosphorus 1080p1TB-Run170140210280350SE +/- 0.85, N = 20337.641. (CXX) g++ options: -march=native -mno-avx

LAMMPS Molecular Dynamics Simulator

LAMMPS is a classical molecular dynamics code, and an acronym for Large-scale Atomic/Molecular Massively Parallel Simulator. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgns/day, More Is BetterLAMMPS Molecular Dynamics Simulator 23Jun2022Model: Rhodopsin Protein1TB-Run1714212835SE +/- 0.12, N = 631.951. (CXX) g++ options: -O3 -pthread -ldl

Stress-NG

Test: IO_uring

1TB-Run1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Test: MEMFD

1TB-Run1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

OpenCV

This is a benchmark of the OpenCV (Computer Vision) library's built-in performance tests. Learn more via the OpenBenchmarking.org test page.

Test: Graph API

1TB-Run1: The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. E: opencv: line 4: ./opencv_perf_gapi: No such file or directory

128 Results Shown

oneDNN:
  Recurrent Neural Network Inference - u8s8f32 - CPU
  Recurrent Neural Network Inference - bf16bf16bf16 - CPU
Xcompact3d Incompact3d
OpenCV
Zstd Compression:
  8 - Decompression Speed
  8 - Compression Speed
  8, Long Mode - Decompression Speed
  8, Long Mode - Compression Speed
  12 - Decompression Speed
  12 - Compression Speed
John The Ripper:
  MD5
  HMAC-SHA512
OpenCV
CloverLeaf
LAMMPS Molecular Dynamics Simulator
OpenCV:
  Features 2D
  Core
Stress-NG:
  Atomic
  Pthread
  Mutex
  Context Switching
  Zlib
John The Ripper
Stress-NG
John The Ripper
Stress-NG:
  System V Message Passing
  Function Call
  SENDFILE
  Glibc C String Functions
  CPU Cache
  Socket Activity
  MMAP
  Forking
  x86_64 RdRand
Zstd Compression:
  3, Long Mode - Decompression Speed
  3, Long Mode - Compression Speed
oneDNN
Stress-NG
NCNN:
  CPU - FastestDet
  CPU - vision_transformer
  CPU - regnety_400m
  CPU - squeezenet_ssd
  CPU - yolov4-tiny
  CPU - resnet50
  CPU - alexnet
  CPU - resnet18
  CPU - vgg16
  CPU - googlenet
  CPU - blazeface
  CPU - efficientnet-b0
  CPU - mnasnet
  CPU - shufflenet-v2
  CPU-v3-v3 - mobilenet-v3
  CPU-v2-v2 - mobilenet-v2
  CPU - mobilenet
Stress-NG:
  Memory Copying
  Vector Math
  Hash
  Futex
OpenCV
oneDNN:
  Deconvolution Batch shapes_1d - f32 - CPU
  Deconvolution Batch shapes_1d - u8s8f32 - CPU
miniBUDE:
  OpenMP - BM2:
    Billion Interactions/s
    GFInst/s
Numpy Benchmark
Zstd Compression:
  19, Long Mode - Decompression Speed
  19, Long Mode - Compression Speed
oneDNN:
  Recurrent Neural Network Training - f32 - CPU
  Recurrent Neural Network Training - u8s8f32 - CPU
  Recurrent Neural Network Inference - f32 - CPU
  IP Shapes 1D - u8s8f32 - CPU
SVT-AV1
SVT-HEVC
Zstd Compression:
  19 - Decompression Speed
  19 - Compression Speed
  3 - Decompression Speed
  3 - Compression Speed
DaCapo Benchmark
oneDNN:
  Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU
  IP Shapes 3D - f32 - CPU
  IP Shapes 3D - bf16bf16bf16 - CPU
  IP Shapes 3D - u8s8f32 - CPU
SVT-AV1
SVT-HEVC
NAMD
SVT-AV1
PyBench
oneDNN:
  Convolution Batch Shapes Auto - f32 - CPU
  Convolution Batch Shapes Auto - u8s8f32 - CPU
  Convolution Batch Shapes Auto - bf16bf16bf16 - CPU
Algebraic Multi-Grid Benchmark
PHPBench
John The Ripper
Stress-NG:
  Semaphores
  Poll
  Matrix Math
  Malloc
  CPU Stress
  Crypto
oneDNN
SVT-AV1
Cython Benchmark
Xcompact3d Incompact3d
libjpeg-turbo tjbench
DaCapo Benchmark
Dolfyn
SVT-HEVC
miniBUDE:
  OpenMP - BM1:
    Billion Interactions/s
    GFInst/s
oneDNN:
  Deconvolution Batch shapes_3d - bf16bf16bf16 - CPU
  IP Shapes 1D - f32 - CPU
  IP Shapes 1D - bf16bf16bf16 - CPU
LULESH
OpenCV
oneDNN:
  Matrix Multiply Batch Shapes Transformer - bf16bf16bf16 - CPU
  Matrix Multiply Batch Shapes Transformer - f32 - CPU
SVT-AV1
Xcompact3d Incompact3d
SVT-HEVC
SVT-AV1
DaCapo Benchmark
SVT-HEVC
SVT-AV1
oneDNN
SVT-HEVC
oneDNN
SVT-AV1
LAMMPS Molecular Dynamics Simulator