AMD Ryzen 9 5950X 16-Core testing with a ASUS ROG CROSSHAIR VIII HERO (WI-FI) (3202 BIOS) and AMD Radeon RX 5600 OEM/5600 XT / 5700/5700 8GB on Ubuntu 20.10 via the Phoronix Test Suite.
3003 Processor: AMD Ryzen 9 5950X 16-Core @ 3.40GHz (16 Cores / 32 Threads), Motherboard: ASUS ROG CROSSHAIR VIII HERO (WI-FI) (3003 BIOS), Chipset: AMD Starship/Matisse, Memory: 32GB, Disk: 2000GB Corsair Force MP600 + 2000GB, Graphics: AMD Radeon RX 5600 OEM/5600 XT / 5700/5700 8GB (2100/875MHz), Audio: AMD Navi 10 HDMI Audio, Monitor: ASUS MG28U, Network: Realtek RTL8125 2.5GbE + Intel I211 + Intel Wi-Fi 6 AX200
OS: Ubuntu 20.10, Kernel: 5.11.0-051100rc2daily20210108-generic (x86_64) 20210107, Desktop: GNOME Shell 3.38.1, Display Server: X Server 1.20.9, Display Driver: amdgpu 19.1.0, OpenGL: 4.6 Mesa 21.0.0-devel (git-f01bca8 2021-01-08 groovy-oibaf-ppa) (LLVM 11.0.1), Vulkan: 1.2.164, Compiler: GCC 10.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0xa201009Graphics Notes: GLAMORPython Notes: Python 3.8.6Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional IBRS_FW STIBP: always-on RSB filling + srbds: Not affected + tsx_async_abort: Not affected
3202 Changed Motherboard to ASUS ROG CROSSHAIR VIII HERO (WI-FI) (3202 BIOS) .
Disk Notes: NONE / errors=remount-ro,relatime,rw / Block Size: 4096
yquake2 This is a test of Yamagi Quake II. Yamagi Quake II is an enhanced client for id Software's Quake II with focus on offline and coop gameplay. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better yquake2 7.45 Renderer: OpenGL 3.x - Resolution: 3840 x 2160 3003 3202 200 400 600 800 1000 SE +/- 1.03, N = 3 SE +/- 1.35, N = 3 979.3 986.5 1. (CC) gcc options: -lm -ldl -rdynamic -shared -lSDL2 -O2 -pipe -fomit-frame-pointer -std=gnu99 -fno-strict-aliasing -fwrapv -fvisibility=hidden -MMD -mfpmath=sse -fPIC
BRL-CAD BRL-CAD 7.28.0 is a cross-platform, open-source solid modeling system with built-in benchmark mode. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org VGR Performance Metric, More Is Better BRL-CAD 7.30.8 VGR Performance Metric 3202 3003 60K 120K 180K 240K 300K 262742 265419 1. (CXX) g++ options: -std=c++11 -pipe -fno-strict-aliasing -fno-common -fexceptions -ftemplate-depth-128 -m64 -ggdb3 -O3 -fipa-pta -fstrength-reduce -finline-functions -flto -pedantic -rdynamic -lSM -lICE -lXi -lGLU -lGL -lGLdispatch -lX11 -lXext -lXrender -lpthread -ldl -luuid -lm
Opus Codec Encoding Opus is an open audio codec. Opus is a lossy audio compression format designed primarily for interactive real-time applications over the Internet. This test uses Opus-Tools and measures the time required to encode a WAV file to Opus. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Opus Codec Encoding 1.3.1 WAV To Opus Encode 3202 3003 2 4 6 8 10 SE +/- 0.042, N = 5 SE +/- 0.032, N = 5 6.150 6.126 1. (CXX) g++ options: -fvisibility=hidden -logg -lm
ASTC Encoder ASTC Encoder (astcenc) is for the Adaptive Scalable Texture Compression (ASTC) format commonly used with OpenGL, OpenGL ES, and Vulkan graphics APIs. This test profile does a coding test of both compression/decompression. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better ASTC Encoder 2.0 Preset: Fast 3003 3202 0.9248 1.8496 2.7744 3.6992 4.624 SE +/- 0.01, N = 3 SE +/- 0.03, N = 3 4.11 4.10 1. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better ASTC Encoder 2.0 Preset: Medium 3202 3003 1.2218 2.4436 3.6654 4.8872 6.109 SE +/- 0.02, N = 3 SE +/- 0.04, N = 3 5.43 5.35 1. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better ASTC Encoder 2.0 Preset: Thorough 3202 3003 3 6 9 12 15 SE +/- 0.02, N = 3 SE +/- 0.03, N = 3 12.67 12.47 1. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better ASTC Encoder 2.0 Preset: Exhaustive 3202 3003 20 40 60 80 100 SE +/- 0.15, N = 3 SE +/- 0.10, N = 3 100.90 99.00 1. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
Etcpak Etcpack is the self-proclaimed "fastest ETC compressor on the planet" with focused on providing open-source, very fast ETC and S3 texture compression support. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Mpx/s, More Is Better Etcpak 0.7 Configuration: DXT1 3003 3202 300 600 900 1200 1500 SE +/- 2.32, N = 3 SE +/- 21.94, N = 3 1532.41 1562.89 1. (CXX) g++ options: -O3 -march=native -std=c++11 -lpthread
OpenBenchmarking.org Mpx/s, More Is Better Etcpak 0.7 Configuration: ETC1 3003 3202 80 160 240 320 400 SE +/- 1.98, N = 3 SE +/- 1.60, N = 3 382.93 387.14 1. (CXX) g++ options: -O3 -march=native -std=c++11 -lpthread
OpenBenchmarking.org Mpx/s, More Is Better Etcpak 0.7 Configuration: ETC2 3003 3202 50 100 150 200 250 SE +/- 0.59, N = 3 SE +/- 1.66, N = 3 236.51 245.00 1. (CXX) g++ options: -O3 -march=native -std=c++11 -lpthread
OpenBenchmarking.org Mpx/s, More Is Better Etcpak 0.7 Configuration: ETC1 + Dithering 3003 3202 80 160 240 320 400 SE +/- 0.36, N = 3 SE +/- 3.87, N = 3 349.70 355.44 1. (CXX) g++ options: -O3 -march=native -std=c++11 -lpthread
WebP Image Encode This is a test of Google's libwebp with the cwebp image encode utility and using a sample 6000x4000 pixel JPEG image as the input. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Encode Time - Seconds, Fewer Is Better WebP Image Encode 1.1 Encode Settings: Quality 100 3003 3202 0.3917 0.7834 1.1751 1.5668 1.9585 SE +/- 0.001, N = 3 SE +/- 0.004, N = 3 1.741 1.729 1. (CC) gcc options: -fvisibility=hidden -O2 -pthread -lm -ljpeg -lpng16 -ltiff
OpenBenchmarking.org Encode Time - Seconds, Fewer Is Better WebP Image Encode 1.1 Encode Settings: Quality 100, Lossless 3003 3202 3 6 9 12 15 SE +/- 0.01, N = 3 SE +/- 0.05, N = 3 13.04 12.87 1. (CC) gcc options: -fvisibility=hidden -O2 -pthread -lm -ljpeg -lpng16 -ltiff
OpenBenchmarking.org Encode Time - Seconds, Fewer Is Better WebP Image Encode 1.1 Encode Settings: Quality 100, Highest Compression 3003 3202 1.2296 2.4592 3.6888 4.9184 6.148 SE +/- 0.059, N = 3 SE +/- 0.062, N = 3 5.465 5.360 1. (CC) gcc options: -fvisibility=hidden -O2 -pthread -lm -ljpeg -lpng16 -ltiff
OpenBenchmarking.org Encode Time - Seconds, Fewer Is Better WebP Image Encode 1.1 Encode Settings: Quality 100, Lossless, Highest Compression 3202 3003 6 12 18 24 30 SE +/- 0.04, N = 3 SE +/- 0.05, N = 3 27.52 27.23 1. (CC) gcc options: -fvisibility=hidden -O2 -pthread -lm -ljpeg -lpng16 -ltiff
Google SynthMark SynthMark is a cross platform tool for benchmarking CPU performance under a variety of real-time audio workloads. It uses a polyphonic synthesizer model to provide standardized tests for latency, jitter and computational throughput. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Voices, More Is Better Google SynthMark 20201109 Test: VoiceMark_100 3202 3003 200 400 600 800 1000 SE +/- 4.96, N = 3 SE +/- 3.00, N = 3 957.95 958.66 1. (CXX) g++ options: -lm -lpthread -std=c++11 -Ofast
Xonotic This is a benchmark of Xonotic, which is a fork of the DarkPlaces-based Nexuiz game. Development began in March of 2010 on the Xonotic game. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better Xonotic 0.8.2 Resolution: 3840 x 2160 - Effects Quality: Ultimate 3003 3202 60 120 180 240 300 SE +/- 4.83, N = 15 SE +/- 3.79, N = 3 257.46 288.97 MIN: 55 / MAX: 623 MIN: 60 / MAX: 571
RELION RELION - REgularised LIkelihood OptimisatioN - is a stand-alone computer program for Maximum A Posteriori refinement of (multiple) 3D reconstructions or 2D class averages in cryo-electron microscopy (cryo-EM). It is developed in the research group of Sjors Scheres at the MRC Laboratory of Molecular Biology. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better RELION 3.1.1 Test: Basic - Device: CPU 3003 3202 400 800 1200 1600 2000 SE +/- 3.84, N = 3 SE +/- 6.24, N = 3 1892.61 1875.48 1. (CXX) g++ options: -fopenmp -std=c++0x -O3 -rdynamic -ldl -ltiff -lfftw3f -lfftw3 -lpng -pthread -lmpi_cxx -lmpi
HPC Challenge HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org GFLOPS, More Is Better HPC Challenge 1.5.0 Test / Class: G-HPL 3003 3202 12 24 36 48 60 SE +/- 0.08, N = 3 SE +/- 0.05, N = 3 53.17 53.17 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. OpenBLAS + Open MPI 4.0.3
OpenBenchmarking.org GFLOPS, More Is Better HPC Challenge 1.5.0 Test / Class: G-Ffte 3003 3202 2 4 6 8 10 SE +/- 0.10692, N = 3 SE +/- 0.02613, N = 3 6.23452 6.55050 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. OpenBLAS + Open MPI 4.0.3
OpenBenchmarking.org GFLOPS, More Is Better HPC Challenge 1.5.0 Test / Class: EP-DGEMM 3202 3003 4 8 12 16 20 SE +/- 0.17, N = 3 SE +/- 0.11, N = 3 16.57 16.86 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. OpenBLAS + Open MPI 4.0.3
OpenBenchmarking.org GB/s, More Is Better HPC Challenge 1.5.0 Test / Class: G-Ptrans 3003 3202 0.5576 1.1152 1.6728 2.2304 2.788 SE +/- 0.00427, N = 3 SE +/- 0.01009, N = 3 2.45326 2.47841 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. OpenBLAS + Open MPI 4.0.3
OpenBenchmarking.org GB/s, More Is Better HPC Challenge 1.5.0 Test / Class: EP-STREAM Triad 3003 3202 0.3329 0.6658 0.9987 1.3316 1.6645 SE +/- 0.00070, N = 3 SE +/- 0.00089, N = 3 1.42468 1.47956 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. OpenBLAS + Open MPI 4.0.3
OpenBenchmarking.org GUP/s, More Is Better HPC Challenge 1.5.0 Test / Class: G-Random Access 3202 3003 0.0112 0.0224 0.0336 0.0448 0.056 SE +/- 0.00017, N = 3 SE +/- 0.00046, N = 3 0.04973 0.04998 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. OpenBLAS + Open MPI 4.0.3
OpenBenchmarking.org usecs, Fewer Is Better HPC Challenge 1.5.0 Test / Class: Random Ring Latency 3202 3003 0.1107 0.2214 0.3321 0.4428 0.5535 SE +/- 0.00234, N = 3 SE +/- 0.00404, N = 3 0.49204 0.48933 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. OpenBLAS + Open MPI 4.0.3
OpenBenchmarking.org GB/s, More Is Better HPC Challenge 1.5.0 Test / Class: Random Ring Bandwidth 3003 3202 0.4556 0.9112 1.3668 1.8224 2.278 SE +/- 0.02644, N = 3 SE +/- 0.02919, N = 3 1.92562 2.02474 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. OpenBLAS + Open MPI 4.0.3
OpenBenchmarking.org MB/s, More Is Better HPC Challenge 1.5.0 Test / Class: Max Ping Pong Bandwidth 3202 3003 7K 14K 21K 28K 35K SE +/- 130.59, N = 3 SE +/- 146.47, N = 3 34166.84 34205.28 1. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops 2. OpenBLAS + Open MPI 4.0.3
CloverLeaf CloverLeaf is a Lagrangian-Eulerian hydrodynamics benchmark. This test profile currently makes use of CloverLeaf's OpenMP version and benchmarked with the clover_bm.in input file (Problem 5). Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better CloverLeaf Lagrangian-Eulerian Hydrodynamics 3003 3202 30 60 90 120 150 SE +/- 0.03, N = 3 SE +/- 0.24, N = 3 136.80 134.74 1. (F9X) gfortran options: -O3 -march=native -funroll-loops -fopenmp
Dolfyn Dolfyn is a Computational Fluid Dynamics (CFD) code of modern numerical simulation techniques. The Dolfyn test profile measures the execution time of the bundled computational fluid dynamics demos that are bundled with Dolfyn. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Dolfyn 0.527 Computational Fluid Dynamics 3202 3003 3 6 9 12 15 SE +/- 0.02, N = 3 SE +/- 0.10, N = 3 13.29 12.90
DeepSpeech Mozilla DeepSpeech is a speech-to-text engine powered by TensorFlow for machine learning and derived from Baidu's Deep Speech research paper. This test profile times the speech-to-text process for a roughly three minute audio recording. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better DeepSpeech 0.6 Acceleration: CPU 3202 3003 16 32 48 64 80 SE +/- 0.04, N = 3 SE +/- 0.10, N = 3 70.53 70.48
RNNoise RNNoise is a recurrent neural network for audio noise reduction developed by Mozilla and Xiph.Org. This test profile is a single-threaded test measuring the time to denoise a sample 26 minute long 16-bit RAW audio file using this recurrent neural network noise suppression library. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better RNNoise 2020-06-28 3202 3003 4 8 12 16 20 SE +/- 0.18, N = 3 SE +/- 0.07, N = 3 15.23 15.18 1. (CC) gcc options: -O2 -pedantic -fvisibility=hidden
Mobile Neural Network MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.1.1 Model: SqueezeNetV1.0 3003 3202 1.1772 2.3544 3.5316 4.7088 5.886 SE +/- 0.062, N = 3 SE +/- 0.022, N = 3 5.232 5.153 MIN: 5.02 / MAX: 8.4 MIN: 5.02 / MAX: 14.32 1. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.1.1 Model: resnet-v2-50 3003 3202 6 12 18 24 30 SE +/- 0.25, N = 3 SE +/- 0.19, N = 3 24.04 23.63 MIN: 21.95 / MAX: 33.08 MIN: 22.31 / MAX: 33.12 1. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.1.1 Model: MobileNetV2_224 3003 3202 0.7655 1.531 2.2965 3.062 3.8275 SE +/- 0.047, N = 3 SE +/- 0.039, N = 3 3.402 3.325 MIN: 3.23 / MAX: 5.81 MIN: 3.16 / MAX: 4.02 1. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.1.1 Model: mobilenet-v1-1.0 3003 3202 0.5627 1.1254 1.6881 2.2508 2.8135 SE +/- 0.094, N = 3 SE +/- 0.027, N = 3 2.501 2.481 MIN: 2.32 / MAX: 4.53 MIN: 2.42 / MAX: 2.71 1. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.1.1 Model: inception-v3 3003 3202 7 14 21 28 35 SE +/- 0.18, N = 3 SE +/- 0.26, N = 3 31.01 30.33 MIN: 29.92 / MAX: 38.55 MIN: 29.28 / MAX: 56.48 1. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
ONNX Runtime ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.6 Model: yolov4 - Device: OpenMP CPU 3202 3003 90 180 270 360 450 SE +/- 3.35, N = 3 SE +/- 1.17, N = 3 428 438 1. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.6 Model: bertsquad-10 - Device: OpenMP CPU 3202 3003 150 300 450 600 750 SE +/- 11.02, N = 12 SE +/- 1.32, N = 3 665 705 1. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt
IOR IOR is a parallel I/O storage benchmark making use of MPI with a particular focus on HPC (High Performance Computing) systems. IOR is developed at the Lawrence Livermore National Laboratory (LLNL). Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MB/s, More Is Better IOR 3.3.0 Block Size: 4MB - Disk Target: Default Test Directory 3202 3003 300 600 900 1200 1500 SE +/- 11.11, N = 11 SE +/- 5.96, N = 3 1482.95 1580.25 MIN: 955.9 / MAX: 2484.92 MIN: 1161.4 / MAX: 2244.12 1. (CC) gcc options: -O2 -lm -pthread -lmpi
OpenBenchmarking.org MB/s, More Is Better IOR 3.3.0 Block Size: 256MB - Disk Target: Default Test Directory 3202 3003 300 600 900 1200 1500 SE +/- 14.35, N = 9 SE +/- 19.35, N = 9 1251.65 1339.91 MIN: 354.68 / MAX: 2107.13 MIN: 282.98 / MAX: 2236.64 1. (CC) gcc options: -O2 -lm -pthread -lmpi
ONNX Runtime ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.6 Model: fcn-resnet101-11 - Device: OpenMP CPU 3202 3003 20 40 60 80 100 SE +/- 0.33, N = 3 98 99 1. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.6 Model: shufflenet-v2-10 - Device: OpenMP CPU 3202 3003 3K 6K 9K 12K 15K SE +/- 68.50, N = 3 SE +/- 20.67, N = 3 15863 16013 1. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.6 Model: super-resolution-10 - Device: OpenMP CPU 3003 3202 1600 3200 4800 6400 8000 SE +/- 202.78, N = 12 SE +/- 175.44, N = 12 7056 7324 1. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt
TNN TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better TNN 0.2.3 Target: CPU - Model: MobileNet v2 3202 3003 50 100 150 200 250 SE +/- 0.67, N = 3 SE +/- 0.32, N = 3 220.33 218.29 MIN: 216.95 / MAX: 261.2 MIN: 208.52 / MAX: 289.05 1. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -O3 -rdynamic -ldl
OpenBenchmarking.org ms, Fewer Is Better TNN 0.2.3 Target: CPU - Model: SqueezeNet v1.1 3202 3003 50 100 150 200 250 SE +/- 0.34, N = 3 SE +/- 0.43, N = 3 212.62 211.46 MIN: 211.89 / MAX: 213.24 MIN: 210.71 / MAX: 212.37 1. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -O3 -rdynamic -ldl
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: mobilenet 3003 3202 3 6 9 12 15 SE +/- 0.15, N = 3 SE +/- 0.01, N = 3 12.04 11.99 MIN: 11.72 / MAX: 14.17 MIN: 11.79 / MAX: 12.19 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU-v2-v2 - Model: mobilenet-v2 3202 3003 1.0035 2.007 3.0105 4.014 5.0175 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 4.46 4.38 MIN: 4.31 / MAX: 5.3 MIN: 4.24 / MAX: 7.32 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU-v3-v3 - Model: mobilenet-v3 3202 3003 0.9338 1.8676 2.8014 3.7352 4.669 SE +/- 0.00, N = 3 SE +/- 0.01, N = 3 4.15 4.07 MIN: 4.11 / MAX: 5.03 MIN: 4.03 / MAX: 5.2 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: shufflenet-v2 3202 3003 0.9945 1.989 2.9835 3.978 4.9725 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 4.42 4.38 MIN: 4.35 / MAX: 5.28 MIN: 4.33 / MAX: 4.89 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: mnasnet 3202 3003 0.891 1.782 2.673 3.564 4.455 SE +/- 0.01, N = 3 SE +/- 0.00, N = 3 3.96 3.90 MIN: 3.84 / MAX: 5.26 MIN: 3.78 / MAX: 4.83 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: efficientnet-b0 3202 3003 1.2083 2.4166 3.6249 4.8332 6.0415 SE +/- 0.01, N = 3 SE +/- 0.00, N = 3 5.37 5.29 MIN: 5.31 / MAX: 7.18 MIN: 5.24 / MAX: 5.79 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: blazeface 3202 3003 0.4095 0.819 1.2285 1.638 2.0475 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 1.82 1.80 MIN: 1.79 / MAX: 2.65 MIN: 1.78 / MAX: 2.28 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: googlenet 3202 3003 3 6 9 12 15 SE +/- 0.02, N = 3 SE +/- 0.00, N = 3 13.00 12.99 MIN: 12.64 / MAX: 20.99 MIN: 12.62 / MAX: 13.57 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: vgg16 3003 3202 14 28 42 56 70 SE +/- 0.13, N = 3 SE +/- 0.09, N = 3 60.60 59.90 MIN: 59.43 / MAX: 62.32 MIN: 58.71 / MAX: 61.76 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: resnet18 3202 3003 4 8 12 16 20 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 14.57 14.51 MIN: 14.45 / MAX: 23.09 MIN: 14.39 / MAX: 15.06 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: alexnet 3202 3003 3 6 9 12 15 SE +/- 0.24, N = 3 SE +/- 0.01, N = 3 11.26 11.04 MIN: 10.91 / MAX: 19.04 MIN: 10.95 / MAX: 11.89 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: resnet50 3003 3202 6 12 18 24 30 SE +/- 0.13, N = 3 SE +/- 0.17, N = 3 25.05 25.00 MIN: 24.65 / MAX: 26.23 MIN: 24.58 / MAX: 26.31 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: yolov4-tiny 3202 3003 5 10 15 20 25 SE +/- 0.08, N = 3 SE +/- 0.11, N = 3 21.29 21.14 MIN: 20.98 / MAX: 21.78 MIN: 20.72 / MAX: 29.47 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: squeezenet_ssd 3003 3202 4 8 12 16 20 SE +/- 0.03, N = 3 SE +/- 0.04, N = 3 14.64 14.60 MIN: 14.31 / MAX: 15.31 MIN: 14.21 / MAX: 15.07 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: regnety_400m 3202 3003 4 8 12 16 20 SE +/- 0.13, N = 3 SE +/- 0.04, N = 3 17.95 17.67 MIN: 17.66 / MAX: 19.3 MIN: 17.47 / MAX: 18.02 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NAMD NAMD is a parallel molecular dynamics code designed for high-performance simulation of large biomolecular systems. NAMD was developed by the Theoretical and Computational Biophysics Group in the Beckman Institute for Advanced Science and Technology at the University of Illinois at Urbana-Champaign. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org days/ns, Fewer Is Better NAMD 2.14 ATPase Simulation - 327,506 Atoms 3202 3003 0.2447 0.4894 0.7341 0.9788 1.2235 SE +/- 0.00500, N = 3 SE +/- 0.00258, N = 3 1.08736 1.07519
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU 3003 3202 0.8955 1.791 2.6865 3.582 4.4775 SE +/- 0.01644, N = 3 SE +/- 0.00985, N = 3 3.98022 3.97992 MIN: 3.72 MIN: 3.76 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU 3202 3003 3 6 9 12 15 SE +/- 0.00745, N = 3 SE +/- 0.01431, N = 3 9.51868 9.48452 MIN: 9.44 MIN: 9.38 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU 3202 3003 0.1873 0.3746 0.5619 0.7492 0.9365 SE +/- 0.001536, N = 3 SE +/- 0.001756, N = 3 0.832251 0.816147 MIN: 0.75 MIN: 0.74 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU 3202 3003 0.109 0.218 0.327 0.436 0.545 SE +/- 0.003401, N = 15 SE +/- 0.002170, N = 3 0.484368 0.477138 MIN: 0.43 MIN: 0.44 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU 3202 3003 4 8 12 16 20 SE +/- 0.02, N = 3 SE +/- 0.01, N = 3 17.29 17.23 MIN: 16.66 MIN: 16.81 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU 3202 3003 0.5472 1.0944 1.6416 2.1888 2.736 SE +/- 0.00987, N = 3 SE +/- 0.00251, N = 3 2.43213 2.38690 MIN: 2.3 MIN: 2.28 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU 3202 3003 0.7947 1.5894 2.3841 3.1788 3.9735 SE +/- 0.00421, N = 3 SE +/- 0.00613, N = 3 3.53182 3.50445 MIN: 3.41 MIN: 3.38 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU 3202 3003 5 10 15 20 25 SE +/- 0.04, N = 3 SE +/- 0.02, N = 3 19.26 19.13 MIN: 18.8 MIN: 18.77 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU 3202 3003 0.6737 1.3474 2.0211 2.6948 3.3685 SE +/- 0.00911, N = 3 SE +/- 0.00671, N = 3 2.99402 2.95002 MIN: 2.83 MIN: 2.81 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU 3202 3003 0.3596 0.7192 1.0788 1.4384 1.798 SE +/- 0.00175, N = 3 SE +/- 0.00153, N = 3 1.59818 1.53431 MIN: 1.48 MIN: 1.41 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU 3003 3202 600 1200 1800 2400 3000 SE +/- 13.63, N = 3 SE +/- 2.89, N = 3 2753.77 2749.23 MIN: 2727.8 MIN: 2735.56 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU 3202 3003 400 800 1200 1600 2000 SE +/- 9.60, N = 3 SE +/- 15.33, N = 3 1823.19 1809.76 MIN: 1798.09 MIN: 1773.21 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU 3202 3003 600 1200 1800 2400 3000 SE +/- 9.00, N = 3 SE +/- 11.23, N = 3 2763.90 2752.88 MIN: 2736.15 MIN: 2722.06 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU 3202 3003 400 800 1200 1600 2000 SE +/- 17.01, N = 3 SE +/- 8.25, N = 3 1789.23 1783.66 MIN: 1761.64 MIN: 1765.38 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU 3202 3003 0.1433 0.2866 0.4299 0.5732 0.7165 SE +/- 0.000300, N = 3 SE +/- 0.000749, N = 3 0.636735 0.625862 MIN: 0.6 MIN: 0.6 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU 3003 3202 600 1200 1800 2400 3000 SE +/- 9.46, N = 3 SE +/- 4.41, N = 3 2769.09 2761.96 MIN: 2739.1 MIN: 2740.57 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU 3003 3202 400 800 1200 1600 2000 SE +/- 20.56, N = 3 SE +/- 19.66, N = 4 1818.11 1795.37 MIN: 1764.89 MIN: 1755.37 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU 3202 3003 0.3348 0.6696 1.0044 1.3392 1.674 SE +/- 0.00214, N = 3 SE +/- 0.00059, N = 3 1.48786 1.45970 MIN: 1.39 MIN: 1.39 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
Algebraic Multi-Grid Benchmark AMG is a parallel algebraic multigrid solver for linear systems arising from problems on unstructured grids. The driver provided with AMG builds linear systems for various 3-dimensional problems. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Figure Of Merit, More Is Better Algebraic Multi-Grid Benchmark 1.2 3003 3202 50M 100M 150M 200M 250M SE +/- 2159740.57, N = 3 SE +/- 569155.55, N = 3 207261167 210165200 1. (CC) gcc options: -lparcsr_ls -lparcsr_mv -lseq_mv -lIJ_mv -lkrylov -lHYPRE_utilities -lm -fopenmp -pthread -lmpi
Kripke Kripke is a simple, scalable, 3D Sn deterministic particle transport code. Its primary purpose is to research how data layout, programming paradigms and architectures effect the implementation and performance of Sn transport. Kripke is developed by LLNL. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Throughput FoM, More Is Better Kripke 1.2.4 3003 3202 16M 32M 48M 64M 80M SE +/- 1023469.37, N = 3 SE +/- 319563.42, N = 3 72385553 72580800 1. (CXX) g++ options: -O3 -fopenmp
OpenFOAM OpenFOAM is the leading free, open source software for computational fluid dynamics (CFD). Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better OpenFOAM 8 Input: Motorbike 30M 3003 3202 20 40 60 80 100 SE +/- 0.09, N = 3 SE +/- 0.15, N = 3 104.48 97.87 1. (CXX) g++ options: -std=c++11 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm
OpenBenchmarking.org Seconds, Fewer Is Better OpenFOAM 8 Input: Motorbike 60M 3003 3202 300 600 900 1200 1500 SE +/- 0.57, N = 3 SE +/- 0.41, N = 3 1391.39 1380.20 1. (CXX) g++ options: -std=c++11 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm
QMCPACK QMCPACK is a modern high-performance open-source Quantum Monte Carlo (QMC) simulation code making use of MPI for this benchmark of the H20 example code. QMCPACK is an open-source production level many-body ab initio Quantum Monte Carlo code for computing the electronic structure of atoms, molecules, and solids. QMCPACK is supported by the U.S. Department of Energy. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Total Execution Time - Seconds, Fewer Is Better QMCPACK 3.10 Input: simple-H2O 3202 3003 5 10 15 20 25 SE +/- 0.14, N = 3 SE +/- 0.21, N = 7 22.37 22.31 1. (CXX) g++ options: -fopenmp -finline-limit=1000 -fstrict-aliasing -funroll-all-loops -march=native -O3 -fomit-frame-pointer -ffast-math -pthread -lm
Quantum ESPRESSO Quantum ESPRESSO is an integrated suite of Open-Source computer codes for electronic-structure calculations and materials modeling at the nanoscale. It is based on density-functional theory, plane waves, and pseudopotentials. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Quantum ESPRESSO 6.7 Input: AUSURF112 3202 3003 300 600 900 1200 1500 SE +/- 3.58, N = 3 SE +/- 0.92, N = 3 1221.19 1199.39 1. (F9X) gfortran options: -lopenblas -lFoX_dom -lFoX_sax -lFoX_wxml -lFoX_common -lFoX_utils -lFoX_fsys -lfftw3 -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz
dav1d Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org FPS, More Is Better dav1d 0.8.1 Video Input: Chimera 1080p 3003 3202 130 260 390 520 650 SE +/- 0.69, N = 3 SE +/- 0.77, N = 3 590.32 592.56 MIN: 447.67 / MAX: 749.27 MIN: 447.8 / MAX: 754.79 1. (CC) gcc options: -pthread
OpenBenchmarking.org FPS, More Is Better dav1d 0.8.1 Video Input: Summer Nature 4K 3003 3202 50 100 150 200 250 SE +/- 0.49, N = 3 SE +/- 0.36, N = 3 224.40 228.62 MIN: 172.58 / MAX: 234.67 MIN: 172.54 / MAX: 238.82 1. (CC) gcc options: -pthread
OpenBenchmarking.org FPS, More Is Better dav1d 0.8.1 Video Input: Summer Nature 1080p 3003 3202 120 240 360 480 600 SE +/- 4.29, N = 3 SE +/- 1.59, N = 3 534.95 535.52 MIN: 432.57 / MAX: 611.74 MIN: 453.14 / MAX: 589.94 1. (CC) gcc options: -pthread
x265 This is a simple test of the x265 encoder run on the CPU with 1080p and 4K options for H.265 video encode performance with x265. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better x265 3.4 Video Input: Bosphorus 4K 3202 3003 6 12 18 24 30 SE +/- 0.26, N = 4 SE +/- 0.35, N = 3 24.18 24.57 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
OpenBenchmarking.org Frames Per Second, More Is Better x265 3.4 Video Input: Bosphorus 1080p 3202 3003 11 22 33 44 55 SE +/- 0.22, N = 3 SE +/- 0.08, N = 3 47.62 47.87 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
OpenBenchmarking.org Frames Per Second, More Is Better rav1e 0.4 Speed: 10 3003 3202 0.7754 1.5508 2.3262 3.1016 3.877 SE +/- 0.037, N = 3 SE +/- 0.037, N = 15 3.362 3.446
PHPBench PHPBench is a benchmark suite for PHP. It performs a large number of simple tests in order to bench various aspects of the PHP interpreter. PHPBench can be used to compare hardware, operating systems, PHP versions, PHP accelerators and caches, compiler options, etc. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Score, More Is Better PHPBench 0.8.1 PHP Benchmark Suite 3003 3202 200K 400K 600K 800K 1000K SE +/- 8175.01, N = 3 SE +/- 7522.35, N = 3 831939 834019
IOR IOR is a parallel I/O storage benchmark making use of MPI with a particular focus on HPC (High Performance Computing) systems. IOR is developed at the Lawrence Livermore National Laboratory (LLNL). Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MB/s, More Is Better IOR 3.3.0 Block Size: 2MB - Disk Target: Default Test Directory 3202 3003 300 600 900 1200 1500 SE +/- 14.36, N = 3 SE +/- 2.40, N = 3 1388.54 1540.02 MIN: 890.97 / MAX: 2113.75 MIN: 1034.78 / MAX: 2149.91 1. (CC) gcc options: -O2 -lm -pthread -lmpi
dav1d Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org FPS, More Is Better dav1d 0.8.1 Video Input: Chimera 1080p 10-bit 3003 3202 20 40 60 80 100 SE +/- 0.06, N = 3 SE +/- 0.05, N = 3 96.35 96.66 MIN: 61.49 / MAX: 217.11 MIN: 61.56 / MAX: 221.22 1. (CC) gcc options: -pthread
IOR IOR is a parallel I/O storage benchmark making use of MPI with a particular focus on HPC (High Performance Computing) systems. IOR is developed at the Lawrence Livermore National Laboratory (LLNL). Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MB/s, More Is Better IOR 3.3.0 Block Size: 8MB - Disk Target: Default Test Directory 3202 3003 300 600 900 1200 1500 SE +/- 28.69, N = 13 SE +/- 6.28, N = 3 1461.18 1601.21 MIN: 491.21 / MAX: 2711.8 MIN: 1005.82 / MAX: 2534.37 1. (CC) gcc options: -O2 -lm -pthread -lmpi
IOR IOR is a parallel I/O storage benchmark making use of MPI with a particular focus on HPC (High Performance Computing) systems. IOR is developed at the Lawrence Livermore National Laboratory (LLNL). Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MB/s, More Is Better IOR 3.3.0 Block Size: 512MB - Disk Target: Default Test Directory 3202 3003 400 800 1200 1600 2000 SE +/- 22.61, N = 9 SE +/- 11.67, N = 3 1682.82 1748.21 MIN: 251.69 / MAX: 2253.72 MIN: 534.9 / MAX: 2360.08 1. (CC) gcc options: -O2 -lm -pthread -lmpi
3003 Processor: AMD Ryzen 9 5950X 16-Core @ 3.40GHz (16 Cores / 32 Threads), Motherboard: ASUS ROG CROSSHAIR VIII HERO (WI-FI) (3003 BIOS), Chipset: AMD Starship/Matisse, Memory: 32GB, Disk: 2000GB Corsair Force MP600 + 2000GB, Graphics: AMD Radeon RX 5600 OEM/5600 XT / 5700/5700 8GB (2100/875MHz), Audio: AMD Navi 10 HDMI Audio, Monitor: ASUS MG28U, Network: Realtek RTL8125 2.5GbE + Intel I211 + Intel Wi-Fi 6 AX200
OS: Ubuntu 20.10, Kernel: 5.11.0-051100rc2daily20210108-generic (x86_64) 20210107, Desktop: GNOME Shell 3.38.1, Display Server: X Server 1.20.9, Display Driver: amdgpu 19.1.0, OpenGL: 4.6 Mesa 21.0.0-devel (git-f01bca8 2021-01-08 groovy-oibaf-ppa) (LLVM 11.0.1), Vulkan: 1.2.164, Compiler: GCC 10.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0xa201009Graphics Notes: GLAMORPython Notes: Python 3.8.6Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional IBRS_FW STIBP: always-on RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 18 January 2021 13:46 by user phoronix.
3202 Processor: AMD Ryzen 9 5950X 16-Core @ 3.40GHz (16 Cores / 32 Threads), Motherboard: ASUS ROG CROSSHAIR VIII HERO (WI-FI) (3202 BIOS), Chipset: AMD Starship/Matisse, Memory: 32GB, Disk: 2000GB Corsair Force MP600 + 2000GB, Graphics: AMD Radeon RX 5600 OEM/5600 XT / 5700/5700 8GB (2100/875MHz), Audio: AMD Navi 10 HDMI Audio, Monitor: ASUS MG28U, Network: Realtek RTL8125 2.5GbE + Intel I211 + Intel Wi-Fi 6 AX200
OS: Ubuntu 20.10, Kernel: 5.11.0-051100rc2daily20210108-generic (x86_64) 20210107, Desktop: GNOME Shell 3.38.1, Display Server: X Server 1.20.9, Display Driver: amdgpu 19.1.0, OpenGL: 4.6 Mesa 21.0.0-devel (git-f01bca8 2021-01-08 groovy-oibaf-ppa) (LLVM 11.0.1), Vulkan: 1.2.164, Compiler: GCC 10.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vDisk Notes: NONE / errors=remount-ro,relatime,rw / Block Size: 4096Processor Notes: Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0xa201009Graphics Notes: GLAMORPython Notes: Python 3.8.6Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional IBRS_FW STIBP: always-on RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 19 January 2021 12:59 by user phoronix.