AMD Ryzen 3 2200G testing with a ASUS PRIME B350M-E (5220 BIOS) and ASUS AMD Radeon Vega / Mobile 2GB on Ubuntu 20.10 via the Phoronix Test Suite.
1 Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: acpi-cpufreq ondemand (Boost: Enabled) - CPU Microcode: 0x8101016Graphics Notes: GLAMORJava Notes: OpenJDK Runtime Environment (build 11.0.9.1+1-Ubuntu-0ubuntu1.20.10)Python Notes: Python 3.8.6Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional STIBP: disabled RSB filling + srbds: Not affected + tsx_async_abort: Not affected
2 3 Processor: AMD Ryzen 3 2200G @ 3.50GHz (4 Cores), Motherboard: ASUS PRIME B350M-E (5220 BIOS), Chipset: AMD Raven/Raven2, Memory: 6GB, Disk: Samsung SSD 970 EVO 250GB, Graphics: ASUS AMD Radeon Vega / Mobile 2GB (1100/1600MHz), Audio: AMD Raven/Raven2/Fenghuang, Monitor: G237HL, Network: Realtek RTL8111/8168/8411
OS: Ubuntu 20.10, Kernel: 5.8.0-38-generic (x86_64), Desktop: GNOME Shell 3.38.1, Display Server: X Server 1.20.9, Display Driver: modesetting 1.20.9, OpenGL: 4.6 Mesa 20.2.6 (LLVM 11.0.0), Vulkan: 1.2.131, Compiler: GCC 10.2.0, File-System: ext4, Screen Resolution: 1920x1080
yquake2 This is a test of Yamagi Quake II. Yamagi Quake II is an enhanced client for id Software's Quake II with focus on offline and coop gameplay. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better yquake2 7.45 Renderer: OpenGL 3.x - Resolution: 1920 x 1080 1 2 3 200 400 600 800 1000 SE +/- 4.11, N = 3 SE +/- 3.33, N = 3 SE +/- 4.89, N = 3 814.1 807.2 807.9 1. (CC) gcc options: -lm -ldl -rdynamic -shared -lSDL2 -O2 -pipe -fomit-frame-pointer -std=gnu99 -fno-strict-aliasing -fwrapv -fvisibility=hidden -MMD -mfpmath=sse -fPIC
OpenBenchmarking.org Frames Per Second, More Is Better yquake2 7.45 Renderer: Software CPU - Resolution: 1920 x 1080 1 2 3 20 40 60 80 100 SE +/- 0.38, N = 3 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 92.9 93.3 93.3 1. (CC) gcc options: -lm -ldl -rdynamic -shared -lSDL2 -O2 -pipe -fomit-frame-pointer -std=gnu99 -fno-strict-aliasing -fwrapv -fvisibility=hidden -MMD -mfpmath=sse -fPIC
OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.3 Compression Level: 3 - Compression Speed 1 2 3 10 20 30 40 50 SE +/- 0.58, N = 3 SE +/- 0.43, N = 15 SE +/- 0.65, N = 15 42.77 42.34 41.81 1. (CC) gcc options: -O3
OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.3 Compression Level: 3 - Decompression Speed 1 2 3 2K 4K 6K 8K 10K SE +/- 26.74, N = 3 SE +/- 8.28, N = 15 SE +/- 5.83, N = 15 8554.3 8547.8 8547.2 1. (CC) gcc options: -O3
OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.3 Compression Level: 9 - Compression Speed 1 2 3 10 20 30 40 50 SE +/- 0.73, N = 13 SE +/- 0.55, N = 15 SE +/- 0.47, N = 15 42.22 41.02 41.23 1. (CC) gcc options: -O3
OpenBenchmarking.org MB/s, More Is Better LZ4 Compression 1.9.3 Compression Level: 9 - Decompression Speed 1 2 3 2K 4K 6K 8K 10K SE +/- 6.38, N = 13 SE +/- 5.73, N = 15 SE +/- 3.43, N = 15 8565.2 8552.1 8562.8 1. (CC) gcc options: -O3
CLOMP CLOMP is the C version of the Livermore OpenMP benchmark developed to measure OpenMP overheads and other performance impacts due to threading in order to influence future system designs. This particular test profile configuration is currently set to look at the OpenMP static schedule speed-up across all available CPU cores using the recommended test configuration. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Speedup, More Is Better CLOMP 1.2 Static OMP Speedup 1 2 3 0.45 0.9 1.35 1.8 2.25 SE +/- 0.03, N = 3 2.0 2.0 2.0 1. (CC) gcc options: -fopenmp -O3 -lm
OpenBenchmarking.org usec, Fewer Is Better Sockperf 3.4 Test: Latency Ping Pong 1 2 3 2 4 6 8 10 SE +/- 0.065, N = 5 SE +/- 0.049, N = 5 SE +/- 0.074, N = 5 6.927 6.751 6.790 1. (CXX) g++ options: --param -O3 -rdynamic -ldl -lpthread
OpenBenchmarking.org usec, Fewer Is Better Sockperf 3.4 Test: Latency Under Load 1 2 3 12 24 36 48 60 SE +/- 2.25, N = 20 SE +/- 1.88, N = 25 SE +/- 1.93, N = 25 53.78 52.76 50.70 1. (CXX) g++ options: --param -O3 -rdynamic -ldl -lpthread
Basis Universal Basis Universal is a GPU texture codoec. This test times how long it takes to convert sRGB PNGs into Basis Univeral assets with various settings. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Basis Universal 1.12 Settings: ETC1S 1 2 3 20 40 60 80 100 SE +/- 0.05, N = 3 SE +/- 0.14, N = 3 SE +/- 0.07, N = 3 82.06 82.18 82.08 1. (CXX) g++ options: -std=c++11 -fvisibility=hidden -fPIC -fno-strict-aliasing -O3 -rdynamic -lm -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better Basis Universal 1.12 Settings: UASTC Level 0 1 2 3 3 6 9 12 15 SE +/- 0.02, N = 3 SE +/- 0.09, N = 3 SE +/- 0.00, N = 3 11.90 12.05 11.86 1. (CXX) g++ options: -std=c++11 -fvisibility=hidden -fPIC -fno-strict-aliasing -O3 -rdynamic -lm -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better Basis Universal 1.12 Settings: UASTC Level 2 1 2 3 20 40 60 80 100 SE +/- 0.24, N = 3 SE +/- 0.18, N = 3 SE +/- 0.16, N = 3 86.48 86.54 86.34 1. (CXX) g++ options: -std=c++11 -fvisibility=hidden -fPIC -fno-strict-aliasing -O3 -rdynamic -lm -lpthread
Darktable Darktable is an open-source photography / workflow application this will use any system-installed Darktable program or on Windows will automatically download the pre-built binary from the project. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Darktable 3.2.1 Test: Boat - Acceleration: CPU-only 1 2 3 6 12 18 24 30 SE +/- 0.09, N = 3 SE +/- 0.25, N = 13 SE +/- 0.09, N = 3 25.21 25.90 25.45
OpenBenchmarking.org Seconds, Fewer Is Better Darktable 3.2.1 Test: Masskrug - Acceleration: CPU-only 1 2 3 6 12 18 24 30 SE +/- 0.03, N = 3 SE +/- 0.04, N = 3 SE +/- 0.08, N = 3 24.17 24.52 24.19
OpenBenchmarking.org Seconds, Fewer Is Better Darktable 3.2.1 Test: Server Rack - Acceleration: CPU-only 1 2 3 0.0774 0.1548 0.2322 0.3096 0.387 SE +/- 0.001, N = 3 SE +/- 0.001, N = 3 SE +/- 0.004, N = 3 0.339 0.342 0.344
OpenBenchmarking.org Seconds, Fewer Is Better Darktable 3.2.1 Test: Server Room - Acceleration: CPU-only 1 2 3 5 10 15 20 25 SE +/- 0.21, N = 3 SE +/- 0.12, N = 3 SE +/- 0.19, N = 3 20.70 21.01 20.74
Opus Codec Encoding Opus is an open audio codec. Opus is a lossy audio compression format designed primarily for interactive real-time applications over the Internet. This test uses Opus-Tools and measures the time required to encode a WAV file to Opus. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Opus Codec Encoding 1.3.1 WAV To Opus Encode 1 2 3 2 4 6 8 10 SE +/- 0.033, N = 5 SE +/- 0.030, N = 5 SE +/- 0.021, N = 5 8.936 8.923 8.915 1. (CXX) g++ options: -fvisibility=hidden -logg -lm
ASTC Encoder ASTC Encoder (astcenc) is for the Adaptive Scalable Texture Compression (ASTC) format commonly used with OpenGL, OpenGL ES, and Vulkan graphics APIs. This test profile does a coding test of both compression/decompression. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better ASTC Encoder 2.0 Preset: Fast 1 2 3 3 6 9 12 15 SE +/- 0.07, N = 3 SE +/- 0.04, N = 3 SE +/- 0.06, N = 3 9.75 9.74 9.80 1. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better ASTC Encoder 2.0 Preset: Medium 1 2 3 3 6 9 12 15 SE +/- 0.03, N = 3 SE +/- 0.04, N = 3 SE +/- 0.03, N = 3 12.77 12.83 12.75 1. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better ASTC Encoder 2.0 Preset: Thorough 1 2 3 20 40 60 80 100 SE +/- 0.21, N = 3 SE +/- 0.21, N = 3 SE +/- 0.04, N = 3 84.33 84.61 84.49 1. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better ASTC Encoder 2.0 Preset: Exhaustive 1 2 3 150 300 450 600 750 SE +/- 1.18, N = 3 SE +/- 0.48, N = 3 SE +/- 0.20, N = 3 696.05 697.50 695.24 1. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
Hugin Hugin is an open-source, cross-platform panorama photo stitcher software package. This test profile times how long it takes to run the assistant and panorama photo stitching on a set of images. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Hugin Panorama Photo Assistant + Stitching Time 1 2 3 20 40 60 80 100 SE +/- 0.57, N = 3 SE +/- 0.26, N = 3 SE +/- 0.13, N = 3 82.01 83.58 82.19
WebP Image Encode This is a test of Google's libwebp with the cwebp image encode utility and using a sample 6000x4000 pixel JPEG image as the input. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Encode Time - Seconds, Fewer Is Better WebP Image Encode 1.1 Encode Settings: Default 1 2 3 0.374 0.748 1.122 1.496 1.87 SE +/- 0.002, N = 3 SE +/- 0.003, N = 3 SE +/- 0.009, N = 3 1.648 1.662 1.657 1. (CC) gcc options: -fvisibility=hidden -O2 -pthread -lm -ljpeg -lpng16 -ltiff
OpenBenchmarking.org Encode Time - Seconds, Fewer Is Better WebP Image Encode 1.1 Encode Settings: Quality 100 1 2 3 0.5852 1.1704 1.7556 2.3408 2.926 SE +/- 0.009, N = 3 SE +/- 0.007, N = 3 SE +/- 0.004, N = 3 2.595 2.601 2.590 1. (CC) gcc options: -fvisibility=hidden -O2 -pthread -lm -ljpeg -lpng16 -ltiff
OpenBenchmarking.org Encode Time - Seconds, Fewer Is Better WebP Image Encode 1.1 Encode Settings: Quality 100, Lossless 1 2 3 6 12 18 24 30 SE +/- 0.03, N = 3 SE +/- 0.10, N = 3 SE +/- 0.21, N = 3 24.89 24.96 24.90 1. (CC) gcc options: -fvisibility=hidden -O2 -pthread -lm -ljpeg -lpng16 -ltiff
OpenBenchmarking.org Encode Time - Seconds, Fewer Is Better WebP Image Encode 1.1 Encode Settings: Quality 100, Highest Compression 1 2 3 2 4 6 8 10 SE +/- 0.001, N = 3 SE +/- 0.011, N = 3 SE +/- 0.013, N = 3 8.872 8.871 8.869 1. (CC) gcc options: -fvisibility=hidden -O2 -pthread -lm -ljpeg -lpng16 -ltiff
OpenBenchmarking.org Encode Time - Seconds, Fewer Is Better WebP Image Encode 1.1 Encode Settings: Quality 100, Lossless, Highest Compression 1 2 3 13 26 39 52 65 SE +/- 0.26, N = 3 SE +/- 0.02, N = 3 SE +/- 0.09, N = 3 57.67 57.22 57.45 1. (CC) gcc options: -fvisibility=hidden -O2 -pthread -lm -ljpeg -lpng16 -ltiff
OCRMyPDF OCRMyPDF is an optical character recognition (OCR) text layer to scanned PDF files, producing new PDFs with the text now selectable/searchable/copy-paste capable. OCRMyPDF leverages the Tesseract OCR engine and is written in Python. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better OCRMyPDF 10.3.1+dfsg Processing 60 Page PDF Document 1 2 3 12 24 36 48 60 SE +/- 0.13, N = 3 SE +/- 0.08, N = 3 SE +/- 0.10, N = 3 52.68 52.74 52.99
Google SynthMark SynthMark is a cross platform tool for benchmarking CPU performance under a variety of real-time audio workloads. It uses a polyphonic synthesizer model to provide standardized tests for latency, jitter and computational throughput. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Voices, More Is Better Google SynthMark 20201109 Test: VoiceMark_100 1 2 3 130 260 390 520 650 SE +/- 1.56, N = 3 SE +/- 0.89, N = 3 SE +/- 1.20, N = 3 596.25 596.62 593.91 1. (CXX) g++ options: -lm -lpthread -std=c++11 -Ofast
CloverLeaf CloverLeaf is a Lagrangian-Eulerian hydrodynamics benchmark. This test profile currently makes use of CloverLeaf's OpenMP version and benchmarked with the clover_bm.in input file (Problem 5). Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better CloverLeaf Lagrangian-Eulerian Hydrodynamics 1 2 3 40 80 120 160 200 SE +/- 0.19, N = 3 SE +/- 0.07, N = 3 SE +/- 0.07, N = 3 191.41 191.45 191.01 1. (F9X) gfortran options: -O3 -march=native -funroll-loops -fopenmp
Dolfyn Dolfyn is a Computational Fluid Dynamics (CFD) code of modern numerical simulation techniques. The Dolfyn test profile measures the execution time of the bundled computational fluid dynamics demos that are bundled with Dolfyn. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Dolfyn 0.527 Computational Fluid Dynamics 1 2 3 5 10 15 20 25 SE +/- 0.07, N = 3 SE +/- 0.05, N = 3 SE +/- 0.04, N = 3 21.07 21.14 20.99
RNNoise RNNoise is a recurrent neural network for audio noise reduction developed by Mozilla and Xiph.Org. This test profile is a single-threaded test measuring the time to denoise a sample 26 minute long 16-bit RAW audio file using this recurrent neural network noise suppression library. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better RNNoise 2020-06-28 1 2 3 5 10 15 20 25 SE +/- 0.03, N = 3 SE +/- 0.23, N = 8 SE +/- 0.35, N = 3 22.24 22.69 22.58 1. (CC) gcc options: -O2 -pedantic -fvisibility=hidden
Mobile Neural Network MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.1.1 Model: SqueezeNetV1.0 1 2 3 3 6 9 12 15 SE +/- 0.127, N = 3 SE +/- 0.039, N = 3 SE +/- 0.066, N = 3 9.732 9.613 9.867 MIN: 8.67 / MAX: 18.76 MIN: 8.69 / MAX: 20.7 MIN: 8.73 / MAX: 39.42 1. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.1.1 Model: resnet-v2-50 1 2 3 11 22 33 44 55 SE +/- 0.51, N = 3 SE +/- 0.33, N = 3 SE +/- 0.27, N = 3 50.22 50.27 50.49 MIN: 47.21 / MAX: 83.29 MIN: 47.5 / MAX: 72.85 MIN: 47.85 / MAX: 147.84 1. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.1.1 Model: MobileNetV2_224 1 2 3 1.2204 2.4408 3.6612 4.8816 6.102 SE +/- 0.019, N = 3 SE +/- 0.038, N = 3 SE +/- 0.037, N = 3 5.424 5.419 5.398 MIN: 4.8 / MAX: 14.75 MIN: 4.88 / MAX: 15.69 MIN: 4.83 / MAX: 15.64 1. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.1.1 Model: mobilenet-v1-1.0 1 2 3 2 4 6 8 10 SE +/- 0.021, N = 3 SE +/- 0.065, N = 3 SE +/- 0.030, N = 3 7.395 7.526 7.313 MIN: 6.57 / MAX: 16.47 MIN: 6.6 / MAX: 20.05 MIN: 6.61 / MAX: 17.32 1. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.1.1 Model: inception-v3 1 2 3 14 28 42 56 70 SE +/- 0.18, N = 3 SE +/- 0.32, N = 3 SE +/- 0.19, N = 3 63.42 64.00 63.27 MIN: 60.02 / MAX: 120.02 MIN: 60.33 / MAX: 93.06 MIN: 60.45 / MAX: 98.45 1. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
TNN TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better TNN 0.2.3 Target: CPU - Model: MobileNet v2 1 2 3 60 120 180 240 300 SE +/- 0.05, N = 3 SE +/- 0.44, N = 3 SE +/- 0.29, N = 3 279.34 279.63 279.28 MIN: 276.5 / MAX: 293.72 MIN: 276.69 / MAX: 295.47 MIN: 276.39 / MAX: 296.43 1. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -O3 -rdynamic -ldl
OpenBenchmarking.org ms, Fewer Is Better TNN 0.2.3 Target: CPU - Model: SqueezeNet v1.1 1 2 3 60 120 180 240 300 SE +/- 0.12, N = 3 SE +/- 0.28, N = 3 SE +/- 0.02, N = 3 287.26 287.06 286.14 MIN: 286.22 / MAX: 288.27 MIN: 286.17 / MAX: 287.91 MIN: 285.41 / MAX: 286.85 1. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -O3 -rdynamic -ldl
Caffe This is a benchmark of the Caffe deep learning framework and currently supports the AlexNet and Googlenet model and execution on both CPUs and NVIDIA GPUs. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Milli-Seconds, Fewer Is Better Caffe 2020-02-13 Model: AlexNet - Acceleration: CPU - Iterations: 100 1 2 3 9K 18K 27K 36K 45K SE +/- 193.32, N = 3 SE +/- 90.86, N = 3 SE +/- 137.35, N = 3 41877 41672 41573 1. (CXX) g++ options: -fPIC -O3 -rdynamic -lglog -lgflags -lprotobuf -lpthread -lsz -lz -ldl -lm -llmdb -lopenblas
OpenBenchmarking.org Milli-Seconds, Fewer Is Better Caffe 2020-02-13 Model: GoogleNet - Acceleration: CPU - Iterations: 100 1 2 3 20K 40K 60K 80K 100K SE +/- 106.80, N = 3 SE +/- 38.96, N = 3 SE +/- 223.03, N = 3 110084 110320 110157 1. (CXX) g++ options: -fPIC -O3 -rdynamic -lglog -lgflags -lprotobuf -lpthread -lsz -lz -ldl -lm -llmdb -lopenblas
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: mobilenet 1 2 3 11 22 33 44 55 SE +/- 0.46, N = 3 SE +/- 0.07, N = 3 SE +/- 0.05, N = 3 46.83 46.49 46.32 MIN: 42.34 / MAX: 64.41 MIN: 42.72 / MAX: 62.2 MIN: 43.57 / MAX: 62.16 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU-v2-v2 - Model: mobilenet-v2 1 2 3 3 6 9 12 15 SE +/- 0.13, N = 3 SE +/- 0.08, N = 3 SE +/- 0.05, N = 3 11.20 10.91 11.04 MIN: 8.97 / MAX: 20.5 MIN: 8.91 / MAX: 18.07 MIN: 8.96 / MAX: 21.86 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU-v3-v3 - Model: mobilenet-v3 1 2 3 3 6 9 12 15 SE +/- 0.07, N = 3 SE +/- 0.05, N = 3 SE +/- 0.03, N = 3 9.67 9.49 9.59 MIN: 7.8 / MAX: 15.65 MIN: 7.78 / MAX: 14.76 MIN: 7.81 / MAX: 16.29 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: shufflenet-v2 1 2 3 3 6 9 12 15 SE +/- 0.12, N = 3 SE +/- 0.14, N = 3 SE +/- 0.14, N = 3 12.65 12.89 12.70 MIN: 10.41 / MAX: 23.3 MIN: 10.42 / MAX: 26.77 MIN: 10.47 / MAX: 19.71 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: mnasnet 1 2 3 3 6 9 12 15 SE +/- 0.05, N = 3 SE +/- 0.10, N = 3 SE +/- 0.06, N = 3 10.50 10.34 10.25 MIN: 8.45 / MAX: 18.4 MIN: 8.42 / MAX: 16.24 MIN: 8.43 / MAX: 24.69 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: efficientnet-b0 1 2 3 4 8 12 16 20 SE +/- 0.15, N = 3 SE +/- 0.03, N = 3 SE +/- 0.12, N = 3 17.07 16.76 16.89 MIN: 14.1 / MAX: 30.1 MIN: 13.96 / MAX: 31.87 MIN: 14.11 / MAX: 30.24 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: blazeface 1 2 3 0.7493 1.4986 2.2479 2.9972 3.7465 SE +/- 0.01, N = 3 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 3.31 3.33 3.29 MIN: 2.64 / MAX: 9.91 MIN: 2.62 / MAX: 5.11 MIN: 2.73 / MAX: 4.72 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: googlenet 1 2 3 8 16 24 32 40 SE +/- 0.18, N = 3 SE +/- 0.05, N = 3 SE +/- 0.18, N = 3 32.59 32.79 32.43 MIN: 28.72 / MAX: 51.96 MIN: 28.77 / MAX: 48.75 MIN: 28.44 / MAX: 46.34 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: vgg16 1 2 3 30 60 90 120 150 SE +/- 0.18, N = 3 SE +/- 0.23, N = 3 SE +/- 0.17, N = 3 117.19 119.38 117.45 MIN: 112.25 / MAX: 143.19 MIN: 113.9 / MAX: 142.52 MIN: 112.54 / MAX: 135.06 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: resnet18 1 2 3 7 14 21 28 35 SE +/- 0.16, N = 3 SE +/- 0.12, N = 3 SE +/- 0.33, N = 3 29.04 29.04 29.18 MIN: 25.89 / MAX: 42.02 MIN: 25.69 / MAX: 36.07 MIN: 25.64 / MAX: 44.22 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: alexnet 1 2 3 6 12 18 24 30 SE +/- 0.11, N = 3 SE +/- 0.07, N = 3 SE +/- 0.07, N = 3 23.30 23.42 23.51 MIN: 21.24 / MAX: 38.11 MIN: 21.27 / MAX: 37.51 MIN: 21.21 / MAX: 37.48 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: resnet50 1 2 3 16 32 48 64 80 SE +/- 0.82, N = 3 SE +/- 0.39, N = 3 SE +/- 0.19, N = 3 72.55 73.14 71.82 MIN: 66.75 / MAX: 91.97 MIN: 66.24 / MAX: 103.74 MIN: 65.77 / MAX: 87.35 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: yolov4-tiny 1 2 3 13 26 39 52 65 SE +/- 0.05, N = 3 SE +/- 0.08, N = 3 SE +/- 0.06, N = 3 59.28 59.42 59.28 MIN: 55.69 / MAX: 71.74 MIN: 55.13 / MAX: 75.95 MIN: 55.36 / MAX: 72.94 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: squeezenet_ssd 1 2 3 13 26 39 52 65 SE +/- 0.05, N = 3 SE +/- 0.20, N = 3 SE +/- 0.27, N = 3 59.39 59.71 59.34 MIN: 53.07 / MAX: 79.98 MIN: 52.9 / MAX: 78.52 MIN: 52.64 / MAX: 71.17 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: CPU - Model: regnety_400m 1 2 3 5 10 15 20 25 SE +/- 0.07, N = 3 SE +/- 0.10, N = 3 SE +/- 0.18, N = 3 19.11 18.87 18.75 MIN: 16.69 / MAX: 35.81 MIN: 16.81 / MAX: 33.23 MIN: 16.84 / MAX: 32.85 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU - Model: mobilenet 1 2 3 11 22 33 44 55 SE +/- 0.03, N = 3 SE +/- 0.66, N = 4 SE +/- 0.07, N = 3 46.40 47.01 46.32 MIN: 42.8 / MAX: 59.97 MIN: 42.66 / MAX: 60.93 MIN: 42.77 / MAX: 61.8 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU-v2-v2 - Model: mobilenet-v2 1 2 3 3 6 9 12 15 SE +/- 0.01, N = 3 SE +/- 0.21, N = 4 SE +/- 0.38, N = 3 10.88 11.30 10.68 MIN: 8.93 / MAX: 27.73 MIN: 8.86 / MAX: 23.26 MIN: 8.89 / MAX: 17.41 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU-v3-v3 - Model: mobilenet-v3 1 2 3 3 6 9 12 15 SE +/- 0.13, N = 3 SE +/- 0.14, N = 4 SE +/- 0.09, N = 3 9.59 9.69 9.60 MIN: 7.78 / MAX: 22.63 MIN: 7.81 / MAX: 18.31 MIN: 7.73 / MAX: 19.55 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU - Model: shufflenet-v2 1 2 3 3 6 9 12 15 SE +/- 0.09, N = 3 SE +/- 0.16, N = 4 SE +/- 0.20, N = 3 12.63 12.98 12.75 MIN: 10.39 / MAX: 25.47 MIN: 10.26 / MAX: 24.78 MIN: 10.36 / MAX: 21.98 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU - Model: mnasnet 1 2 3 3 6 9 12 15 SE +/- 0.08, N = 3 SE +/- 0.20, N = 4 SE +/- 0.08, N = 3 10.38 10.44 10.43 MIN: 8.43 / MAX: 17.05 MIN: 8.36 / MAX: 26.87 MIN: 8.39 / MAX: 21.7 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU - Model: efficientnet-b0 1 2 3 4 8 12 16 20 SE +/- 0.07, N = 3 SE +/- 0.26, N = 4 SE +/- 0.14, N = 3 16.98 16.95 16.90 MIN: 14.01 / MAX: 31.1 MIN: 14.04 / MAX: 27.15 MIN: 13.97 / MAX: 30.89 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU - Model: blazeface 1 2 3 0.7493 1.4986 2.2479 2.9972 3.7465 SE +/- 0.01, N = 3 SE +/- 0.02, N = 4 SE +/- 0.03, N = 3 3.25 3.31 3.33 MIN: 2.61 / MAX: 14.28 MIN: 2.6 / MAX: 4.93 MIN: 2.62 / MAX: 5.7 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU - Model: googlenet 1 2 3 8 16 24 32 40 SE +/- 0.20, N = 3 SE +/- 0.12, N = 4 SE +/- 0.13, N = 3 32.56 32.64 32.59 MIN: 28.77 / MAX: 47.43 MIN: 28.94 / MAX: 42.38 MIN: 28.42 / MAX: 45.73 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU - Model: vgg16 1 2 3 30 60 90 120 150 SE +/- 0.37, N = 3 SE +/- 0.24, N = 4 SE +/- 0.16, N = 3 117.46 118.91 118.04 MIN: 111.97 / MAX: 149.37 MIN: 113.22 / MAX: 141.78 MIN: 112.22 / MAX: 141.41 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU - Model: resnet18 1 2 3 7 14 21 28 35 SE +/- 0.26, N = 3 SE +/- 0.14, N = 4 SE +/- 0.02, N = 3 29.17 29.08 29.28 MIN: 26 / MAX: 40.97 MIN: 25.74 / MAX: 44.35 MIN: 25.57 / MAX: 39.41 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU - Model: alexnet 1 2 3 6 12 18 24 30 SE +/- 0.08, N = 3 SE +/- 0.03, N = 4 SE +/- 0.02, N = 3 23.46 23.24 23.48 MIN: 21.29 / MAX: 37.36 MIN: 21.11 / MAX: 37.35 MIN: 21.25 / MAX: 36.41 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU - Model: resnet50 1 2 3 16 32 48 64 80 SE +/- 0.20, N = 3 SE +/- 0.63, N = 4 SE +/- 0.29, N = 3 71.68 74.10 71.69 MIN: 65.87 / MAX: 90.56 MIN: 66.26 / MAX: 110.26 MIN: 66.47 / MAX: 91.18 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU - Model: yolov4-tiny 1 2 3 13 26 39 52 65 SE +/- 0.04, N = 3 SE +/- 0.06, N = 4 SE +/- 0.16, N = 3 59.00 59.40 59.28 MIN: 55.05 / MAX: 74.18 MIN: 55.12 / MAX: 75.24 MIN: 54.59 / MAX: 74.87 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU - Model: squeezenet_ssd 1 2 3 13 26 39 52 65 SE +/- 0.08, N = 3 SE +/- 0.14, N = 4 SE +/- 0.16, N = 3 59.28 59.16 59.50 MIN: 53.06 / MAX: 77.54 MIN: 51.95 / MAX: 72.82 MIN: 52.65 / MAX: 71.96 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20201218 Target: Vulkan GPU - Model: regnety_400m 1 2 3 5 10 15 20 25 SE +/- 0.09, N = 3 SE +/- 0.15, N = 4 SE +/- 0.01, N = 3 18.88 19.06 19.07 MIN: 16.76 / MAX: 26.52 MIN: 16.61 / MAX: 34.07 MIN: 16.77 / MAX: 34.77 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
GROMACS The GROMACS (GROningen MAchine for Chemical Simulations) molecular dynamics package testing on the CPU with the water_GMX50 data. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Ns Per Day, More Is Better GROMACS 2020.3 Water Benchmark 1 2 3 0.0749 0.1498 0.2247 0.2996 0.3745 SE +/- 0.002, N = 3 SE +/- 0.002, N = 3 SE +/- 0.005, N = 3 0.333 0.330 0.326 1. (CXX) g++ options: -O3 -pthread -lrt -lpthread -lm
NAMD NAMD is a parallel molecular dynamics code designed for high-performance simulation of large biomolecular systems. NAMD was developed by the Theoretical and Computational Biophysics Group in the Beckman Institute for Advanced Science and Technology at the University of Illinois at Urbana-Champaign. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org days/ns, Fewer Is Better NAMD 2.14 ATPase Simulation - 327,506 Atoms 1 2 3 2 4 6 8 10 SE +/- 0.01425, N = 3 SE +/- 0.03865, N = 3 SE +/- 0.08887, N = 5 6.75407 6.79902 6.83284
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU 1 2 3 4 8 12 16 20 SE +/- 0.17, N = 3 SE +/- 0.19, N = 3 SE +/- 0.24, N = 3 16.54 16.56 16.61 MIN: 13.54 MIN: 13.67 MIN: 13.5 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU 1 2 3 3 6 9 12 15 SE +/- 0.18, N = 15 SE +/- 0.17, N = 15 SE +/- 0.06, N = 3 13.07 13.17 13.37 MIN: 10.63 MIN: 10.78 MIN: 12.28 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU 1 2 3 4 8 12 16 20 SE +/- 0.11, N = 3 SE +/- 0.12, N = 3 SE +/- 0.12, N = 3 14.63 14.87 15.20 MIN: 13.38 MIN: 13.38 MIN: 13.47 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU 1 2 3 1.3143 2.6286 3.9429 5.2572 6.5715 SE +/- 0.02077, N = 3 SE +/- 0.01362, N = 3 SE +/- 0.01488, N = 3 5.81417 5.84129 5.79119 MIN: 5.17 MIN: 5.26 MIN: 5.24 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU 1 2 3 6 12 18 24 30 SE +/- 0.34, N = 3 SE +/- 0.20, N = 3 SE +/- 0.14, N = 3 23.02 23.68 23.67 MIN: 19.03 MIN: 20.06 MIN: 20.01 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU 1 2 3 5 10 15 20 25 SE +/- 0.26, N = 3 SE +/- 0.18, N = 15 SE +/- 0.14, N = 3 22.54 22.31 22.66 MIN: 17.75 MIN: 17.69 MIN: 17.69 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU 1 2 3 7 14 21 28 35 SE +/- 0.15, N = 3 SE +/- 0.28, N = 3 SE +/- 0.14, N = 3 30.81 31.41 30.82 MIN: 22.57 MIN: 22.58 MIN: 22.62 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU 1 2 3 9 18 27 36 45 SE +/- 0.18, N = 3 SE +/- 0.23, N = 3 SE +/- 0.51, N = 3 38.84 38.97 38.51 MIN: 35.67 MIN: 35.93 MIN: 35.63 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU 1 2 3 8 16 24 32 40 SE +/- 1.59, N = 15 SE +/- 1.77, N = 15 SE +/- 1.44, N = 12 36.56 35.85 33.06 MIN: 27.16 MIN: 27.16 MIN: 26.99 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU 1 2 3 7 14 21 28 35 SE +/- 0.09, N = 3 SE +/- 0.14, N = 3 SE +/- 0.29, N = 3 29.71 29.89 29.12 MIN: 26.4 MIN: 26.42 MIN: 26.35 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU 1 2 3 2K 4K 6K 8K 10K SE +/- 133.51, N = 3 SE +/- 86.65, N = 3 SE +/- 61.07, N = 3 8437.12 8277.53 8342.73 MIN: 7874.47 MIN: 7767.68 MIN: 7929.12 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU 1 2 3 2K 4K 6K 8K 10K SE +/- 11.25, N = 3 SE +/- 85.05, N = 3 SE +/- 30.28, N = 3 7721.13 7794.13 7837.72 MIN: 7547.72 MIN: 7534.49 MIN: 7613.07 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU 1 2 3 2K 4K 6K 8K 10K SE +/- 128.90, N = 3 SE +/- 144.66, N = 3 SE +/- 28.77, N = 3 8193.61 8356.19 8419.82 MIN: 7671.11 MIN: 7776.27 MIN: 8053.44 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU 1 2 3 1700 3400 5100 6800 8500 SE +/- 38.84, N = 3 SE +/- 27.04, N = 3 SE +/- 16.23, N = 3 7746.87 7725.63 7701.75 MIN: 7556.29 MIN: 7520.66 MIN: 7494.48 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU 1 2 3 2 4 6 8 10 SE +/- 0.02489, N = 3 SE +/- 0.00711, N = 3 SE +/- 0.00751, N = 3 7.35744 7.36061 7.31351 MIN: 6.35 MIN: 6.34 MIN: 6.35 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU 1 2 3 2K 4K 6K 8K 10K SE +/- 99.84, N = 5 SE +/- 66.22, N = 15 SE +/- 46.50, N = 3 8195.20 8438.64 8426.84 MIN: 7505 MIN: 7752.96 MIN: 8003.45 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU 1 2 3 2K 4K 6K 8K 10K SE +/- 101.76, N = 3 SE +/- 22.70, N = 3 SE +/- 13.61, N = 3 7913.61 7667.94 7750.49 MIN: 7617.25 MIN: 7509.56 MIN: 7562.51 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.0 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU 1 2 3 4 8 12 16 20 SE +/- 0.22, N = 15 SE +/- 0.18, N = 15 SE +/- 0.09, N = 3 14.82 15.08 15.56 MIN: 11.81 MIN: 12.35 MIN: 13.12 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
Algebraic Multi-Grid Benchmark AMG is a parallel algebraic multigrid solver for linear systems arising from problems on unstructured grids. The driver provided with AMG builds linear systems for various 3-dimensional problems. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Figure Of Merit, More Is Better Algebraic Multi-Grid Benchmark 1.2 1 2 3 50M 100M 150M 200M 250M SE +/- 326563.65, N = 3 SE +/- 543041.48, N = 3 SE +/- 483314.65, N = 3 213491533 214232233 214072633 1. (CC) gcc options: -lparcsr_ls -lparcsr_mv -lseq_mv -lIJ_mv -lkrylov -lHYPRE_utilities -lm -fopenmp -pthread -lmpi
FFTE FFTE is a package by Daisuke Takahashi to compute Discrete Fourier Transforms of 1-, 2- and 3- dimensional sequences of length (2^p)*(3^q)*(5^r). Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MFLOPS, More Is Better FFTE 7.0 N=256, 3D Complex FFT Routine 1 2 3 3K 6K 9K 12K 15K SE +/- 120.97, N = 3 SE +/- 110.35, N = 3 SE +/- 161.49, N = 3 15392.81 15755.59 15437.47 1. (F9X) gfortran options: -O3 -fomit-frame-pointer -fopenmp
Kripke Kripke is a simple, scalable, 3D Sn deterministic particle transport code. Its primary purpose is to research how data layout, programming paradigms and architectures effect the implementation and performance of Sn transport. Kripke is developed by LLNL. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Throughput FoM, More Is Better Kripke 1.2.4 1 2 1000K 2000K 3000K 4000K 5000K SE +/- 36406.50, N = 2 SE +/- 35494.54, N = 3 4811563 3117717 1. (CXX) g++ options: -O3 -fopenmp
OpenFOAM OpenFOAM is the leading free, open source software for computational fluid dynamics (CFD). Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better OpenFOAM 8 Input: Motorbike 30M 1 2 3 70 140 210 280 350 SE +/- 1.66, N = 3 SE +/- 0.27, N = 3 SE +/- 2.23, N = 3 342.98 339.54 338.27 1. (CXX) g++ options: -std=c++11 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -ldynamicMesh -lgenericPatchFields -lOpenFOAM -ldl -lm
Monte Carlo Simulations of Ionised Nebulae Mocassin is the Monte Carlo Simulations of Ionised Nebulae. MOCASSIN is a fully 3D or 2D photoionisation and dust radiative transfer code which employs a Monte Carlo approach to the transfer of radiation through media of arbitrary geometry and density distribution. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Monte Carlo Simulations of Ionised Nebulae 2019-03-24 Input: Dust 2D tau100.0 1 2 3 70 140 210 280 350 SE +/- 1.76, N = 3 SE +/- 0.67, N = 3 342 340 341 1. (F9X) gfortran options: -cpp -Jsource/ -ffree-line-length-0 -lm -std=legacy -O3 -O2 -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lrt -lz
Incompact3D Incompact3d is a Fortran-MPI based, finite difference high-performance code for solving the incompressible Navier-Stokes equation and as many as you need scalar transport equations. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Incompact3D 2020-09-17 Input: Cylinder 1 2 3 200 400 600 800 1000 SE +/- 3.54, N = 3 SE +/- 10.03, N = 3 SE +/- 2.19, N = 3 810.95 821.06 820.32 1. (F9X) gfortran options: -cpp -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz
Sunflow Rendering System This test runs benchmarks of the Sunflow Rendering System. The Sunflow Rendering System is an open-source render engine for photo-realistic image synthesis with a ray-tracing core. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Sunflow Rendering System 0.07.2 Global Illumination + Image Synthesis 1 2 3 0.743 1.486 2.229 2.972 3.715 SE +/- 0.041, N = 3 SE +/- 0.032, N = 3 SE +/- 0.028, N = 15 3.206 3.148 3.302 MIN: 2.88 / MAX: 3.79 MIN: 2.89 / MAX: 3.84 MIN: 2.87 / MAX: 4.18
OSBench OSBench is a collection of micro-benchmarks for measuring operating system primitives like time to create threads/processes, launching programs, creating files, and memory allocation. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org us Per Event, Fewer Is Better OSBench Test: Create Files 1 2 3 5 10 15 20 25 SE +/- 0.24, N = 3 SE +/- 0.21, N = 3 SE +/- 0.12, N = 3 18.25 18.32 18.44 1. (CC) gcc options: -lm
OpenBenchmarking.org us Per Event, Fewer Is Better OSBench Test: Create Threads 1 2 3 4 8 12 16 20 SE +/- 0.03, N = 3 SE +/- 0.09, N = 3 SE +/- 0.14, N = 3 14.92 14.82 14.91 1. (CC) gcc options: -lm
OpenBenchmarking.org us Per Event, Fewer Is Better OSBench Test: Launch Programs 1 2 3 20 40 60 80 100 SE +/- 0.27, N = 3 SE +/- 0.05, N = 3 SE +/- 0.09, N = 3 81.52 81.97 82.07 1. (CC) gcc options: -lm
OpenBenchmarking.org us Per Event, Fewer Is Better OSBench Test: Create Processes 1 2 3 6 12 18 24 30 SE +/- 0.10, N = 3 SE +/- 0.03, N = 3 SE +/- 0.21, N = 3 26.19 26.48 26.86 1. (CC) gcc options: -lm
OpenBenchmarking.org Ns Per Event, Fewer Is Better OSBench Test: Memory Allocations 1 2 3 20 40 60 80 100 SE +/- 0.02, N = 3 SE +/- 0.10, N = 3 SE +/- 1.45, N = 3 81.74 82.00 85.63 1. (CC) gcc options: -lm
Stockfish This is a test of Stockfish, an advanced C++11 chess benchmark that can scale up to 128 CPU cores. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Nodes Per Second, More Is Better Stockfish 12 Total Time 1 2 3 1.2M 2.4M 3.6M 4.8M 6M SE +/- 48644.36, N = 3 SE +/- 74806.22, N = 3 SE +/- 39149.01, N = 3 5718169 5628220 5648589 1. (CXX) g++ options: -m64 -lpthread -fno-exceptions -std=c++17 -pedantic -O3 -msse -msse3 -mpopcnt -msse4.1 -mssse3 -msse2 -flto -flto=jobserver
Kvazaar This is a test of Kvazaar as a CPU-based H.265 video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 4K - Video Preset: Medium 1 2 3 0.3375 0.675 1.0125 1.35 1.6875 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 1.49 1.49 1.50 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 1080p - Video Preset: Medium 1 2 3 2 4 6 8 10 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 SE +/- 0.02, N = 3 6.51 6.51 6.50 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 4K - Video Preset: Very Fast 1 2 3 0.8888 1.7776 2.6664 3.5552 4.444 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 SE +/- 0.01, N = 3 3.94 3.94 3.95 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 4K - Video Preset: Ultra Fast 1 2 3 2 4 6 8 10 SE +/- 0.02, N = 3 SE +/- 0.04, N = 3 SE +/- 0.02, N = 3 6.84 6.85 6.83 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 1080p - Video Preset: Very Fast 1 2 3 4 8 12 16 20 SE +/- 0.06, N = 3 SE +/- 0.06, N = 3 SE +/- 0.03, N = 3 15.60 15.52 15.62 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 1080p - Video Preset: Ultra Fast 1 2 3 6 12 18 24 30 SE +/- 0.16, N = 3 SE +/- 0.05, N = 3 SE +/- 0.02, N = 3 27.01 27.07 27.05 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 2.0 Encoder Mode: Speed 6 Realtime 1 2 3 3 6 9 12 15 SE +/- 0.04, N = 3 SE +/- 0.08, N = 3 SE +/- 0.13, N = 3 10.13 10.12 10.25 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 2.0 Encoder Mode: Speed 6 Two-Pass 1 2 3 0.504 1.008 1.512 2.016 2.52 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 SE +/- 0.01, N = 3 2.22 2.22 2.24 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 2.0 Encoder Mode: Speed 8 Realtime 1 2 3 6 12 18 24 30 SE +/- 0.02, N = 3 SE +/- 0.03, N = 3 SE +/- 0.02, N = 3 27.20 27.18 27.29 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
dav1d Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org FPS, More Is Better dav1d 0.8.1 Video Input: Chimera 1080p 1 2 3 40 80 120 160 200 SE +/- 0.04, N = 3 SE +/- 1.76, N = 3 SE +/- 1.89, N = 3 184.25 184.17 183.88 MIN: 129.28 / MAX: 331.16 MIN: 127.79 / MAX: 333.51 MIN: 127.66 / MAX: 340.88 1. (CC) gcc options: -pthread -ldl -lm
OpenBenchmarking.org FPS, More Is Better dav1d 0.8.1 Video Input: Summer Nature 4K 1 2 3 12 24 36 48 60 SE +/- 0.35, N = 3 SE +/- 0.37, N = 3 SE +/- 0.44, N = 3 51.92 52.00 52.09 MIN: 48.08 / MAX: 61.73 MIN: 48.01 / MAX: 61.72 MIN: 48.07 / MAX: 61.57 1. (CC) gcc options: -pthread -ldl -lm
OpenBenchmarking.org FPS, More Is Better dav1d 0.8.1 Video Input: Summer Nature 1080p 1 2 3 40 80 120 160 200 SE +/- 0.82, N = 3 SE +/- 0.30, N = 3 SE +/- 0.36, N = 3 182.89 183.50 184.81 MIN: 167.96 / MAX: 203.4 MIN: 169.65 / MAX: 201.98 MIN: 171.93 / MAX: 203.27 1. (CC) gcc options: -pthread -ldl -lm
OpenBenchmarking.org FPS, More Is Better dav1d 0.8.1 Video Input: Chimera 1080p 10-bit 1 2 3 12 24 36 48 60 SE +/- 0.17, N = 3 SE +/- 0.21, N = 3 SE +/- 0.31, N = 3 52.55 52.39 53.51 MIN: 35.45 / MAX: 124.71 MIN: 35.47 / MAX: 120.48 MIN: 35.6 / MAX: 125.13 1. (CC) gcc options: -pthread -ldl -lm
x265 This is a simple test of the x265 encoder run on the CPU with 1080p and 4K options for H.265 video encode performance with x265. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better x265 3.4 Video Input: Bosphorus 4K 1 2 3 1.0868 2.1736 3.2604 4.3472 5.434 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 SE +/- 0.02, N = 3 4.81 4.83 4.83 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
OpenBenchmarking.org Frames Per Second, More Is Better x265 3.4 Video Input: Bosphorus 1080p 1 2 3 5 10 15 20 25 SE +/- 0.16, N = 3 SE +/- 0.07, N = 3 SE +/- 0.11, N = 3 19.49 19.60 19.71 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
OpenBenchmarking.org Frames Per Second, More Is Better rav1e 0.4 Speed: 6 1 2 3 0.245 0.49 0.735 0.98 1.225 SE +/- 0.005, N = 3 SE +/- 0.001, N = 3 SE +/- 0.003, N = 3 1.082 1.083 1.089
OpenBenchmarking.org Frames Per Second, More Is Better rav1e 0.4 Speed: 10 1 2 3 0.5938 1.1876 1.7814 2.3752 2.969 SE +/- 0.015, N = 3 SE +/- 0.008, N = 3 SE +/- 0.007, N = 3 2.573 2.566 2.639
Embree Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better Embree 3.9.0 Binary: Pathtracer - Model: Crown 1 2 3 0.625 1.25 1.875 2.5 3.125 SE +/- 0.0145, N = 3 SE +/- 0.0043, N = 3 SE +/- 0.0087, N = 3 2.7601 2.7659 2.7779 MIN: 2.71 / MAX: 2.86 MIN: 2.73 / MAX: 2.83 MIN: 2.75 / MAX: 2.87
OpenBenchmarking.org Frames Per Second, More Is Better Embree 3.9.0 Binary: Pathtracer ISPC - Model: Crown 1 2 3 0.5811 1.1622 1.7433 2.3244 2.9055 SE +/- 0.0040, N = 3 SE +/- 0.0105, N = 3 SE +/- 0.0165, N = 3 2.5819 2.5670 2.5828 MIN: 2.55 / MAX: 2.62 MIN: 2.52 / MAX: 2.63 MIN: 2.51 / MAX: 2.65
OpenBenchmarking.org Frames Per Second, More Is Better Embree 3.9.0 Binary: Pathtracer - Model: Asian Dragon 1 2 3 0.7522 1.5044 2.2566 3.0088 3.761 SE +/- 0.0186, N = 3 SE +/- 0.0143, N = 3 SE +/- 0.0299, N = 3 3.3140 3.3113 3.3432 MIN: 3.25 / MAX: 3.4 MIN: 3.26 / MAX: 3.4 MIN: 3.25 / MAX: 3.45
OpenBenchmarking.org Frames Per Second, More Is Better Embree 3.9.0 Binary: Pathtracer - Model: Asian Dragon Obj 1 2 3 0.6728 1.3456 2.0184 2.6912 3.364 SE +/- 0.0237, N = 3 SE +/- 0.0135, N = 3 SE +/- 0.0217, N = 3 2.9903 2.9682 2.9782 MIN: 2.9 / MAX: 3.08 MIN: 2.9 / MAX: 3.07 MIN: 2.91 / MAX: 3.08
OpenBenchmarking.org Frames Per Second, More Is Better Embree 3.9.0 Binary: Pathtracer ISPC - Model: Asian Dragon 1 2 3 0.7313 1.4626 2.1939 2.9252 3.6565 SE +/- 0.0133, N = 3 SE +/- 0.0130, N = 3 SE +/- 0.0025, N = 3 3.2431 3.2504 3.2482 MIN: 3.18 / MAX: 3.32 MIN: 3.19 / MAX: 3.32 MIN: 3.2 / MAX: 3.32
OpenBenchmarking.org Frames Per Second, More Is Better Embree 3.9.0 Binary: Pathtracer ISPC - Model: Asian Dragon Obj 1 2 3 0.6383 1.2766 1.9149 2.5532 3.1915 SE +/- 0.0119, N = 3 SE +/- 0.0157, N = 3 SE +/- 0.0152, N = 3 2.8199 2.8371 2.8151 MIN: 2.75 / MAX: 2.92 MIN: 2.77 / MAX: 2.9 MIN: 2.75 / MAX: 2.89
OpenBenchmarking.org M samples/s, More Is Better IndigoBench 4.4 Acceleration: CPU - Scene: Supercar 1 2 3 0.2491 0.4982 0.7473 0.9964 1.2455 SE +/- 0.004, N = 3 SE +/- 0.009, N = 3 SE +/- 0.002, N = 3 1.107 1.098 1.106
Build2 This test profile measures the time to bootstrap/install the build2 C++ build toolchain from source. Build2 is a cross-platform build toolchain for C/C++ code and features Cargo-like features. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Build2 0.13 Time To Compile 1 2 3 110 220 330 440 550 SE +/- 0.43, N = 3 SE +/- 2.15, N = 3 SE +/- 1.15, N = 3 514.48 516.52 514.80
RealSR-NCNN RealSR-NCNN is an NCNN neural network implementation of the RealSR project and accelerated using the Vulkan API. RealSR is the Real-World Super Resolution via Kernel Estimation and Noise Injection. NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. This test profile times how long it takes to increase the resolution of a sample image by a scale of 4x with Vulkan. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better RealSR-NCNN 20200818 Scale: 4x - TAA: No 1 2 3 14 28 42 56 70 SE +/- 0.02, N = 3 SE +/- 0.03, N = 3 SE +/- 0.02, N = 3 63.03 63.02 63.00
Waifu2x-NCNN Vulkan Waifu2x-NCNN is an NCNN neural network implementation of the Waifu2x converter project and accelerated using the Vulkan API. NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. This test profile times how long it takes to increase the resolution of a sample image with Vulkan. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Waifu2x-NCNN Vulkan 20200818 Scale: 2x - Denoise: 3 - TAA: No 1 2 3 0.9259 1.8518 2.7777 3.7036 4.6295 SE +/- 0.023, N = 3 SE +/- 0.005, N = 3 SE +/- 0.005, N = 3 4.115 4.110 4.097
PHPBench PHPBench is a benchmark suite for PHP. It performs a large number of simple tests in order to bench various aspects of the PHP interpreter. PHPBench can be used to compare hardware, operating systems, PHP versions, PHP accelerators and caches, compiler options, etc. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Score, More Is Better PHPBench 0.8.1 PHP Benchmark Suite 1 2 3 110K 220K 330K 440K 550K SE +/- 423.52, N = 3 SE +/- 1952.23, N = 3 SE +/- 2233.09, N = 3 508106 506055 504159
InfluxDB This is a benchmark of the InfluxDB open-source time-series database optimized for fast, high-availability storage for IoT and other use-cases. The InfluxDB test profile makes use of InfluxDB Inch for facilitating the benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org val/sec, More Is Better InfluxDB 1.8.2 Concurrent Streams: 4 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000 1 2 3 150K 300K 450K 600K 750K SE +/- 8641.46, N = 3 SE +/- 6558.55, N = 3 SE +/- 5594.95, N = 3 706035.5 696009.5 700554.6
OpenBenchmarking.org val/sec, More Is Better InfluxDB 1.8.2 Concurrent Streams: 64 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000 1 2 3 160K 320K 480K 640K 800K SE +/- 1902.33, N = 3 SE +/- 1666.30, N = 3 SE +/- 3366.02, N = 3 721428.3 725224.2 723222.1
KeyDB A benchmark of KeyDB as a multi-threaded fork of the Redis server. The KeyDB benchmark is conducted using memtier-benchmark. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Ops/sec, More Is Better KeyDB 6.0.16 1 2 3 60K 120K 180K 240K 300K SE +/- 3138.20, N = 3 SE +/- 2051.97, N = 3 SE +/- 1852.38, N = 3 265074.45 267212.95 269044.48 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
OpenBenchmarking.org Requests Per Second, More Is Better Redis 6.0.9 Test: SADD 1 2 3 400K 800K 1200K 1600K 2000K SE +/- 11595.14, N = 3 SE +/- 21162.12, N = 3 SE +/- 4337.84, N = 3 1735687.33 1758200.50 1734495.83 1. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
OpenBenchmarking.org Requests Per Second, More Is Better Redis 6.0.9 Test: LPUSH 1 2 3 300K 600K 900K 1200K 1500K SE +/- 16215.78, N = 3 SE +/- 2985.46, N = 3 SE +/- 4396.42, N = 3 1216336.46 1213155.04 1223284.42 1. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
OpenBenchmarking.org Requests Per Second, More Is Better Redis 6.0.9 Test: GET 1 2 3 400K 800K 1200K 1600K 2000K SE +/- 35016.53, N = 3 SE +/- 23617.43, N = 5 SE +/- 22292.08, N = 3 2064794.83 1931045.20 1930168.38 1. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
OpenBenchmarking.org Requests Per Second, More Is Better Redis 6.0.9 Test: SET 1 2 3 300K 600K 900K 1200K 1500K SE +/- 5888.95, N = 3 SE +/- 19097.96, N = 3 SE +/- 15253.86, N = 8 1489969.25 1486411.63 1472539.67 1. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
Node.js V8 Web Tooling Benchmark Running the V8 project's Web-Tooling-Benchmark under Node.js. The Web-Tooling-Benchmark stresses JavaScript-related workloads common to web developers like Babel and TypeScript and Babylon. This test profile can test the system's JavaScript performance with Node.js. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org runs/s, More Is Better Node.js V8 Web Tooling Benchmark 1 2 3 2 4 6 8 10 SE +/- 0.08, N = 3 SE +/- 0.03, N = 3 SE +/- 0.09, N = 4 7.38 7.74 7.38 1. Nodejs
v12.18.2
simdjson This is a benchmark of SIMDJSON, a high performance JSON parser. SIMDJSON aims to be the fastest JSON parser and is used by projects like Microsoft FishStore, Yandex ClickHouse, Shopify, and others. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org GB/s, More Is Better simdjson 0.7.1 Throughput Test: Kostya 1 2 3 0.0855 0.171 0.2565 0.342 0.4275 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 0.38 0.38 0.38 1. (CXX) g++ options: -O3 -pthread
OpenBenchmarking.org GB/s, More Is Better simdjson 0.7.1 Throughput Test: LargeRandom 1 2 3 0.0788 0.1576 0.2364 0.3152 0.394 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 0.35 0.35 0.35 1. (CXX) g++ options: -O3 -pthread
OpenBenchmarking.org GB/s, More Is Better simdjson 0.7.1 Throughput Test: PartialTweets 1 2 3 0.1013 0.2026 0.3039 0.4052 0.5065 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 0.45 0.45 0.45 1. (CXX) g++ options: -O3 -pthread
OpenBenchmarking.org GB/s, More Is Better simdjson 0.7.1 Throughput Test: DistinctUserID 1 2 3 0.1035 0.207 0.3105 0.414 0.5175 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 0.46 0.46 0.46 1. (CXX) g++ options: -O3 -pthread
1 Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: acpi-cpufreq ondemand (Boost: Enabled) - CPU Microcode: 0x8101016Graphics Notes: GLAMORJava Notes: OpenJDK Runtime Environment (build 11.0.9.1+1-Ubuntu-0ubuntu1.20.10)Python Notes: Python 3.8.6Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional STIBP: disabled RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 16 January 2021 13:11 by user phoronix.
2 Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: acpi-cpufreq ondemand (Boost: Enabled) - CPU Microcode: 0x8101016Graphics Notes: GLAMORJava Notes: OpenJDK Runtime Environment (build 11.0.9.1+1-Ubuntu-0ubuntu1.20.10)Python Notes: Python 3.8.6Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional STIBP: disabled RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 17 January 2021 05:40 by user phoronix.
3 Processor: AMD Ryzen 3 2200G @ 3.50GHz (4 Cores), Motherboard: ASUS PRIME B350M-E (5220 BIOS), Chipset: AMD Raven/Raven2, Memory: 6GB, Disk: Samsung SSD 970 EVO 250GB, Graphics: ASUS AMD Radeon Vega / Mobile 2GB (1100/1600MHz), Audio: AMD Raven/Raven2/Fenghuang, Monitor: G237HL, Network: Realtek RTL8111/8168/8411
OS: Ubuntu 20.10, Kernel: 5.8.0-38-generic (x86_64), Desktop: GNOME Shell 3.38.1, Display Server: X Server 1.20.9, Display Driver: modesetting 1.20.9, OpenGL: 4.6 Mesa 20.2.6 (LLVM 11.0.0), Vulkan: 1.2.131, Compiler: GCC 10.2.0, File-System: ext4, Screen Resolution: 1920x1080
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: acpi-cpufreq ondemand (Boost: Enabled) - CPU Microcode: 0x8101016Graphics Notes: GLAMORJava Notes: OpenJDK Runtime Environment (build 11.0.9.1+1-Ubuntu-0ubuntu1.20.10)Python Notes: Python 3.8.6Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional STIBP: disabled RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 18 January 2021 05:43 by user phoronix.