E3-1260L 2021

Intel Xeon E3-1260L v5 testing with a ASRock E3V5 WS (P7.10 BIOS) and XFX NVIDIA GeForce GT 220 on Ubuntu 20.10 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2102050-HA-E31260L2045
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results

Limit displaying results to tests within:

Audio Encoding 3 Tests
AV1 3 Tests
C++ Boost Tests 3 Tests
Timed Code Compilation 3 Tests
C/C++ Compiler Tests 2 Tests
Compression Tests 2 Tests
CPU Massive 7 Tests
Creator Workloads 11 Tests
Cryptography 3 Tests
Encoding 6 Tests
Fortran Tests 3 Tests
Game Development 2 Tests
HPC - High Performance Computing 14 Tests
Imaging 2 Tests
Machine Learning 5 Tests
Molecular Dynamics 4 Tests
MPI Benchmarks 4 Tests
Multi-Core 11 Tests
OpenMPI Tests 7 Tests
Programmer / Developer System Benchmarks 5 Tests
Python Tests 2 Tests
Scientific Computing 8 Tests
Server CPU Tests 4 Tests
Single-Threaded 3 Tests
Video Encoding 3 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
1
February 03 2021
  11 Hours, 32 Minutes
2
February 04 2021
  12 Hours, 16 Minutes
3
February 04 2021
  2 Hours, 38 Minutes
4
February 04 2021
  11 Hours, 48 Minutes
Invert Hiding All Results Option
  9 Hours, 34 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


E3-1260L 2021ProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerDisplay DriverCompilerFile-SystemScreen Resolution1234Intel Xeon E3-1260L v5 @ 3.90GHz (4 Cores / 8 Threads)ASRock E3V5 WS (P7.10 BIOS)Intel Xeon E3-1200 v5/E3-15008GB120GB INTEL SSDSC2BW12XFX NVIDIA GeForce GT 220Realtek ALC892LG Ultra HDIntel I219-LMUbuntu 20.105.8.0-33-generic (x86_64)GNOME Shell 3.38.1X Server 1.20.8xfxGCC 10.2.0ext41920x1080OpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-JvwpWM/gcc-10-10.2.0/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- Scaling Governor: intel_pstate powersave - CPU Microcode: 0xe2 - Thermald 2.3Python Details- Python 3.8.6Security Details- itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Mitigation of PTE Inversion; VMX: conditional cache flushes SMT vulnerable + mds: Mitigation of Clear buffers; SMT vulnerable + meltdown: Mitigation of PTI + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + srbds: Mitigation of Microcode + tsx_async_abort: Mitigation of Clear buffers; SMT vulnerable

1234Result OverviewPhoronix Test Suite100%100%101%101%102%QMCPACKFFTELAMMPS Molecular Dynamics SimulatorOpenFOAMQuantLiboneDNNLULESHPennantNAS Parallel BenchmarksLZ4 CompressionlzbenchAlgebraic Multi-Grid Benchmark

E3-1260L 2021redis: LPOPonednn: IP Shapes 3D - f32 - CPUredis: GETonednn: IP Shapes 3D - u8s8f32 - CPUncnn: CPU - blazefacerav1e: 10lzbench: Crush 0 - Compressiononednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUqmcpack: simple-H2Oredis: LPUSHnpb: FT.Cmnn: inception-v3ffte: N=256, 3D Complex FFT Routineencode-opus: WAV To Opus Encodenpb: LU.Cmnn: MobileNetV2_224npb: EP.Credis: SADDlzbench: Zstd 8 - Compressiononednn: Deconvolution Batch shapes_3d - f32 - CPUastcenc: Fastrav1e: 6rav1e: 5onednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUmnn: resnet-v2-50onednn: Deconvolution Batch shapes_1d - f32 - CPUbuild2: Time To Compilecompress-lz4: 1 - Decompression Speedunpack-firefox: firefox-84.0.source.tar.xznpb: CG.Conednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUavifenc: 2dav1d: Chimera 1080pncnn: CPU - mnasnetlammps: Rhodopsin Proteindav1d: Summer Nature 1080ponednn: Convolution Batch Shapes Auto - f32 - CPUavifenc: 10lzbench: Zstd 1 - Decompressionquantlib: dav1d: Summer Nature 4Konednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUlzbench: Zstd 8 - Decompressionncnn: CPU - efficientnet-b0encode-ape: WAV To APElulesh: lzbench: Brotli 2 - Compressiononednn: IP Shapes 1D - f32 - CPUavifenc: 8pennant: leblancbigmnn: SqueezeNetV1.0npb: EP.Dwebp2: Defaultnpb: SP.Bonnx: yolov4 - OpenMP CPUlzbench: Libdeflate 1 - Compressionmnn: mobilenet-v1-1.0compress-lz4: 9 - Compression Speedcoremark: CoreMark Size 666 - Iterations Per Secondcompress-lz4: 3 - Decompression Speedopenfoam: Motorbike 30Mlzbench: Zstd 1 - Compressionwebp2: Quality 75, Compression Effort 7ncnn: CPU - vgg16encode-wavpack: WAV To WavPackdav1d: Chimera 1080p 10-bitbuild-eigen: Time To Compilencnn: CPU - shufflenet-v2npb: BT.Cpennant: sedovbigonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUcryptsetup: Twofish-XTS 512b Decryptionkripke: cryptsetup: AES-XTS 256b Decryptiononednn: Recurrent Neural Network Inference - u8s8f32 - CPUcryptsetup: AES-XTS 256b Encryptioncryptsetup: PBKDF2-sha512tnn: CPU - MobileNet v2compress-lz4: 1 - Compression Speedonnx: bertsquad-10 - OpenMP CPUonnx: super-resolution-10 - OpenMP CPUlzbench: Brotli 2 - Decompressioncompress-lz4: 3 - Compression Speedonednn: Recurrent Neural Network Training - u8s8f32 - CPUncnn: CPU - mobilenetcryptsetup: AES-XTS 512b Encryptionavifenc: 0onednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUrav1e: 1gcrypt: gnupg: 2.7GB Sample File Encryptiononednn: Recurrent Neural Network Training - f32 - CPUcryptsetup: AES-XTS 512b Decryptiononednn: IP Shapes 1D - u8s8f32 - CPUncnn: CPU - yolov4-tinyastcenc: Mediumredis: SETlzbench: Brotli 0 - Compressionbuild-godot: Time To Compilencnn: CPU - regnety_400monednn: Recurrent Neural Network Inference - f32 - CPUncnn: CPU - squeezenet_ssdcryptsetup: Twofish-XTS 256b Encryptiononednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUwebp2: Quality 95, Compression Effort 7lzbench: Brotli 0 - Decompressionnpb: MG.Cncnn: CPU-v3-v3 - mobilenet-v3cryptsetup: Serpent-XTS 256b Decryptioncryptsetup: Serpent-XTS 256b Encryptionncnn: CPU - resnet50ncnn: CPU-v2-v2 - mobilenet-v2astcenc: Thoroughwebp2: Quality 100, Lossless Compressioncryptsetup: Serpent-XTS 512b Decryptioncompress-lz4: 9 - Decompression Speedncnn: CPU - alexnetcryptsetup: Twofish-XTS 256b Decryptioncryptsetup: Serpent-XTS 512b Encryptiononnx: shufflenet-v2-10 - OpenMP CPUncnn: CPU - resnet18tnn: CPU - SqueezeNet v1.1cryptsetup: PBKDF2-whirlpoolsynthmark: VoiceMark_100cryptsetup: Twofish-XTS 512b Encryptionamg: astcenc: Exhaustivewebp2: Quality 100, Compression Effort 5ncnn: CPU - googlenetonnx: fcn-resnet101-11 - OpenMP CPUlzbench: Crush 0 - Decompressionlzbench: XZ 0 - Decompressionlzbench: XZ 0 - Compression12342447471.5811.36392281791.253.067972.702.8079012.596934.1891507373.506871.1960.43219035.0914023989.78914764.504.389514.121917948.348316.41339.421.3771.0385.3889448.88912.4974363.9097915.223.6683438.098.25057119.201299.546.742.615275.7122.60477.81416052181.878.947.41513174910.9113.2681172.90021699.449458.474105.06397.811504.2710.5334409.822072085.07144.52137084.9803667869.7386.71453639.61089.2217.00565.1089.8889.7813485.45172.986221.3841393.0248348632051.04631.892043.41579979363.1146691.28312223164845.788917.5230.441834.6203.2504640.830.362233.10073.3018915.851835.94.2865940.9512.301717799.75427363.21317.564632.0231.30390.18927.631178.6935587148.356.54720.0704.846.627.9980.272398.273719.67847.419.51394.1705.01410421.88345.555665903603.253390.8180621100643.4537.43721.7533447105381501169.1311.01872189234.252.944282.702.8659012.474934.8141486920.076902.7559.99918773.2668944959.76414562.224.362507.421934230.878416.29719.531.3731.0505.3772048.88912.5719367.3097837.523.7183463.428.22961120.281302.256.732.638277.6722.42767.86716002168.379.537.42722175010.9113.1861169.99781709.396978.512105.63997.789503.5510.5084387.592072095.07044.42136732.1515997834.3387.68451642.37188.8616.93865.0490.2599.8013518.04173.692221.3000394.4249155072055.04628.542049.51576054362.7246682.99311222464845.748925.4530.451838.3203.8464630.470.362232.51173.1068919.791838.14.2849840.9612.311721937.79427363.91617.524628.6731.23390.98933.081180.0195587153.826.54720.1705.846.667.9980.322400.663719.47844.019.53394.5705.71410321.90345.251665904603.132390.5180582833643.8837.44321.75334471053811.80982.966079012.478534.3956986.4019060.18895772314704.44513.268316.48825.4368312.44427870.73448.628.281092.62922.419415932167.07.4444917491165.74441709.43266105.4640502.034391.7420944.617840.2388.4545213518.01173.686821.35404645.086668.8164845.708944.354643.818939.184.291284274635.338945.605587142.147852.5180576900447105381465596.1611.46322165661.082.964632.632.8709212.747634.7521513254.556982.3159.49419047.0219170179.90914707.684.330510.531942305.798316.38029.531.3891.0495.4240448.36512.4693367.5877867.123.4863472.038.20667119.623301.046.792.622275.3422.43507.80716042165.579.467.38944176110.9813.2241166.66831709.414278.521105.57367.771501.6910.5624391.952082085.04744.40137374.6056007851.0388.02453641.28588.8416.93365.3190.1879.7613540.52173.527121.3246394.5249255532047.64632.412042.31581562363.9786669.47311222964645.648938.3730.361832.9203.6804640.340.361232.47773.1248931.431833.44.2804540.8612.281721260.04428364.05817.534639.2131.24390.88936.311177.8285577142.466.53719.0705.746.687.9880.222397.749718.87843.819.53394.4705.61411721.90345.338665340602.746390.8180705033643.5737.41921.763344710538OpenBenchmarking.org

Redis

Redis is an open-source in-memory data structure store, used as a database, cache, and message broker. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: LPOP124500K1000K1500K2000K2500KSE +/- 7282.72, N = 3SE +/- 21480.12, N = 3SE +/- 20957.17, N = 42447471.581501169.131465596.161. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: LPOP124400K800K1200K1600K2000KMin: 2434875.25 / Avg: 2447471.58 / Max: 2460103.25Min: 1459455.88 / Avg: 1501169.13 / Max: 1530924.62Min: 1425973.5 / Avg: 1465596.16 / Max: 1519564.881. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU12343691215SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 311.3611.0211.8111.46MIN: 11.21MIN: 10.88MIN: 11.65MIN: 11.31. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU12343691215Min: 11.34 / Avg: 11.36 / Max: 11.38Min: 11.01 / Avg: 11.02 / Max: 11.03Min: 11.79 / Avg: 11.81 / Max: 11.83Min: 11.45 / Avg: 11.46 / Max: 11.481. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

Redis

Redis is an open-source in-memory data structure store, used as a database, cache, and message broker. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: GET124500K1000K1500K2000K2500KSE +/- 32237.70, N = 3SE +/- 20266.35, N = 3SE +/- 19501.96, N = 32281791.252189234.252165661.081. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: GET124400K800K1200K1600K2000KMin: 2230720.5 / Avg: 2281791.25 / Max: 2341409.5Min: 2148765 / Avg: 2189234.25 / Max: 2211432.25Min: 2130454.25 / Avg: 2165661.08 / Max: 21978021. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU12340.69031.38062.07092.76123.4515SE +/- 0.01345, N = 3SE +/- 0.00215, N = 3SE +/- 0.00090, N = 3SE +/- 0.01115, N = 33.067972.944282.966072.96463MIN: 2.96MIN: 2.85MIN: 2.89MIN: 2.871. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU1234246810Min: 3.04 / Avg: 3.07 / Max: 3.09Min: 2.94 / Avg: 2.94 / Max: 2.95Min: 2.96 / Avg: 2.97 / Max: 2.97Min: 2.95 / Avg: 2.96 / Max: 2.991. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: blazeface1240.60751.2151.82252.433.0375SE +/- 0.07, N = 3SE +/- 0.08, N = 3SE +/- 0.00, N = 32.702.702.63MIN: 2.6 / MAX: 11.85MIN: 2.58 / MAX: 11.85MIN: 2.6 / MAX: 3.381. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: blazeface124246810Min: 2.63 / Avg: 2.7 / Max: 2.84Min: 2.62 / Avg: 2.7 / Max: 2.85Min: 2.63 / Avg: 2.63 / Max: 2.631. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

rav1e

Xiph rav1e is a Rust-written AV1 video encoder. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4Speed: 101240.64581.29161.93742.58323.229SE +/- 0.022, N = 3SE +/- 0.024, N = 3SE +/- 0.024, N = 32.8072.8652.870
OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4Speed: 10124246810Min: 2.77 / Avg: 2.81 / Max: 2.84Min: 2.82 / Avg: 2.87 / Max: 2.9Min: 2.84 / Avg: 2.87 / Max: 2.92

lzbench

lzbench is an in-memory benchmark of various compressors. The file used for compression is a Linux kernel source tree tarball. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Crush 0 - Process: Compression123420406080100SE +/- 0.33, N = 3SE +/- 0.58, N = 3SE +/- 0.33, N = 3909090921. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Crush 0 - Process: Compression123420406080100Min: 89 / Avg: 89.67 / Max: 90Min: 89 / Avg: 90 / Max: 91Min: 91 / Avg: 91.67 / Max: 921. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU12343691215SE +/- 0.10, N = 3SE +/- 0.02, N = 3SE +/- 0.03, N = 3SE +/- 0.11, N = 312.6012.4712.4812.75MIN: 11.11MIN: 11.03MIN: 11.03MIN: 11.151. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU123448121620Min: 12.5 / Avg: 12.6 / Max: 12.79Min: 12.44 / Avg: 12.47 / Max: 12.52Min: 12.43 / Avg: 12.48 / Max: 12.52Min: 12.52 / Avg: 12.75 / Max: 12.891. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

QMCPACK

QMCPACK is a modern high-performance open-source Quantum Monte Carlo (QMC) simulation code making use of MPI for this benchmark of the H20 example code. QMCPACK is an open-source production level many-body ab initio Quantum Monte Carlo code for computing the electronic structure of atoms, molecules, and solids. QMCPACK is supported by the U.S. Department of Energy. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Execution Time - Seconds, Fewer Is BetterQMCPACK 3.10Input: simple-H2O1234816243240SE +/- 0.16, N = 3SE +/- 0.26, N = 3SE +/- 0.26, N = 3SE +/- 0.29, N = 334.1934.8134.4034.751. (CXX) g++ options: -fopenmp -finline-limit=1000 -fstrict-aliasing -funroll-all-loops -march=native -O3 -fomit-frame-pointer -ffast-math -pthread -lm
OpenBenchmarking.orgTotal Execution Time - Seconds, Fewer Is BetterQMCPACK 3.10Input: simple-H2O1234714212835Min: 33.93 / Avg: 34.19 / Max: 34.47Min: 34.29 / Avg: 34.81 / Max: 35.11Min: 34.07 / Avg: 34.4 / Max: 34.9Min: 34.32 / Avg: 34.75 / Max: 35.311. (CXX) g++ options: -fopenmp -finline-limit=1000 -fstrict-aliasing -funroll-all-loops -march=native -O3 -fomit-frame-pointer -ffast-math -pthread -lm

Redis

Redis is an open-source in-memory data structure store, used as a database, cache, and message broker. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: LPUSH124300K600K900K1200K1500KSE +/- 13446.73, N = 3SE +/- 14467.44, N = 9SE +/- 6781.65, N = 31507373.501486920.071513254.551. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: LPUSH124300K600K900K1200K1500KMin: 1480894.5 / Avg: 1507373.5 / Max: 1524686.12Min: 1398605.88 / Avg: 1486920.07 / Max: 1529315.12Min: 1499714.38 / Avg: 1513254.55 / Max: 1520710.381. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: FT.C123415003000450060007500SE +/- 72.82, N = 3SE +/- 15.03, N = 3SE +/- 12.69, N = 3SE +/- 15.55, N = 36871.196902.756986.406982.311. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: FT.C123412002400360048006000Min: 6725.79 / Avg: 6871.19 / Max: 6951.18Min: 6886.32 / Avg: 6902.75 / Max: 6932.76Min: 6961.04 / Avg: 6986.4 / Max: 7000.08Min: 6951.8 / Avg: 6982.31 / Max: 7002.781. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.1Model: inception-v31241428425670SE +/- 0.13, N = 3SE +/- 0.21, N = 3SE +/- 0.97, N = 360.4360.0059.49MIN: 59.98 / MAX: 66.21MIN: 59.43 / MAX: 64.63MIN: 57.28 / MAX: 89.731. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.1Model: inception-v31241224364860Min: 60.22 / Avg: 60.43 / Max: 60.68Min: 59.58 / Avg: 60 / Max: 60.22Min: 57.57 / Avg: 59.49 / Max: 60.651. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

FFTE

FFTE is a package by Daisuke Takahashi to compute Discrete Fourier Transforms of 1-, 2- and 3- dimensional sequences of length (2^p)*(3^q)*(5^r). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMFLOPS, More Is BetterFFTE 7.0N=256, 3D Complex FFT Routine12344K8K12K16K20KSE +/- 29.86, N = 3SE +/- 16.02, N = 3SE +/- 44.69, N = 3SE +/- 42.48, N = 319035.0918773.2719060.1919047.021. (F9X) gfortran options: -O3 -fomit-frame-pointer -fopenmp
OpenBenchmarking.orgMFLOPS, More Is BetterFFTE 7.0N=256, 3D Complex FFT Routine12343K6K9K12K15KMin: 18989.46 / Avg: 19035.09 / Max: 19091.28Min: 18756.89 / Avg: 18773.27 / Max: 18805.3Min: 18981.61 / Avg: 19060.19 / Max: 19136.36Min: 18999.87 / Avg: 19047.02 / Max: 19131.81. (F9X) gfortran options: -O3 -fomit-frame-pointer -fopenmp

Opus Codec Encoding

Opus is an open audio codec. Opus is a lossy audio compression format designed primarily for interactive real-time applications over the Internet. This test uses Opus-Tools and measures the time required to encode a WAV file to Opus. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpus Codec Encoding 1.3.1WAV To Opus Encode1243691215SE +/- 0.037, N = 5SE +/- 0.068, N = 5SE +/- 0.044, N = 59.7899.7649.9091. (CXX) g++ options: -fvisibility=hidden -logg -lm
OpenBenchmarking.orgSeconds, Fewer Is BetterOpus Codec Encoding 1.3.1WAV To Opus Encode1243691215Min: 9.65 / Avg: 9.79 / Max: 9.85Min: 9.64 / Avg: 9.76 / Max: 10.02Min: 9.83 / Avg: 9.91 / Max: 10.031. (CXX) g++ options: -fvisibility=hidden -logg -lm

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: LU.C12343K6K9K12K15KSE +/- 14.35, N = 3SE +/- 134.72, N = 10SE +/- 10.10, N = 3SE +/- 15.78, N = 314764.5014562.2214704.4414707.681. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: LU.C12343K6K9K12K15KMin: 14741.66 / Avg: 14764.5 / Max: 14790.96Min: 13350.59 / Avg: 14562.22 / Max: 14725.94Min: 14689.17 / Avg: 14704.44 / Max: 14723.54Min: 14679.19 / Avg: 14707.68 / Max: 14733.691. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.1Model: MobileNetV2_2241240.98751.9752.96253.954.9375SE +/- 0.010, N = 3SE +/- 0.025, N = 3SE +/- 0.049, N = 34.3894.3624.330MIN: 4.29 / MAX: 6.77MIN: 4.01 / MAX: 4.84MIN: 4.15 / MAX: 4.531. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.1Model: MobileNetV2_224124246810Min: 4.38 / Avg: 4.39 / Max: 4.41Min: 4.31 / Avg: 4.36 / Max: 4.4Min: 4.23 / Avg: 4.33 / Max: 4.391. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.C1234110220330440550SE +/- 0.45, N = 3SE +/- 5.06, N = 8SE +/- 0.41, N = 3SE +/- 1.81, N = 3514.12507.42513.26510.531. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.C123490180270360450Min: 513.25 / Avg: 514.12 / Max: 514.77Min: 472.84 / Avg: 507.42 / Max: 517.38Min: 512.49 / Avg: 513.26 / Max: 513.88Min: 506.95 / Avg: 510.53 / Max: 512.791. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3

Redis

Redis is an open-source in-memory data structure store, used as a database, cache, and message broker. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: SADD124400K800K1200K1600K2000KSE +/- 24215.71, N = 4SE +/- 8591.83, N = 3SE +/- 13080.73, N = 31917948.341934230.871942305.791. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: SADD124300K600K900K1200K1500KMin: 1849124.25 / Avg: 1917948.34 / Max: 1958115.62Min: 1921285 / Avg: 1934230.87 / Max: 1950489.62Min: 1917177.88 / Avg: 1942305.79 / Max: 1961175.121. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

lzbench

lzbench is an in-memory benchmark of various compressors. The file used for compression is a Linux kernel source tree tarball. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Zstd 8 - Process: Compression123420406080100838483831. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU123448121620SE +/- 0.02, N = 3SE +/- 0.02, N = 3SE +/- 0.01, N = 3SE +/- 0.03, N = 316.4116.3016.4916.38MIN: 15.36MIN: 15.26MIN: 15.39MIN: 15.261. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU123448121620Min: 16.4 / Avg: 16.41 / Max: 16.45Min: 16.27 / Avg: 16.3 / Max: 16.33Min: 16.47 / Avg: 16.49 / Max: 16.51Min: 16.35 / Avg: 16.38 / Max: 16.431. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

ASTC Encoder

ASTC Encoder (astcenc) is for the Adaptive Scalable Texture Compression (ASTC) format commonly used with OpenGL, OpenGL ES, and Vulkan graphics APIs. This test profile does a coding test of both compression/decompression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Fast1243691215SE +/- 0.07, N = 3SE +/- 0.07, N = 3SE +/- 0.07, N = 39.429.539.531. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Fast1243691215Min: 9.32 / Avg: 9.42 / Max: 9.55Min: 9.39 / Avg: 9.53 / Max: 9.63Min: 9.39 / Avg: 9.53 / Max: 9.621. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

rav1e

Xiph rav1e is a Rust-written AV1 video encoder. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4Speed: 61240.31250.6250.93751.251.5625SE +/- 0.006, N = 3SE +/- 0.008, N = 3SE +/- 0.010, N = 31.3771.3731.389
OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4Speed: 6124246810Min: 1.37 / Avg: 1.38 / Max: 1.38Min: 1.36 / Avg: 1.37 / Max: 1.39Min: 1.37 / Avg: 1.39 / Max: 1.4

OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4Speed: 51240.23630.47260.70890.94521.1815SE +/- 0.007, N = 3SE +/- 0.002, N = 3SE +/- 0.017, N = 31.0381.0501.049
OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4Speed: 5124246810Min: 1.03 / Avg: 1.04 / Max: 1.05Min: 1.05 / Avg: 1.05 / Max: 1.05Min: 1.02 / Avg: 1.05 / Max: 1.07

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU12341.22332.44663.66994.89326.1165SE +/- 0.00277, N = 3SE +/- 0.00432, N = 3SE +/- 0.00933, N = 3SE +/- 0.00360, N = 35.388945.377205.436835.42404MIN: 5.3MIN: 5.28MIN: 5.35MIN: 5.351. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU1234246810Min: 5.38 / Avg: 5.39 / Max: 5.39Min: 5.37 / Avg: 5.38 / Max: 5.39Min: 5.42 / Avg: 5.44 / Max: 5.45Min: 5.42 / Avg: 5.42 / Max: 5.431. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.1Model: resnet-v2-501241122334455SE +/- 0.14, N = 3SE +/- 0.14, N = 3SE +/- 0.75, N = 348.8948.8948.37MIN: 48.41 / MAX: 50.57MIN: 48.4 / MAX: 141.21MIN: 46.64 / MAX: 52.551. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.1Model: resnet-v2-501241020304050Min: 48.66 / Avg: 48.89 / Max: 49.15Min: 48.74 / Avg: 48.89 / Max: 49.16Min: 46.87 / Avg: 48.37 / Max: 49.261. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU12343691215SE +/- 0.06, N = 3SE +/- 0.03, N = 3SE +/- 0.01, N = 3SE +/- 0.05, N = 312.5012.5712.4412.47MIN: 11.34MIN: 11.44MIN: 11.38MIN: 11.011. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU123448121620Min: 12.44 / Avg: 12.5 / Max: 12.61Min: 12.53 / Avg: 12.57 / Max: 12.61Min: 12.43 / Avg: 12.44 / Max: 12.47Min: 12.41 / Avg: 12.47 / Max: 12.571. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

Build2

This test profile measures the time to bootstrap/install the build2 C++ build toolchain from source. Build2 is a cross-platform build toolchain for C/C++ code and features Cargo-like features. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBuild2 0.13Time To Compile12480160240320400SE +/- 0.47, N = 3SE +/- 4.00, N = 3SE +/- 4.15, N = 3363.91367.31367.59
OpenBenchmarking.orgSeconds, Fewer Is BetterBuild2 0.13Time To Compile12470140210280350Min: 362.96 / Avg: 363.91 / Max: 364.44Min: 361.19 / Avg: 367.31 / Max: 374.82Min: 363.27 / Avg: 367.59 / Max: 375.88

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 1 - Decompression Speed12342K4K6K8K10KSE +/- 33.42, N = 3SE +/- 34.62, N = 3SE +/- 2.99, N = 3SE +/- 10.98, N = 37915.27837.57870.77867.11. (CC) gcc options: -O3
OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 1 - Decompression Speed123414002800420056007000Min: 7869.9 / Avg: 7915.17 / Max: 7980.4Min: 7769.1 / Avg: 7837.53 / Max: 7880.9Min: 7866.3 / Avg: 7870.67 / Max: 7876.4Min: 7848.7 / Avg: 7867.13 / Max: 7886.71. (CC) gcc options: -O3

Unpacking Firefox

This simple test profile measures how long it takes to extract the .tar.xz source package of the Mozilla Firefox Web Browser. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterUnpacking Firefox 84.0Extracting: firefox-84.0.source.tar.xz124612182430SE +/- 0.07, N = 4SE +/- 0.05, N = 4SE +/- 0.12, N = 423.6723.7223.49
OpenBenchmarking.orgSeconds, Fewer Is BetterUnpacking Firefox 84.0Extracting: firefox-84.0.source.tar.xz124612182430Min: 23.51 / Avg: 23.67 / Max: 23.78Min: 23.64 / Avg: 23.72 / Max: 23.86Min: 23.24 / Avg: 23.49 / Max: 23.77

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: CG.C12347001400210028003500SE +/- 0.98, N = 3SE +/- 3.60, N = 3SE +/- 3.85, N = 3SE +/- 1.69, N = 33438.093463.423448.623472.031. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: CG.C12346001200180024003000Min: 3436.27 / Avg: 3438.09 / Max: 3439.61Min: 3458.94 / Avg: 3463.42 / Max: 3470.54Min: 3441.53 / Avg: 3448.62 / Max: 3454.77Min: 3468.87 / Avg: 3472.03 / Max: 3474.671. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU1234246810SE +/- 0.00369, N = 3SE +/- 0.00963, N = 3SE +/- 0.03920, N = 3SE +/- 0.00650, N = 38.250578.229618.281098.20667MIN: 7.78MIN: 7.72MIN: 7.77MIN: 7.721. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU12343691215Min: 8.25 / Avg: 8.25 / Max: 8.26Min: 8.21 / Avg: 8.23 / Max: 8.24Min: 8.24 / Avg: 8.28 / Max: 8.36Min: 8.2 / Avg: 8.21 / Max: 8.221. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 2124306090120150SE +/- 0.12, N = 3SE +/- 0.44, N = 3SE +/- 0.11, N = 3119.20120.28119.621. (CXX) g++ options: -O3 -fPIC
OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 212420406080100Min: 119.03 / Avg: 119.2 / Max: 119.43Min: 119.59 / Avg: 120.28 / Max: 121.11Min: 119.49 / Avg: 119.62 / Max: 119.841. (CXX) g++ options: -O3 -fPIC

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.8.1Video Input: Chimera 1080p12470140210280350SE +/- 4.00, N = 3SE +/- 3.63, N = 3SE +/- 3.45, N = 3299.54302.25301.04MIN: 212.6 / MAX: 499.39MIN: 213.91 / MAX: 505.1MIN: 213.89 / MAX: 507.671. (CC) gcc options: -pthread
OpenBenchmarking.orgFPS, More Is Betterdav1d 0.8.1Video Input: Chimera 1080p12450100150200250Min: 292.51 / Avg: 299.54 / Max: 306.36Min: 295.3 / Avg: 302.25 / Max: 307.54Min: 295.49 / Avg: 301.04 / Max: 307.371. (CC) gcc options: -pthread

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: mnasnet124246810SE +/- 0.06, N = 3SE +/- 0.07, N = 3SE +/- 0.01, N = 36.746.736.79MIN: 6.58 / MAX: 7.21MIN: 6.56 / MAX: 7.06MIN: 6.72 / MAX: 6.941. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: mnasnet1243691215Min: 6.61 / Avg: 6.74 / Max: 6.8Min: 6.59 / Avg: 6.73 / Max: 6.81Min: 6.77 / Avg: 6.79 / Max: 6.81. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

LAMMPS Molecular Dynamics Simulator

LAMMPS is a classical molecular dynamics code, and an acronym for Large-scale Atomic/Molecular Massively Parallel Simulator. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgns/day, More Is BetterLAMMPS Molecular Dynamics Simulator 29Oct2020Model: Rhodopsin Protein12340.59361.18721.78082.37442.968SE +/- 0.030, N = 3SE +/- 0.001, N = 3SE +/- 0.007, N = 3SE +/- 0.006, N = 32.6152.6382.6292.6221. (CXX) g++ options: -O3 -pthread -lm
OpenBenchmarking.orgns/day, More Is BetterLAMMPS Molecular Dynamics Simulator 29Oct2020Model: Rhodopsin Protein1234246810Min: 2.56 / Avg: 2.62 / Max: 2.66Min: 2.64 / Avg: 2.64 / Max: 2.64Min: 2.62 / Avg: 2.63 / Max: 2.64Min: 2.61 / Avg: 2.62 / Max: 2.631. (CXX) g++ options: -O3 -pthread -lm

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.8.1Video Input: Summer Nature 1080p12460120180240300SE +/- 2.99, N = 3SE +/- 2.75, N = 3SE +/- 3.22, N = 6275.71277.67275.34MIN: 249.12 / MAX: 309.54MIN: 250.8 / MAX: 311.34MIN: 211.52 / MAX: 312.591. (CC) gcc options: -pthread
OpenBenchmarking.orgFPS, More Is Betterdav1d 0.8.1Video Input: Summer Nature 1080p12450100150200250Min: 272.12 / Avg: 275.71 / Max: 281.66Min: 274.74 / Avg: 277.67 / Max: 283.16Min: 261.19 / Avg: 275.34 / Max: 281.641. (CC) gcc options: -pthread

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU1234510152025SE +/- 0.01, N = 3SE +/- 0.02, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 322.6022.4322.4222.44MIN: 22.5MIN: 22.31MIN: 22.33MIN: 22.331. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU1234510152025Min: 22.58 / Avg: 22.6 / Max: 22.62Min: 22.39 / Avg: 22.43 / Max: 22.45Min: 22.4 / Avg: 22.42 / Max: 22.44Min: 22.42 / Avg: 22.44 / Max: 22.451. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 10124246810SE +/- 0.002, N = 3SE +/- 0.053, N = 3SE +/- 0.027, N = 37.8147.8677.8071. (CXX) g++ options: -O3 -fPIC
OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 101243691215Min: 7.81 / Avg: 7.81 / Max: 7.82Min: 7.81 / Avg: 7.87 / Max: 7.97Min: 7.76 / Avg: 7.81 / Max: 7.851. (CXX) g++ options: -O3 -fPIC

lzbench

lzbench is an in-memory benchmark of various compressors. The file used for compression is a Linux kernel source tree tarball. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Zstd 1 - Process: Decompression123430060090012001500SE +/- 2.65, N = 3SE +/- 0.67, N = 3SE +/- 7.97, N = 3SE +/- 2.85, N = 316051600159316041. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Zstd 1 - Process: Decompression123430060090012001500Min: 1600 / Avg: 1605 / Max: 1609Min: 1599 / Avg: 1600.33 / Max: 1601Min: 1577 / Avg: 1592.67 / Max: 1603Min: 1598 / Avg: 1603.67 / Max: 16071. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3

QuantLib

QuantLib is an open-source library/framework around quantitative finance for modeling, trading and risk management scenarios. QuantLib is written in C++ with Boost and its built-in benchmark used reports the QuantLib Benchmark Index benchmark score. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMFLOPS, More Is BetterQuantLib 1.2112345001000150020002500SE +/- 6.62, N = 3SE +/- 27.08, N = 5SE +/- 19.28, N = 12SE +/- 27.91, N = 52181.82168.32167.02165.51. (CXX) g++ options: -O3 -march=native -rdynamic
OpenBenchmarking.orgMFLOPS, More Is BetterQuantLib 1.211234400800120016002000Min: 2168.6 / Avg: 2181.83 / Max: 2188.7Min: 2060.2 / Avg: 2168.34 / Max: 2198.7Min: 1964.2 / Avg: 2167 / Max: 2199.3Min: 2053.9 / Avg: 2165.52 / Max: 2195.31. (CXX) g++ options: -O3 -march=native -rdynamic

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.8.1Video Input: Summer Nature 4K12420406080100SE +/- 0.77, N = 3SE +/- 0.27, N = 3SE +/- 0.42, N = 378.9479.5379.46MIN: 71.73 / MAX: 96.06MIN: 72.51 / MAX: 96.2MIN: 72.52 / MAX: 95.621. (CC) gcc options: -pthread
OpenBenchmarking.orgFPS, More Is Betterdav1d 0.8.1Video Input: Summer Nature 4K1241530456075Min: 77.56 / Avg: 78.94 / Max: 80.22Min: 79.12 / Avg: 79.53 / Max: 80.05Min: 78.73 / Avg: 79.46 / Max: 80.181. (CC) gcc options: -pthread

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU1234246810SE +/- 0.00175, N = 3SE +/- 0.00949, N = 3SE +/- 0.01684, N = 3SE +/- 0.02622, N = 37.415137.427227.444497.38944MIN: 6.64MIN: 6.63MIN: 6.63MIN: 6.621. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU12343691215Min: 7.41 / Avg: 7.42 / Max: 7.42Min: 7.41 / Avg: 7.43 / Max: 7.45Min: 7.42 / Avg: 7.44 / Max: 7.48Min: 7.34 / Avg: 7.39 / Max: 7.431. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

lzbench

lzbench is an in-memory benchmark of various compressors. The file used for compression is a Linux kernel source tree tarball. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Zstd 8 - Process: Decompression1234400800120016002000SE +/- 3.28, N = 3SE +/- 4.26, N = 3SE +/- 5.03, N = 3SE +/- 3.38, N = 317491750174917611. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Zstd 8 - Process: Decompression123430060090012001500Min: 1744 / Avg: 1748.67 / Max: 1755Min: 1742 / Avg: 1750.33 / Max: 1756Min: 1739 / Avg: 1749 / Max: 1755Min: 1754 / Avg: 1760.67 / Max: 17651. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: efficientnet-b01243691215SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.09, N = 310.9110.9110.98MIN: 10.85 / MAX: 11.97MIN: 10.67 / MAX: 11.08MIN: 10.67 / MAX: 59.781. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: efficientnet-b01243691215Min: 10.89 / Avg: 10.91 / Max: 10.93Min: 10.88 / Avg: 10.91 / Max: 10.92Min: 10.88 / Avg: 10.98 / Max: 11.161. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Monkey Audio Encoding

This test times how long it takes to encode a sample WAV file to Monkey's Audio APE format. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterMonkey Audio Encoding 3.99.6WAV To APE1243691215SE +/- 0.09, N = 5SE +/- 0.04, N = 5SE +/- 0.05, N = 513.2713.1913.221. (CXX) g++ options: -O3 -pedantic -rdynamic -lrt
OpenBenchmarking.orgSeconds, Fewer Is BetterMonkey Audio Encoding 3.99.6WAV To APE12448121620Min: 13 / Avg: 13.27 / Max: 13.54Min: 13.05 / Avg: 13.19 / Max: 13.3Min: 13.05 / Avg: 13.22 / Max: 13.31. (CXX) g++ options: -O3 -pedantic -rdynamic -lrt

LULESH

LULESH is the Livermore Unstructured Lagrangian Explicit Shock Hydrodynamics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgz/s, More Is BetterLULESH 2.0.3123430060090012001500SE +/- 0.58, N = 3SE +/- 0.89, N = 3SE +/- 2.48, N = 3SE +/- 0.65, N = 31172.901170.001165.741166.671. (CXX) g++ options: -O3 -fopenmp -lm -pthread -lmpi_cxx -lmpi
OpenBenchmarking.orgz/s, More Is BetterLULESH 2.0.312342004006008001000Min: 1171.9 / Avg: 1172.9 / Max: 1173.92Min: 1168.24 / Avg: 1170 / Max: 1171.12Min: 1160.8 / Avg: 1165.74 / Max: 1168.42Min: 1165.72 / Avg: 1166.67 / Max: 1167.921. (CXX) g++ options: -O3 -fopenmp -lm -pthread -lmpi_cxx -lmpi

lzbench

lzbench is an in-memory benchmark of various compressors. The file used for compression is a Linux kernel source tree tarball. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Brotli 2 - Process: Compression12344080120160200SE +/- 0.33, N = 31691701701701. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Brotli 2 - Process: Compression1234306090120150Min: 169 / Avg: 169.67 / Max: 1701. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU12343691215SE +/- 0.03159, N = 3SE +/- 0.02455, N = 3SE +/- 0.01670, N = 3SE +/- 0.01698, N = 39.449459.396979.432669.41427MIN: 8.38MIN: 8.24MIN: 8.21MIN: 8.171. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU12343691215Min: 9.4 / Avg: 9.45 / Max: 9.51Min: 9.36 / Avg: 9.4 / Max: 9.44Min: 9.4 / Avg: 9.43 / Max: 9.45Min: 9.39 / Avg: 9.41 / Max: 9.451. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 8124246810SE +/- 0.004, N = 3SE +/- 0.011, N = 3SE +/- 0.014, N = 38.4748.5128.5211. (CXX) g++ options: -O3 -fPIC
OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 81243691215Min: 8.47 / Avg: 8.47 / Max: 8.48Min: 8.49 / Avg: 8.51 / Max: 8.53Min: 8.5 / Avg: 8.52 / Max: 8.551. (CXX) g++ options: -O3 -fPIC

Pennant

Pennant is an application focused on hydrodynamics on general unstructured meshes in 2D. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgHydro Cycle Time - Seconds, Fewer Is BetterPennant 1.0.1Test: leblancbig123420406080100SE +/- 0.05, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.03, N = 3105.06105.64105.46105.571. (CXX) g++ options: -fopenmp -pthread -lmpi_cxx -lmpi
OpenBenchmarking.orgHydro Cycle Time - Seconds, Fewer Is BetterPennant 1.0.1Test: leblancbig123420406080100Min: 105.01 / Avg: 105.06 / Max: 105.16Min: 105.63 / Avg: 105.64 / Max: 105.65Min: 105.44 / Avg: 105.46 / Max: 105.49Min: 105.53 / Avg: 105.57 / Max: 105.621. (CXX) g++ options: -fopenmp -pthread -lmpi_cxx -lmpi

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.1Model: SqueezeNetV1.0124246810SE +/- 0.010, N = 3SE +/- 0.025, N = 3SE +/- 0.076, N = 37.8117.7897.771MIN: 7.14 / MAX: 13.45MIN: 7.12 / MAX: 8.86MIN: 7.02 / MAX: 8.971. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.1Model: SqueezeNetV1.01243691215Min: 7.8 / Avg: 7.81 / Max: 7.83Min: 7.74 / Avg: 7.79 / Max: 7.82Min: 7.62 / Avg: 7.77 / Max: 7.851. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.D1234110220330440550SE +/- 6.77, N = 4SE +/- 5.28, N = 9SE +/- 8.69, N = 3SE +/- 5.41, N = 7504.27503.55502.03501.691. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.D123490180270360450Min: 483.96 / Avg: 504.27 / Max: 511.22Min: 467.97 / Avg: 503.55 / Max: 512.61Min: 484.65 / Avg: 502.03 / Max: 511.23Min: 480.64 / Avg: 501.69 / Max: 512.761. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20210126Encode Settings: Default1243691215SE +/- 0.02, N = 3SE +/- 0.03, N = 3SE +/- 0.01, N = 310.5310.5110.561. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -rdynamic -lpthread -ljpeg
OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20210126Encode Settings: Default1243691215Min: 10.5 / Avg: 10.53 / Max: 10.57Min: 10.47 / Avg: 10.51 / Max: 10.56Min: 10.54 / Avg: 10.56 / Max: 10.581. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -rdynamic -lpthread -ljpeg

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: SP.B12349001800270036004500SE +/- 0.56, N = 3SE +/- 0.54, N = 3SE +/- 1.21, N = 3SE +/- 2.68, N = 34409.824387.594391.744391.951. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: SP.B12348001600240032004000Min: 4408.75 / Avg: 4409.82 / Max: 4410.64Min: 4386.57 / Avg: 4387.59 / Max: 4388.38Min: 4390.33 / Avg: 4391.74 / Max: 4394.15Min: 4387.69 / Avg: 4391.95 / Max: 4396.911. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.6Model: yolov4 - Device: OpenMP CPU12450100150200250SE +/- 0.17, N = 3SE +/- 0.29, N = 3SE +/- 0.29, N = 32072072081. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt
OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.6Model: yolov4 - Device: OpenMP CPU1244080120160200Min: 207 / Avg: 207.17 / Max: 207.5Min: 206.5 / Avg: 207 / Max: 207.5Min: 207 / Avg: 207.5 / Max: 2081. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt

lzbench

lzbench is an in-memory benchmark of various compressors. The file used for compression is a Linux kernel source tree tarball. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Libdeflate 1 - Process: Compression123450100150200250SE +/- 0.33, N = 3SE +/- 0.33, N = 3SE +/- 0.33, N = 32082092092081. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Libdeflate 1 - Process: Compression12344080120160200Min: 208 / Avg: 208.67 / Max: 209Min: 208 / Avg: 208.67 / Max: 209Min: 208 / Avg: 208.33 / Max: 2091. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.1Model: mobilenet-v1-1.01241.1412.2823.4234.5645.705SE +/- 0.015, N = 3SE +/- 0.018, N = 3SE +/- 0.029, N = 35.0715.0705.047MIN: 4.59 / MAX: 6.1MIN: 4.58 / MAX: 5.75MIN: 4.56 / MAX: 5.571. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.1Model: mobilenet-v1-1.0124246810Min: 5.04 / Avg: 5.07 / Max: 5.09Min: 5.04 / Avg: 5.07 / Max: 5.1Min: 4.99 / Avg: 5.05 / Max: 5.081. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 9 - Compression Speed12341020304050SE +/- 0.09, N = 3SE +/- 0.17, N = 3SE +/- 0.02, N = 3SE +/- 0.13, N = 344.5244.4244.6144.401. (CC) gcc options: -O3
OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 9 - Compression Speed1234918273645Min: 44.34 / Avg: 44.52 / Max: 44.65Min: 44.09 / Avg: 44.42 / Max: 44.59Min: 44.57 / Avg: 44.61 / Max: 44.65Min: 44.15 / Avg: 44.4 / Max: 44.61. (CC) gcc options: -O3

Coremark

This is a test of EEMBC CoreMark processor benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations/Sec, More Is BetterCoremark 1.0CoreMark Size 666 - Iterations Per Second12430K60K90K120K150KSE +/- 574.00, N = 3SE +/- 313.64, N = 3SE +/- 645.39, N = 3137084.98136732.15137374.611. (CC) gcc options: -O2 -lrt" -lrt
OpenBenchmarking.orgIterations/Sec, More Is BetterCoremark 1.0CoreMark Size 666 - Iterations Per Second12420K40K60K80K100KMin: 136384.95 / Avg: 137084.98 / Max: 138222.97Min: 136158.62 / Avg: 136732.15 / Max: 137238.92Min: 136204.99 / Avg: 137374.61 / Max: 138432.251. (CC) gcc options: -O2 -lrt" -lrt

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 3 - Decompression Speed12342K4K6K8K10KSE +/- 11.01, N = 3SE +/- 9.12, N = 3SE +/- 6.77, N = 3SE +/- 14.70, N = 37869.77834.37840.27851.01. (CC) gcc options: -O3
OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 3 - Decompression Speed123414002800420056007000Min: 7853.7 / Avg: 7869.7 / Max: 7890.8Min: 7819.2 / Avg: 7834.27 / Max: 7850.7Min: 7831 / Avg: 7840.2 / Max: 7853.4Min: 7833.5 / Avg: 7851 / Max: 7880.21. (CC) gcc options: -O3

OpenFOAM

OpenFOAM is the leading free, open source software for computational fluid dynamics (CFD). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenFOAM 8Input: Motorbike 30M123480160240320400SE +/- 0.17, N = 3SE +/- 0.21, N = 3SE +/- 0.53, N = 3SE +/- 0.33, N = 3386.71387.68388.45388.021. (CXX) g++ options: -std=c++11 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lspecie -lfiniteVolume -lfvOptions -lgenericPatchFields -lmeshTools -lsampling -lOpenFOAM -ldl -lm
OpenBenchmarking.orgSeconds, Fewer Is BetterOpenFOAM 8Input: Motorbike 30M123470140210280350Min: 386.48 / Avg: 386.71 / Max: 387.05Min: 387.46 / Avg: 387.68 / Max: 388.09Min: 387.67 / Avg: 388.45 / Max: 389.45Min: 387.44 / Avg: 388.02 / Max: 388.581. (CXX) g++ options: -std=c++11 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lspecie -lfiniteVolume -lfvOptions -lgenericPatchFields -lmeshTools -lsampling -lOpenFOAM -ldl -lm

lzbench

lzbench is an in-memory benchmark of various compressors. The file used for compression is a Linux kernel source tree tarball. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Zstd 1 - Process: Compression1234100200300400500SE +/- 1.15, N = 3SE +/- 1.33, N = 3SE +/- 0.33, N = 34534514524531. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Zstd 1 - Process: Compression123480160240320400Min: 451 / Avg: 453 / Max: 455Min: 448 / Avg: 450.67 / Max: 452Min: 453 / Avg: 453.33 / Max: 4541. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20210126Encode Settings: Quality 75, Compression Effort 7124140280420560700SE +/- 0.26, N = 3SE +/- 0.75, N = 3SE +/- 1.61, N = 3639.61642.37641.291. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -rdynamic -lpthread -ljpeg
OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20210126Encode Settings: Quality 75, Compression Effort 7124110220330440550Min: 639.1 / Avg: 639.61 / Max: 639.94Min: 640.87 / Avg: 642.37 / Max: 643.24Min: 638.83 / Avg: 641.29 / Max: 644.311. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -rdynamic -lpthread -ljpeg

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: vgg1612420406080100SE +/- 0.11, N = 3SE +/- 0.08, N = 3SE +/- 0.10, N = 389.2288.8688.84MIN: 88.84 / MAX: 148.87MIN: 88.59 / MAX: 91.44MIN: 88.59 / MAX: 89.771. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: vgg1612420406080100Min: 89.09 / Avg: 89.22 / Max: 89.43Min: 88.75 / Avg: 88.86 / Max: 89.02Min: 88.72 / Avg: 88.84 / Max: 89.041. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

WavPack Audio Encoding

This test times how long it takes to encode a sample WAV file to WavPack format with very high quality settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWavPack Audio Encoding 5.3WAV To WavPack12448121620SE +/- 0.07, N = 5SE +/- 0.05, N = 5SE +/- 0.04, N = 517.0116.9416.931. (CXX) g++ options: -rdynamic
OpenBenchmarking.orgSeconds, Fewer Is BetterWavPack Audio Encoding 5.3WAV To WavPack12448121620Min: 16.82 / Avg: 17.01 / Max: 17.23Min: 16.83 / Avg: 16.94 / Max: 17.03Min: 16.83 / Avg: 16.93 / Max: 17.021. (CXX) g++ options: -rdynamic

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.8.1Video Input: Chimera 1080p 10-bit1241530456075SE +/- 0.57, N = 3SE +/- 0.58, N = 3SE +/- 0.39, N = 365.1065.0465.31MIN: 44.13 / MAX: 149.8MIN: 44.01 / MAX: 149.22MIN: 44.34 / MAX: 148.991. (CC) gcc options: -pthread
OpenBenchmarking.orgFPS, More Is Betterdav1d 0.8.1Video Input: Chimera 1080p 10-bit1241326395265Min: 64.35 / Avg: 65.1 / Max: 66.21Min: 64.37 / Avg: 65.04 / Max: 66.19Min: 64.86 / Avg: 65.31 / Max: 66.091. (CC) gcc options: -pthread

Timed Eigen Compilation

This test times how long it takes to build all Eigen examples. The Eigen examples are compiled serially. Eigen is a C++ template library for linear algebra. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Eigen Compilation 3.3.9Time To Compile12420406080100SE +/- 0.03, N = 3SE +/- 0.13, N = 3SE +/- 0.08, N = 389.8990.2690.19
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Eigen Compilation 3.3.9Time To Compile12420406080100Min: 89.84 / Avg: 89.89 / Max: 89.94Min: 90.09 / Avg: 90.26 / Max: 90.52Min: 90.03 / Avg: 90.19 / Max: 90.3

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: shufflenet-v21243691215SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 39.789.809.76MIN: 9.42 / MAX: 10.68MIN: 9.44 / MAX: 10.88MIN: 9.43 / MAX: 10.041. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: shufflenet-v21243691215Min: 9.76 / Avg: 9.78 / Max: 9.8Min: 9.78 / Avg: 9.8 / Max: 9.82Min: 9.75 / Avg: 9.76 / Max: 9.771. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: BT.C12343K6K9K12K15KSE +/- 8.84, N = 3SE +/- 19.55, N = 3SE +/- 6.18, N = 3SE +/- 5.42, N = 313485.4513518.0413518.0113540.521. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: BT.C12342K4K6K8K10KMin: 13469.98 / Avg: 13485.45 / Max: 13500.61Min: 13496.09 / Avg: 13518.04 / Max: 13557.03Min: 13507.62 / Avg: 13518.01 / Max: 13528.99Min: 13531.82 / Avg: 13540.52 / Max: 13550.471. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3

Pennant

Pennant is an application focused on hydrodynamics on general unstructured meshes in 2D. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgHydro Cycle Time - Seconds, Fewer Is BetterPennant 1.0.1Test: sedovbig12344080120160200SE +/- 0.13, N = 3SE +/- 0.04, N = 3SE +/- 0.21, N = 3SE +/- 0.17, N = 3172.99173.69173.69173.531. (CXX) g++ options: -fopenmp -pthread -lmpi_cxx -lmpi
OpenBenchmarking.orgHydro Cycle Time - Seconds, Fewer Is BetterPennant 1.0.1Test: sedovbig1234306090120150Min: 172.74 / Avg: 172.99 / Max: 173.13Min: 173.62 / Avg: 173.69 / Max: 173.77Min: 173.4 / Avg: 173.69 / Max: 174.11Min: 173.2 / Avg: 173.53 / Max: 173.71. (CXX) g++ options: -fopenmp -pthread -lmpi_cxx -lmpi

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU1234510152025SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 321.3821.3021.3521.32MIN: 21.12MIN: 21.2MIN: 21.27MIN: 21.231. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU1234510152025Min: 21.36 / Avg: 21.38 / Max: 21.4Min: 21.29 / Avg: 21.3 / Max: 21.31Min: 21.34 / Avg: 21.35 / Max: 21.37Min: 21.31 / Avg: 21.32 / Max: 21.341. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

Cryptsetup

This is a test profile for running the cryptsetup benchmark to report on the system's cryptography performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupTwofish-XTS 512b Decryption12490180270360450SE +/- 1.38, N = 3SE +/- 0.26, N = 3SE +/- 0.25, N = 2393.0394.4394.5
OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupTwofish-XTS 512b Decryption12470140210280350Min: 390.2 / Avg: 392.97 / Max: 394.4Min: 393.9 / Avg: 394.4 / Max: 394.8Min: 394.2 / Avg: 394.45 / Max: 394.7

Kripke

Kripke is a simple, scalable, 3D Sn deterministic particle transport code. Its primary purpose is to research how data layout, programming paradigms and architectures effect the implementation and performance of Sn transport. Kripke is developed by LLNL. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgThroughput FoM, More Is BetterKripke 1.2.41245M10M15M20M25MSE +/- 307740.69, N = 3SE +/- 267442.30, N = 3SE +/- 248353.03, N = 32483486324915507249255531. (CXX) g++ options: -O3 -fopenmp
OpenBenchmarking.orgThroughput FoM, More Is BetterKripke 1.2.41244M8M12M16M20MMin: 24279470 / Avg: 24834863.33 / Max: 25342270Min: 24455540 / Avg: 24915506.67 / Max: 25381920Min: 24505050 / Avg: 24925553.33 / Max: 253647601. (CXX) g++ options: -O3 -fopenmp

Cryptsetup

This is a test profile for running the cryptsetup benchmark to report on the system's cryptography performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupAES-XTS 256b Decryption124400800120016002000SE +/- 2.09, N = 3SE +/- 3.87, N = 3SE +/- 5.93, N = 32051.02055.02047.6
OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupAES-XTS 256b Decryption124400800120016002000Min: 2047.8 / Avg: 2050.97 / Max: 2054.9Min: 2047.6 / Avg: 2055.03 / Max: 2060.6Min: 2039.6 / Avg: 2047.63 / Max: 2059.2

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU123410002000300040005000SE +/- 4.24, N = 3SE +/- 4.30, N = 3SE +/- 2.60, N = 3SE +/- 3.31, N = 34631.894628.544645.084632.41MIN: 4615.88MIN: 4606.88MIN: 4629.78MIN: 4614.331. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU12348001600240032004000Min: 4627.46 / Avg: 4631.89 / Max: 4640.37Min: 4619.93 / Avg: 4628.54 / Max: 4632.88Min: 4641.37 / Avg: 4645.08 / Max: 4650.09Min: 4628.75 / Avg: 4632.41 / Max: 4639.021. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

Cryptsetup

This is a test profile for running the cryptsetup benchmark to report on the system's cryptography performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupAES-XTS 256b Encryption124400800120016002000SE +/- 3.45, N = 3SE +/- 2.61, N = 3SE +/- 3.88, N = 32043.42049.52042.3
OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupAES-XTS 256b Encryption124400800120016002000Min: 2036.5 / Avg: 2043.4 / Max: 2047.1Min: 2046.3 / Avg: 2049.53 / Max: 2054.7Min: 2037.6 / Avg: 2042.3 / Max: 2050

OpenBenchmarking.orgIterations Per Second, More Is BetterCryptsetupPBKDF2-sha512124300K600K900K1200K1500KSE +/- 2100.92, N = 3SE +/- 5508.00, N = 3157997915760541581562
OpenBenchmarking.orgIterations Per Second, More Is BetterCryptsetupPBKDF2-sha512124300K600K900K1200K1500KMin: 1576806 / Avg: 1579979 / Max: 1583951Min: 1565038 / Avg: 1576054 / Max: 1581562

TNN

TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.2.3Target: CPU - Model: MobileNet v212480160240320400SE +/- 0.30, N = 3SE +/- 0.12, N = 3SE +/- 0.79, N = 3363.11362.72363.98MIN: 361.8 / MAX: 364.48MIN: 361.88 / MAX: 365.57MIN: 362.08 / MAX: 366.51. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -O3 -rdynamic -ldl
OpenBenchmarking.orgms, Fewer Is BetterTNN 0.2.3Target: CPU - Model: MobileNet v212470140210280350Min: 362.58 / Avg: 363.11 / Max: 363.63Min: 362.52 / Avg: 362.72 / Max: 362.94Min: 363.03 / Avg: 363.98 / Max: 365.541. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -O3 -rdynamic -ldl

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 1 - Compression Speed123414002800420056007000SE +/- 4.14, N = 3SE +/- 9.69, N = 3SE +/- 2.97, N = 3SE +/- 5.66, N = 36691.286682.996668.816669.471. (CC) gcc options: -O3
OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 1 - Compression Speed123412002400360048006000Min: 6683.29 / Avg: 6691.28 / Max: 6697.15Min: 6666.73 / Avg: 6682.99 / Max: 6700.26Min: 6663.22 / Avg: 6668.81 / Max: 6673.36Min: 6658.23 / Avg: 6669.47 / Max: 6676.221. (CC) gcc options: -O3

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.6Model: bertsquad-10 - Device: OpenMP CPU12470140210280350SE +/- 0.29, N = 3SE +/- 0.44, N = 3SE +/- 0.50, N = 33123113111. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt
OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.6Model: bertsquad-10 - Device: OpenMP CPU12460120180240300Min: 311 / Avg: 311.5 / Max: 312Min: 310 / Avg: 310.83 / Max: 311.5Min: 309.5 / Avg: 310.5 / Max: 3111. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.6Model: super-resolution-10 - Device: OpenMP CPU1245001000150020002500SE +/- 4.48, N = 3SE +/- 2.17, N = 3SE +/- 1.48, N = 32231222422291. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt
OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.6Model: super-resolution-10 - Device: OpenMP CPU124400800120016002000Min: 2222.5 / Avg: 2231.17 / Max: 2237.5Min: 2220 / Avg: 2223.67 / Max: 2227.5Min: 2226.5 / Avg: 2229.33 / Max: 2231.51. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt

lzbench

lzbench is an in-memory benchmark of various compressors. The file used for compression is a Linux kernel source tree tarball. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Brotli 2 - Process: Decompression1234140280420560700SE +/- 0.58, N = 36486486486461. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Brotli 2 - Process: Decompression1234110220330440550Min: 647 / Avg: 648 / Max: 6491. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 3 - Compression Speed12341020304050SE +/- 0.00, N = 3SE +/- 0.03, N = 3SE +/- 0.07, N = 3SE +/- 0.09, N = 345.7845.7445.7045.641. (CC) gcc options: -O3
OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 3 - Compression Speed1234918273645Min: 45.77 / Avg: 45.78 / Max: 45.78Min: 45.7 / Avg: 45.74 / Max: 45.8Min: 45.57 / Avg: 45.7 / Max: 45.8Min: 45.47 / Avg: 45.64 / Max: 45.741. (CC) gcc options: -O3

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU12342K4K6K8K10KSE +/- 5.71, N = 3SE +/- 2.41, N = 3SE +/- 2.21, N = 3SE +/- 1.31, N = 38917.528925.458944.358938.37MIN: 8892.85MIN: 8909.78MIN: 8929.69MIN: 8921.681. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU123416003200480064008000Min: 8906.35 / Avg: 8917.52 / Max: 8925.17Min: 8922.65 / Avg: 8925.45 / Max: 8930.26Min: 8940.58 / Avg: 8944.35 / Max: 8948.24Min: 8935.76 / Avg: 8938.37 / Max: 8939.771. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: mobilenet124714212835SE +/- 0.07, N = 3SE +/- 0.10, N = 3SE +/- 0.06, N = 330.4430.4530.36MIN: 29.84 / MAX: 31.45MIN: 29.81 / MAX: 39.56MIN: 29.76 / MAX: 31.321. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: mobilenet124714212835Min: 30.31 / Avg: 30.44 / Max: 30.56Min: 30.29 / Avg: 30.45 / Max: 30.63Min: 30.24 / Avg: 30.36 / Max: 30.431. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Cryptsetup

This is a test profile for running the cryptsetup benchmark to report on the system's cryptography performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupAES-XTS 512b Encryption124400800120016002000SE +/- 1.43, N = 3SE +/- 2.73, N = 3SE +/- 4.40, N = 31834.61838.31832.9
OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupAES-XTS 512b Encryption12430060090012001500Min: 1831.7 / Avg: 1834.57 / Max: 1836.1Min: 1833.9 / Avg: 1838.3 / Max: 1843.3Min: 1826.2 / Avg: 1832.93 / Max: 1841.2

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 01244080120160200SE +/- 0.71, N = 3SE +/- 0.20, N = 3SE +/- 0.48, N = 3203.25203.85203.681. (CXX) g++ options: -O3 -fPIC
OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 01244080120160200Min: 201.84 / Avg: 203.25 / Max: 204.08Min: 203.61 / Avg: 203.85 / Max: 204.24Min: 203.07 / Avg: 203.68 / Max: 204.621. (CXX) g++ options: -O3 -fPIC

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU123410002000300040005000SE +/- 7.28, N = 3SE +/- 5.57, N = 3SE +/- 5.16, N = 3SE +/- 3.98, N = 34640.834630.474643.814640.34MIN: 4620.66MIN: 4609.21MIN: 4625.36MIN: 4617.821. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU12348001600240032004000Min: 4631.78 / Avg: 4640.83 / Max: 4655.24Min: 4624.52 / Avg: 4630.47 / Max: 4641.6Min: 4636.18 / Avg: 4643.81 / Max: 4653.63Min: 4636.15 / Avg: 4640.34 / Max: 4648.31. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

rav1e

Xiph rav1e is a Rust-written AV1 video encoder. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4Speed: 11240.08150.1630.24450.3260.4075SE +/- 0.001, N = 3SE +/- 0.000, N = 3SE +/- 0.001, N = 30.3620.3620.361
OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4Speed: 112412345Min: 0.36 / Avg: 0.36 / Max: 0.36Min: 0.36 / Avg: 0.36 / Max: 0.36Min: 0.36 / Avg: 0.36 / Max: 0.36

Gcrypt Library

Libgcrypt is a general purpose cryptographic library developed as part of the GnuPG project. This is a benchmark of libgcrypt's integrated benchmark and is measuring the time to run the benchmark command with a cipher/mac/hash repetition count set for 50 times as simple, high level look at the overall crypto performance of the system under test. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterGcrypt Library 1.912450100150200250SE +/- 0.59, N = 3SE +/- 0.86, N = 3SE +/- 0.63, N = 3233.10232.51232.481. (CC) gcc options: -O2 -fvisibility=hidden -lgpg-error
OpenBenchmarking.orgSeconds, Fewer Is BetterGcrypt Library 1.91244080120160200Min: 231.93 / Avg: 233.1 / Max: 233.85Min: 231.22 / Avg: 232.51 / Max: 234.13Min: 231.43 / Avg: 232.48 / Max: 233.621. (CC) gcc options: -O2 -fvisibility=hidden -lgpg-error

GnuPG

This test times how long it takes to encrypt a sample file using GnuPG. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterGnuPG 2.2.272.7GB Sample File Encryption1241632486480SE +/- 0.81, N = 3SE +/- 0.55, N = 3SE +/- 0.59, N = 373.3073.1173.121. (CC) gcc options: -O2
OpenBenchmarking.orgSeconds, Fewer Is BetterGnuPG 2.2.272.7GB Sample File Encryption1241428425670Min: 72.33 / Avg: 73.3 / Max: 74.91Min: 72.47 / Avg: 73.11 / Max: 74.19Min: 72.49 / Avg: 73.12 / Max: 74.311. (CC) gcc options: -O2

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU12342K4K6K8K10KSE +/- 2.73, N = 3SE +/- 2.10, N = 3SE +/- 4.02, N = 3SE +/- 2.31, N = 38915.858919.798939.188931.43MIN: 8864.25MIN: 8896.59MIN: 8893.94MIN: 8883.841. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU123416003200480064008000Min: 8910.5 / Avg: 8915.85 / Max: 8919.43Min: 8915.75 / Avg: 8919.79 / Max: 8922.81Min: 8934.81 / Avg: 8939.18 / Max: 8947.2Min: 8926.98 / Avg: 8931.43 / Max: 8934.741. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

Cryptsetup

This is a test profile for running the cryptsetup benchmark to report on the system's cryptography performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupAES-XTS 512b Decryption124400800120016002000SE +/- 2.21, N = 3SE +/- 2.49, N = 3SE +/- 5.05, N = 31835.91838.11833.4
OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupAES-XTS 512b Decryption12430060090012001500Min: 1832.6 / Avg: 1835.9 / Max: 1840.1Min: 1834 / Avg: 1838.13 / Max: 1842.6Min: 1826.8 / Avg: 1833.37 / Max: 1843.3

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU12340.96551.9312.89653.8624.8275SE +/- 0.00558, N = 3SE +/- 0.00482, N = 3SE +/- 0.00537, N = 3SE +/- 0.00642, N = 34.286594.284984.291284.28045MIN: 3.8MIN: 3.79MIN: 3.8MIN: 3.81. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU1234246810Min: 4.28 / Avg: 4.29 / Max: 4.3Min: 4.28 / Avg: 4.28 / Max: 4.29Min: 4.28 / Avg: 4.29 / Max: 4.3Min: 4.27 / Avg: 4.28 / Max: 4.291. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: yolov4-tiny124918273645SE +/- 0.04, N = 3SE +/- 0.08, N = 3SE +/- 0.02, N = 340.9540.9640.86MIN: 40.63 / MAX: 41.77MIN: 40.51 / MAX: 41.85MIN: 40.54 / MAX: 41.981. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: yolov4-tiny124918273645Min: 40.91 / Avg: 40.95 / Max: 41.02Min: 40.86 / Avg: 40.96 / Max: 41.11Min: 40.83 / Avg: 40.86 / Max: 40.881. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

ASTC Encoder

ASTC Encoder (astcenc) is for the Adaptive Scalable Texture Compression (ASTC) format commonly used with OpenGL, OpenGL ES, and Vulkan graphics APIs. This test profile does a coding test of both compression/decompression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Medium1243691215SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 312.3012.3112.281. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Medium12448121620Min: 12.28 / Avg: 12.3 / Max: 12.31Min: 12.29 / Avg: 12.31 / Max: 12.32Min: 12.26 / Avg: 12.28 / Max: 12.291. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

Redis

Redis is an open-source in-memory data structure store, used as a database, cache, and message broker. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: SET124400K800K1200K1600K2000KSE +/- 11364.87, N = 3SE +/- 11294.99, N = 3SE +/- 7748.68, N = 31717799.751721937.791721260.041. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: SET124300K600K900K1200K1500KMin: 1704180 / Avg: 1717799.75 / Max: 1740369Min: 1699840.25 / Avg: 1721937.79 / Max: 1737049.12Min: 1712920.88 / Avg: 1721260.04 / Max: 17367421. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

lzbench

lzbench is an in-memory benchmark of various compressors. The file used for compression is a Linux kernel source tree tarball. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Brotli 0 - Process: Compression123490180270360450SE +/- 0.33, N = 34274274274281. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Brotli 0 - Process: Compression123480160240320400Min: 427 / Avg: 427.67 / Max: 4281. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3

Timed Godot Game Engine Compilation

This test times how long it takes to compile the Godot Game Engine. Godot is a popular, open-source, cross-platform 2D/3D game engine and is built using the SCons build system and targeting the X11 platform. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Godot Game Engine Compilation 3.2.3Time To Compile12480160240320400SE +/- 0.35, N = 3SE +/- 0.23, N = 3SE +/- 0.06, N = 3363.21363.92364.06
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Godot Game Engine Compilation 3.2.3Time To Compile12470140210280350Min: 362.67 / Avg: 363.21 / Max: 363.88Min: 363.47 / Avg: 363.92 / Max: 364.24Min: 363.95 / Avg: 364.06 / Max: 364.14

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: regnety_400m12448121620SE +/- 0.03, N = 3SE +/- 0.04, N = 3SE +/- 0.01, N = 317.5617.5217.53MIN: 17.44 / MAX: 18.56MIN: 17.39 / MAX: 18.59MIN: 17.46 / MAX: 18.51. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: regnety_400m12448121620Min: 17.5 / Avg: 17.56 / Max: 17.59Min: 17.44 / Avg: 17.52 / Max: 17.59Min: 17.52 / Avg: 17.53 / Max: 17.541. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU123410002000300040005000SE +/- 3.08, N = 3SE +/- 3.81, N = 3SE +/- 1.60, N = 3SE +/- 2.68, N = 34632.024628.674635.334639.21MIN: 4614.43MIN: 4609.9MIN: 4621.25MIN: 4621.741. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU12348001600240032004000Min: 4626.99 / Avg: 4632.02 / Max: 4637.62Min: 4622.17 / Avg: 4628.67 / Max: 4635.38Min: 4632.26 / Avg: 4635.33 / Max: 4637.64Min: 4634.32 / Avg: 4639.21 / Max: 4643.541. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: squeezenet_ssd124714212835SE +/- 0.05, N = 3SE +/- 0.02, N = 3SE +/- 0.01, N = 331.3031.2331.24MIN: 31.14 / MAX: 32.35MIN: 31.12 / MAX: 32.11MIN: 31.14 / MAX: 32.241. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: squeezenet_ssd124714212835Min: 31.22 / Avg: 31.3 / Max: 31.38Min: 31.2 / Avg: 31.23 / Max: 31.27Min: 31.22 / Avg: 31.24 / Max: 31.261. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Cryptsetup

This is a test profile for running the cryptsetup benchmark to report on the system's cryptography performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupTwofish-XTS 256b Encryption12480160240320400SE +/- 0.84, N = 3SE +/- 0.26, N = 3SE +/- 0.13, N = 3390.1390.9390.8
OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupTwofish-XTS 256b Encryption12470140210280350Min: 388.4 / Avg: 390.07 / Max: 391Min: 390.4 / Avg: 390.87 / Max: 391.3Min: 390.7 / Avg: 390.83 / Max: 391.1

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU12342K4K6K8K10KSE +/- 6.77, N = 3SE +/- 3.96, N = 3SE +/- 2.35, N = 3SE +/- 1.00, N = 38927.638933.088945.608936.31MIN: 8896MIN: 8916.93MIN: 8926.38MIN: 8921.311. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU123416003200480064008000Min: 8914.39 / Avg: 8927.63 / Max: 8936.7Min: 8927.3 / Avg: 8933.08 / Max: 8940.67Min: 8942.53 / Avg: 8945.6 / Max: 8950.22Min: 8934.61 / Avg: 8936.31 / Max: 8938.061. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20210126Encode Settings: Quality 95, Compression Effort 712430060090012001500SE +/- 0.76, N = 3SE +/- 0.85, N = 3SE +/- 1.46, N = 31178.691180.021177.831. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -rdynamic -lpthread -ljpeg
OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20210126Encode Settings: Quality 95, Compression Effort 71242004006008001000Min: 1177.73 / Avg: 1178.69 / Max: 1180.2Min: 1178.33 / Avg: 1180.02 / Max: 1181.08Min: 1174.93 / Avg: 1177.83 / Max: 1179.541. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -rdynamic -lpthread -ljpeg

lzbench

lzbench is an in-memory benchmark of various compressors. The file used for compression is a Linux kernel source tree tarball. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Brotli 0 - Process: Decompression1234120240360480600SE +/- 0.67, N = 3SE +/- 0.67, N = 35585585585571. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Brotli 0 - Process: Decompression1234100200300400500Min: 557 / Avg: 557.67 / Max: 559Min: 557 / Avg: 557.67 / Max: 5591. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: MG.C123415003000450060007500SE +/- 4.14, N = 3SE +/- 3.90, N = 3SE +/- 1.13, N = 3SE +/- 2.21, N = 37148.357153.827142.147142.461. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: MG.C123412002400360048006000Min: 7143.38 / Avg: 7148.35 / Max: 7156.58Min: 7148.21 / Avg: 7153.82 / Max: 7161.33Min: 7139.9 / Avg: 7142.14 / Max: 7143.52Min: 7138.06 / Avg: 7142.46 / Max: 7145.081. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.0.3

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU-v3-v3 - Model: mobilenet-v3124246810SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 36.546.546.53MIN: 6.44 / MAX: 6.79MIN: 6.45 / MAX: 6.8MIN: 6.42 / MAX: 9.191. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU-v3-v3 - Model: mobilenet-v31243691215Min: 6.52 / Avg: 6.54 / Max: 6.55Min: 6.53 / Avg: 6.54 / Max: 6.55Min: 6.51 / Avg: 6.53 / Max: 6.551. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Cryptsetup

This is a test profile for running the cryptsetup benchmark to report on the system's cryptography performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupSerpent-XTS 256b Decryption124160320480640800SE +/- 0.32, N = 3SE +/- 0.70, N = 3SE +/- 0.64, N = 3720.0720.1719.0
OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupSerpent-XTS 256b Decryption124130260390520650Min: 719.4 / Avg: 720.03 / Max: 720.4Min: 719 / Avg: 720.1 / Max: 721.4Min: 718 / Avg: 719.03 / Max: 720.2

OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupSerpent-XTS 256b Encryption124150300450600750SE +/- 1.02, N = 3SE +/- 0.22, N = 3SE +/- 0.67, N = 3704.8705.8705.7
OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupSerpent-XTS 256b Encryption124120240360480600Min: 702.8 / Avg: 704.83 / Max: 706Min: 705.4 / Avg: 705.83 / Max: 706.1Min: 704.9 / Avg: 705.67 / Max: 707

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: resnet501241122334455SE +/- 0.04, N = 3SE +/- 0.03, N = 3SE +/- 0.02, N = 346.6246.6646.68MIN: 45.3 / MAX: 48.43MIN: 45.36 / MAX: 47.7MIN: 45.41 / MAX: 48.981. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: resnet501241020304050Min: 46.54 / Avg: 46.62 / Max: 46.68Min: 46.6 / Avg: 46.66 / Max: 46.71Min: 46.65 / Avg: 46.68 / Max: 46.711. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU-v2-v2 - Model: mobilenet-v2124246810SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 37.997.997.98MIN: 7.82 / MAX: 8.88MIN: 7.82 / MAX: 8.7MIN: 7.83 / MAX: 8.111. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU-v2-v2 - Model: mobilenet-v21243691215Min: 7.97 / Avg: 7.99 / Max: 8.01Min: 7.97 / Avg: 7.99 / Max: 8.01Min: 7.97 / Avg: 7.98 / Max: 7.991. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

ASTC Encoder

ASTC Encoder (astcenc) is for the Adaptive Scalable Texture Compression (ASTC) format commonly used with OpenGL, OpenGL ES, and Vulkan graphics APIs. This test profile does a coding test of both compression/decompression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Thorough12420406080100SE +/- 0.05, N = 3SE +/- 0.05, N = 3SE +/- 0.06, N = 380.2780.3280.221. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Thorough1241530456075Min: 80.19 / Avg: 80.27 / Max: 80.35Min: 80.22 / Avg: 80.32 / Max: 80.38Min: 80.12 / Avg: 80.22 / Max: 80.321. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20210126Encode Settings: Quality 100, Lossless Compression1245001000150020002500SE +/- 1.91, N = 3SE +/- 0.23, N = 3SE +/- 0.76, N = 32398.272400.662397.751. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -rdynamic -lpthread -ljpeg
OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20210126Encode Settings: Quality 100, Lossless Compression124400800120016002000Min: 2395.71 / Avg: 2398.27 / Max: 2402Min: 2400.32 / Avg: 2400.66 / Max: 2401.1Min: 2396.25 / Avg: 2397.75 / Max: 2398.751. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -rdynamic -lpthread -ljpeg

Cryptsetup

This is a test profile for running the cryptsetup benchmark to report on the system's cryptography performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupSerpent-XTS 512b Decryption124160320480640800SE +/- 0.33, N = 3SE +/- 2.15, N = 2SE +/- 0.90, N = 3719.6719.4718.8
OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupSerpent-XTS 512b Decryption124130260390520650Min: 719 / Avg: 719.63 / Max: 720.1Min: 717.2 / Avg: 719.35 / Max: 721.5Min: 717.5 / Avg: 718.77 / Max: 720.5

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 9 - Decompression Speed12342K4K6K8K10KSE +/- 12.26, N = 3SE +/- 2.80, N = 3SE +/- 9.72, N = 3SE +/- 7.13, N = 37847.47844.07852.57843.81. (CC) gcc options: -O3
OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 9 - Decompression Speed123414002800420056007000Min: 7824.7 / Avg: 7847.4 / Max: 7866.8Min: 7839.2 / Avg: 7844 / Max: 7848.9Min: 7833.2 / Avg: 7852.5 / Max: 7864.1Min: 7829.5 / Avg: 7843.77 / Max: 78511. (CC) gcc options: -O3

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: alexnet124510152025SE +/- 0.02, N = 3SE +/- 0.00, N = 3SE +/- 0.02, N = 319.5119.5319.53MIN: 19.38 / MAX: 20.06MIN: 19.41 / MAX: 20.29MIN: 19.4 / MAX: 20.431. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: alexnet124510152025Min: 19.48 / Avg: 19.51 / Max: 19.53Min: 19.52 / Avg: 19.53 / Max: 19.53Min: 19.49 / Avg: 19.53 / Max: 19.561. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Cryptsetup

This is a test profile for running the cryptsetup benchmark to report on the system's cryptography performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupTwofish-XTS 256b Decryption12490180270360450SE +/- 0.09, N = 3SE +/- 0.23, N = 3SE +/- 0.24, N = 3394.1394.5394.4
OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupTwofish-XTS 256b Decryption12470140210280350Min: 393.9 / Avg: 394.07 / Max: 394.2Min: 394.1 / Avg: 394.47 / Max: 394.9Min: 393.9 / Avg: 394.37 / Max: 394.7

OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupSerpent-XTS 512b Encryption124150300450600750SE +/- 0.57, N = 3SE +/- 0.99, N = 3SE +/- 0.64, N = 3705.0705.7705.6
OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupSerpent-XTS 512b Encryption124120240360480600Min: 703.9 / Avg: 705 / Max: 705.8Min: 703.9 / Avg: 705.73 / Max: 707.3Min: 704.5 / Avg: 705.57 / Max: 706.7

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.6Model: shufflenet-v2-10 - Device: OpenMP CPU1243K6K9K12K15KSE +/- 43.67, N = 3SE +/- 12.85, N = 3SE +/- 40.07, N = 31410414103141171. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt
OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.6Model: shufflenet-v2-10 - Device: OpenMP CPU1242K4K6K8K10KMin: 14056 / Avg: 14104.33 / Max: 14191.5Min: 14087 / Avg: 14102.5 / Max: 14128Min: 14041 / Avg: 14117 / Max: 141771. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: resnet18124510152025SE +/- 0.03, N = 3SE +/- 0.01, N = 3SE +/- 0.03, N = 321.8821.9021.90MIN: 21.61 / MAX: 22.98MIN: 21.62 / MAX: 22.75MIN: 21.58 / MAX: 23.11. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: resnet18124510152025Min: 21.85 / Avg: 21.88 / Max: 21.94Min: 21.89 / Avg: 21.9 / Max: 21.92Min: 21.85 / Avg: 21.9 / Max: 21.961. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

TNN

TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.2.3Target: CPU - Model: SqueezeNet v1.112480160240320400SE +/- 0.06, N = 3SE +/- 0.23, N = 3SE +/- 0.13, N = 3345.56345.25345.34MIN: 344.76 / MAX: 347.02MIN: 344.1 / MAX: 347.01MIN: 344.17 / MAX: 346.641. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -O3 -rdynamic -ldl
OpenBenchmarking.orgms, Fewer Is BetterTNN 0.2.3Target: CPU - Model: SqueezeNet v1.112460120180240300Min: 345.49 / Avg: 345.55 / Max: 345.67Min: 344.99 / Avg: 345.25 / Max: 345.7Min: 345.07 / Avg: 345.34 / Max: 345.511. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -O3 -rdynamic -ldl

Cryptsetup

This is a test profile for running the cryptsetup benchmark to report on the system's cryptography performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Second, More Is BetterCryptsetupPBKDF2-whirlpool124140K280K420K560K700KSE +/- 281.67, N = 3SE +/- 564.33, N = 3665903665904665340
OpenBenchmarking.orgIterations Per Second, More Is BetterCryptsetupPBKDF2-whirlpool124120K240K360K480K600KMin: 665340 / Avg: 665903.33 / Max: 666185Min: 665340 / Avg: 665904.33 / Max: 667033

Google SynthMark

SynthMark is a cross platform tool for benchmarking CPU performance under a variety of real-time audio workloads. It uses a polyphonic synthesizer model to provide standardized tests for latency, jitter and computational throughput. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgVoices, More Is BetterGoogle SynthMark 20201109Test: VoiceMark_100124130260390520650SE +/- 0.60, N = 3SE +/- 1.48, N = 3SE +/- 1.02, N = 3603.25603.13602.751. (CXX) g++ options: -lm -lpthread -std=c++11 -Ofast
OpenBenchmarking.orgVoices, More Is BetterGoogle SynthMark 20201109Test: VoiceMark_100124110220330440550Min: 602.51 / Avg: 603.25 / Max: 604.45Min: 601.03 / Avg: 603.13 / Max: 606Min: 600.74 / Avg: 602.75 / Max: 604.041. (CXX) g++ options: -lm -lpthread -std=c++11 -Ofast

Cryptsetup

This is a test profile for running the cryptsetup benchmark to report on the system's cryptography performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupTwofish-XTS 512b Encryption12480160240320400SE +/- 0.15, N = 2SE +/- 0.15, N = 3SE +/- 0.20, N = 3390.8390.5390.8
OpenBenchmarking.orgMiB/s, More Is BetterCryptsetupTwofish-XTS 512b Encryption12470140210280350Min: 390.6 / Avg: 390.75 / Max: 390.9Min: 390.3 / Avg: 390.5 / Max: 390.8Min: 390.5 / Avg: 390.83 / Max: 391.2

Algebraic Multi-Grid Benchmark

AMG is a parallel algebraic multigrid solver for linear systems arising from problems on unstructured grids. The driver provided with AMG builds linear systems for various 3-dimensional problems. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFigure Of Merit, More Is BetterAlgebraic Multi-Grid Benchmark 1.2123440M80M120M160M200MSE +/- 27815.64, N = 3SE +/- 72288.18, N = 3SE +/- 87081.74, N = 3SE +/- 65783.92, N = 31806211001805828331805769001807050331. (CC) gcc options: -lparcsr_ls -lparcsr_mv -lseq_mv -lIJ_mv -lkrylov -lHYPRE_utilities -lm -fopenmp -pthread -lmpi
OpenBenchmarking.orgFigure Of Merit, More Is BetterAlgebraic Multi-Grid Benchmark 1.2123430M60M90M120M150MMin: 180572000 / Avg: 180621100 / Max: 180668300Min: 180469100 / Avg: 180582833.33 / Max: 180717000Min: 180467200 / Avg: 180576900 / Max: 180748900Min: 180573700 / Avg: 180705033.33 / Max: 1807775001. (CC) gcc options: -lparcsr_ls -lparcsr_mv -lseq_mv -lIJ_mv -lkrylov -lHYPRE_utilities -lm -fopenmp -pthread -lmpi

ASTC Encoder

ASTC Encoder (astcenc) is for the Adaptive Scalable Texture Compression (ASTC) format commonly used with OpenGL, OpenGL ES, and Vulkan graphics APIs. This test profile does a coding test of both compression/decompression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Exhaustive124140280420560700SE +/- 0.13, N = 3SE +/- 0.27, N = 3SE +/- 0.29, N = 3643.45643.88643.571. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Exhaustive124110220330440550Min: 643.23 / Avg: 643.45 / Max: 643.68Min: 643.45 / Avg: 643.88 / Max: 644.37Min: 643.22 / Avg: 643.57 / Max: 644.151. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20210126Encode Settings: Quality 100, Compression Effort 5124918273645SE +/- 0.00, N = 3SE +/- 0.02, N = 3SE +/- 0.01, N = 337.4437.4437.421. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -rdynamic -lpthread -ljpeg
OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20210126Encode Settings: Quality 100, Compression Effort 5124816243240Min: 37.43 / Avg: 37.44 / Max: 37.44Min: 37.4 / Avg: 37.44 / Max: 37.49Min: 37.41 / Avg: 37.42 / Max: 37.441. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -rdynamic -lpthread -ljpeg

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: googlenet124510152025SE +/- 0.04, N = 3SE +/- 0.05, N = 3SE +/- 0.01, N = 321.7521.7521.76MIN: 21.16 / MAX: 22.66MIN: 21.14 / MAX: 22.58MIN: 21.15 / MAX: 22.671. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: googlenet124510152025Min: 21.67 / Avg: 21.75 / Max: 21.8Min: 21.64 / Avg: 21.75 / Max: 21.8Min: 21.74 / Avg: 21.76 / Max: 21.781. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.6Model: fcn-resnet101-11 - Device: OpenMP CPU124816243240SE +/- 0.00, N = 3SE +/- 0.00, N = 3SE +/- 0.00, N = 33333331. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt
OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.6Model: fcn-resnet101-11 - Device: OpenMP CPU124714212835Min: 32.5 / Avg: 32.5 / Max: 32.5Min: 32.5 / Avg: 32.5 / Max: 32.5Min: 32.5 / Avg: 32.5 / Max: 32.51. (CXX) g++ options: -fopenmp -ffunction-sections -fdata-sections -O3 -ldl -lrt

lzbench

lzbench is an in-memory benchmark of various compressors. The file used for compression is a Linux kernel source tree tarball. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Crush 0 - Process: Decompression1234100200300400500SE +/- 0.33, N = 3SE +/- 0.33, N = 34474474474471. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3
OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: Crush 0 - Process: Decompression123480160240320400Min: 447 / Avg: 447.33 / Max: 448Min: 447 / Avg: 447.33 / Max: 4481. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3

OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: XZ 0 - Process: Decompression1234204060801001051051051051. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3

OpenBenchmarking.orgMB/s, More Is Betterlzbench 1.8Test: XZ 0 - Process: Compression1234918273645383838381. (CXX) g++ options: -pthread -fomit-frame-pointer -fstrict-aliasing -ffast-math -O3

133 Results Shown

Redis
oneDNN
Redis
oneDNN
NCNN
rav1e
lzbench
oneDNN
QMCPACK
Redis
NAS Parallel Benchmarks
Mobile Neural Network
FFTE
Opus Codec Encoding
NAS Parallel Benchmarks
Mobile Neural Network
NAS Parallel Benchmarks
Redis
lzbench
oneDNN
ASTC Encoder
rav1e:
  6
  5
oneDNN
Mobile Neural Network
oneDNN
Build2
LZ4 Compression
Unpacking Firefox
NAS Parallel Benchmarks
oneDNN
libavif avifenc
dav1d
NCNN
LAMMPS Molecular Dynamics Simulator
dav1d
oneDNN
libavif avifenc
lzbench
QuantLib
dav1d
oneDNN
lzbench
NCNN
Monkey Audio Encoding
LULESH
lzbench
oneDNN
libavif avifenc
Pennant
Mobile Neural Network
NAS Parallel Benchmarks
WebP2 Image Encode
NAS Parallel Benchmarks
ONNX Runtime
lzbench
Mobile Neural Network
LZ4 Compression
Coremark
LZ4 Compression
OpenFOAM
lzbench
WebP2 Image Encode
NCNN
WavPack Audio Encoding
dav1d
Timed Eigen Compilation
NCNN
NAS Parallel Benchmarks
Pennant
oneDNN
Cryptsetup
Kripke
Cryptsetup
oneDNN
Cryptsetup:
  AES-XTS 256b Encryption
  PBKDF2-sha512
TNN
LZ4 Compression
ONNX Runtime:
  bertsquad-10 - OpenMP CPU
  super-resolution-10 - OpenMP CPU
lzbench
LZ4 Compression
oneDNN
NCNN
Cryptsetup
libavif avifenc
oneDNN
rav1e
Gcrypt Library
GnuPG
oneDNN
Cryptsetup
oneDNN
NCNN
ASTC Encoder
Redis
lzbench
Timed Godot Game Engine Compilation
NCNN
oneDNN
NCNN
Cryptsetup
oneDNN
WebP2 Image Encode
lzbench
NAS Parallel Benchmarks
NCNN
Cryptsetup:
  Serpent-XTS 256b Decryption
  Serpent-XTS 256b Encryption
NCNN:
  CPU - resnet50
  CPU-v2-v2 - mobilenet-v2
ASTC Encoder
WebP2 Image Encode
Cryptsetup
LZ4 Compression
NCNN
Cryptsetup:
  Twofish-XTS 256b Decryption
  Serpent-XTS 512b Encryption
ONNX Runtime
NCNN
TNN
Cryptsetup
Google SynthMark
Cryptsetup
Algebraic Multi-Grid Benchmark
ASTC Encoder
WebP2 Image Encode
NCNN
ONNX Runtime
lzbench:
  Crush 0 - Decompression
  XZ 0 - Decompression
  XZ 0 - Compression