Core i7 5775C EOY 2020

Intel Core i7-5775C testing with a MSI Z97-G45 GAMING (MS-7821) v1.0 (V2.9 BIOS) and MSI Intel Iris Pro 6200 3GB on Ubuntu 18.10 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2012292-HA-COREI757721
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results

Limit displaying results to tests within:

Audio Encoding 4 Tests
Bioinformatics 2 Tests
BLAS (Basic Linear Algebra Sub-Routine) Tests 2 Tests
Chess Test Suite 3 Tests
Timed Code Compilation 3 Tests
C/C++ Compiler Tests 12 Tests
CPU Massive 16 Tests
Creator Workloads 16 Tests
Database Test Suite 3 Tests
Encoding 7 Tests
Fortran Tests 4 Tests
Game Development 3 Tests
HPC - High Performance Computing 13 Tests
Machine Learning 6 Tests
Molecular Dynamics 3 Tests
MPI Benchmarks 3 Tests
Multi-Core 14 Tests
NVIDIA GPU Compute 5 Tests
Intel oneAPI 2 Tests
OpenMPI Tests 3 Tests
Programmer / Developer System Benchmarks 5 Tests
Python 2 Tests
Scientific Computing 7 Tests
Server 4 Tests
Server CPU Tests 7 Tests
Single-Threaded 7 Tests
Speech 2 Tests
Telephony 2 Tests
Texture Compression 3 Tests
Video Encoding 3 Tests
Vulkan Compute 2 Tests
Common Workstation Benchmarks 2 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
1
December 28 2020
  9 Hours, 24 Minutes
2
December 28 2020
  9 Hours, 10 Minutes
3
December 29 2020
  9 Hours, 29 Minutes
Invert Hiding All Results Option
  9 Hours, 21 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


Core i7 5775C EOY 2020ProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerDisplay DriverOpenGLVulkanCompilerFile-SystemScreen Resolution123Intel Core i7-5775C @ 3.70GHz (4 Cores / 8 Threads)MSI Z97-G45 GAMING (MS-7821) v1.0 (V2.9 BIOS)Intel Broadwell-U DMI16GB120GB CT120BX100SSD1MSI Intel Iris Pro 6200 3GB (1150MHz)Intel Broadwell-U AudioVA2431Qualcomm Atheros Killer E220xUbuntu 18.105.0.0-999-generic (x86_64) 20190223GNOME Shell 3.30.2X Server 1.20.1modesetting 1.20.14.5 Mesa 19.2.0-devel (git-2631fd3 2019-07-24 cosmic-oibaf-ppa)1.1.102GCC 8.3.0ext41920x1080OpenBenchmarking.orgCompiler Details- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v Processor Details- Scaling Governor: intel_pstate powersave - CPU Microcode: 0x20 - Thermald 1.7Java Details- OpenJDK Runtime Environment (build 11.0.3+7-Ubuntu-1ubuntu218.10.1)Python Details- Python 2.7.16 + Python 3.6.8Security Details- l1tf: Mitigation of PTE Inversion; VMX: conditional cache flushes SMT vulnerable + meltdown: Mitigation of PTI + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling

123Result OverviewPhoronix Test Suite100%103%106%109%112%RedisBetsy GPU CompressorUnpacking FirefoxSunflow Rendering SystemBuild2oneDNNeSpeak-NG Speech EngineMlpack BenchmarkNumpy BenchmarkasmFishFFTELAMMPS Molecular Dynamics Simulatoryquake2Timed FFmpeg CompilationHPC ChallengeDolfynCraftyStockfishTimed MAFFT AlignmentKeyDBSQLite SpeedtestEmbreeCoremarkrav1eGROMACSMonkey Audio Encodingx265BRL-CADHierarchical INTegrationLZ4 CompressionBasis UniversalNCNNBYTE Unix BenchmarkIndigoBenchTimed Eigen CompilationKvazaarTimed HMMer SearchOpus Codec EncodingPHPBenchOgg Audio EncodingWavPack Audio EncodingRNNoiseCaffeASTC EncoderCLOMP

Core i7 5775C EOY 2020onednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUonednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: IP Shapes 1D - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUmlpack: scikit_qdayquake2: OpenGL 1.x - 1920 x 1080mlpack: scikit_icaonednn: IP Shapes 1D - u8s8f32 - CPUredis: LPUSHunpack-firefox: firefox-84.0.source.tar.xzmlpack: scikit_linearridgeregressionsunflow: Global Illumination + Image Synthesishpcc: G-HPLyquake2: OpenGL 3.x - 1920 x 1080embree: Pathtracer ISPC - Crownbuild2: Time To Compileonednn: Convolution Batch Shapes Auto - f32 - CPUcaffe: AlexNet - CPU - 100onednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUespeak: Text-To-Speech Synthesisncnn: CPU-v2-v2 - mobilenet-v2compress-lz4: 9 - Compression Speednumpy: asmfish: 1024 Hash Memory, 26 Depthffte: N=256, 3D Complex FFT Routinecompress-lz4: 3 - Decompression Speedncnn: CPU - mnasnethpcc: EP-DGEMMbasis: ETC1Slammps: Rhodopsin Proteinncnn: CPU - blazefacebuild-ffmpeg: Time To Compiledolfyn: Computational Fluid Dynamicscrafty: Elapsed Timencnn: CPU - shufflenet-v2embree: Pathtracer ISPC - Asian Dragonncnn: CPU - mobilenethpcc: Max Ping Pong Bandwidthx265: Bosphorus 4Khpcc: G-Rand Accessonednn: Recurrent Neural Network Inference - u8s8f32 - CPUstockfish: Total Timencnn: CPU - efficientnet-b0ncnn: CPU - yolov4-tinymafft: Multiple Sequence Alignment - LSU RNAncnn: CPU - squeezenet_ssdredis: SETncnn: CPU - googlenetrav1e: 10ncnn: CPU - regnety_400monednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUkeydb: x265: Bosphorus 1080pcompress-lz4: 1 - Decompression Speedcompress-lz4: 3 - Compression Speedhpcc: G-Ffteyquake2: Software CPU - 1920 x 1080onednn: Recurrent Neural Network Inference - f32 - CPUsqlite-speedtest: Timed Time - Size 1,000kvazaar: Bosphorus 1080p - Slowcoremark: CoreMark Size 666 - Iterations Per Secondrav1e: 5hpcc: G-Ptransncnn: CPU - resnet18hpcc: Rand Ring Latencycompress-lz4: 9 - Decompression Speedgromacs: Water Benchmarkencode-ape: WAV To APEcompress-lz4: 1 - Compression Speedncnn: CPU - alexnetbrl-cad: VGR Performance Metrichint: FLOATkvazaar: Bosphorus 4K - Very Fastkvazaar: Bosphorus 1080p - Very Fastrav1e: 1caffe: GoogleNet - CPU - 100ncnn: CPU-v3-v3 - mobilenet-v3kvazaar: Bosphorus 1080p - Mediumembree: Pathtracer - Crownonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUbyte: Dhrystone 2kvazaar: Bosphorus 4K - Ultra Fastindigobench: CPU - Bedroomncnn: CPU - resnet50kvazaar: Bosphorus 1080p - Ultra Fastindigobench: CPU - Supercarbuild-eigen: Time To Compilehpcc: EP-STREAM Triadncnn: CPU - vgg16embree: Pathtracer - Asian Dragonrav1e: 6mlpack: scikit_svmastcenc: Fastbasis: UASTC Level 0onednn: Recurrent Neural Network Training - f32 - CPUencode-opus: WAV To Opus Encodephpbench: PHP Benchmark Suiteencode-ogg: WAV To Ogghmmer: Pfam Database Searchonednn: Recurrent Neural Network Training - u8s8f32 - CPUencode-wavpack: WAV To WavPackrnnoise: astcenc: Thoroughbasis: UASTC Level 2astcenc: Exhaustivebasis: UASTC Level 2 + RDO Post-Processingbasis: UASTC Level 3astcenc: Mediumkvazaar: Bosphorus 4K - Mediumkvazaar: Bosphorus 4K - Slowclomp: Static OMP Speedupredis: GETredis: SADDredis: LPOPhpcc: Rand Ring Bandwidthbetsy: ETC2 RGB - Highestbetsy: ETC1 - Highest1233.508973.1275417.992713.143312.484321.28689.006016.7002899.00518.361.605.775781421632.9220.7235.252.50582.76190553.85.4516242.55420.79086703513.013140.4797.0842.62279.091114438219803.7561242165408.76.6836.0171779.2602.8783.09121.02723.148682494411.396.801726.4517277.7386.110.032933898.31741948810.2037.6412.52431.071660500.8719.602.63617.303897.98411366.5728.065489.243.282.79950114.33879.9772.2748.75140877.1848230.8781.2372523.820.211095451.80.48214.5034787.4721.9645294381176866.348385.6222.680.2881381935.999.024.71267399.7711.228637136699.510.180.72146.9940.391.65482.3064.0814481.265.66571.18313.358.4212.0027404.849.41257464324.080114.0017400.5116.30825.67072.6381.244588.791001.729158.99810.932.062.013.12215813.51661804.392368415.095.4635216.12715.3974.273313.2460616.958513.950811.726320.60528.593586.3374998.58510.360.425.614391444723.5421.0235.392.48781.58543547.15.3988240.87620.80086687712.806441.1237.0742.68275.031122918219947.5619998635450.26.6035.8683778.3392.8493.09122.19923.227686063911.296.859726.6817131.1806.140.033193884.54744451810.2037.6512.61031.151671609.5019.492.65017.383882.87411227.1627.905497.543.512.79007114.13895.4272.6288.77141109.5710450.8821.2396923.820.210925458.70.48414.4764805.5922.0145275381158594.086395.6222.720.2891386696.009.044.72767396.6311.195137026672.110.160.72346.9840.451.65582.4824.0738081.405.67531.18513.348.4311.9897396.949.42257521324.079114.0697394.5916.29825.65772.6081.226588.801001.520158.99510.932.062.013.11939149.381761393.521438792.175.8148715.31414.6983.710112.6783416.090714.107511.688621.90848.495556.3226095.58527.659.835.622321461476.5421.3035.282.44580.82890542.55.3503238.06421.15906593312.889740.9396.9742.05277.381130283819694.2903269455477.86.6236.2917378.7572.8653.06121.20823.020688550711.316.799826.6817216.0676.090.033123869.51739287010.1337.9012.54330.941668121.3719.472.65317.273873.60408847.8927.945519.643.382.78478114.73876.3672.5338.79141520.5795440.8821.2343523.720.211805474.50.48414.4474792.6122.0445131382533367.815535.6422.760.2881382495.989.054.72477419.2011.213637053570.510.190.72347.1140.491.65882.3344.0737781.335.66951.18513.368.4311.9987398.389.41657479924.102114.0937399.5916.30325.65972.6281.253588.641001.727158.98010.932.062.013.12041822.961687783.912024710.525.3604215.71615.956OpenBenchmarking.org

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU1230.96151.9232.88453.8464.8075SE +/- 0.03532, N = 3SE +/- 0.00958, N = 3SE +/- 0.06500, N = 33.508974.273313.71011MIN: 3.36MIN: 4.17MIN: 3.471. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU123246810Min: 3.44 / Avg: 3.51 / Max: 3.55Min: 4.26 / Avg: 4.27 / Max: 4.29Min: 3.58 / Avg: 3.71 / Max: 3.781. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU1230.73041.46082.19122.92163.652SE +/- 0.01583, N = 3SE +/- 0.03533, N = 3SE +/- 0.01863, N = 33.127543.246062.67834MIN: 3.06MIN: 3.13MIN: 2.61. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU123246810Min: 3.11 / Avg: 3.13 / Max: 3.16Min: 3.2 / Avg: 3.25 / Max: 3.32Min: 2.65 / Avg: 2.68 / Max: 2.711. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU12348121620SE +/- 0.21, N = 3SE +/- 0.07, N = 3SE +/- 0.03, N = 317.9916.9616.09MIN: 17.6MIN: 16.49MIN: 15.861. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU123510152025Min: 17.76 / Avg: 17.99 / Max: 18.41Min: 16.85 / Avg: 16.96 / Max: 17.08Min: 16.04 / Avg: 16.09 / Max: 16.131. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU12348121620SE +/- 0.10, N = 3SE +/- 0.07, N = 3SE +/- 0.01, N = 313.1413.9514.11MIN: 12.91MIN: 13.68MIN: 13.891. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU12348121620Min: 12.99 / Avg: 13.14 / Max: 13.32Min: 13.84 / Avg: 13.95 / Max: 14.09Min: 14.09 / Avg: 14.11 / Max: 14.141. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU1233691215SE +/- 0.11, N = 3SE +/- 0.03, N = 3SE +/- 0.03, N = 312.4811.7311.69MIN: 12.21MIN: 11.59MIN: 11.581. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU12348121620Min: 12.32 / Avg: 12.48 / Max: 12.68Min: 11.69 / Avg: 11.73 / Max: 11.79Min: 11.65 / Avg: 11.69 / Max: 11.751. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU123510152025SE +/- 0.05, N = 3SE +/- 0.03, N = 3SE +/- 0.02, N = 321.2920.6121.91MIN: 21MIN: 20.41MIN: 21.581. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU123510152025Min: 21.23 / Avg: 21.29 / Max: 21.38Min: 20.56 / Avg: 20.61 / Max: 20.67Min: 21.87 / Avg: 21.91 / Max: 21.941. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU1233691215SE +/- 0.07657, N = 3SE +/- 0.03663, N = 3SE +/- 0.00624, N = 39.006018.593588.49555MIN: 8.67MIN: 8.22MIN: 8.071. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU1233691215Min: 8.86 / Avg: 9.01 / Max: 9.13Min: 8.54 / Avg: 8.59 / Max: 8.66Min: 8.48 / Avg: 8.5 / Max: 8.511. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU123246810SE +/- 0.01488, N = 3SE +/- 0.00312, N = 3SE +/- 0.01493, N = 36.700286.337496.32260MIN: 6.39MIN: 6.17MIN: 6.081. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU1233691215Min: 6.67 / Avg: 6.7 / Max: 6.72Min: 6.33 / Avg: 6.34 / Max: 6.34Min: 6.29 / Avg: 6.32 / Max: 6.341. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

Mlpack Benchmark

Mlpack benchmark scripts for machine learning libraries Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_qda12320406080100SE +/- 0.67, N = 3SE +/- 1.16, N = 3SE +/- 0.02, N = 399.0098.5895.58
OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_qda12320406080100Min: 97.96 / Avg: 99 / Max: 100.24Min: 96.52 / Avg: 98.58 / Max: 100.53Min: 95.55 / Avg: 95.58 / Max: 95.62

yquake2

This is a test of Yamagi Quake II. Yamagi Quake II is an enhanced client for id Software's Quake II with focus on offline and coop gameplay. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteryquake2 7.45Renderer: OpenGL 1.x - Resolution: 1920 x 1080123110220330440550SE +/- 9.01, N = 4SE +/- 5.25, N = 15SE +/- 3.86, N = 3518.3510.3527.61. (CC) gcc options: -lm -ldl -rdynamic -shared -lSDL2 -O2 -pipe -fomit-frame-pointer -std=gnu99 -fno-strict-aliasing -fwrapv -fvisibility=hidden -MMD -mfpmath=sse -fPIC
OpenBenchmarking.orgFrames Per Second, More Is Betteryquake2 7.45Renderer: OpenGL 1.x - Resolution: 1920 x 108012390180270360450Min: 491.4 / Avg: 518.3 / Max: 528.9Min: 470.9 / Avg: 510.32 / Max: 530.2Min: 522.3 / Avg: 527.6 / Max: 535.11. (CC) gcc options: -lm -ldl -rdynamic -shared -lSDL2 -O2 -pipe -fomit-frame-pointer -std=gnu99 -fno-strict-aliasing -fwrapv -fvisibility=hidden -MMD -mfpmath=sse -fPIC

Mlpack Benchmark

Mlpack benchmark scripts for machine learning libraries Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_ica1231428425670SE +/- 0.75, N = 3SE +/- 0.28, N = 3SE +/- 0.08, N = 361.6060.4259.83
OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_ica1231224364860Min: 60.62 / Avg: 61.6 / Max: 63.07Min: 59.9 / Avg: 60.42 / Max: 60.87Min: 59.69 / Avg: 59.83 / Max: 59.96

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU1231.29962.59923.89885.19846.498SE +/- 0.00787, N = 3SE +/- 0.00125, N = 3SE +/- 0.00754, N = 35.775785.614395.62232MIN: 5.72MIN: 5.59MIN: 5.591. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU123246810Min: 5.76 / Avg: 5.78 / Max: 5.78Min: 5.61 / Avg: 5.61 / Max: 5.62Min: 5.61 / Avg: 5.62 / Max: 5.641. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

Redis

Redis is an open-source data structure server. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: LPUSH123300K600K900K1200K1500KSE +/- 23763.10, N = 3SE +/- 10846.13, N = 3SE +/- 3122.80, N = 31421632.921444723.541461476.541. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: LPUSH123300K600K900K1200K1500KMin: 1379884.12 / Avg: 1421632.92 / Max: 1462175.38Min: 1426624.88 / Avg: 1444723.54 / Max: 1464128.75Min: 1455790.5 / Avg: 1461476.54 / Max: 1466557.251. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Unpacking Firefox

This simple test profile measures how long it takes to extract the .tar.xz source package of the Mozilla Firefox Web Browser. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterUnpacking Firefox 84.0Extracting: firefox-84.0.source.tar.xz123510152025SE +/- 0.19, N = 4SE +/- 0.16, N = 4SE +/- 0.22, N = 1120.7221.0221.30
OpenBenchmarking.orgSeconds, Fewer Is BetterUnpacking Firefox 84.0Extracting: firefox-84.0.source.tar.xz123510152025Min: 20.36 / Avg: 20.72 / Max: 21.25Min: 20.6 / Avg: 21.02 / Max: 21.26Min: 20.71 / Avg: 21.3 / Max: 23.11

Mlpack Benchmark

Mlpack benchmark scripts for machine learning libraries Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_linearridgeregression1231.21282.42563.63844.85126.064SE +/- 0.02, N = 3SE +/- 0.08, N = 4SE +/- 0.06, N = 35.255.395.28
OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_linearridgeregression123246810Min: 5.21 / Avg: 5.25 / Max: 5.28Min: 5.28 / Avg: 5.39 / Max: 5.61Min: 5.16 / Avg: 5.28 / Max: 5.39

Sunflow Rendering System

This test runs benchmarks of the Sunflow Rendering System. The Sunflow Rendering System is an open-source render engine for photo-realistic image synthesis with a ray-tracing core. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSunflow Rendering System 0.07.2Global Illumination + Image Synthesis1230.56361.12721.69082.25442.818SE +/- 0.008, N = 3SE +/- 0.020, N = 3SE +/- 0.009, N = 32.5052.4872.445MIN: 2.38 / MAX: 3.2MIN: 2.36 / MAX: 3.31MIN: 2.34 / MAX: 3.15
OpenBenchmarking.orgSeconds, Fewer Is BetterSunflow Rendering System 0.07.2Global Illumination + Image Synthesis123246810Min: 2.49 / Avg: 2.51 / Max: 2.52Min: 2.45 / Avg: 2.49 / Max: 2.52Min: 2.43 / Avg: 2.44 / Max: 2.46

HPC Challenge

HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFLOPS, More Is BetterHPC Challenge 1.5.0Test / Class: G-HPL12320406080100SE +/- 1.43, N = 3SE +/- 0.40, N = 3SE +/- 0.62, N = 382.7681.5980.831. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2
OpenBenchmarking.orgGFLOPS, More Is BetterHPC Challenge 1.5.0Test / Class: G-HPL1231632486480Min: 80.33 / Avg: 82.76 / Max: 85.28Min: 80.91 / Avg: 81.59 / Max: 82.29Min: 79.69 / Avg: 80.83 / Max: 81.821. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2

yquake2

This is a test of Yamagi Quake II. Yamagi Quake II is an enhanced client for id Software's Quake II with focus on offline and coop gameplay. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteryquake2 7.45Renderer: OpenGL 3.x - Resolution: 1920 x 1080123120240360480600SE +/- 2.95, N = 3SE +/- 4.68, N = 3SE +/- 3.26, N = 3553.8547.1542.51. (CC) gcc options: -lm -ldl -rdynamic -shared -lSDL2 -O2 -pipe -fomit-frame-pointer -std=gnu99 -fno-strict-aliasing -fwrapv -fvisibility=hidden -MMD -mfpmath=sse -fPIC
OpenBenchmarking.orgFrames Per Second, More Is Betteryquake2 7.45Renderer: OpenGL 3.x - Resolution: 1920 x 1080123100200300400500Min: 548.6 / Avg: 553.77 / Max: 558.8Min: 539.2 / Avg: 547.1 / Max: 555.4Min: 536.5 / Avg: 542.53 / Max: 547.71. (CC) gcc options: -lm -ldl -rdynamic -shared -lSDL2 -O2 -pipe -fomit-frame-pointer -std=gnu99 -fno-strict-aliasing -fwrapv -fvisibility=hidden -MMD -mfpmath=sse -fPIC

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.9.0Binary: Pathtracer ISPC - Model: Crown1231.22662.45323.67984.90646.133SE +/- 0.0048, N = 3SE +/- 0.0109, N = 3SE +/- 0.0687, N = 35.45165.39885.3503MIN: 5.43 / MAX: 5.51MIN: 5.36 / MAX: 5.47MIN: 5.16 / MAX: 5.47
OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.9.0Binary: Pathtracer ISPC - Model: Crown123246810Min: 5.45 / Avg: 5.45 / Max: 5.46Min: 5.39 / Avg: 5.4 / Max: 5.42Min: 5.21 / Avg: 5.35 / Max: 5.43

Build2

This test profile measures the time to bootstrap/install the build2 C++ build toolchain from source. Build2 is a cross-platform build toolchain for C/C++ code and features Cargo-like features. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBuild2 0.13Time To Compile12350100150200250SE +/- 3.92, N = 3SE +/- 0.21, N = 3SE +/- 0.19, N = 3242.55240.88238.06
OpenBenchmarking.orgSeconds, Fewer Is BetterBuild2 0.13Time To Compile1234080120160200Min: 237.64 / Avg: 242.55 / Max: 250.3Min: 240.47 / Avg: 240.88 / Max: 241.16Min: 237.7 / Avg: 238.06 / Max: 238.36

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU123510152025SE +/- 0.27, N = 3SE +/- 0.06, N = 3SE +/- 0.03, N = 320.7920.8021.16MIN: 19.99MIN: 20.59MIN: 20.851. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU123510152025Min: 20.26 / Avg: 20.79 / Max: 21.13Min: 20.73 / Avg: 20.8 / Max: 20.92Min: 21.09 / Avg: 21.16 / Max: 21.211. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

Caffe

This is a benchmark of the Caffe deep learning framework and currently supports the AlexNet and Googlenet model and execution on both CPUs and NVIDIA GPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMilli-Seconds, Fewer Is BetterCaffe 2020-02-13Model: AlexNet - Acceleration: CPU - Iterations: 10012314K28K42K56K70KSE +/- 296.71, N = 3SE +/- 202.49, N = 3SE +/- 171.69, N = 36703566877659331. (CXX) g++ options: -fPIC -O3 -rdynamic -lboost_system -lboost_thread -lboost_filesystem -lboost_chrono -lboost_date_time -lboost_atomic -lpthread -lglog -lgflags -lsz -lz -ldl -lm -lprotobuf -llmdb -lopenblas
OpenBenchmarking.orgMilli-Seconds, Fewer Is BetterCaffe 2020-02-13Model: AlexNet - Acceleration: CPU - Iterations: 10012312K24K36K48K60KMin: 66589 / Avg: 67035 / Max: 67597Min: 66508 / Avg: 66877 / Max: 67206Min: 65590 / Avg: 65933.33 / Max: 661101. (CXX) g++ options: -fPIC -O3 -rdynamic -lboost_system -lboost_thread -lboost_filesystem -lboost_chrono -lboost_date_time -lboost_atomic -lpthread -lglog -lgflags -lsz -lz -ldl -lm -lprotobuf -llmdb -lopenblas

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU1233691215SE +/- 0.14, N = 3SE +/- 0.02, N = 3SE +/- 0.04, N = 313.0112.8112.89MIN: 12.66MIN: 12.64MIN: 12.741. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU12348121620Min: 12.85 / Avg: 13.01 / Max: 13.29Min: 12.78 / Avg: 12.81 / Max: 12.84Min: 12.83 / Avg: 12.89 / Max: 12.961. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

eSpeak-NG Speech Engine

This test times how long it takes the eSpeak speech synthesizer to read Project Gutenberg's The Outline of Science and output to a WAV file. This test profile is now tracking the eSpeak-NG version of eSpeak. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BettereSpeak-NG Speech Engine 20200907Text-To-Speech Synthesis123918273645SE +/- 0.35, N = 16SE +/- 0.48, N = 4SE +/- 0.56, N = 640.4841.1240.941. (CC) gcc options: -O2 -std=c99
OpenBenchmarking.orgSeconds, Fewer Is BettereSpeak-NG Speech Engine 20200907Text-To-Speech Synthesis123918273645Min: 36.26 / Avg: 40.48 / Max: 43.12Min: 40.3 / Avg: 41.12 / Max: 42.13Min: 40.3 / Avg: 40.94 / Max: 43.741. (CC) gcc options: -O2 -std=c99

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU-v2-v2 - Model: mobilenet-v2123246810SE +/- 0.10, N = 3SE +/- 0.02, N = 3SE +/- 0.13, N = 37.087.076.97MIN: 6.91 / MAX: 8.17MIN: 6.97 / MAX: 8.06MIN: 6.76 / MAX: 8.011. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU-v2-v2 - Model: mobilenet-v21233691215Min: 6.98 / Avg: 7.08 / Max: 7.28Min: 7.03 / Avg: 7.07 / Max: 7.1Min: 6.82 / Avg: 6.97 / Max: 7.231. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 9 - Compression Speed1231020304050SE +/- 0.07, N = 3SE +/- 0.02, N = 3SE +/- 0.42, N = 342.6242.6842.051. (CC) gcc options: -O3
OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 9 - Compression Speed123918273645Min: 42.49 / Avg: 42.62 / Max: 42.72Min: 42.64 / Avg: 42.68 / Max: 42.71Min: 41.21 / Avg: 42.05 / Max: 42.471. (CC) gcc options: -O3

Numpy Benchmark

This is a test to obtain the general Numpy performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgScore, More Is BetterNumpy Benchmark12360120180240300SE +/- 1.04, N = 3SE +/- 0.54, N = 3SE +/- 0.46, N = 3279.09275.03277.38
OpenBenchmarking.orgScore, More Is BetterNumpy Benchmark12350100150200250Min: 277.65 / Avg: 279.09 / Max: 281.11Min: 274.14 / Avg: 275.03 / Max: 276.01Min: 276.62 / Avg: 277.38 / Max: 278.22

asmFish

This is a test of asmFish, an advanced chess benchmark written in Assembly. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNodes/second, More Is BetterasmFish 2018-07-231024 Hash Memory, 26 Depth1232M4M6M8M10MSE +/- 69549.32, N = 3SE +/- 81955.72, N = 3SE +/- 85928.55, N = 3111443821122918211302838
OpenBenchmarking.orgNodes/second, More Is BetterasmFish 2018-07-231024 Hash Memory, 26 Depth1232M4M6M8M10MMin: 11029781 / Avg: 11144381.67 / Max: 11269956Min: 11084789 / Avg: 11229181.67 / Max: 11368559Min: 11206887 / Avg: 11302838 / Max: 11474289

FFTE

FFTE is a package by Daisuke Takahashi to compute Discrete Fourier Transforms of 1-, 2- and 3- dimensional sequences of length (2^p)*(3^q)*(5^r). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMFLOPS, More Is BetterFFTE 7.0N=256, 3D Complex FFT Routine1234K8K12K16K20KSE +/- 49.39, N = 3SE +/- 34.77, N = 3SE +/- 59.28, N = 319803.7619947.5619694.291. (F9X) gfortran options: -O3 -fomit-frame-pointer -fopenmp
OpenBenchmarking.orgMFLOPS, More Is BetterFFTE 7.0N=256, 3D Complex FFT Routine1233K6K9K12K15KMin: 19731.05 / Avg: 19803.76 / Max: 19898.01Min: 19891.76 / Avg: 19947.56 / Max: 20011.42Min: 19611.14 / Avg: 19694.29 / Max: 19809.051. (F9X) gfortran options: -O3 -fomit-frame-pointer -fopenmp

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 3 - Decompression Speed12312002400360048006000SE +/- 4.57, N = 3SE +/- 4.97, N = 3SE +/- 9.93, N = 35408.75450.25477.81. (CC) gcc options: -O3
OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 3 - Decompression Speed12310002000300040005000Min: 5400.6 / Avg: 5408.7 / Max: 5416.4Min: 5441.8 / Avg: 5450.17 / Max: 5459Min: 5460.4 / Avg: 5477.77 / Max: 5494.81. (CC) gcc options: -O3

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: mnasnet123246810SE +/- 0.08, N = 3SE +/- 0.03, N = 3SE +/- 0.04, N = 36.686.606.62MIN: 6.56 / MAX: 6.86MIN: 6.55 / MAX: 6.74MIN: 6.5 / MAX: 6.811. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: mnasnet1233691215Min: 6.58 / Avg: 6.68 / Max: 6.83Min: 6.57 / Avg: 6.6 / Max: 6.65Min: 6.54 / Avg: 6.62 / Max: 6.681. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

HPC Challenge

HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFLOPS, More Is BetterHPC Challenge 1.5.0Test / Class: EP-DGEMM123816243240SE +/- 0.20, N = 3SE +/- 0.08, N = 3SE +/- 0.40, N = 336.0235.8736.291. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2
OpenBenchmarking.orgGFLOPS, More Is BetterHPC Challenge 1.5.0Test / Class: EP-DGEMM123816243240Min: 35.79 / Avg: 36.02 / Max: 36.41Min: 35.71 / Avg: 35.87 / Max: 36Min: 35.86 / Avg: 36.29 / Max: 37.091. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2

Basis Universal

Basis Universal is a GPU texture codoec. This test times how long it takes to convert sRGB PNGs into Basis Univeral assets with various settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBasis Universal 1.12Settings: ETC1S12320406080100SE +/- 0.52, N = 3SE +/- 0.06, N = 3SE +/- 0.11, N = 379.2678.3478.761. (CXX) g++ options: -std=c++11 -fvisibility=hidden -fPIC -fno-strict-aliasing -O3 -rdynamic -lm -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterBasis Universal 1.12Settings: ETC1S1231530456075Min: 78.37 / Avg: 79.26 / Max: 80.15Min: 78.26 / Avg: 78.34 / Max: 78.47Min: 78.64 / Avg: 78.76 / Max: 78.981. (CXX) g++ options: -std=c++11 -fvisibility=hidden -fPIC -fno-strict-aliasing -O3 -rdynamic -lm -lpthread

LAMMPS Molecular Dynamics Simulator

LAMMPS is a classical molecular dynamics code, and an acronym for Large-scale Atomic/Molecular Massively Parallel Simulator. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgns/day, More Is BetterLAMMPS Molecular Dynamics Simulator 29Oct2020Model: Rhodopsin Protein1230.64761.29521.94282.59043.238SE +/- 0.006, N = 3SE +/- 0.007, N = 3SE +/- 0.010, N = 32.8782.8492.8651. (CXX) g++ options: -O3 -pthread -lm
OpenBenchmarking.orgns/day, More Is BetterLAMMPS Molecular Dynamics Simulator 29Oct2020Model: Rhodopsin Protein123246810Min: 2.87 / Avg: 2.88 / Max: 2.89Min: 2.84 / Avg: 2.85 / Max: 2.86Min: 2.86 / Avg: 2.87 / Max: 2.891. (CXX) g++ options: -O3 -pthread -lm

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: blazeface1230.69531.39062.08592.78123.4765SE +/- 0.03, N = 3SE +/- 0.04, N = 3SE +/- 0.01, N = 33.093.093.06MIN: 3.03 / MAX: 3.29MIN: 3.03 / MAX: 7.35MIN: 3.03 / MAX: 3.141. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: blazeface123246810Min: 3.05 / Avg: 3.09 / Max: 3.15Min: 3.04 / Avg: 3.09 / Max: 3.16Min: 3.05 / Avg: 3.06 / Max: 3.071. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Timed FFmpeg Compilation

This test times how long it takes to build the FFmpeg multimedia library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed FFmpeg Compilation 4.2.2Time To Compile123306090120150SE +/- 0.08, N = 3SE +/- 0.47, N = 3SE +/- 0.17, N = 3121.03122.20121.21
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed FFmpeg Compilation 4.2.2Time To Compile12320406080100Min: 120.88 / Avg: 121.03 / Max: 121.14Min: 121.27 / Avg: 122.2 / Max: 122.75Min: 120.89 / Avg: 121.21 / Max: 121.45

Dolfyn

Dolfyn is a Computational Fluid Dynamics (CFD) code of modern numerical simulation techniques. The Dolfyn test profile measures the execution time of the bundled computational fluid dynamics demos that are bundled with Dolfyn. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterDolfyn 0.527Computational Fluid Dynamics123612182430SE +/- 0.02, N = 3SE +/- 0.07, N = 3SE +/- 0.01, N = 323.1523.2323.02
OpenBenchmarking.orgSeconds, Fewer Is BetterDolfyn 0.527Computational Fluid Dynamics123510152025Min: 23.12 / Avg: 23.15 / Max: 23.19Min: 23.09 / Avg: 23.23 / Max: 23.32Min: 23.01 / Avg: 23.02 / Max: 23.04

Crafty

This is a performance test of Crafty, an advanced open-source chess engine. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNodes Per Second, More Is BetterCrafty 25.2Elapsed Time1231.5M3M4.5M6M7.5MSE +/- 29965.63, N = 3SE +/- 16848.33, N = 3SE +/- 4260.63, N = 36824944686063968855071. (CC) gcc options: -pthread -lstdc++ -fprofile-use -lm
OpenBenchmarking.orgNodes Per Second, More Is BetterCrafty 25.2Elapsed Time1231.2M2.4M3.6M4.8M6MMin: 6780714 / Avg: 6824944.33 / Max: 6882082Min: 6829242 / Avg: 6860639.33 / Max: 6886933Min: 6877103 / Avg: 6885507 / Max: 68909291. (CC) gcc options: -pthread -lstdc++ -fprofile-use -lm

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: shufflenet-v21233691215SE +/- 0.02, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 311.3911.2911.31MIN: 11.31 / MAX: 12.43MIN: 11.24 / MAX: 12.08MIN: 11.22 / MAX: 24.441. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: shufflenet-v21233691215Min: 11.34 / Avg: 11.39 / Max: 11.42Min: 11.26 / Avg: 11.29 / Max: 11.3Min: 11.29 / Avg: 11.31 / Max: 11.331. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.9.0Binary: Pathtracer ISPC - Model: Asian Dragon123246810SE +/- 0.0262, N = 3SE +/- 0.0255, N = 3SE +/- 0.0223, N = 36.80176.85976.7998MIN: 6.73 / MAX: 6.95MIN: 6.77 / MAX: 6.98MIN: 6.73 / MAX: 6.95
OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.9.0Binary: Pathtracer ISPC - Model: Asian Dragon1233691215Min: 6.77 / Avg: 6.8 / Max: 6.85Min: 6.81 / Avg: 6.86 / Max: 6.89Min: 6.77 / Avg: 6.8 / Max: 6.84

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: mobilenet123612182430SE +/- 0.12, N = 3SE +/- 0.30, N = 3SE +/- 0.17, N = 326.4526.6826.68MIN: 26.18 / MAX: 39.6MIN: 26.23 / MAX: 28.16MIN: 26.16 / MAX: 37.31. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: mobilenet123612182430Min: 26.28 / Avg: 26.45 / Max: 26.67Min: 26.36 / Avg: 26.68 / Max: 27.27Min: 26.34 / Avg: 26.68 / Max: 26.91. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

HPC Challenge

HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterHPC Challenge 1.5.0Test / Class: Max Ping Pong Bandwidth1234K8K12K16K20KSE +/- 43.07, N = 3SE +/- 272.70, N = 3SE +/- 105.67, N = 317277.7417131.1817216.071. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2
OpenBenchmarking.orgMB/s, More Is BetterHPC Challenge 1.5.0Test / Class: Max Ping Pong Bandwidth1233K6K9K12K15KMin: 17202.23 / Avg: 17277.74 / Max: 17351.4Min: 16586.43 / Avg: 17131.18 / Max: 17426.46Min: 17018.24 / Avg: 17216.07 / Max: 17379.371. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2

x265

This is a simple test of the x265 encoder run on the CPU with 1080p and 4K options for H.265 video encode performance with x265. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betterx265 3.4Video Input: Bosphorus 4K123246810SE +/- 0.01, N = 3SE +/- 0.03, N = 3SE +/- 0.02, N = 36.116.146.091. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
OpenBenchmarking.orgFrames Per Second, More Is Betterx265 3.4Video Input: Bosphorus 4K123246810Min: 6.1 / Avg: 6.11 / Max: 6.13Min: 6.09 / Avg: 6.14 / Max: 6.19Min: 6.04 / Avg: 6.09 / Max: 6.121. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

HPC Challenge

HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGUP/s, More Is BetterHPC Challenge 1.5.0Test / Class: G-Random Access1230.00750.0150.02250.030.0375SE +/- 0.00010, N = 3SE +/- 0.00018, N = 3SE +/- 0.00022, N = 30.032930.033190.033121. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2
OpenBenchmarking.orgGUP/s, More Is BetterHPC Challenge 1.5.0Test / Class: G-Random Access12312345Min: 0.03 / Avg: 0.03 / Max: 0.03Min: 0.03 / Avg: 0.03 / Max: 0.03Min: 0.03 / Avg: 0.03 / Max: 0.031. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU1238001600240032004000SE +/- 11.68, N = 3SE +/- 3.15, N = 3SE +/- 6.34, N = 33898.313884.543869.51MIN: 3877.01MIN: 3876.92MIN: 3857.361. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU1237001400210028003500Min: 3881.34 / Avg: 3898.31 / Max: 3920.71Min: 3878.74 / Avg: 3884.54 / Max: 3889.57Min: 3859.78 / Avg: 3869.51 / Max: 3881.421. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

Stockfish

This is a test of Stockfish, an advanced C++11 chess benchmark that can scale up to 128 CPU cores. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNodes Per Second, More Is BetterStockfish 12Total Time1231.6M3.2M4.8M6.4M8MSE +/- 44996.61, N = 3SE +/- 60261.05, N = 3SE +/- 74914.67, N = 37419488744451873928701. (CXX) g++ options: -m64 -lpthread -fno-exceptions -std=c++17 -pedantic -O3 -msse -msse3 -mpopcnt -msse4.1 -mssse3 -msse2 -flto -flto=jobserver
OpenBenchmarking.orgNodes Per Second, More Is BetterStockfish 12Total Time1231.3M2.6M3.9M5.2M6.5MMin: 7357974 / Avg: 7419488.33 / Max: 7507132Min: 7324031 / Avg: 7444518.33 / Max: 7507269Min: 7245595 / Avg: 7392869.67 / Max: 74903661. (CXX) g++ options: -m64 -lpthread -fno-exceptions -std=c++17 -pedantic -O3 -msse -msse3 -mpopcnt -msse4.1 -mssse3 -msse2 -flto -flto=jobserver

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: efficientnet-b01233691215SE +/- 0.02, N = 3SE +/- 0.03, N = 3SE +/- 0.11, N = 310.2010.2010.13MIN: 10.13 / MAX: 10.32MIN: 9.86 / MAX: 21.1MIN: 9.89 / MAX: 11.321. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: efficientnet-b01233691215Min: 10.16 / Avg: 10.2 / Max: 10.24Min: 10.14 / Avg: 10.2 / Max: 10.24Min: 9.92 / Avg: 10.13 / Max: 10.291. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: yolov4-tiny123918273645SE +/- 0.11, N = 3SE +/- 0.18, N = 3SE +/- 0.04, N = 337.6437.6537.90MIN: 37.09 / MAX: 39.96MIN: 37.19 / MAX: 39.03MIN: 37.49 / MAX: 40.471. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: yolov4-tiny123816243240Min: 37.43 / Avg: 37.64 / Max: 37.8Min: 37.43 / Avg: 37.65 / Max: 38Min: 37.84 / Avg: 37.9 / Max: 37.981. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Timed MAFFT Alignment

This test performs an alignment of 100 pyruvate decarboxylase sequences. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed MAFFT Alignment 7.471Multiple Sequence Alignment - LSU RNA1233691215SE +/- 0.02, N = 3SE +/- 0.08, N = 3SE +/- 0.06, N = 312.5212.6112.541. (CC) gcc options: -std=c99 -O3 -lm -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed MAFFT Alignment 7.471Multiple Sequence Alignment - LSU RNA12348121620Min: 12.5 / Avg: 12.52 / Max: 12.56Min: 12.46 / Avg: 12.61 / Max: 12.74Min: 12.44 / Avg: 12.54 / Max: 12.621. (CC) gcc options: -std=c99 -O3 -lm -lpthread

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: squeezenet_ssd123714212835SE +/- 0.13, N = 3SE +/- 0.05, N = 3SE +/- 0.10, N = 331.0731.1530.94MIN: 30.77 / MAX: 40.23MIN: 30.98 / MAX: 33.13MIN: 30.7 / MAX: 32.441. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: squeezenet_ssd123714212835Min: 30.81 / Avg: 31.07 / Max: 31.21Min: 31.08 / Avg: 31.15 / Max: 31.24Min: 30.77 / Avg: 30.94 / Max: 31.11. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Redis

Redis is an open-source data structure server. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: SET123400K800K1200K1600K2000KSE +/- 30456.48, N = 3SE +/- 10885.22, N = 3SE +/- 19626.30, N = 31660500.871671609.501668121.371. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: SET123300K600K900K1200K1500KMin: 1600153.62 / Avg: 1660500.87 / Max: 1697847.25Min: 1655682.25 / Avg: 1671609.5 / Max: 1692426.38Min: 1642141.12 / Avg: 1668121.37 / Max: 1706593.751. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: googlenet123510152025SE +/- 0.02, N = 3SE +/- 0.14, N = 3SE +/- 0.09, N = 319.6019.4919.47MIN: 19.49 / MAX: 19.99MIN: 19.14 / MAX: 19.91MIN: 19.19 / MAX: 20.461. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: googlenet123510152025Min: 19.57 / Avg: 19.6 / Max: 19.63Min: 19.22 / Avg: 19.49 / Max: 19.68Min: 19.3 / Avg: 19.47 / Max: 19.561. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

rav1e

Xiph rav1e is a Rust-written AV1 video encoder. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4 AlphaSpeed: 101230.59691.19381.79072.38762.9845SE +/- 0.002, N = 3SE +/- 0.009, N = 3SE +/- 0.008, N = 32.6362.6502.653
OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4 AlphaSpeed: 10123246810Min: 2.63 / Avg: 2.64 / Max: 2.64Min: 2.64 / Avg: 2.65 / Max: 2.67Min: 2.64 / Avg: 2.65 / Max: 2.67

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: regnety_400m12348121620SE +/- 0.03, N = 3SE +/- 0.22, N = 3SE +/- 0.05, N = 317.3017.3817.27MIN: 17.18 / MAX: 17.63MIN: 17.04 / MAX: 98.97MIN: 17.12 / MAX: 18.451. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: regnety_400m12348121620Min: 17.24 / Avg: 17.3 / Max: 17.36Min: 17.09 / Avg: 17.38 / Max: 17.8Min: 17.17 / Avg: 17.27 / Max: 17.321. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU1238001600240032004000SE +/- 11.47, N = 3SE +/- 0.98, N = 3SE +/- 8.25, N = 33897.983882.873873.60MIN: 3871.77MIN: 3878.39MIN: 3858.231. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU1237001400210028003500Min: 3875.44 / Avg: 3897.98 / Max: 3912.99Min: 3881.15 / Avg: 3882.87 / Max: 3884.55Min: 3861.31 / Avg: 3873.6 / Max: 3889.271. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

KeyDB

A benchmark of KeyDB as a multi-threaded fork of the Redis server. The KeyDB benchmark is conducted using memtier-benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps/sec, More Is BetterKeyDB 6.0.1612390K180K270K360K450KSE +/- 1505.49, N = 3SE +/- 1060.79, N = 3SE +/- 57.06, N = 3411366.57411227.16408847.891. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
OpenBenchmarking.orgOps/sec, More Is BetterKeyDB 6.0.1612370K140K210K280K350KMin: 409227.52 / Avg: 411366.57 / Max: 414271.26Min: 409146.42 / Avg: 411227.16 / Max: 412626.26Min: 408736.83 / Avg: 408847.89 / Max: 408926.171. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

x265

This is a simple test of the x265 encoder run on the CPU with 1080p and 4K options for H.265 video encode performance with x265. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betterx265 3.4Video Input: Bosphorus 1080p123714212835SE +/- 0.09, N = 3SE +/- 0.15, N = 3SE +/- 0.02, N = 328.0627.9027.941. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
OpenBenchmarking.orgFrames Per Second, More Is Betterx265 3.4Video Input: Bosphorus 1080p123612182430Min: 27.91 / Avg: 28.06 / Max: 28.22Min: 27.67 / Avg: 27.9 / Max: 28.19Min: 27.9 / Avg: 27.94 / Max: 27.981. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 1 - Decompression Speed12312002400360048006000SE +/- 33.51, N = 3SE +/- 14.56, N = 3SE +/- 19.26, N = 35489.25497.55519.61. (CC) gcc options: -O3
OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 1 - Decompression Speed12310002000300040005000Min: 5450.9 / Avg: 5489.23 / Max: 5556Min: 5471.6 / Avg: 5497.47 / Max: 5522Min: 5481.2 / Avg: 5519.57 / Max: 5541.71. (CC) gcc options: -O3

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 3 - Compression Speed1231020304050SE +/- 0.02, N = 3SE +/- 0.00, N = 3SE +/- 0.07, N = 343.2843.5143.381. (CC) gcc options: -O3
OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 3 - Compression Speed123918273645Min: 43.25 / Avg: 43.28 / Max: 43.31Min: 43.51 / Avg: 43.51 / Max: 43.52Min: 43.31 / Avg: 43.38 / Max: 43.521. (CC) gcc options: -O3

HPC Challenge

HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFLOPS, More Is BetterHPC Challenge 1.5.0Test / Class: G-Ffte1230.62991.25981.88972.51963.1495SE +/- 0.00194, N = 3SE +/- 0.00441, N = 3SE +/- 0.00289, N = 32.799502.790072.784781. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2
OpenBenchmarking.orgGFLOPS, More Is BetterHPC Challenge 1.5.0Test / Class: G-Ffte123246810Min: 2.8 / Avg: 2.8 / Max: 2.8Min: 2.78 / Avg: 2.79 / Max: 2.8Min: 2.78 / Avg: 2.78 / Max: 2.791. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2

yquake2

This is a test of Yamagi Quake II. Yamagi Quake II is an enhanced client for id Software's Quake II with focus on offline and coop gameplay. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteryquake2 7.45Renderer: Software CPU - Resolution: 1920 x 1080123306090120150SE +/- 0.15, N = 3SE +/- 0.68, N = 3SE +/- 0.24, N = 3114.3114.1114.71. (CC) gcc options: -lm -ldl -rdynamic -shared -lSDL2 -O2 -pipe -fomit-frame-pointer -std=gnu99 -fno-strict-aliasing -fwrapv -fvisibility=hidden -MMD -mfpmath=sse -fPIC
OpenBenchmarking.orgFrames Per Second, More Is Betteryquake2 7.45Renderer: Software CPU - Resolution: 1920 x 108012320406080100Min: 114.1 / Avg: 114.33 / Max: 114.6Min: 112.8 / Avg: 114.1 / Max: 115.1Min: 114.2 / Avg: 114.67 / Max: 1151. (CC) gcc options: -lm -ldl -rdynamic -shared -lSDL2 -O2 -pipe -fomit-frame-pointer -std=gnu99 -fno-strict-aliasing -fwrapv -fvisibility=hidden -MMD -mfpmath=sse -fPIC

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU1238001600240032004000SE +/- 13.83, N = 3SE +/- 3.39, N = 3SE +/- 11.46, N = 33879.973895.423876.36MIN: 3849.25MIN: 3886.76MIN: 3849.851. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU1237001400210028003500Min: 3852.31 / Avg: 3879.97 / Max: 3894.06Min: 3888.82 / Avg: 3895.42 / Max: 3900.06Min: 3855.16 / Avg: 3876.36 / Max: 3894.481. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

SQLite Speedtest

This is a benchmark of SQLite's speedtest1 benchmark program with an increased problem size of 1,000. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSQLite Speedtest 3.30Timed Time - Size 1,0001231632486480SE +/- 0.13, N = 3SE +/- 0.06, N = 3SE +/- 0.19, N = 372.2772.6372.531. (CC) gcc options: -O2 -ldl -lz -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterSQLite Speedtest 3.30Timed Time - Size 1,0001231428425670Min: 72.08 / Avg: 72.27 / Max: 72.51Min: 72.55 / Avg: 72.63 / Max: 72.74Min: 72.16 / Avg: 72.53 / Max: 72.81. (CC) gcc options: -O2 -ldl -lz -lpthread

Kvazaar

This is a test of Kvazaar as a CPU-based H.265 video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 1080p - Video Preset: Slow123246810SE +/- 0.01, N = 3SE +/- 0.00, N = 3SE +/- 0.01, N = 38.758.778.791. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 1080p - Video Preset: Slow1233691215Min: 8.73 / Avg: 8.75 / Max: 8.77Min: 8.77 / Avg: 8.77 / Max: 8.77Min: 8.77 / Avg: 8.79 / Max: 8.81. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

Coremark

This is a test of EEMBC CoreMark processor benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations/Sec, More Is BetterCoremark 1.0CoreMark Size 666 - Iterations Per Second12330K60K90K120K150KSE +/- 280.09, N = 3SE +/- 334.19, N = 3SE +/- 224.91, N = 3140877.18141109.57141520.581. (CC) gcc options: -O2 -lrt" -lrt
OpenBenchmarking.orgIterations/Sec, More Is BetterCoremark 1.0CoreMark Size 666 - Iterations Per Second12320K40K60K80K100KMin: 140523.45 / Avg: 140877.18 / Max: 141430.21Min: 140721.2 / Avg: 141109.57 / Max: 141774.84Min: 141112.14 / Avg: 141520.58 / Max: 1418881. (CC) gcc options: -O2 -lrt" -lrt

rav1e

Xiph rav1e is a Rust-written AV1 video encoder. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4 AlphaSpeed: 51230.19850.3970.59550.7940.9925SE +/- 0.002, N = 3SE +/- 0.001, N = 3SE +/- 0.002, N = 30.8780.8820.882
OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4 AlphaSpeed: 5123246810Min: 0.87 / Avg: 0.88 / Max: 0.88Min: 0.88 / Avg: 0.88 / Max: 0.89Min: 0.88 / Avg: 0.88 / Max: 0.89

HPC Challenge

HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGB/s, More Is BetterHPC Challenge 1.5.0Test / Class: G-Ptrans1230.27890.55780.83671.11561.3945SE +/- 0.01118, N = 3SE +/- 0.00266, N = 3SE +/- 0.00515, N = 31.237251.239691.234351. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2
OpenBenchmarking.orgGB/s, More Is BetterHPC Challenge 1.5.0Test / Class: G-Ptrans123246810Min: 1.22 / Avg: 1.24 / Max: 1.26Min: 1.24 / Avg: 1.24 / Max: 1.24Min: 1.23 / Avg: 1.23 / Max: 1.241. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: resnet18123612182430SE +/- 0.05, N = 3SE +/- 0.08, N = 3SE +/- 0.06, N = 323.8223.8223.72MIN: 23.6 / MAX: 36.25MIN: 23.47 / MAX: 25.66MIN: 23.51 / MAX: 36.611. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: resnet18123612182430Min: 23.73 / Avg: 23.82 / Max: 23.9Min: 23.66 / Avg: 23.82 / Max: 23.91Min: 23.61 / Avg: 23.72 / Max: 23.811. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

HPC Challenge

HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgusecs, Fewer Is BetterHPC Challenge 1.5.0Test / Class: Random Ring Latency1230.04770.09540.14310.19080.2385SE +/- 0.00015, N = 3SE +/- 0.00022, N = 3SE +/- 0.00054, N = 30.211090.210920.211801. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2
OpenBenchmarking.orgusecs, Fewer Is BetterHPC Challenge 1.5.0Test / Class: Random Ring Latency12312345Min: 0.21 / Avg: 0.21 / Max: 0.21Min: 0.21 / Avg: 0.21 / Max: 0.21Min: 0.21 / Avg: 0.21 / Max: 0.211. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 9 - Decompression Speed12312002400360048006000SE +/- 6.01, N = 3SE +/- 11.99, N = 3SE +/- 5.41, N = 35451.85458.75474.51. (CC) gcc options: -O3
OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 9 - Decompression Speed12310002000300040005000Min: 5443.3 / Avg: 5451.8 / Max: 5463.4Min: 5444.7 / Avg: 5458.73 / Max: 5482.6Min: 5465.5 / Avg: 5474.47 / Max: 5484.21. (CC) gcc options: -O3

GROMACS

The GROMACS (GROningen MAchine for Chemical Simulations) molecular dynamics package testing on the CPU with the water_GMX50 data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNs Per Day, More Is BetterGROMACS 2020.3Water Benchmark1230.10890.21780.32670.43560.5445SE +/- 0.002, N = 3SE +/- 0.002, N = 3SE +/- 0.001, N = 30.4820.4840.4841. (CXX) g++ options: -O3 -pthread -lrt -lpthread -lm
OpenBenchmarking.orgNs Per Day, More Is BetterGROMACS 2020.3Water Benchmark123246810Min: 0.48 / Avg: 0.48 / Max: 0.49Min: 0.48 / Avg: 0.48 / Max: 0.49Min: 0.48 / Avg: 0.48 / Max: 0.491. (CXX) g++ options: -O3 -pthread -lrt -lpthread -lm

Monkey Audio Encoding

This test times how long it takes to encode a sample WAV file to Monkey's Audio APE format. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterMonkey Audio Encoding 3.99.6WAV To APE12348121620SE +/- 0.11, N = 5SE +/- 0.06, N = 5SE +/- 0.06, N = 514.5014.4814.451. (CXX) g++ options: -O3 -pedantic -rdynamic -lrt
OpenBenchmarking.orgSeconds, Fewer Is BetterMonkey Audio Encoding 3.99.6WAV To APE12348121620Min: 14.34 / Avg: 14.5 / Max: 14.92Min: 14.37 / Avg: 14.48 / Max: 14.7Min: 14.35 / Avg: 14.45 / Max: 14.641. (CXX) g++ options: -O3 -pedantic -rdynamic -lrt

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 1 - Compression Speed12310002000300040005000SE +/- 14.35, N = 3SE +/- 18.70, N = 3SE +/- 5.40, N = 34787.474805.594792.611. (CC) gcc options: -O3
OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 1 - Compression Speed1238001600240032004000Min: 4773.08 / Avg: 4787.47 / Max: 4816.16Min: 4768.3 / Avg: 4805.59 / Max: 4826.63Min: 4785.25 / Avg: 4792.61 / Max: 4803.131. (CC) gcc options: -O3

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: alexnet123510152025SE +/- 0.09, N = 3SE +/- 0.02, N = 3SE +/- 0.02, N = 321.9622.0122.04MIN: 21.61 / MAX: 23.2MIN: 21.84 / MAX: 22.4MIN: 21.86 / MAX: 22.471. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: alexnet123510152025Min: 21.81 / Avg: 21.96 / Max: 22.11Min: 21.98 / Avg: 22.01 / Max: 22.03Min: 22.02 / Avg: 22.04 / Max: 22.081. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

BRL-CAD

BRL-CAD 7.28.0 is a cross-platform, open-source solid modeling system with built-in benchmark mode. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgVGR Performance Metric, More Is BetterBRL-CAD 7.30.8VGR Performance Metric12310K20K30K40K50K4529445275451311. (CXX) g++ options: -std=c++11 -pipe -fno-strict-aliasing -fno-common -fexceptions -ftemplate-depth-128 -m64 -ggdb3 -O3 -fipa-pta -fstrength-reduce -finline-functions -flto -pedantic -rdynamic -lSM -lICE -lXi -lGLU -lGL -lGLdispatch -lX11 -lXext -lXrender -lpthread -ldl -luuid -lm

Hierarchical INTegration

This test runs the U.S. Department of Energy's Ames Laboratory Hierarchical INTegration (HINT) benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQUIPs, More Is BetterHierarchical INTegration 1.0Test: FLOAT12380M160M240M320M400MSE +/- 977507.07, N = 3SE +/- 786027.10, N = 3SE +/- 82458.49, N = 3381176866.35381158594.09382533367.821. (CC) gcc options: -O3 -march=native -lm
OpenBenchmarking.orgQUIPs, More Is BetterHierarchical INTegration 1.0Test: FLOAT12370M140M210M280M350MMin: 379271479.41 / Avg: 381176866.35 / Max: 382508618.76Min: 379721923.67 / Avg: 381158594.09 / Max: 382429649.29Min: 382370243.4 / Avg: 382533367.82 / Max: 382635930.711. (CC) gcc options: -O3 -march=native -lm

Kvazaar

This is a test of Kvazaar as a CPU-based H.265 video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 4K - Video Preset: Very Fast1231.2692.5383.8075.0766.345SE +/- 0.00, N = 3SE +/- 0.01, N = 3SE +/- 0.00, N = 35.625.625.641. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 4K - Video Preset: Very Fast123246810Min: 5.62 / Avg: 5.62 / Max: 5.63Min: 5.61 / Avg: 5.62 / Max: 5.63Min: 5.64 / Avg: 5.64 / Max: 5.651. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 1080p - Video Preset: Very Fast123510152025SE +/- 0.02, N = 3SE +/- 0.02, N = 3SE +/- 0.02, N = 322.6822.7222.761. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 1080p - Video Preset: Very Fast123510152025Min: 22.66 / Avg: 22.68 / Max: 22.71Min: 22.68 / Avg: 22.72 / Max: 22.75Min: 22.72 / Avg: 22.76 / Max: 22.81. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

rav1e

Xiph rav1e is a Rust-written AV1 video encoder. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4 AlphaSpeed: 11230.0650.130.1950.260.325SE +/- 0.000, N = 3SE +/- 0.000, N = 3SE +/- 0.000, N = 30.2880.2890.288
OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4 AlphaSpeed: 112312345Min: 0.29 / Avg: 0.29 / Max: 0.29Min: 0.29 / Avg: 0.29 / Max: 0.29Min: 0.29 / Avg: 0.29 / Max: 0.29

Caffe

This is a benchmark of the Caffe deep learning framework and currently supports the AlexNet and Googlenet model and execution on both CPUs and NVIDIA GPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMilli-Seconds, Fewer Is BetterCaffe 2020-02-13Model: GoogleNet - Acceleration: CPU - Iterations: 10012330K60K90K120K150KSE +/- 306.29, N = 3SE +/- 50.29, N = 3SE +/- 303.12, N = 31381931386691382491. (CXX) g++ options: -fPIC -O3 -rdynamic -lboost_system -lboost_thread -lboost_filesystem -lboost_chrono -lboost_date_time -lboost_atomic -lpthread -lglog -lgflags -lsz -lz -ldl -lm -lprotobuf -llmdb -lopenblas
OpenBenchmarking.orgMilli-Seconds, Fewer Is BetterCaffe 2020-02-13Model: GoogleNet - Acceleration: CPU - Iterations: 10012320K40K60K80K100KMin: 137682 / Avg: 138192.67 / Max: 138741Min: 138591 / Avg: 138669 / Max: 138763Min: 137659 / Avg: 138249.33 / Max: 1386641. (CXX) g++ options: -fPIC -O3 -rdynamic -lboost_system -lboost_thread -lboost_filesystem -lboost_chrono -lboost_date_time -lboost_atomic -lpthread -lglog -lgflags -lsz -lz -ldl -lm -lprotobuf -llmdb -lopenblas

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU-v3-v3 - Model: mobilenet-v3123246810SE +/- 0.02, N = 3SE +/- 0.01, N = 3SE +/- 0.03, N = 35.996.005.98MIN: 5.93 / MAX: 6.86MIN: 5.95 / MAX: 7.09MIN: 5.93 / MAX: 6.941. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU-v3-v3 - Model: mobilenet-v3123246810Min: 5.95 / Avg: 5.99 / Max: 6.03Min: 5.98 / Avg: 6 / Max: 6.02Min: 5.95 / Avg: 5.98 / Max: 6.041. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Kvazaar

This is a test of Kvazaar as a CPU-based H.265 video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 1080p - Video Preset: Medium1233691215SE +/- 0.01, N = 3SE +/- 0.00, N = 3SE +/- 0.00, N = 39.029.049.051. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 1080p - Video Preset: Medium1233691215Min: 9.01 / Avg: 9.02 / Max: 9.03Min: 9.03 / Avg: 9.04 / Max: 9.04Min: 9.05 / Avg: 9.05 / Max: 9.051. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.9.0Binary: Pathtracer - Model: Crown1231.06372.12743.19114.25485.3185SE +/- 0.0107, N = 3SE +/- 0.0150, N = 3SE +/- 0.0010, N = 34.71264.72764.7247MIN: 4.68 / MAX: 4.76MIN: 4.69 / MAX: 4.79MIN: 4.71 / MAX: 4.76
OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.9.0Binary: Pathtracer - Model: Crown123246810Min: 4.69 / Avg: 4.71 / Max: 4.72Min: 4.7 / Avg: 4.73 / Max: 4.75Min: 4.72 / Avg: 4.72 / Max: 4.73

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU12316003200480064008000SE +/- 5.89, N = 3SE +/- 3.99, N = 3SE +/- 5.09, N = 37399.777396.637419.20MIN: 7384.45MIN: 7381.31MIN: 7404.381. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU12313002600390052006500Min: 7389.6 / Avg: 7399.77 / Max: 7410.02Min: 7390.08 / Avg: 7396.63 / Max: 7403.84Min: 7409.86 / Avg: 7419.2 / Max: 7427.381. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU1233691215SE +/- 0.01, N = 3SE +/- 0.00, N = 3SE +/- 0.02, N = 311.2311.2011.21MIN: 11.19MIN: 11.17MIN: 11.161. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU1233691215Min: 11.22 / Avg: 11.23 / Max: 11.24Min: 11.19 / Avg: 11.2 / Max: 11.2Min: 11.18 / Avg: 11.21 / Max: 11.261. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

BYTE Unix Benchmark

This is a test of BYTE. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgLPS, More Is BetterBYTE Unix Benchmark 3.6Computational Test: Dhrystone 21238M16M24M32M40MSE +/- 250280.80, N = 3SE +/- 317702.14, N = 3SE +/- 365197.62, N = 337136699.537026672.137053570.5
OpenBenchmarking.orgLPS, More Is BetterBYTE Unix Benchmark 3.6Computational Test: Dhrystone 21236M12M18M24M30MMin: 36814964 / Avg: 37136699.47 / Max: 37629660.8Min: 36402067.7 / Avg: 37026672.1 / Max: 37439998.8Min: 36362553.6 / Avg: 37053570.47 / Max: 37603968.3

Kvazaar

This is a test of Kvazaar as a CPU-based H.265 video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 4K - Video Preset: Ultra Fast1233691215SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 310.1810.1610.191. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 4K - Video Preset: Ultra Fast1233691215Min: 10.15 / Avg: 10.18 / Max: 10.19Min: 10.15 / Avg: 10.16 / Max: 10.18Min: 10.17 / Avg: 10.19 / Max: 10.21. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

IndigoBench

This is a test of Indigo Renderer's IndigoBench benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgM samples/s, More Is BetterIndigoBench 4.4Acceleration: CPU - Scene: Bedroom1230.16270.32540.48810.65080.8135SE +/- 0.001, N = 3SE +/- 0.001, N = 3SE +/- 0.002, N = 30.7210.7230.723
OpenBenchmarking.orgM samples/s, More Is BetterIndigoBench 4.4Acceleration: CPU - Scene: Bedroom123246810Min: 0.72 / Avg: 0.72 / Max: 0.72Min: 0.72 / Avg: 0.72 / Max: 0.72Min: 0.72 / Avg: 0.72 / Max: 0.73

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: resnet501231122334455SE +/- 0.06, N = 3SE +/- 0.17, N = 3SE +/- 0.21, N = 346.9946.9847.11MIN: 46.57 / MAX: 49.08MIN: 46.59 / MAX: 60.2MIN: 46.64 / MAX: 50.761. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: resnet501231020304050Min: 46.88 / Avg: 46.99 / Max: 47.1Min: 46.75 / Avg: 46.98 / Max: 47.32Min: 46.78 / Avg: 47.11 / Max: 47.491. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Kvazaar

This is a test of Kvazaar as a CPU-based H.265 video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 1080p - Video Preset: Ultra Fast123918273645SE +/- 0.01, N = 3SE +/- 0.02, N = 3SE +/- 0.01, N = 340.3940.4540.491. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 1080p - Video Preset: Ultra Fast123816243240Min: 40.38 / Avg: 40.39 / Max: 40.41Min: 40.42 / Avg: 40.45 / Max: 40.49Min: 40.46 / Avg: 40.49 / Max: 40.51. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

IndigoBench

This is a test of Indigo Renderer's IndigoBench benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgM samples/s, More Is BetterIndigoBench 4.4Acceleration: CPU - Scene: Supercar1230.37310.74621.11931.49241.8655SE +/- 0.005, N = 3SE +/- 0.006, N = 3SE +/- 0.002, N = 31.6541.6551.658
OpenBenchmarking.orgM samples/s, More Is BetterIndigoBench 4.4Acceleration: CPU - Scene: Supercar123246810Min: 1.64 / Avg: 1.65 / Max: 1.66Min: 1.64 / Avg: 1.66 / Max: 1.66Min: 1.66 / Avg: 1.66 / Max: 1.66

Timed Eigen Compilation

This test times how long it takes to build all Eigen examples. The Eigen examples are compiled serially. Eigen is a C++ template library for linear algebra. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Eigen Compilation 3.3.9Time To Compile12320406080100SE +/- 0.06, N = 3SE +/- 0.11, N = 3SE +/- 0.02, N = 382.3182.4882.33
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Eigen Compilation 3.3.9Time To Compile1231632486480Min: 82.19 / Avg: 82.31 / Max: 82.38Min: 82.33 / Avg: 82.48 / Max: 82.68Min: 82.3 / Avg: 82.33 / Max: 82.36

HPC Challenge

HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGB/s, More Is BetterHPC Challenge 1.5.0Test / Class: EP-STREAM Triad1230.91831.83662.75493.67324.5915SE +/- 0.00157, N = 3SE +/- 0.00200, N = 3SE +/- 0.00055, N = 34.081444.073804.073771. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2
OpenBenchmarking.orgGB/s, More Is BetterHPC Challenge 1.5.0Test / Class: EP-STREAM Triad123246810Min: 4.08 / Avg: 4.08 / Max: 4.08Min: 4.07 / Avg: 4.07 / Max: 4.08Min: 4.07 / Avg: 4.07 / Max: 4.071. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: vgg1612320406080100SE +/- 0.03, N = 3SE +/- 0.09, N = 3SE +/- 0.13, N = 381.2681.4081.33MIN: 80.69 / MAX: 92.8MIN: 80.9 / MAX: 92.48MIN: 80.91 / MAX: 85.211. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20201218Target: CPU - Model: vgg161231632486480Min: 81.23 / Avg: 81.26 / Max: 81.31Min: 81.22 / Avg: 81.4 / Max: 81.49Min: 81.11 / Avg: 81.33 / Max: 81.571. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.9.0Binary: Pathtracer - Model: Asian Dragon1231.27692.55383.83075.10766.3845SE +/- 0.0264, N = 3SE +/- 0.0277, N = 3SE +/- 0.0224, N = 35.66575.67535.6695MIN: 5.59 / MAX: 5.79MIN: 5.59 / MAX: 5.79MIN: 5.62 / MAX: 5.79
OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.9.0Binary: Pathtracer - Model: Asian Dragon123246810Min: 5.63 / Avg: 5.67 / Max: 5.72Min: 5.63 / Avg: 5.68 / Max: 5.73Min: 5.64 / Avg: 5.67 / Max: 5.71

rav1e

Xiph rav1e is a Rust-written AV1 video encoder. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4 AlphaSpeed: 61230.26660.53320.79981.06641.333SE +/- 0.004, N = 3SE +/- 0.004, N = 3SE +/- 0.001, N = 31.1831.1851.185
OpenBenchmarking.orgFrames Per Second, More Is Betterrav1e 0.4 AlphaSpeed: 6123246810Min: 1.18 / Avg: 1.18 / Max: 1.19Min: 1.18 / Avg: 1.18 / Max: 1.19Min: 1.18 / Avg: 1.19 / Max: 1.19

Mlpack Benchmark

Mlpack benchmark scripts for machine learning libraries Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_svm1233691215SE +/- 0.00, N = 3SE +/- 0.00, N = 3SE +/- 0.01, N = 313.3513.3413.36
OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_svm12348121620Min: 13.35 / Avg: 13.35 / Max: 13.36Min: 13.34 / Avg: 13.34 / Max: 13.35Min: 13.35 / Avg: 13.36 / Max: 13.37

ASTC Encoder

ASTC Encoder (astcenc) is for the Adaptive Scalable Texture Compression (ASTC) format commonly used with OpenGL, OpenGL ES, and Vulkan graphics APIs. This test profile does a coding test of both compression/decompression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Fast123246810SE +/- 0.00, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 38.428.438.431. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Fast1233691215Min: 8.41 / Avg: 8.42 / Max: 8.42Min: 8.42 / Avg: 8.43 / Max: 8.44Min: 8.41 / Avg: 8.43 / Max: 8.451. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

Basis Universal

Basis Universal is a GPU texture codoec. This test times how long it takes to convert sRGB PNGs into Basis Univeral assets with various settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBasis Universal 1.12Settings: UASTC Level 01233691215SE +/- 0.00, N = 3SE +/- 0.00, N = 3SE +/- 0.00, N = 312.0011.9912.001. (CXX) g++ options: -std=c++11 -fvisibility=hidden -fPIC -fno-strict-aliasing -O3 -rdynamic -lm -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterBasis Universal 1.12Settings: UASTC Level 01233691215Min: 11.99 / Avg: 12 / Max: 12.01Min: 11.98 / Avg: 11.99 / Max: 12Min: 11.99 / Avg: 12 / Max: 121. (CXX) g++ options: -std=c++11 -fvisibility=hidden -fPIC -fno-strict-aliasing -O3 -rdynamic -lm -lpthread

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU12316003200480064008000SE +/- 2.74, N = 3SE +/- 4.24, N = 3SE +/- 5.94, N = 37404.847396.947398.38MIN: 7395.03MIN: 7384.58MIN: 7383.671. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU12313002600390052006500Min: 7399.35 / Avg: 7404.84 / Max: 7407.66Min: 7389.01 / Avg: 7396.94 / Max: 7403.49Min: 7388.4 / Avg: 7398.38 / Max: 7408.961. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

Opus Codec Encoding

Opus is an open audio codec. Opus is a lossy audio compression format designed primarily for interactive real-time applications over the Internet. This test uses Opus-Tools and measures the time required to encode a WAV file to Opus. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpus Codec Encoding 1.3.1WAV To Opus Encode1233691215SE +/- 0.019, N = 5SE +/- 0.014, N = 5SE +/- 0.011, N = 59.4129.4229.4161. (CXX) g++ options: -fvisibility=hidden -logg -lm
OpenBenchmarking.orgSeconds, Fewer Is BetterOpus Codec Encoding 1.3.1WAV To Opus Encode1233691215Min: 9.39 / Avg: 9.41 / Max: 9.49Min: 9.39 / Avg: 9.42 / Max: 9.47Min: 9.4 / Avg: 9.42 / Max: 9.461. (CXX) g++ options: -fvisibility=hidden -logg -lm

PHPBench

PHPBench is a benchmark suite for PHP. It performs a large number of simple tests in order to bench various aspects of the PHP interpreter. PHPBench can be used to compare hardware, operating systems, PHP versions, PHP accelerators and caches, compiler options, etc. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgScore, More Is BetterPHPBench 0.8.1PHP Benchmark Suite123120K240K360K480K600KSE +/- 503.51, N = 3SE +/- 1210.22, N = 3SE +/- 1507.55, N = 3574643575213574799
OpenBenchmarking.orgScore, More Is BetterPHPBench 0.8.1PHP Benchmark Suite123100K200K300K400K500KMin: 573747 / Avg: 574643.33 / Max: 575489Min: 573707 / Avg: 575213 / Max: 577607Min: 572000 / Avg: 574799.33 / Max: 577169

Ogg Audio Encoding

This test times how long it takes to encode a sample WAV file to Ogg format using the reference Xiph.org tools/libraries. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOgg Audio Encoding 1.3.4WAV To Ogg123612182430SE +/- 0.02, N = 3SE +/- 0.03, N = 3SE +/- 0.07, N = 324.0824.0824.101. (CC) gcc options: -O2 -ffast-math -fsigned-char
OpenBenchmarking.orgSeconds, Fewer Is BetterOgg Audio Encoding 1.3.4WAV To Ogg123612182430Min: 24.05 / Avg: 24.08 / Max: 24.13Min: 24.03 / Avg: 24.08 / Max: 24.12Min: 24.03 / Avg: 24.1 / Max: 24.241. (CC) gcc options: -O2 -ffast-math -fsigned-char

Timed HMMer Search

This test searches through the Pfam database of profile hidden markov models. The search finds the domain structure of Drosophila Sevenless protein. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed HMMer Search 3.3.1Pfam Database Search123306090120150SE +/- 0.02, N = 3SE +/- 0.05, N = 3SE +/- 0.00, N = 3114.00114.07114.091. (CC) gcc options: -O3 -pthread -lhmmer -leasel -lm
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed HMMer Search 3.3.1Pfam Database Search12320406080100Min: 113.97 / Avg: 114 / Max: 114.04Min: 113.97 / Avg: 114.07 / Max: 114.15Min: 114.09 / Avg: 114.09 / Max: 114.11. (CC) gcc options: -O3 -pthread -lhmmer -leasel -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU12316003200480064008000SE +/- 4.52, N = 3SE +/- 3.54, N = 3SE +/- 7.47, N = 37400.517394.597399.59MIN: 7385.65MIN: 7380.26MIN: 7382.131. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread
OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.0Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU12313002600390052006500Min: 7392.25 / Avg: 7400.51 / Max: 7407.83Min: 7387.52 / Avg: 7394.59 / Max: 7398.23Min: 7389.05 / Avg: 7399.59 / Max: 7414.031. (CXX) g++ options: -O3 -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread

WavPack Audio Encoding

This test times how long it takes to encode a sample WAV file to WavPack format with very high quality settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWavPack Audio Encoding 5.3WAV To WavPack12348121620SE +/- 0.01, N = 5SE +/- 0.00, N = 5SE +/- 0.01, N = 516.3116.3016.301. (CXX) g++ options: -rdynamic
OpenBenchmarking.orgSeconds, Fewer Is BetterWavPack Audio Encoding 5.3WAV To WavPack12348121620Min: 16.29 / Avg: 16.31 / Max: 16.34Min: 16.28 / Avg: 16.3 / Max: 16.31Min: 16.29 / Avg: 16.3 / Max: 16.321. (CXX) g++ options: -rdynamic

RNNoise

RNNoise is a recurrent neural network for audio noise reduction developed by Mozilla and Xiph.Org. This test profile is a single-threaded test measuring the time to denoise a sample 26 minute long 16-bit RAW audio file using this recurrent neural network noise suppression library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterRNNoise 2020-06-28123612182430SE +/- 0.22, N = 3SE +/- 0.22, N = 3SE +/- 0.20, N = 325.6725.6625.661. (CC) gcc options: -O2 -pedantic -fvisibility=hidden
OpenBenchmarking.orgSeconds, Fewer Is BetterRNNoise 2020-06-28123612182430Min: 25.22 / Avg: 25.67 / Max: 25.9Min: 25.21 / Avg: 25.66 / Max: 25.9Min: 25.27 / Avg: 25.66 / Max: 25.861. (CC) gcc options: -O2 -pedantic -fvisibility=hidden

ASTC Encoder

ASTC Encoder (astcenc) is for the Adaptive Scalable Texture Compression (ASTC) format commonly used with OpenGL, OpenGL ES, and Vulkan graphics APIs. This test profile does a coding test of both compression/decompression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Thorough1231632486480SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.02, N = 372.6372.6072.621. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Thorough1231428425670Min: 72.62 / Avg: 72.63 / Max: 72.64Min: 72.58 / Avg: 72.6 / Max: 72.63Min: 72.59 / Avg: 72.62 / Max: 72.651. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

Basis Universal

Basis Universal is a GPU texture codoec. This test times how long it takes to convert sRGB PNGs into Basis Univeral assets with various settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBasis Universal 1.12Settings: UASTC Level 212320406080100SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 381.2481.2381.251. (CXX) g++ options: -std=c++11 -fvisibility=hidden -fPIC -fno-strict-aliasing -O3 -rdynamic -lm -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterBasis Universal 1.12Settings: UASTC Level 21231632486480Min: 81.23 / Avg: 81.24 / Max: 81.26Min: 81.21 / Avg: 81.23 / Max: 81.24Min: 81.24 / Avg: 81.25 / Max: 81.271. (CXX) g++ options: -std=c++11 -fvisibility=hidden -fPIC -fno-strict-aliasing -O3 -rdynamic -lm -lpthread

ASTC Encoder

ASTC Encoder (astcenc) is for the Adaptive Scalable Texture Compression (ASTC) format commonly used with OpenGL, OpenGL ES, and Vulkan graphics APIs. This test profile does a coding test of both compression/decompression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Exhaustive123130260390520650SE +/- 0.01, N = 3SE +/- 0.04, N = 3SE +/- 0.08, N = 3588.79588.80588.641. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Exhaustive123100200300400500Min: 588.77 / Avg: 588.79 / Max: 588.81Min: 588.76 / Avg: 588.8 / Max: 588.88Min: 588.47 / Avg: 588.64 / Max: 588.721. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

Basis Universal

Basis Universal is a GPU texture codoec. This test times how long it takes to convert sRGB PNGs into Basis Univeral assets with various settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBasis Universal 1.12Settings: UASTC Level 2 + RDO Post-Processing1232004006008001000SE +/- 0.26, N = 3SE +/- 0.24, N = 3SE +/- 0.27, N = 31001.731001.521001.731. (CXX) g++ options: -std=c++11 -fvisibility=hidden -fPIC -fno-strict-aliasing -O3 -rdynamic -lm -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterBasis Universal 1.12Settings: UASTC Level 2 + RDO Post-Processing1232004006008001000Min: 1001.21 / Avg: 1001.73 / Max: 1002.02Min: 1001.05 / Avg: 1001.52 / Max: 1001.84Min: 1001.41 / Avg: 1001.73 / Max: 1002.271. (CXX) g++ options: -std=c++11 -fvisibility=hidden -fPIC -fno-strict-aliasing -O3 -rdynamic -lm -lpthread

OpenBenchmarking.orgSeconds, Fewer Is BetterBasis Universal 1.12Settings: UASTC Level 31234080120160200SE +/- 0.03, N = 3SE +/- 0.02, N = 3SE +/- 0.01, N = 3159.00159.00158.981. (CXX) g++ options: -std=c++11 -fvisibility=hidden -fPIC -fno-strict-aliasing -O3 -rdynamic -lm -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterBasis Universal 1.12Settings: UASTC Level 3123306090120150Min: 158.95 / Avg: 159 / Max: 159.05Min: 158.95 / Avg: 158.99 / Max: 159.02Min: 158.97 / Avg: 158.98 / Max: 1591. (CXX) g++ options: -std=c++11 -fvisibility=hidden -fPIC -fno-strict-aliasing -O3 -rdynamic -lm -lpthread

ASTC Encoder

ASTC Encoder (astcenc) is for the Adaptive Scalable Texture Compression (ASTC) format commonly used with OpenGL, OpenGL ES, and Vulkan graphics APIs. This test profile does a coding test of both compression/decompression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Medium1233691215SE +/- 0.00, N = 3SE +/- 0.00, N = 3SE +/- 0.00, N = 310.9310.9310.931. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread
OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: Medium1233691215Min: 10.92 / Avg: 10.93 / Max: 10.93Min: 10.92 / Avg: 10.93 / Max: 10.93Min: 10.93 / Avg: 10.93 / Max: 10.931. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

Kvazaar

This is a test of Kvazaar as a CPU-based H.265 video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 4K - Video Preset: Medium1230.46350.9271.39051.8542.3175SE +/- 0.00, N = 3SE +/- 0.00, N = 3SE +/- 0.00, N = 32.062.062.061. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 4K - Video Preset: Medium123246810Min: 2.06 / Avg: 2.06 / Max: 2.06Min: 2.06 / Avg: 2.06 / Max: 2.06Min: 2.06 / Avg: 2.06 / Max: 2.071. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 4K - Video Preset: Slow1230.45230.90461.35691.80922.2615SE +/- 0.00, N = 3SE +/- 0.00, N = 3SE +/- 0.00, N = 32.012.012.011. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.0Video Input: Bosphorus 4K - Video Preset: Slow123246810Min: 2.01 / Avg: 2.01 / Max: 2.01Min: 2.01 / Avg: 2.01 / Max: 2.01Min: 2.01 / Avg: 2.01 / Max: 2.011. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

CLOMP

CLOMP is the C version of the Livermore OpenMP benchmark developed to measure OpenMP overheads and other performance impacts due to threading in order to influence future system designs. This particular test profile configuration is currently set to look at the OpenMP static schedule speed-up across all available CPU cores using the recommended test configuration. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSpeedup, More Is BetterCLOMP 1.2Static OMP Speedup1230.69751.3952.09252.793.4875SE +/- 0.03, N = 12SE +/- 0.03, N = 3SE +/- 0.03, N = 153.13.13.11. (CC) gcc options: -fopenmp -O3 -lm
OpenBenchmarking.orgSpeedup, More Is BetterCLOMP 1.2Static OMP Speedup123246810Min: 2.9 / Avg: 3.07 / Max: 3.3Min: 3 / Avg: 3.07 / Max: 3.1Min: 2.9 / Avg: 3.1 / Max: 3.31. (CC) gcc options: -fopenmp -O3 -lm

Redis

Redis is an open-source data structure server. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: GET123500K1000K1500K2000K2500KSE +/- 41608.51, N = 3SE +/- 51226.02, N = 15SE +/- 26746.34, N = 32215813.501939149.382041822.961. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: GET123400K800K1200K1600K2000KMin: 2136889 / Avg: 2215813.5 / Max: 2278123Min: 1508874.88 / Avg: 1939149.38 / Max: 2141327.5Min: 1992159.5 / Avg: 2041822.96 / Max: 2083866.751. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: SADD123400K800K1200K1600K2000KSE +/- 57333.88, N = 15SE +/- 50997.94, N = 15SE +/- 56308.56, N = 151661804.391761393.521687783.911. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: SADD123300K600K900K1200K1500KMin: 1422520.62 / Avg: 1661804.39 / Max: 1919631.38Min: 1449692.75 / Avg: 1761393.52 / Max: 1923200.12Min: 1418893.62 / Avg: 1687783.91 / Max: 1920122.751. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: LPOP123500K1000K1500K2000K2500KSE +/- 28970.62, N = 8SE +/- 20132.66, N = 3SE +/- 116386.47, N = 122368415.091438792.172024710.521. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: LPOP123400K800K1200K1600K2000KMin: 2183685.5 / Avg: 2368415.09 / Max: 2433323.5Min: 1398691 / Avg: 1438792.17 / Max: 1461988.25Min: 1364343.88 / Avg: 2024710.52 / Max: 2365049.51. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

HPC Challenge

HPC Challenge (HPCC) is a cluster-focused benchmark consisting of the HPL Linpack TPP benchmark, DGEMM, STREAM, PTRANS, RandomAccess, FFT, and communication bandwidth and latency. This HPC Challenge test profile attempts to ship with standard yet versatile configuration/input files though they can be modified. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGB/s, More Is BetterHPC Challenge 1.5.0Test / Class: Random Ring Bandwidth1231.30832.61663.92495.23326.5415SE +/- 0.18268, N = 3SE +/- 0.02203, N = 3SE +/- 0.18898, N = 35.463525.814875.360421. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2
OpenBenchmarking.orgGB/s, More Is BetterHPC Challenge 1.5.0Test / Class: Random Ring Bandwidth123246810Min: 5.21 / Avg: 5.46 / Max: 5.82Min: 5.77 / Avg: 5.81 / Max: 5.84Min: 5.16 / Avg: 5.36 / Max: 5.741. (CC) gcc options: -lblas -lm -pthread -lmpi -fomit-frame-pointer -funroll-loops2. ATLAS + Open MPI 3.1.2

Betsy GPU Compressor

Betsy is an open-source GPU compressor of various GPU compression techniques. Betsy is written in GLSL for Vulkan/OpenGL (compute shader) support for GPU-based texture compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBetsy GPU Compressor 1.1 BetaCodec: ETC2 RGB - Quality: Highest12348121620SE +/- 0.51, N = 13SE +/- 0.32, N = 15SE +/- 0.19, N = 1516.1315.3115.721. (CXX) g++ options: -O3 -O2 -lpthread -ldl
OpenBenchmarking.orgSeconds, Fewer Is BetterBetsy GPU Compressor 1.1 BetaCodec: ETC2 RGB - Quality: Highest12348121620Min: 13.98 / Avg: 16.13 / Max: 21.66Min: 12 / Avg: 15.31 / Max: 16.01Min: 13.73 / Avg: 15.72 / Max: 16.021. (CXX) g++ options: -O3 -O2 -lpthread -ldl

OpenBenchmarking.orgSeconds, Fewer Is BetterBetsy GPU Compressor 1.1 BetaCodec: ETC1 - Quality: Highest12348121620SE +/- 0.55, N = 15SE +/- 0.45, N = 14SE +/- 0.04, N = 315.4014.7015.961. (CXX) g++ options: -O3 -O2 -lpthread -ldl
OpenBenchmarking.orgSeconds, Fewer Is BetterBetsy GPU Compressor 1.1 BetaCodec: ETC1 - Quality: Highest12348121620Min: 11.98 / Avg: 15.4 / Max: 21Min: 11.99 / Avg: 14.7 / Max: 16.01Min: 15.89 / Avg: 15.96 / Max: 161. (CXX) g++ options: -O3 -O2 -lpthread -ldl

122 Results Shown

oneDNN:
  Matrix Multiply Batch Shapes Transformer - f32 - CPU
  IP Shapes 3D - u8s8f32 - CPU
  Deconvolution Batch shapes_3d - f32 - CPU
  IP Shapes 3D - f32 - CPU
  Deconvolution Batch shapes_1d - f32 - CPU
  Convolution Batch Shapes Auto - u8s8f32 - CPU
  IP Shapes 1D - f32 - CPU
  Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU
Mlpack Benchmark
yquake2
Mlpack Benchmark
oneDNN
Redis
Unpacking Firefox
Mlpack Benchmark
Sunflow Rendering System
HPC Challenge
yquake2
Embree
Build2
oneDNN
Caffe
oneDNN
eSpeak-NG Speech Engine
NCNN
LZ4 Compression
Numpy Benchmark
asmFish
FFTE
LZ4 Compression
NCNN
HPC Challenge
Basis Universal
LAMMPS Molecular Dynamics Simulator
NCNN
Timed FFmpeg Compilation
Dolfyn
Crafty
NCNN
Embree
NCNN
HPC Challenge
x265
HPC Challenge
oneDNN
Stockfish
NCNN:
  CPU - efficientnet-b0
  CPU - yolov4-tiny
Timed MAFFT Alignment
NCNN
Redis
NCNN
rav1e
NCNN
oneDNN
KeyDB
x265
LZ4 Compression:
  1 - Decompression Speed
  3 - Compression Speed
HPC Challenge
yquake2
oneDNN
SQLite Speedtest
Kvazaar
Coremark
rav1e
HPC Challenge
NCNN
HPC Challenge
LZ4 Compression
GROMACS
Monkey Audio Encoding
LZ4 Compression
NCNN
BRL-CAD
Hierarchical INTegration
Kvazaar:
  Bosphorus 4K - Very Fast
  Bosphorus 1080p - Very Fast
rav1e
Caffe
NCNN
Kvazaar
Embree
oneDNN:
  Recurrent Neural Network Training - bf16bf16bf16 - CPU
  Deconvolution Batch shapes_3d - u8s8f32 - CPU
BYTE Unix Benchmark
Kvazaar
IndigoBench
NCNN
Kvazaar
IndigoBench
Timed Eigen Compilation
HPC Challenge
NCNN
Embree
rav1e
Mlpack Benchmark
ASTC Encoder
Basis Universal
oneDNN
Opus Codec Encoding
PHPBench
Ogg Audio Encoding
Timed HMMer Search
oneDNN
WavPack Audio Encoding
RNNoise
ASTC Encoder
Basis Universal
ASTC Encoder
Basis Universal:
  UASTC Level 2 + RDO Post-Processing
  UASTC Level 3
ASTC Encoder
Kvazaar:
  Bosphorus 4K - Medium
  Bosphorus 4K - Slow
CLOMP
Redis:
  GET
  SADD
  LPOP
HPC Challenge
Betsy GPU Compressor:
  ETC2 RGB - Highest
  ETC1 - Highest