Cascade Lake Clear Linux 2021

2 x Intel Xeon Platinum 8280 testing with a GIGABYTE MD61-SC2-00 v01000100 (T15 BIOS) and llvmpipe on Ubuntu 21.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2103273-IB-2103266IB69
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts

Limit displaying results to tests within:

AV1 3 Tests
C/C++ Compiler Tests 6 Tests
Compression Tests 2 Tests
CPU Massive 12 Tests
Creator Workloads 14 Tests
Encoding 5 Tests
Fortran Tests 2 Tests
Game Development 3 Tests
HPC - High Performance Computing 9 Tests
Machine Learning 4 Tests
Molecular Dynamics 3 Tests
MPI Benchmarks 3 Tests
Multi-Core 19 Tests
NVIDIA GPU Compute 5 Tests
Intel oneAPI 5 Tests
OpenMPI Tests 4 Tests
Programmer / Developer System Benchmarks 2 Tests
Renderers 4 Tests
Scientific Computing 3 Tests
Server CPU Tests 11 Tests
Video Encoding 5 Tests
Common Workstation Benchmarks 2 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
Clear Linux 34420
March 26 2021
  5 Hours, 57 Minutes
Ubuntu 21.04 Dev
March 26 2021
  8 Hours, 35 Minutes
Invert Hiding All Results Option
  7 Hours, 16 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


Cascade Lake Clear Linux 2021ProcessorMotherboardChipsetMemoryDiskGraphicsMonitorNetworkOSKernelDesktopDisplay ServerOpenGLCompilerFile-SystemScreen ResolutionClear Linux 34420Ubuntu 21.04 Dev2 x Intel Xeon Platinum 8280 @ 4.00GHz (56 Cores / 112 Threads)GIGABYTE MD61-SC2-00 v01000100 (T15 BIOS)Intel Sky Lake-E DMI3 Registers378GB280GB INTEL SSDPED1D280GAllvmpipeVE2282 x Intel X722 for 1GbE + 2 x QLogic FastLinQ QL41000 10/25/40/50GbEClear Linux OS 344205.10.19-1032.native (x86_64)GNOME Shell 3.38.4X Server 1.20.104.5 Mesa 20.3.4 (LLVM 10.0.1 256 bits)GCC 10.2.1 20210324 releases/gcc-10.2.0-1013-g592388d4f6 + Clang 10.0.1 + LLVM 10.0.1ext41920x1080Ubuntu 21.045.11.0-11-generic (x86_64)GNOME Shell 3.38.3X Server4.5 Mesa 21.0.0 (LLVM 11.0.1 256 bits)GCC 10.2.1 20210320OpenBenchmarking.orgKernel Details- Clear Linux 34420: Transparent Huge Pages: always- Ubuntu 21.04 Dev: Transparent Huge Pages: madviseEnvironment Details- Clear Linux 34420: FFLAGS="-g -O3 -feliminate-unused-debug-types -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=32 -m64 -fasynchronous-unwind-tables -Wp,-D_REENTRANT -ftree-loop-distribute-patterns -Wl,-z -Wl,now -Wl,-z -Wl,relro -malign-data=abi -fno-semantic-interposition -ftree-vectorize -ftree-loop-vectorize -Wl,--enable-new-dtags -Wa,-mbranches-within-32B-boundaries" CXXFLAGS="-g -O3 -feliminate-unused-debug-types -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=32 -Wformat -Wformat-security -m64 -fasynchronous-unwind-tables -Wp,-D_REENTRANT -ftree-loop-distribute-patterns -Wl,-z -Wl,now -Wl,-z -Wl,relro -fno-semantic-interposition -ffat-lto-objects -fno-trapping-math -Wl,-sort-common -Wl,--enable-new-dtags -mtune=skylake -Wa,-mbranches-within-32B-boundaries -fvisibility-inlines-hidden -Wl,--enable-new-dtags" MESA_GLSL_CACHE_DISABLE=0 FCFLAGS="-g -O3 -feliminate-unused-debug-types -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=32 -m64 -fasynchronous-unwind-tables -Wp,-D_REENTRANT -ftree-loop-distribute-patterns -Wl,-z -Wl,now -Wl,-z -Wl,relro -malign-data=abi -fno-semantic-interposition -ftree-vectorize -ftree-loop-vectorize -Wl,-sort-common -Wl,--enable-new-dtags" CFLAGS="-g -O3 -feliminate-unused-debug-types -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=32 -Wformat -Wformat-security -m64 -fasynchronous-unwind-tables -Wp,-D_REENTRANT -ftree-loop-distribute-patterns -Wl,-z -Wl,now -Wl,-z -Wl,relro -fno-semantic-interposition -ffat-lto-objects -fno-trapping-math -Wl,-sort-common -Wl,--enable-new-dtags -mtune=skylake -Wa,-mbranches-within-32B-boundaries" THEANO_FLAGS="floatX=float32,openmp=true,gcc.cxxflags="-ftree-vectorize -mavx"" Compiler Details- Clear Linux 34420: --build=x86_64-generic-linux --disable-libmpx --disable-libunwind-exceptions --disable-multiarch --disable-vtable-verify --disable-werror --enable-__cxa_atexit --enable-bootstrap --enable-cet --enable-clocale=gnu --enable-default-pie --enable-gnu-indirect-function --enable-languages=c,c++,fortran,go --enable-ld=default --enable-libstdcxx-pch --enable-lto --enable-multilib --enable-plugin --enable-shared --enable-threads=posix --exec-prefix=/usr --includedir=/usr/include --target=x86_64-generic-linux --with-arch=westmere --with-gcc-major-version-only --with-glibc-version=2.19 --with-gnu-ld --with-isl --with-ppl=yes --with-tune=haswell - Ubuntu 21.04 Dev: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-mutex --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-DjbZbO/gcc-10-10.2.1/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-DjbZbO/gcc-10-10.2.1/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- Clear Linux 34420: Scaling Governor: intel_pstate performance - CPU Microcode: 0x5003006- Ubuntu 21.04 Dev: Scaling Governor: intel_pstate powersave - CPU Microcode: 0x5003003Python Details- Clear Linux 34420: Python 3.9.2Security Details- itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Mitigation of TSX disabled

Clear Linux 34420 vs. Ubuntu 21.04 Dev ComparisonPhoronix Test SuiteBaseline+47.5%+47.5%+95%+95%+142.5%+142.5%22.1%10.1%8.7%4.1%3.3%3.3%2.8%NASNet Mobile189.8%8, Long Mode - Compression Speed156.5%Chimera 1080p147.4%C.1.1.b136.1%S.N.1123.9%Fishy Cat - CPU-Only72.4%SqueezeNet70.3%Mobilenet Quant67.8%Mobilenet Float63.3%Summer Nature 4K62.3%1050.2%10 - Bosphorus 1080p46%Inception V444.8%IP Shapes 3D - bf16bf16bf16 - CPU42.5%I.R.V41.9%V.Q.O - Bosphorus 1080p38.7%Enc Mode 8 - 1080p35.7%10, Lossless35.4%P.S.O - Bosphorus 1080p34%CPU - MobileNet v231.4%VMAF Optimized - Bosphorus 1080p29.6%R.N.N.I - bf16bf16bf16 - CPU28.8%R.N.N.I - u8s8f32 - CPU28.6%628.6%vklBenchmark26.5%M.M.B.S.T - u8s8f32 - CPU24.8%D.B.s - bf16bf16bf16 - CPU23.9%D.B.s - u8s8f32 - CPU23.8%19 - D.S7 - Bosphorus 1080p21.5%R.N.N.T - bf16bf16bf16 - CPU21.4%R.N.N.T - u8s8f32 - CPU21.3%mobilenet-v1-1.021.2%OpenMP Leukocyte21%112 - 256 - 5720.7%6, Lossless20.7%1 - 256 - 5720%OpenMP CFD Solver18.2%16 - 256 - 5717.4%Enc Mode 4 - 1080p17.2%D.B.s - u8s8f32 - CPU17.1%DistinctUserID16.6%Enc Mode 0 - 1080p16.1%215.7%32 - 256 - 5715%PartialTweets14.2%resnet-v2-5013.9%19 - Compression Speed13.3%M.M.B.S.T - bf16bf16bf16 - CPU13.1%VoiceMark_10012.2%IP Shapes 1D - bf16bf16bf16 - CPU12.1%C.B.S.A - bf16bf16bf16 - CPU10.1%19, Long Mode - D.SC.B.S.A - u8s8f32 - CPU9.3%EP.D8.9%LU.CLargeRand8.6%64 - 256 - 578.6%OpenMP LavaMD8.4%ATPase Simulation - 327,506 Atoms7.8%inception-v37.4%O.S6.4%D.B.s - bf16bf16bf16 - CPU5.5%SqueezeNetV1.05.4%1 - Bosphorus 1080p5.4%CPU5.4%Memorial4.7%9 - Compression Speed4.4%OpenMP HotSpot3D4.4%IP Shapes 1D - u8s8f32 - CPUCPU - SqueezeNet v1.14%Kostya3.7%Pathtracer - Asian Dragon3.4%CPU - Bedroom3.4%19, Long Mode - Compression Speed3 - D.SCPU - Supercar3.2%9 - D.SMobileNetV2_2242.7%Pathtracer - Crown2.6%Total Time2.6%X.b.i.i2.1%San Miguel - Path Tracer2%TensorFlow LiteZstd Compressiondav1ddav1ddav1dBlenderTensorFlow LiteTensorFlow LiteTensorFlow Litedav1dlibavif avifencSVT-HEVCTensorFlow LiteoneDNNTensorFlow LiteSVT-VP9SVT-AV1libavif avifencSVT-VP9TNNSVT-VP9oneDNNoneDNNlibavif avifencOpenVKLoneDNNoneDNNoneDNNZstd CompressionSVT-HEVConeDNNoneDNNMobile Neural NetworkRodiniaLiquid-DSPlibavif avifencLiquid-DSPRodiniaLiquid-DSPSVT-AV1oneDNNsimdjsonSVT-AV1libavif avifencLiquid-DSPsimdjsonMobile Neural NetworkZstd CompressiononeDNNGoogle SynthMarkoneDNNoneDNNZstd CompressiononeDNNNAS Parallel BenchmarksNAS Parallel BenchmarkssimdjsonLiquid-DSPRodiniaNAMDMobile Neural NetworkRodiniaoneDNNMobile Neural NetworkSVT-HEVCChaos Group V-RAYIntel Open Image DenoiseLZ4 CompressionRodiniaoneDNNTNNsimdjsonEmbreeIndigoBenchZstd CompressionLZ4 CompressionIndigoBenchLZ4 CompressionMobile Neural NetworkEmbreeStockfishXcompact3d Incompact3dOSPrayClear Linux 34420Ubuntu 21.04 Dev

Cascade Lake Clear Linux 2021compress-zstd: 8, Long Mode - Compression Speeddav1d: Chimera 1080pdav1d: Chimera 1080p 10-bitdav1d: Summer Nature 1080ptensorflow-lite: SqueezeNetdav1d: Summer Nature 4Kavifenc: 10svt-hevc: 10 - Bosphorus 1080ptensorflow-lite: Inception V4tensorflow-lite: Inception ResNet V2svt-vp9: Visual Quality Optimized - Bosphorus 1080psvt-av1: Enc Mode 8 - 1080pavifenc: 10, Losslesssvt-vp9: PSNR/SSIM Optimized - Bosphorus 1080ptnn: CPU - MobileNet v2svt-vp9: VMAF Optimized - Bosphorus 1080ponednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUavifenc: 6openvkl: vklBenchmarkonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUonednn: Deconvolution Batch shapes_1d - bf16bf16bf16 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUcompress-zstd: 19 - Decompression Speedsvt-hevc: 7 - Bosphorus 1080ponednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUmnn: mobilenet-v1-1.0liquid-dsp: 112 - 256 - 57avifenc: 6, Losslessliquid-dsp: 1 - 256 - 57liquid-dsp: 16 - 256 - 57svt-av1: Enc Mode 4 - 1080ponednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUsimdjson: DistinctUserIDsvt-av1: Enc Mode 0 - 1080pavifenc: 2liquid-dsp: 32 - 256 - 57simdjson: PartialTweetsmnn: resnet-v2-50compress-zstd: 19 - Compression Speedonednn: Matrix Multiply Batch Shapes Transformer - bf16bf16bf16 - CPUsynthmark: VoiceMark_100onednn: IP Shapes 1D - bf16bf16bf16 - CPUonednn: Convolution Batch Shapes Auto - bf16bf16bf16 - CPUcompress-zstd: 19, Long Mode - Decompression Speedonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUnpb: EP.Dnpb: LU.Csimdjson: LargeRandliquid-dsp: 64 - 256 - 57rodinia: OpenMP LavaMDnamd: ATPase Simulation - 327,506 Atomsmnn: inception-v3onednn: Deconvolution Batch shapes_3d - bf16bf16bf16 - CPUmnn: SqueezeNetV1.0svt-hevc: 1 - Bosphorus 1080pv-ray: CPUoidn: Memorialcompress-lz4: 9 - Compression Speedrodinia: OpenMP HotSpot3Donednn: IP Shapes 1D - u8s8f32 - CPUtnn: CPU - SqueezeNet v1.1simdjson: Kostyaembree: Pathtracer - Asian Dragonindigobench: CPU - Bedroomcompress-zstd: 19, Long Mode - Compression Speedcompress-lz4: 3 - Decompression Speedindigobench: CPU - Supercarcompress-lz4: 9 - Decompression Speedmnn: MobileNetV2_224embree: Pathtracer - Crownstockfish: Total Timeincompact3d: X3D-benchmarking input.i3dospray: San Miguel - Path Tracerospray: NASA Streamlines - Path Tracerblender: BMW27 - CPU-Onlyincompact3d: input.i3d 193 Cells Per Directiononednn: IP Shapes 3D - u8s8f32 - CPUembree: Pathtracer ISPC - Crowngromacs: water_GMX50_bareospray: XFrog Forest - Path Tracerincompact3d: input.i3d 129 Cells Per Directioncompress-lz4: 3 - Compression Speedembree: Pathtracer ISPC - Asian Dragoncompress-zstd: 8, Long Mode - Decompression Speedospray: Magnetic Reconnection - Path Tracerblender: Fishy Cat - CPU-Onlytensorflow-lite: Mobilenet Quanttensorflow-lite: Mobilenet Floattensorflow-lite: NASNet Mobileonednn: IP Shapes 3D - bf16bf16bf16 - CPUrodinia: OpenMP Streamclusterrodinia: OpenMP CFD Solverrodinia: OpenMP LeukocyteClear Linux 34420Ubuntu 21.04 Dev834.81038.26240.091010.3861863.5396.463.669480.50883114811638291.2474.6326.442356.90330.630348.26477.715476.65611.7345110.3302184.039060.3093971889.2270.36895.445900.3392.672285363333328.713587123338228633339.0090.3837724.210.14430.12115186333334.0125.73384.30.833080614.4573.623753.195912087.93.818936759.15125567.330.88231590000053.5470.3717032.4774.428308.15730.094908237.4446.81111.1451.71888330.1592.5261.21178.57638.96096.920.2026135.44.75051.3869136190317458.1748256.6218.5237.8419.33319661.03871454.70895.7596.674.4892463748.0070.482050055.7149894.151318.397445.72.1106112.9016.54152.679325.5419.62101.70451.23105347244.275.512329.1612790191151590209.9354.9948.725266.40434.391268.71615.292613.07215.0854040.4122645.005840.3830702306.2222.571087.271092.423.239236340000034.659489456677011933337.6880.4492783.610.12434.86113203666673.5129.32074.40.942426547.4984.063723.517892297.94.174846207.05136533.760.81213316666758.0530.4008534.8814.670698.59928.554657335.7544.83116.0081.65179343.3452.4359.18838.29740.26297.019.5796308.04.88050.0961132800942467.6321116.4918.1838.5219.01316641.0554654.13375.7106.624.4697863348.1970.71582814.250096.0383725.483813.32824173.0080513.7267.73463.719OpenBenchmarking.org

Zstd Compression

This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.4.9Compression Level: 8, Long Mode - Compression SpeedUbuntu 21.04 DevClear Linux 344202004006008001000SE +/- 2.07, N = 15SE +/- 2.10, N = 3325.5834.8-llzma-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CC) gcc options: -O3 -pthread -lz

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.8.2Video Input: Chimera 1080pUbuntu 21.04 DevClear Linux 344202004006008001000SE +/- 1.21, N = 3SE +/- 8.00, N = 10419.621038.26MIN: 231.67 / MAX: 545.48-O3 -pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 367.3 / MAX: 1324.251. (CC) gcc options: -pthread

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.8.2Video Input: Chimera 1080p 10-bitUbuntu 21.04 DevClear Linux 3442050100150200250SE +/- 0.18, N = 3SE +/- 0.21, N = 3101.70240.09MIN: 79.78 / MAX: 152.75-O3 -pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 170.42 / MAX: 370.951. (CC) gcc options: -pthread

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.8.2Video Input: Summer Nature 1080pUbuntu 21.04 DevClear Linux 344202004006008001000SE +/- 0.16, N = 3SE +/- 4.74, N = 3451.231010.38MIN: 169.52 / MAX: 513.73-O3 -pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 375.41 / MAX: 1131.641. (CC) gcc options: -pthread

TensorFlow Lite

This is a benchmark of the TensorFlow Lite implementation. The current Linux support is limited to running on CPUs. This test profile is measuring the average inference time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: SqueezeNetUbuntu 21.04 DevClear Linux 3442020K40K60K80K100KSE +/- 1493.06, N = 3SE +/- 312.31, N = 3105347.061863.5

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.8.2Video Input: Summer Nature 4KUbuntu 21.04 DevClear Linux 3442090180270360450SE +/- 0.46, N = 3SE +/- 1.88, N = 3244.27396.46MIN: 101.68 / MAX: 272.3-O3 -pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 163.78 / MAX: 435.21. (CC) gcc options: -pthread

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.9.0Encoder Speed: 10Ubuntu 21.04 DevClear Linux 344201.24022.48043.72064.96086.201SE +/- 0.042, N = 15SE +/- 0.020, N = 35.5123.6691. (CXX) g++ options: -O3 -fPIC -lm

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 10 - Input: Bosphorus 1080pUbuntu 21.04 DevClear Linux 34420100200300400500SE +/- 2.14, N = 3SE +/- 5.37, N = 3329.16480.50-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

TensorFlow Lite

This is a benchmark of the TensorFlow Lite implementation. The current Linux support is limited to running on CPUs. This test profile is measuring the average inference time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception V4Ubuntu 21.04 DevClear Linux 34420300K600K900K1200K1500KSE +/- 12025.89, N = 7SE +/- 3472.47, N = 31279019883114

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception ResNet V2Ubuntu 21.04 DevClear Linux 34420200K400K600K800K1000KSE +/- 3000.84, N = 3SE +/- 5456.47, N = 31151590811638

SVT-VP9

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample YUV input video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-VP9 0.3Tuning: Visual Quality Optimized - Input: Bosphorus 1080pUbuntu 21.04 DevClear Linux 3442060120180240300SE +/- 1.85, N = 8SE +/- 1.15, N = 3209.93291.24-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm

SVT-AV1

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-AV1 CPU-based multi-threaded video encoder for the AV1 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8Encoder Mode: Enc Mode 8 - Input: 1080pUbuntu 21.04 DevClear Linux 3442020406080100SE +/- 0.35, N = 3SE +/- 0.78, N = 554.9974.631. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.9.0Encoder Speed: 10, LosslessUbuntu 21.04 DevClear Linux 34420246810SE +/- 0.016, N = 3SE +/- 0.039, N = 38.7256.4421. (CXX) g++ options: -O3 -fPIC -lm

SVT-VP9

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample YUV input video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-VP9 0.3Tuning: PSNR/SSIM Optimized - Input: Bosphorus 1080pUbuntu 21.04 DevClear Linux 3442080160240320400SE +/- 2.74, N = 3SE +/- 1.54, N = 3266.40356.90-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm

TNN

TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.2.3Target: CPU - Model: MobileNet v2Ubuntu 21.04 DevClear Linux 3442090180270360450SE +/- 3.34, N = 3SE +/- 0.21, N = 3434.39330.63MIN: 363.56 / MAX: 546.49-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 329.78 / MAX: 333.621. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -O3 -rdynamic -ldl

SVT-VP9

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample YUV input video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-VP9 0.3Tuning: VMAF Optimized - Input: Bosphorus 1080pUbuntu 21.04 DevClear Linux 3442080160240320400SE +/- 2.04, N = 3SE +/- 2.43, N = 3268.71348.26-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUUbuntu 21.04 DevClear Linux 34420130260390520650SE +/- 8.22, N = 3SE +/- 2.69, N = 3615.29477.72MIN: 568.58-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 464.941. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUUbuntu 21.04 DevClear Linux 34420130260390520650SE +/- 7.21, N = 3SE +/- 1.18, N = 3613.07476.66MIN: 556.95-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 4661. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.9.0Encoder Speed: 6Ubuntu 21.04 DevClear Linux 3442048121620SE +/- 0.13, N = 15SE +/- 0.11, N = 715.0911.731. (CXX) g++ options: -O3 -fPIC -lm

OpenVKL

OpenVKL is the Intel Open Volume Kernel Library that offers high-performance volume computation kernels and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 0.9Benchmark: vklBenchmarkUbuntu 21.04 DevClear Linux 34420110220330440550SE +/- 4.45, N = 9SE +/- 2.91, N = 3404511MIN: 1 / MAX: 1613MIN: 1 / MAX: 1812

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUUbuntu 21.04 DevClear Linux 344200.09280.18560.27840.37120.464SE +/- 0.004170, N = 3SE +/- 0.002990, N = 70.4122640.330218MIN: 0.28-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 0.281. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPUUbuntu 21.04 DevClear Linux 344201.12632.25263.37894.50525.6315SE +/- 0.00960, N = 3SE +/- 0.01370, N = 35.005844.03906MIN: 4.26-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 3.941. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUUbuntu 21.04 DevClear Linux 344200.08620.17240.25860.34480.431SE +/- 0.002916, N = 3SE +/- 0.003975, N = 150.3830700.309397MIN: 0.31-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 0.281. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Zstd Compression

This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.4.9Compression Level: 19 - Decompression SpeedClear Linux 34420Ubuntu 21.04 Dev5001000150020002500SE +/- 1.54, N = 15SE +/- 12.81, N = 41889.22306.2-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake-llzma1. (CC) gcc options: -O3 -pthread -lz

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 7 - Input: Bosphorus 1080pUbuntu 21.04 DevClear Linux 3442060120180240300SE +/- 2.15, N = 3SE +/- 2.46, N = 3222.57270.36-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUUbuntu 21.04 DevClear Linux 344202004006008001000SE +/- 5.62, N = 3SE +/- 2.11, N = 31087.27895.45MIN: 1031.43-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 885.581. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUUbuntu 21.04 DevClear Linux 344202004006008001000SE +/- 9.54, N = 7SE +/- 3.62, N = 31092.42900.34MIN: 1002.12-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 883.531. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.3Model: mobilenet-v1-1.0Ubuntu 21.04 DevClear Linux 344200.72881.45762.18642.91523.644SE +/- 0.038, N = 15SE +/- 0.010, N = 153.2392.672MIN: 2.44 / MAX: 9.7-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 2.48 / MAX: 4.131. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 2021.01.31Threads: 112 - Buffer Length: 256 - Filter Length: 57Ubuntu 21.04 DevClear Linux 34420600M1200M1800M2400M3000MSE +/- 5168171.82, N = 3SE +/- 3573202.73, N = 323634000002853633333-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.9.0Encoder Speed: 6, LosslessUbuntu 21.04 DevClear Linux 34420816243240SE +/- 0.06, N = 3SE +/- 0.32, N = 534.6628.711. (CXX) g++ options: -O3 -fPIC -lm

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 2021.01.31Threads: 1 - Buffer Length: 256 - Filter Length: 57Ubuntu 21.04 DevClear Linux 3442013M26M39M52M65MSE +/- 478618.06, N = 6SE +/- 6887.99, N = 34894566758712333-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 2021.01.31Threads: 16 - Buffer Length: 256 - Filter Length: 57Ubuntu 21.04 DevClear Linux 34420200M400M600M800M1000MSE +/- 2162963.50, N = 3SE +/- 1029827.39, N = 3701193333822863333-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

SVT-AV1

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-AV1 CPU-based multi-threaded video encoder for the AV1 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8Encoder Mode: Enc Mode 4 - Input: 1080pUbuntu 21.04 DevClear Linux 344203691215SE +/- 0.060, N = 3SE +/- 0.032, N = 37.6889.0091. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUUbuntu 21.04 DevClear Linux 344200.10110.20220.30330.40440.5055SE +/- 0.006464, N = 3SE +/- 0.001058, N = 30.4492780.383772MIN: 0.34-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 0.341. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

simdjson

This is a benchmark of SIMDJSON, a high performance JSON parser. SIMDJSON aims to be the fastest JSON parser and is used by projects like Microsoft FishStore, Yandex ClickHouse, Shopify, and others. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 0.8.2Throughput Test: DistinctUserIDUbuntu 21.04 DevClear Linux 344200.94731.89462.84193.78924.7365SE +/- 0.01, N = 3SE +/- 0.01, N = 33.614.21-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CXX) g++ options: -O3 -pthread

SVT-AV1

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-AV1 CPU-based multi-threaded video encoder for the AV1 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8Encoder Mode: Enc Mode 0 - Input: 1080pUbuntu 21.04 DevClear Linux 344200.03240.06480.09720.12960.162SE +/- 0.000, N = 3SE +/- 0.001, N = 30.1240.1441. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.9.0Encoder Speed: 2Ubuntu 21.04 DevClear Linux 34420816243240SE +/- 0.20, N = 3SE +/- 0.42, N = 334.8630.121. (CXX) g++ options: -O3 -fPIC -lm

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 2021.01.31Threads: 32 - Buffer Length: 256 - Filter Length: 57Ubuntu 21.04 DevClear Linux 34420300M600M900M1200M1500MSE +/- 9745141.24, N = 3SE +/- 3507293.99, N = 313203666671518633333-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

simdjson

This is a benchmark of SIMDJSON, a high performance JSON parser. SIMDJSON aims to be the fastest JSON parser and is used by projects like Microsoft FishStore, Yandex ClickHouse, Shopify, and others. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 0.8.2Throughput Test: PartialTweetsUbuntu 21.04 DevClear Linux 344200.90231.80462.70693.60924.5115SE +/- 0.01, N = 3SE +/- 0.01, N = 33.514.01-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CXX) g++ options: -O3 -pthread

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.3Model: resnet-v2-50Ubuntu 21.04 DevClear Linux 34420714212835SE +/- 0.13, N = 15SE +/- 0.13, N = 1529.3225.73MIN: 27.43 / MAX: 115.46-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 24.56 / MAX: 67.21. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

Zstd Compression

This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.4.9Compression Level: 19 - Compression SpeedUbuntu 21.04 DevClear Linux 3442020406080100SE +/- 0.86, N = 4SE +/- 0.68, N = 1574.484.3-llzma-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CC) gcc options: -O3 -pthread -lz

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPUUbuntu 21.04 DevClear Linux 344200.2120.4240.6360.8481.06SE +/- 0.004903, N = 3SE +/- 0.006956, N = 30.9424260.833080MIN: 0.76-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 0.771. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Google SynthMark

SynthMark is a cross platform tool for benchmarking CPU performance under a variety of real-time audio workloads. It uses a polyphonic synthesizer model to provide standardized tests for latency, jitter and computational throughput. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgVoices, More Is BetterGoogle SynthMark 20201109Test: VoiceMark_100Ubuntu 21.04 DevClear Linux 34420130260390520650SE +/- 0.98, N = 3SE +/- 0.92, N = 3547.50614.461. (CXX) g++ options: -lm -lpthread -std=c++11 -Ofast

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPUUbuntu 21.04 DevClear Linux 344200.91431.82862.74293.65724.5715SE +/- 0.03832, N = 3SE +/- 0.01053, N = 34.063723.62375MIN: 3.31-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 3.411. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPUUbuntu 21.04 DevClear Linux 344200.79151.5832.37453.1663.9575SE +/- 0.03291, N = 3SE +/- 0.00586, N = 33.517893.19591MIN: 2.94-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 2.991. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Zstd Compression

This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.4.9Compression Level: 19, Long Mode - Decompression SpeedClear Linux 34420Ubuntu 21.04 Dev5001000150020002500SE +/- 2.64, N = 4SE +/- 8.87, N = 152087.92297.9-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake-llzma1. (CC) gcc options: -O3 -pthread -lz

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUUbuntu 21.04 DevClear Linux 344200.93931.87862.81793.75724.6965SE +/- 0.03967, N = 3SE +/- 0.02129, N = 34.174843.81893MIN: 3.55-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 3.721. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.DUbuntu 21.04 DevClear Linux 3442014002800420056007000SE +/- 21.90, N = 3SE +/- 29.20, N = 36207.056759.15-lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi2. Ubuntu 21.04 Dev: Open MPI 4.1.03. Clear Linux 34420: 3.2

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: LU.CClear Linux 34420Ubuntu 21.04 Dev30K60K90K120K150KSE +/- 241.57, N = 3SE +/- 92.89, N = 3125567.33136533.76-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake-lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi2. Clear Linux 34420: 3.23. Ubuntu 21.04 Dev: Open MPI 4.1.0

simdjson

This is a benchmark of SIMDJSON, a high performance JSON parser. SIMDJSON aims to be the fastest JSON parser and is used by projects like Microsoft FishStore, Yandex ClickHouse, Shopify, and others. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 0.8.2Throughput Test: LargeRandomUbuntu 21.04 DevClear Linux 344200.1980.3960.5940.7920.99SE +/- 0.00, N = 3SE +/- 0.00, N = 30.810.88-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CXX) g++ options: -O3 -pthread

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 2021.01.31Threads: 64 - Buffer Length: 256 - Filter Length: 57Ubuntu 21.04 DevClear Linux 34420500M1000M1500M2000M2500MSE +/- 33333.33, N = 3SE +/- 22409893.65, N = 321331666672315900000-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

Rodinia

Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes select OpenCL, NVIDIA CUDA and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP LavaMDUbuntu 21.04 DevClear Linux 344201326395265SE +/- 0.17, N = 3SE +/- 0.10, N = 358.0553.551. (CXX) g++ options: -O2 -lOpenCL

NAMD

NAMD is a parallel molecular dynamics code designed for high-performance simulation of large biomolecular systems. NAMD was developed by the Theoretical and Computational Biophysics Group in the Beckman Institute for Advanced Science and Technology at the University of Illinois at Urbana-Champaign. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgdays/ns, Fewer Is BetterNAMD 2.14ATPase Simulation - 327,506 AtomsUbuntu 21.04 DevClear Linux 344200.09020.18040.27060.36080.451SE +/- 0.00344, N = 15SE +/- 0.00145, N = 30.400850.37170

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.3Model: inception-v3Ubuntu 21.04 DevClear Linux 34420816243240SE +/- 0.13, N = 15SE +/- 0.11, N = 1534.8832.48MIN: 33.07 / MAX: 87.63-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 29.61 / MAX: 94.151. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPUUbuntu 21.04 DevClear Linux 344201.05092.10183.15274.20365.2545SE +/- 0.01675, N = 3SE +/- 0.00032, N = 34.670694.42830MIN: 4.37-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 4.391. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.3Model: SqueezeNetV1.0Ubuntu 21.04 DevClear Linux 34420246810SE +/- 0.073, N = 15SE +/- 0.099, N = 158.5998.157MIN: 7.15 / MAX: 17.81-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 5.82 / MAX: 12.061. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 1 - Input: Bosphorus 1080pUbuntu 21.04 DevClear Linux 34420714212835SE +/- 0.04, N = 3SE +/- 0.10, N = 328.5530.09-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

Chaos Group V-RAY

This is a test of Chaos Group's V-RAY benchmark. V-RAY is a commercial renderer that can integrate with various creator software products like SketchUp and 3ds Max. The V-RAY benchmark is standalone and supports CPU and NVIDIA CUDA/RTX based rendering. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgvsamples, More Is BetterChaos Group V-RAY 5Mode: CPUUbuntu 21.04 DevClear Linux 3442011K22K33K44K55KSE +/- 231.66, N = 3SE +/- 217.08, N = 34657349082

Intel Open Image Denoise

Open Image Denoise is a denoising library for ray-tracing and part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 1.2.0Scene: MemorialUbuntu 21.04 DevClear Linux 34420918273645SE +/- 0.37, N = 3SE +/- 0.28, N = 335.7537.44

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 9 - Compression SpeedUbuntu 21.04 DevClear Linux 344201122334455SE +/- 0.01, N = 3SE +/- 0.16, N = 344.8346.811. (CC) gcc options: -O3

Rodinia

Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes select OpenCL, NVIDIA CUDA and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP HotSpot3DUbuntu 21.04 DevClear Linux 34420306090120150SE +/- 1.17, N = 6SE +/- 0.17, N = 3116.01111.151. (CXX) g++ options: -O2 -lOpenCL

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUClear Linux 34420Ubuntu 21.04 Dev0.38670.77341.16011.54681.9335SE +/- 0.01330, N = 3SE +/- 0.01259, N = 151.718881.65179-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 1.52MIN: 1.181. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

TNN

TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.2.3Target: CPU - Model: SqueezeNet v1.1Ubuntu 21.04 DevClear Linux 3442070140210280350SE +/- 0.22, N = 3SE +/- 0.12, N = 3343.35330.16MIN: 341.6 / MAX: 345.44-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 329.51 / MAX: 332.561. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -O3 -rdynamic -ldl

simdjson

This is a benchmark of SIMDJSON, a high performance JSON parser. SIMDJSON aims to be the fastest JSON parser and is used by projects like Microsoft FishStore, Yandex ClickHouse, Shopify, and others. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 0.8.2Throughput Test: KostyaUbuntu 21.04 DevClear Linux 344200.5671.1341.7012.2682.835SE +/- 0.00, N = 3SE +/- 0.01, N = 32.432.52-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CXX) g++ options: -O3 -pthread

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.9.0Binary: Pathtracer - Model: Asian DragonUbuntu 21.04 DevClear Linux 344201428425670SE +/- 0.80, N = 3SE +/- 0.42, N = 359.1961.21MIN: 56.39 / MAX: 62.22MIN: 59 / MAX: 63.14

IndigoBench

This is a test of Indigo Renderer's IndigoBench benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgM samples/s, More Is BetterIndigoBench 4.4Acceleration: CPU - Scene: BedroomUbuntu 21.04 DevClear Linux 34420246810SE +/- 0.011, N = 3SE +/- 0.013, N = 38.2978.576

Zstd Compression

This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.4.9Compression Level: 19, Long Mode - Compression SpeedClear Linux 34420Ubuntu 21.04 Dev918273645SE +/- 0.43, N = 4SE +/- 0.47, N = 1538.940.2-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake-llzma1. (CC) gcc options: -O3 -pthread -lz

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 3 - Decompression SpeedClear Linux 34420Ubuntu 21.04 Dev13002600390052006500SE +/- 177.02, N = 3SE +/- 121.63, N = 36096.96297.01. (CC) gcc options: -O3

IndigoBench

This is a test of Indigo Renderer's IndigoBench benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgM samples/s, More Is BetterIndigoBench 4.4Acceleration: CPU - Scene: SupercarUbuntu 21.04 DevClear Linux 34420510152025SE +/- 0.04, N = 3SE +/- 0.11, N = 319.5820.20

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 9 - Decompression SpeedClear Linux 34420Ubuntu 21.04 Dev14002800420056007000SE +/- 83.53, N = 3SE +/- 86.84, N = 36135.46308.01. (CC) gcc options: -O3

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.1.3Model: MobileNetV2_224Ubuntu 21.04 DevClear Linux 344201.0982.1963.2944.3925.49SE +/- 0.057, N = 15SE +/- 0.059, N = 154.8804.750MIN: 3.83 / MAX: 16.18-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 3.76 / MAX: 9.41. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.9.0Binary: Pathtracer - Model: CrownUbuntu 21.04 DevClear Linux 344201224364860SE +/- 0.38, N = 3SE +/- 0.28, N = 350.1051.39MIN: 45.6 / MAX: 53.58MIN: 48.91 / MAX: 53.74

Stockfish

This is a test of Stockfish, an advanced open-source C++11 chess benchmark that can scale up to 512 CPU threads. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNodes Per Second, More Is BetterStockfish 13Total TimeUbuntu 21.04 DevClear Linux 3442030M60M90M120M150MSE +/- 1671020.36, N = 15SE +/- 1037933.70, N = 15132800942136190317-pipe -fexceptions -fstack-protector -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CXX) g++ options: -lgcov -m64 -lpthread -fno-exceptions -std=c++17 -fprofile-use -fno-peel-loops -fno-tracer -pedantic -O3 -msse -msse3 -mpopcnt -mavx2 -mavx512f -mavx512bw -mavx512vnni -mavx512dq -mavx512vl -msse4.1 -mssse3 -msse2 -mbmi2 -flto -flto=jobserver

Xcompact3d Incompact3d

Xcompact3d Incompact3d is a Fortran-MPI based, finite difference high-performance code for solving the incompressible Navier-Stokes equation and as many as you need scalar transport equations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterXcompact3d Incompact3d 2021-03-11Input: X3D-benchmarking input.i3dUbuntu 21.04 DevClear Linux 34420100200300400500SE +/- 6.63, N = 3SE +/- 0.16, N = 3467.63458.17-lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz1. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: San Miguel - Renderer: Path TracerUbuntu 21.04 DevClear Linux 34420246810SE +/- 0.00, N = 3SE +/- 0.00, N = 36.496.62MIN: 4.69 / MAX: 6.54MIN: 5.08 / MAX: 6.67

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: NASA Streamlines - Renderer: Path TracerUbuntu 21.04 DevClear Linux 34420510152025SE +/- 0.00, N = 3SE +/- 0.00, N = 318.1818.52MIN: 12.35 / MAX: 18.52MIN: 14.29 / MAX: 19.23

Blender

Blender is an open-source 3D creation and modeling software project. This test is of Blender's Cycles benchmark with various sample files. GPU computing via OpenCL, NVIDIA OptiX, and NVIDIA CUDA is supported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.92Blend File: BMW27 - Compute: CPU-OnlyUbuntu 21.04 DevClear Linux 34420918273645SE +/- 0.14, N = 3SE +/- 0.10, N = 338.5237.84

Xcompact3d Incompact3d

Xcompact3d Incompact3d is a Fortran-MPI based, finite difference high-performance code for solving the incompressible Navier-Stokes equation and as many as you need scalar transport equations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterXcompact3d Incompact3d 2021-03-11Input: input.i3d 193 Cells Per DirectionClear Linux 34420Ubuntu 21.04 Dev510152025SE +/- 0.02, N = 3SE +/- 0.08, N = 319.3319.01-lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz1. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUUbuntu 21.04 DevClear Linux 344200.23750.4750.71250.951.1875SE +/- 0.004748, N = 3SE +/- 0.013211, N = 151.0554601.038714MIN: 0.74-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 0.691. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.9.0Binary: Pathtracer ISPC - Model: CrownUbuntu 21.04 DevClear Linux 344201224364860SE +/- 0.55, N = 3SE +/- 0.20, N = 354.1354.71MIN: 49.99 / MAX: 57.62MIN: 52.61 / MAX: 57.82

GROMACS

The GROMACS (GROningen MAchine for Chemical Simulations) molecular dynamics package testing on the CPU with the water_GMX50 data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNs Per Day, More Is BetterGROMACS 2021Input: water_GMX50_bareUbuntu 21.04 DevClear Linux 344201.29582.59163.88745.18326.479SE +/- 0.018, N = 3SE +/- 0.020, N = 35.7105.759-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake1. (CXX) g++ options: -O3 -pthread

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: XFrog Forest - Renderer: Path TracerUbuntu 21.04 DevClear Linux 34420246810SE +/- 0.00, N = 3SE +/- 0.00, N = 36.626.67MIN: 5.92 / MAX: 6.67MIN: 5.38 / MAX: 6.71

Xcompact3d Incompact3d

Xcompact3d Incompact3d is a Fortran-MPI based, finite difference high-performance code for solving the incompressible Navier-Stokes equation and as many as you need scalar transport equations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterXcompact3d Incompact3d 2021-03-11Input: input.i3d 129 Cells Per DirectionClear Linux 34420Ubuntu 21.04 Dev1.01012.02023.03034.04045.0505SE +/- 0.00386211, N = 3SE +/- 0.04397043, N = 34.489246374.46978633-lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz1. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 3 - Compression SpeedClear Linux 34420Ubuntu 21.04 Dev1122334455SE +/- 0.03, N = 3SE +/- 0.01, N = 348.0048.191. (CC) gcc options: -O3

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.9.0Binary: Pathtracer ISPC - Model: Asian DragonClear Linux 34420Ubuntu 21.04 Dev1632486480SE +/- 0.56, N = 3SE +/- 0.04, N = 370.4870.72MIN: 67.48 / MAX: 74.56MIN: 68 / MAX: 73.5

Zstd Compression

This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.4.9Compression Level: 8, Long Mode - Decompression SpeedUbuntu 21.04 Dev6001200180024003000SE +/- 4.70, N = 152814.21. (CC) gcc options: -O3 -pthread -lz -llzma

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: Magnetic Reconnection - Renderer: Path TracerClear Linux 34420Ubuntu 21.04 Dev110220330440550500500MIN: 250MIN: 111.11

Blender

Blender is an open-source 3D creation and modeling software project. This test is of Blender's Cycles benchmark with various sample files. GPU computing via OpenCL, NVIDIA OptiX, and NVIDIA CUDA is supported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.92Blend File: Fishy Cat - Compute: CPU-OnlyUbuntu 21.04 DevClear Linux 3442020406080100SE +/- 36.38, N = 12SE +/- 0.05, N = 396.0355.71

TensorFlow Lite

This is a benchmark of the TensorFlow Lite implementation. The current Linux support is limited to running on CPUs. This test profile is measuring the average inference time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet QuantUbuntu 21.04 DevClear Linux 3442020K40K60K80K100KSE +/- 1321.28, N = 15SE +/- 584.51, N = 383725.449894.1

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet FloatUbuntu 21.04 DevClear Linux 3442020K40K60K80K100KSE +/- 1637.30, N = 12SE +/- 591.27, N = 383813.351318.3

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: NASNet MobileUbuntu 21.04 DevClear Linux 3442060K120K180K240K300KSE +/- 3037.32, N = 15SE +/- 1598.38, N = 15282417.097445.7

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.1.2Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPUUbuntu 21.04 DevClear Linux 344200.67681.35362.03042.70723.384SE +/- 0.12420, N = 15SE +/- 0.00683, N = 33.008052.11061MIN: 2.04-pipe -fexceptions -fstack-protector -m64 -ffat-lto-objects -fno-trapping-math -mtune=skylake - MIN: 2.021. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Rodinia

Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes select OpenCL, NVIDIA CUDA and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP StreamclusterUbuntu 21.04 DevClear Linux 3442048121620SE +/- 0.29, N = 12SE +/- 0.13, N = 513.7312.901. (CXX) g++ options: -O2 -lOpenCL

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP CFD SolverUbuntu 21.04 DevClear Linux 34420246810SE +/- 0.087, N = 15SE +/- 0.127, N = 157.7346.5411. (CXX) g++ options: -O2 -lOpenCL

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP LeukocyteUbuntu 21.04 DevClear Linux 344201428425670SE +/- 0.49, N = 3SE +/- 1.63, N = 1563.7252.681. (CXX) g++ options: -O2 -lOpenCL

95 Results Shown

Zstd Compression
dav1d:
  Chimera 1080p
  Chimera 1080p 10-bit
  Summer Nature 1080p
TensorFlow Lite
dav1d
libavif avifenc
SVT-HEVC
TensorFlow Lite:
  Inception V4
  Inception ResNet V2
SVT-VP9
SVT-AV1
libavif avifenc
SVT-VP9
TNN
SVT-VP9
oneDNN:
  Recurrent Neural Network Inference - bf16bf16bf16 - CPU
  Recurrent Neural Network Inference - u8s8f32 - CPU
libavif avifenc
OpenVKL
oneDNN:
  Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU
  Deconvolution Batch shapes_1d - bf16bf16bf16 - CPU
  Deconvolution Batch shapes_3d - u8s8f32 - CPU
Zstd Compression
SVT-HEVC
oneDNN:
  Recurrent Neural Network Training - bf16bf16bf16 - CPU
  Recurrent Neural Network Training - u8s8f32 - CPU
Mobile Neural Network
Liquid-DSP
libavif avifenc
Liquid-DSP:
  1 - 256 - 57
  16 - 256 - 57
SVT-AV1
oneDNN
simdjson
SVT-AV1
libavif avifenc
Liquid-DSP
simdjson
Mobile Neural Network
Zstd Compression
oneDNN
Google SynthMark
oneDNN:
  IP Shapes 1D - bf16bf16bf16 - CPU
  Convolution Batch Shapes Auto - bf16bf16bf16 - CPU
Zstd Compression
oneDNN
NAS Parallel Benchmarks:
  EP.D
  LU.C
simdjson
Liquid-DSP
Rodinia
NAMD
Mobile Neural Network
oneDNN
Mobile Neural Network
SVT-HEVC
Chaos Group V-RAY
Intel Open Image Denoise
LZ4 Compression
Rodinia
oneDNN
TNN
simdjson
Embree
IndigoBench
Zstd Compression
LZ4 Compression
IndigoBench
LZ4 Compression
Mobile Neural Network
Embree
Stockfish
Xcompact3d Incompact3d
OSPray:
  San Miguel - Path Tracer
  NASA Streamlines - Path Tracer
Blender
Xcompact3d Incompact3d
oneDNN
Embree
GROMACS
OSPray
Xcompact3d Incompact3d
LZ4 Compression
Embree
Zstd Compression
OSPray
Blender
TensorFlow Lite:
  Mobilenet Quant
  Mobilenet Float
  NASNet Mobile
oneDNN
Rodinia:
  OpenMP Streamcluster
  OpenMP CFD Solver
  OpenMP Leukocyte