AMD EPYC 7601 Ubuntu 19.04

AMD EPYC 7601 32-Core testing with a TYAN B8026T70AE24HR (V1.02.B10 BIOS) and ASPEED on Ubuntu 19.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2009077-FI-AMDEPYC7666
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
Ubuntu 19.04
September 07 2020
  7 Hours, 55 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


AMD EPYC 7601 Ubuntu 19.04OpenBenchmarking.orgPhoronix Test SuiteAMD EPYC 7601 32-Core @ 2.20GHz (32 Cores / 64 Threads)TYAN B8026T70AE24HR (V1.02.B10 BIOS)AMD 17h126GB280GB INTEL SSDPE21D280GAASPEEDVE2282 x Broadcom NetXtreme BCM5720 PCIeUbuntu 19.045.5.0-rc7-phx-k10temp6 (x86_64) 20200123GNOME Shell 3.32.2X Server 1.20.4modesetting 1.20.4GCC 8.3.0ext41920x1080ProcessorMotherboardChipsetMemoryDiskGraphicsMonitorNetworkOSKernelDesktopDisplay ServerDisplay DriverCompilerFile-SystemScreen ResolutionAMD EPYC 7601 Ubuntu 19.04 BenchmarksSystem Logs- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v - Scaling Governor: acpi-cpufreq ondemand - CPU Microcode: 0x8001227- Python 2.7.16 + Python 3.7.3- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional STIBP: disabled RSB filling + tsx_async_abort: Not affected

AMD EPYC 7601 Ubuntu 19.04plaidml: No - Inference - VGG16 - CPUplaidml: No - Inference - VGG19 - CPUplaidml: No - Inference - IMDB LSTM - CPUplaidml: No - Inference - Mobilenet - CPUplaidml: No - Inference - ResNet 50 - CPUplaidml: No - Inference - DenseNet 201 - CPUplaidml: No - Inference - Inception V3 - CPUplaidml: No - Inference - NASNer Large - CPUsvt-av1: Enc Mode 0 - 1080psvt-av1: Enc Mode 4 - 1080psvt-av1: Enc Mode 8 - 1080pperf-bench: Memcpy 1MBperf-bench: Memset 1MBcompress-zstd: 3compress-zstd: 19gromacs: Water Benchmarkperf-bench: Epoll Waitperf-bench: Futex Hashperf-bench: Sched Pipeperf-bench: Futex Lock-Piperf-bench: Syscall Basicnamd: ATPase Simulation - 327,506 Atomstensorflow-lite: SqueezeNettensorflow-lite: Inception V4tensorflow-lite: NASNet Mobiletensorflow-lite: Mobilenet Floattensorflow-lite: Mobilenet Quanttensorflow-lite: Inception ResNet V2onednn: IP Batch 1D - f32 - CPUonednn: IP Batch All - f32 - CPUonednn: IP Batch 1D - u8s8f32 - CPUonednn: IP Batch All - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Deconvolution Batch deconv_1d - f32 - CPUonednn: Deconvolution Batch deconv_3d - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Deconvolution Batch deconv_1d - u8s8f32 - CPUonednn: Deconvolution Batch deconv_3d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUavifenc: 0avifenc: 2avifenc: 8avifenc: 10build-linux-kernel: Time To Compileastcenc: Fastastcenc: Mediumastcenc: Thoroughastcenc: Exhaustivegmic: 2D Function Plotting, 1000 Timesgmic: Plotting Isosurface Of A 3D Volume, 1000 Timesgmic: 3D Elevated Function In Rand Colors, 100 Timeshugin: Panorama Photo Assistant + Stitching Timeocrmypdf: Processing 60 Page PDF Documentrawtherapee: Total Benchmark Timersvg: SVG Files To PNGblender: BMW27 - CPU-Onlyblender: Classroom - CPU-Onlyblender: Fishy Cat - CPU-Onlyblender: Barbershop - CPU-Onlyblender: Pabellon Barcelona - CPU-Onlytesseract-ocr: Time To OCR 7 ImagesUbuntu 19.0415.7813.09716.0411.745.382.357.130.740.0974.23736.91614.99607442.0026424652.961.21.94265112468188263943124149777700.9690099445.7130303712962859349.264549.911513804.2012870.33442.6708236.782918.93153.757918.9365720.55734.402194.29952442.987111.6131.837801.7381491.15355.6147.0847.02143.0896.658.0211.9391.62335.08926.994115.67273.92931.97664.84237.62476.25212.84111.65337.51252.9242.516OpenBenchmarking.org

PlaidML

This test profile uses PlaidML deep learning framework developed by Intel for offering up various benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: VGG16 - Device: CPUUbuntu 19.0448121620SE +/- 0.21, N = 315.78

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: VGG19 - Device: CPUUbuntu 19.043691215SE +/- 0.07, N = 313.09

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: IMDB LSTM - Device: CPUUbuntu 19.04150300450600750SE +/- 1.69, N = 3716.04

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: Mobilenet - Device: CPUUbuntu 19.043691215SE +/- 0.07, N = 311.74

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: ResNet 50 - Device: CPUUbuntu 19.041.21052.4213.63154.8426.0525SE +/- 0.03, N = 35.38

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: DenseNet 201 - Device: CPUUbuntu 19.040.52881.05761.58642.11522.644SE +/- 0.00, N = 32.35

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: Inception V3 - Device: CPUUbuntu 19.04246810SE +/- 0.05, N = 37.13

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: NASNer Large - Device: CPUUbuntu 19.040.16650.3330.49950.6660.8325SE +/- 0.00, N = 30.74

SVT-AV1

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-AV1 CPU-based multi-threaded video encoder for the AV1 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8Encoder Mode: Enc Mode 0 - Input: 1080pUbuntu 19.040.02180.04360.06540.08720.109SE +/- 0.000, N = 30.0971. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8Encoder Mode: Enc Mode 4 - Input: 1080pUbuntu 19.040.95331.90662.85993.81324.7665SE +/- 0.057, N = 34.2371. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8Encoder Mode: Enc Mode 8 - Input: 1080pUbuntu 19.04816243240SE +/- 0.37, N = 336.921. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie

perf-bench

OpenBenchmarking.orgGB/sec, More Is Betterperf-benchBenchmark: Memcpy 1MBUbuntu 19.0448121620SE +/- 0.03, N = 315.001. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

OpenBenchmarking.orgGB/sec, More Is Betterperf-benchBenchmark: Memset 1MBUbuntu 19.041020304050SE +/- 0.22, N = 342.001. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

Zstd Compression

This test measures the time needed to compress a sample file (an Ubuntu ISO) using Zstd compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.4.5Compression Level: 3Ubuntu 19.0410002000300040005000SE +/- 79.97, N = 34652.91. (CC) gcc options: -O3 -pthread -lz -llzma

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.4.5Compression Level: 19Ubuntu 19.041428425670SE +/- 0.58, N = 361.21. (CC) gcc options: -O3 -pthread -lz -llzma

GROMACS

The GROMACS (GROningen MAchine for Chemical Simulations) molecular dynamics package testing on the CPU with the water_GMX50 data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNs Per Day, More Is BetterGROMACS 2020.1Water BenchmarkUbuntu 19.040.4370.8741.3111.7482.185SE +/- 0.017, N = 31.9421. (CXX) g++ options: -O3 -pthread -lrt -lpthread -lm

perf-bench

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Epoll WaitUbuntu 19.0414002800420056007000SE +/- 106.51, N = 365111. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Futex HashUbuntu 19.04500K1000K1500K2000K2500KSE +/- 548.36, N = 324681881. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Sched PipeUbuntu 19.0460K120K180K240K300KSE +/- 2700.16, N = 32639431. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Futex Lock-PiUbuntu 19.04306090120150SE +/- 1.13, N = 151241. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Syscall BasicUbuntu 19.043M6M9M12M15MSE +/- 2524.94, N = 3149777701. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

NAMD

NAMD is a parallel molecular dynamics code designed for high-performance simulation of large biomolecular systems. NAMD was developed by the Theoretical and Computational Biophysics Group in the Beckman Institute for Advanced Science and Technology at the University of Illinois at Urbana-Champaign. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgdays/ns, Fewer Is BetterNAMD 2.14ATPase Simulation - 327,506 AtomsUbuntu 19.040.2180.4360.6540.8721.09SE +/- 0.00501, N = 30.96900

TensorFlow Lite

This is a benchmark of the TensorFlow Lite implementation. The current Linux support is limited to running on CPUs. This test profile is measuring the average inference time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: SqueezeNetUbuntu 19.0420K40K60K80K100KSE +/- 242.10, N = 399445.7

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception V4Ubuntu 19.04300K600K900K1200K1500KSE +/- 7397.48, N = 31303037

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: NASNet MobileUbuntu 19.0430K60K90K120K150KSE +/- 1752.56, N = 3129628

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet FloatUbuntu 19.0413K26K39K52K65KSE +/- 235.78, N = 359349.2

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet QuantUbuntu 19.0414K28K42K56K70KSE +/- 318.17, N = 364549.9

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception ResNet V2Ubuntu 19.04200K400K600K800K1000KSE +/- 5528.66, N = 31151380

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the oneAPI initiative. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch 1D - Data Type: f32 - Engine: CPUUbuntu 19.040.94531.89062.83593.78124.7265SE +/- 0.09463, N = 144.20128MIN: 2.851. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch All - Data Type: f32 - Engine: CPUUbuntu 19.041632486480SE +/- 0.57, N = 370.33MIN: 63.691. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch 1D - Data Type: u8s8f32 - Engine: CPUUbuntu 19.040.60091.20181.80272.40363.0045SE +/- 0.00259, N = 32.67082MIN: 2.591. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch All - Data Type: u8s8f32 - Engine: CPUUbuntu 19.04816243240SE +/- 0.54, N = 1536.78MIN: 32.821. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUUbuntu 19.04510152025SE +/- 0.02, N = 318.93MIN: 17.831. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_1d - Data Type: f32 - Engine: CPUUbuntu 19.040.84551.6912.53653.3824.2275SE +/- 0.02933, N = 33.75791MIN: 3.441. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_3d - Data Type: f32 - Engine: CPUUbuntu 19.04246810SE +/- 0.14257, N = 38.93657MIN: 6.951. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUUbuntu 19.04510152025SE +/- 0.02, N = 320.56MIN: 19.441. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32 - Engine: CPUUbuntu 19.040.99051.9812.97153.9624.9525SE +/- 0.08113, N = 154.40219MIN: 3.981. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32 - Engine: CPUUbuntu 19.040.96741.93482.90223.86964.837SE +/- 0.05187, N = 34.29952MIN: 4.061. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUUbuntu 19.04100200300400500SE +/- 7.32, N = 3442.99MIN: 421.251. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUUbuntu 19.0420406080100SE +/- 0.40, N = 3111.61MIN: 109.391. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUUbuntu 19.040.41350.8271.24051.6542.0675SE +/- 0.06760, N = 151.83780MIN: 11. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUUbuntu 19.040.39110.78221.17331.56441.9555SE +/- 0.00231, N = 31.73814MIN: 1.621. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 0Ubuntu 19.0420406080100SE +/- 0.37, N = 391.151. (CXX) g++ options: -O3 -fPIC

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 2Ubuntu 19.041224364860SE +/- 0.17, N = 355.611. (CXX) g++ options: -O3 -fPIC

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 8Ubuntu 19.04246810SE +/- 0.063, N = 37.0841. (CXX) g++ options: -O3 -fPIC

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 10Ubuntu 19.04246810SE +/- 0.037, N = 37.0211. (CXX) g++ options: -O3 -fPIC

Timed Linux Kernel Compilation

This test times how long it takes to build the Linux kernel in a default configuration. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 5.4Time To CompileUbuntu 19.041020304050SE +/- 0.58, N = 543.09

ASTC Encoder

ASTC Encoder (astcenc) is for the Adaptive Scalable Texture Compression (ASTC) format commonly used with OpenGL, OpenGL ES, and Vulkan graphics APIs. This test profile does a coding test of both compression/decompression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: FastUbuntu 19.04246810SE +/- 0.05, N = 36.651. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: MediumUbuntu 19.04246810SE +/- 0.04, N = 38.021. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: ThoroughUbuntu 19.043691215SE +/- 0.03, N = 311.931. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: ExhaustiveUbuntu 19.0420406080100SE +/- 0.22, N = 391.621. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

G'MIC

G'MIC is an open-source framework for image processing. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterG'MICTest: 2D Function Plotting, 1000 TimesUbuntu 19.0470140210280350SE +/- 0.37, N = 3335.091. Version 2.4.5, Copyright (c) 2008-2019, David Tschumperle.

OpenBenchmarking.orgSeconds, Fewer Is BetterG'MICTest: Plotting Isosurface Of A 3D Volume, 1000 TimesUbuntu 19.04612182430SE +/- 0.02, N = 326.991. Version 2.4.5, Copyright (c) 2008-2019, David Tschumperle.

OpenBenchmarking.orgSeconds, Fewer Is BetterG'MICTest: 3D Elevated Function In Random Colors, 100 TimesUbuntu 19.04306090120150SE +/- 0.08, N = 3115.671. Version 2.4.5, Copyright (c) 2008-2019, David Tschumperle.

Hugin

Hugin is an open-source, cross-platform panorama photo stitcher software package. This test profile times how long it takes to run the assistant and panorama photo stitching on a set of images. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterHuginPanorama Photo Assistant + Stitching TimeUbuntu 19.041632486480SE +/- 0.74, N = 1573.93

OCRMyPDF

OCRMyPDF is an optical character recognition (OCR) text layer to scanned PDF files, producing new PDFs with the text now selectable/searchable/copy-paste capable. OCRMyPDF leverages the Tesseract OCR engine and is written in Python. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOCRMyPDF 8.0.1+dfsgProcessing 60 Page PDF DocumentUbuntu 19.04714212835SE +/- 0.12, N = 331.98

RawTherapee

RawTherapee is a cross-platform, open-source multi-threaded RAW image processing program. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterRawTherapeeTotal Benchmark TimeUbuntu 19.041428425670SE +/- 0.15, N = 364.841. RawTherapee, version 5.5, command line. An advanced, cross-platform program for developing raw photos. Website: http://www.rawtherapee.com/ Documentation: http://rawpedia.rawtherapee.com/ Forum: https://discuss.pixls.us/c/software/rawtherapee Code and bug reports: https://github.com/Beep6581/RawTherapee Symbols: <Chevrons> indicate parameters you can change. [Square brackets] mean the parameter is optional. The pipe symbol | indicates a choice of one or the other. The dash symbol - denotes a range of possible values from one to the other. Usage: rawtherapee-cli -c <dir>|<files> Convert files in batch with default parameters. rawtherapee-cli <other options> -c <dir>|<files> Convert files in batch with your own settings. Options: rawtherapee-cli[-o <output>|-O <output>] [-q] [-a] [-s|-S] [-p <one.pp3> [-p <two.pp3> ...] ] [-d] [ -j[1-100] -js<1-3> | -t[z] -b<8|16|16f|32> | -n -b<8|16> ] [-Y] [-f] -c <input> -c <files> Specify one or more input files or folders. When specifying folders, Rawtherapee will look for image file types which comply with the selected extensions (see also '-a'). -c must be the last option. -o <file>|<dir> Set output file or folder. Saves output file alongside input file if -o is not specified. -O <file>|<dir> Set output file or folder and copy pp3 file into it. Saves output file alongside input file if -O is not specified. -q Quick-start mode. Does not load cached files to speedup start time. -a Process all supported image file types when specifying a folder, even those not currently selected in Preferences > File Browser > Parsed Extensions. -s Use the existing sidecar file to build the processing parameters, e.g. for photo.raw there should be a photo.raw.pp3 file in the same folder. If the sidecar file does not exist, neutral values will be used. -S Like -s but skip if the sidecar file does not exist. -p <file.pp3> Specify processing profile to be used for all conversions. You can specify as many sets of "-p <file.pp3>" options as you like, each will be built on top of the previous one, as explained below. -d Use the default raw or non-raw processing profile as set in Preferences > Image Processing > Default Processing Profile -j[1-100] Specify output to be JPEG (default, if -t and -n are not set). Optionally, specify compression 1-100 (default value: 92). -js<1-3> Specify the JPEG chroma subsampling parameter, where: 1 = Best compression: 2x2, 1x1, 1x1 (4:2:0) Chroma halved vertically and horizontally. 2 = Balanced (default): 2x1, 1x1, 1x1 (4:2:2) Chroma halved horizontally. 3 = Best quality: 1x1, 1x1, 1x1 (4:4:4) No chroma subsampling. -b<8|16|16f|32> Specify bit depth per channel. 8 = 8-bit integer. Applies to JPEG, PNG and TIFF. Default for JPEG and PNG. 16 = 16-bit integer. Applies to TIFF and PNG. Default for TIFF. 16f = 16-bit float. Applies to TIFF. 32 = 32-bit float. Applies to TIFF. -t[z] Specify output to be TIFF. Uncompressed by default, or deflate compression with 'z'. -n Specify output to be compressed PNG. Compression is hard-coded to PNG_FILTER_PAETH, Z_RLE. -Y Overwrite output if present. -f Use the custom fast-export processing pipeline. Your pp3 files can be incomplete, RawTherapee will build the final values as follows: 1- A new processing profile is created using neutral values, 2- If the "-d" option is set, the values are overridden by those found in the default raw or non-raw processing profile. 3- If one or more "-p" options are set, the values are overridden by those found in these processing profiles. 4- If the "-s" or "-S" options are set, the values are finally overridden by those found in the sidecar files. The processing profiles are processed in the order specified on the command line.

librsvg

RSVG/librsvg is an SVG vector graphics library. This test profile times how long it takes to complete various operations by rsvg-convert. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterlibrsvgOperation: SVG Files To PNGUbuntu 19.04918273645SE +/- 0.56, N = 337.621. rsvg-convert version 2.44.10

Blender

Blender is an open-source 3D creation software project. This test is of Blender's Cycles benchmark with various sample files. GPU computing via OpenCL or CUDA is supported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: BMW27 - Compute: CPU-OnlyUbuntu 19.0420406080100SE +/- 0.24, N = 376.25

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: Classroom - Compute: CPU-OnlyUbuntu 19.0450100150200250SE +/- 0.49, N = 3212.84

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: Fishy Cat - Compute: CPU-OnlyUbuntu 19.0420406080100SE +/- 0.25, N = 3111.65

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: Barbershop - Compute: CPU-OnlyUbuntu 19.0470140210280350SE +/- 0.52, N = 3337.51

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: Pabellon Barcelona - Compute: CPU-OnlyUbuntu 19.0460120180240300SE +/- 0.44, N = 3252.92

Tesseract OCR

Tesseract-OCR is the open-source optical character recognition (OCR) engine for the conversion of text within images to raw text output. This test profile relies upon a system-supplied Tesseract installation. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTesseract OCR 4.0.0Time To OCR 7 ImagesUbuntu 19.041020304050SE +/- 0.26, N = 342.52

64 Results Shown

PlaidML:
  No - Inference - VGG16 - CPU
  No - Inference - VGG19 - CPU
  No - Inference - IMDB LSTM - CPU
  No - Inference - Mobilenet - CPU
  No - Inference - ResNet 50 - CPU
  No - Inference - DenseNet 201 - CPU
  No - Inference - Inception V3 - CPU
  No - Inference - NASNer Large - CPU
SVT-AV1:
  Enc Mode 0 - 1080p
  Enc Mode 4 - 1080p
  Enc Mode 8 - 1080p
perf-bench:
  Memcpy 1MB
  Memset 1MB
Zstd Compression:
  3
  19
GROMACS
perf-bench:
  Epoll Wait
  Futex Hash
  Sched Pipe
  Futex Lock-Pi
  Syscall Basic
NAMD
TensorFlow Lite:
  SqueezeNet
  Inception V4
  NASNet Mobile
  Mobilenet Float
  Mobilenet Quant
  Inception ResNet V2
oneDNN:
  IP Batch 1D - f32 - CPU
  IP Batch All - f32 - CPU
  IP Batch 1D - u8s8f32 - CPU
  IP Batch All - u8s8f32 - CPU
  Convolution Batch Shapes Auto - f32 - CPU
  Deconvolution Batch deconv_1d - f32 - CPU
  Deconvolution Batch deconv_3d - f32 - CPU
  Convolution Batch Shapes Auto - u8s8f32 - CPU
  Deconvolution Batch deconv_1d - u8s8f32 - CPU
  Deconvolution Batch deconv_3d - u8s8f32 - CPU
  Recurrent Neural Network Training - f32 - CPU
  Recurrent Neural Network Inference - f32 - CPU
  Matrix Multiply Batch Shapes Transformer - f32 - CPU
  Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU
libavif avifenc:
  0
  2
  8
  10
Timed Linux Kernel Compilation
ASTC Encoder:
  Fast
  Medium
  Thorough
  Exhaustive
G'MIC:
  2D Function Plotting, 1000 Times
  Plotting Isosurface Of A 3D Volume, 1000 Times
  3D Elevated Function In Rand Colors, 100 Times
Hugin
OCRMyPDF
RawTherapee
librsvg
Blender:
  BMW27 - CPU-Only
  Classroom - CPU-Only
  Fishy Cat - CPU-Only
  Barbershop - CPU-Only
  Pabellon Barcelona - CPU-Only
Tesseract OCR