AMD EPYC 7601 Ubuntu 19.04

AMD EPYC 7601 32-Core testing with a TYAN B8026T70AE24HR (V1.02.B10 BIOS) and ASPEED on Ubuntu 19.04 via the Phoronix Test Suite.

HTML result view exported from: https://openbenchmarking.org/result/2009077-FI-AMDEPYC7666.

AMD EPYC 7601 Ubuntu 19.04ProcessorMotherboardChipsetMemoryDiskGraphicsMonitorNetworkOSKernelDesktopDisplay ServerDisplay DriverCompilerFile-SystemScreen ResolutionUbuntu 19.04AMD EPYC 7601 32-Core @ 2.20GHz (32 Cores / 64 Threads)TYAN B8026T70AE24HR (V1.02.B10 BIOS)AMD 17h126GB280GB INTEL SSDPE21D280GAASPEEDVE2282 x Broadcom NetXtreme BCM5720 PCIeUbuntu 19.045.5.0-rc7-phx-k10temp6 (x86_64) 20200123GNOME Shell 3.32.2X Server 1.20.4modesetting 1.20.4GCC 8.3.0ext41920x1080OpenBenchmarking.org- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v - Scaling Governor: acpi-cpufreq ondemand - CPU Microcode: 0x8001227- Python 2.7.16 + Python 3.7.3- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional STIBP: disabled RSB filling + tsx_async_abort: Not affected

AMD EPYC 7601 Ubuntu 19.04namd: ATPase Simulation - 327,506 Atomscompress-zstd: 3compress-zstd: 19onednn: IP Batch 1D - f32 - CPUonednn: IP Batch All - f32 - CPUonednn: IP Batch 1D - u8s8f32 - CPUonednn: IP Batch All - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Deconvolution Batch deconv_1d - f32 - CPUonednn: Deconvolution Batch deconv_3d - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Deconvolution Batch deconv_1d - u8s8f32 - CPUonednn: Deconvolution Batch deconv_3d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUsvt-av1: Enc Mode 0 - 1080psvt-av1: Enc Mode 4 - 1080psvt-av1: Enc Mode 8 - 1080pavifenc: 0avifenc: 2avifenc: 8avifenc: 10build-linux-kernel: Time To Compileperf-bench: Epoll Waitperf-bench: Futex Hashperf-bench: Memcpy 1MBperf-bench: Memset 1MBperf-bench: Sched Pipeperf-bench: Futex Lock-Piperf-bench: Syscall Basicgromacs: Water Benchmarktensorflow-lite: SqueezeNettensorflow-lite: Inception V4tensorflow-lite: NASNet Mobiletensorflow-lite: Mobilenet Floattensorflow-lite: Mobilenet Quanttensorflow-lite: Inception ResNet V2astcenc: Fastastcenc: Mediumastcenc: Thoroughastcenc: Exhaustivegmic: 2D Function Plotting, 1000 Timesgmic: Plotting Isosurface Of A 3D Volume, 1000 Timesgmic: 3D Elevated Function In Rand Colors, 100 Timeshugin: Panorama Photo Assistant + Stitching Timeocrmypdf: Processing 60 Page PDF Documentrawtherapee: Total Benchmark Timersvg: SVG Files To PNGplaidml: No - Inference - VGG16 - CPUplaidml: No - Inference - VGG19 - CPUplaidml: No - Inference - IMDB LSTM - CPUplaidml: No - Inference - Mobilenet - CPUplaidml: No - Inference - ResNet 50 - CPUplaidml: No - Inference - DenseNet 201 - CPUplaidml: No - Inference - Inception V3 - CPUplaidml: No - Inference - NASNer Large - CPUblender: BMW27 - CPU-Onlyblender: Classroom - CPU-Onlyblender: Fishy Cat - CPU-Onlyblender: Barbershop - CPU-Onlyblender: Pabellon Barcelona - CPU-Onlytesseract-ocr: Time To OCR 7 ImagesUbuntu 19.040.969004652.961.24.2012870.33442.6708236.782918.93153.757918.9365720.55734.402194.29952442.987111.6131.837801.738140.0974.23736.91691.15355.6147.0847.02143.0896511246818814.99607442.002642263943124149777701.94299445.7130303712962859349.264549.911513806.658.0211.9391.62335.08926.994115.67273.92931.97664.84237.62415.7813.09716.0411.745.382.357.130.7476.25212.84111.65337.51252.9242.516OpenBenchmarking.org

NAMD

ATPase Simulation - 327,506 Atoms

OpenBenchmarking.orgdays/ns, Fewer Is BetterNAMD 2.14ATPase Simulation - 327,506 AtomsUbuntu 19.040.2180.4360.6540.8721.09SE +/- 0.00501, N = 30.96900

Zstd Compression

Compression Level: 3

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.4.5Compression Level: 3Ubuntu 19.0410002000300040005000SE +/- 79.97, N = 34652.91. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 19

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.4.5Compression Level: 19Ubuntu 19.041428425670SE +/- 0.58, N = 361.21. (CC) gcc options: -O3 -pthread -lz -llzma

oneDNN

Harness: IP Batch 1D - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch 1D - Data Type: f32 - Engine: CPUUbuntu 19.040.94531.89062.83593.78124.7265SE +/- 0.09463, N = 144.20128MIN: 2.851. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: IP Batch All - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch All - Data Type: f32 - Engine: CPUUbuntu 19.041632486480SE +/- 0.57, N = 370.33MIN: 63.691. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: IP Batch 1D - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch 1D - Data Type: u8s8f32 - Engine: CPUUbuntu 19.040.60091.20181.80272.40363.0045SE +/- 0.00259, N = 32.67082MIN: 2.591. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: IP Batch All - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch All - Data Type: u8s8f32 - Engine: CPUUbuntu 19.04816243240SE +/- 0.54, N = 1536.78MIN: 32.821. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUUbuntu 19.04510152025SE +/- 0.02, N = 318.93MIN: 17.831. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch deconv_1d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_1d - Data Type: f32 - Engine: CPUUbuntu 19.040.84551.6912.53653.3824.2275SE +/- 0.02933, N = 33.75791MIN: 3.441. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch deconv_3d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_3d - Data Type: f32 - Engine: CPUUbuntu 19.04246810SE +/- 0.14257, N = 38.93657MIN: 6.951. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUUbuntu 19.04510152025SE +/- 0.02, N = 320.56MIN: 19.441. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32 - Engine: CPUUbuntu 19.040.99051.9812.97153.9624.9525SE +/- 0.08113, N = 154.40219MIN: 3.981. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32 - Engine: CPUUbuntu 19.040.96741.93482.90223.86964.837SE +/- 0.05187, N = 34.29952MIN: 4.061. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUUbuntu 19.04100200300400500SE +/- 7.32, N = 3442.99MIN: 421.251. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUUbuntu 19.0420406080100SE +/- 0.40, N = 3111.61MIN: 109.391. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUUbuntu 19.040.41350.8271.24051.6542.0675SE +/- 0.06760, N = 151.83780MIN: 11. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUUbuntu 19.040.39110.78221.17331.56441.9555SE +/- 0.00231, N = 31.73814MIN: 1.621. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

SVT-AV1

Encoder Mode: Enc Mode 0 - Input: 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8Encoder Mode: Enc Mode 0 - Input: 1080pUbuntu 19.040.02180.04360.06540.08720.109SE +/- 0.000, N = 30.0971. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie

SVT-AV1

Encoder Mode: Enc Mode 4 - Input: 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8Encoder Mode: Enc Mode 4 - Input: 1080pUbuntu 19.040.95331.90662.85993.81324.7665SE +/- 0.057, N = 34.2371. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie

SVT-AV1

Encoder Mode: Enc Mode 8 - Input: 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8Encoder Mode: Enc Mode 8 - Input: 1080pUbuntu 19.04816243240SE +/- 0.37, N = 336.921. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie

libavif avifenc

Encoder Speed: 0

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 0Ubuntu 19.0420406080100SE +/- 0.37, N = 391.151. (CXX) g++ options: -O3 -fPIC

libavif avifenc

Encoder Speed: 2

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 2Ubuntu 19.041224364860SE +/- 0.17, N = 355.611. (CXX) g++ options: -O3 -fPIC

libavif avifenc

Encoder Speed: 8

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 8Ubuntu 19.04246810SE +/- 0.063, N = 37.0841. (CXX) g++ options: -O3 -fPIC

libavif avifenc

Encoder Speed: 10

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 10Ubuntu 19.04246810SE +/- 0.037, N = 37.0211. (CXX) g++ options: -O3 -fPIC

Timed Linux Kernel Compilation

Time To Compile

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 5.4Time To CompileUbuntu 19.041020304050SE +/- 0.58, N = 543.09

perf-bench

Benchmark: Epoll Wait

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Epoll WaitUbuntu 19.0414002800420056007000SE +/- 106.51, N = 365111. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

perf-bench

Benchmark: Futex Hash

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Futex HashUbuntu 19.04500K1000K1500K2000K2500KSE +/- 548.36, N = 324681881. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

perf-bench

Benchmark: Memcpy 1MB

OpenBenchmarking.orgGB/sec, More Is Betterperf-benchBenchmark: Memcpy 1MBUbuntu 19.0448121620SE +/- 0.03, N = 315.001. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

perf-bench

Benchmark: Memset 1MB

OpenBenchmarking.orgGB/sec, More Is Betterperf-benchBenchmark: Memset 1MBUbuntu 19.041020304050SE +/- 0.22, N = 342.001. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

perf-bench

Benchmark: Sched Pipe

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Sched PipeUbuntu 19.0460K120K180K240K300KSE +/- 2700.16, N = 32639431. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

perf-bench

Benchmark: Futex Lock-Pi

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Futex Lock-PiUbuntu 19.04306090120150SE +/- 1.13, N = 151241. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

perf-bench

Benchmark: Syscall Basic

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Syscall BasicUbuntu 19.043M6M9M12M15MSE +/- 2524.94, N = 3149777701. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

GROMACS

Water Benchmark

OpenBenchmarking.orgNs Per Day, More Is BetterGROMACS 2020.1Water BenchmarkUbuntu 19.040.4370.8741.3111.7482.185SE +/- 0.017, N = 31.9421. (CXX) g++ options: -O3 -pthread -lrt -lpthread -lm

TensorFlow Lite

Model: SqueezeNet

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: SqueezeNetUbuntu 19.0420K40K60K80K100KSE +/- 242.10, N = 399445.7

TensorFlow Lite

Model: Inception V4

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception V4Ubuntu 19.04300K600K900K1200K1500KSE +/- 7397.48, N = 31303037

TensorFlow Lite

Model: NASNet Mobile

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: NASNet MobileUbuntu 19.0430K60K90K120K150KSE +/- 1752.56, N = 3129628

TensorFlow Lite

Model: Mobilenet Float

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet FloatUbuntu 19.0413K26K39K52K65KSE +/- 235.78, N = 359349.2

TensorFlow Lite

Model: Mobilenet Quant

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet QuantUbuntu 19.0414K28K42K56K70KSE +/- 318.17, N = 364549.9

TensorFlow Lite

Model: Inception ResNet V2

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception ResNet V2Ubuntu 19.04200K400K600K800K1000KSE +/- 5528.66, N = 31151380

ASTC Encoder

Preset: Fast

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: FastUbuntu 19.04246810SE +/- 0.05, N = 36.651. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

ASTC Encoder

Preset: Medium

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: MediumUbuntu 19.04246810SE +/- 0.04, N = 38.021. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

ASTC Encoder

Preset: Thorough

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: ThoroughUbuntu 19.043691215SE +/- 0.03, N = 311.931. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

ASTC Encoder

Preset: Exhaustive

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: ExhaustiveUbuntu 19.0420406080100SE +/- 0.22, N = 391.621. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

G'MIC

Test: 2D Function Plotting, 1000 Times

OpenBenchmarking.orgSeconds, Fewer Is BetterG'MICTest: 2D Function Plotting, 1000 TimesUbuntu 19.0470140210280350SE +/- 0.37, N = 3335.091. Version 2.4.5, Copyright (c) 2008-2019, David Tschumperle.

G'MIC

Test: Plotting Isosurface Of A 3D Volume, 1000 Times

OpenBenchmarking.orgSeconds, Fewer Is BetterG'MICTest: Plotting Isosurface Of A 3D Volume, 1000 TimesUbuntu 19.04612182430SE +/- 0.02, N = 326.991. Version 2.4.5, Copyright (c) 2008-2019, David Tschumperle.

G'MIC

Test: 3D Elevated Function In Random Colors, 100 Times

OpenBenchmarking.orgSeconds, Fewer Is BetterG'MICTest: 3D Elevated Function In Random Colors, 100 TimesUbuntu 19.04306090120150SE +/- 0.08, N = 3115.671. Version 2.4.5, Copyright (c) 2008-2019, David Tschumperle.

Hugin

Panorama Photo Assistant + Stitching Time

OpenBenchmarking.orgSeconds, Fewer Is BetterHuginPanorama Photo Assistant + Stitching TimeUbuntu 19.041632486480SE +/- 0.74, N = 1573.93

OCRMyPDF

Processing 60 Page PDF Document

OpenBenchmarking.orgSeconds, Fewer Is BetterOCRMyPDF 8.0.1+dfsgProcessing 60 Page PDF DocumentUbuntu 19.04714212835SE +/- 0.12, N = 331.98

RawTherapee

Total Benchmark Time

OpenBenchmarking.orgSeconds, Fewer Is BetterRawTherapeeTotal Benchmark TimeUbuntu 19.041428425670SE +/- 0.15, N = 364.841. RawTherapee, version 5.5, command line. An advanced, cross-platform program for developing raw photos. Website: http://www.rawtherapee.com/ Documentation: http://rawpedia.rawtherapee.com/ Forum: https://discuss.pixls.us/c/software/rawtherapee Code and bug reports: https://github.com/Beep6581/RawTherapee Symbols: <Chevrons> indicate parameters you can change. [Square brackets] mean the parameter is optional. The pipe symbol | indicates a choice of one or the other. The dash symbol - denotes a range of possible values from one to the other. Usage: rawtherapee-cli -c <dir>|<files> Convert files in batch with default parameters. rawtherapee-cli <other options> -c <dir>|<files> Convert files in batch with your own settings. Options: rawtherapee-cli[-o <output>|-O <output>] [-q] [-a] [-s|-S] [-p <one.pp3> [-p <two.pp3> ...] ] [-d] [ -j[1-100] -js<1-3> | -t[z] -b<8|16|16f|32> | -n -b<8|16> ] [-Y] [-f] -c <input> -c <files> Specify one or more input files or folders. When specifying folders, Rawtherapee will look for image file types which comply with the selected extensions (see also '-a'). -c must be the last option. -o <file>|<dir> Set output file or folder. Saves output file alongside input file if -o is not specified. -O <file>|<dir> Set output file or folder and copy pp3 file into it. Saves output file alongside input file if -O is not specified. -q Quick-start mode. Does not load cached files to speedup start time. -a Process all supported image file types when specifying a folder, even those not currently selected in Preferences > File Browser > Parsed Extensions. -s Use the existing sidecar file to build the processing parameters, e.g. for photo.raw there should be a photo.raw.pp3 file in the same folder. If the sidecar file does not exist, neutral values will be used. -S Like -s but skip if the sidecar file does not exist. -p <file.pp3> Specify processing profile to be used for all conversions. You can specify as many sets of "-p <file.pp3>" options as you like, each will be built on top of the previous one, as explained below. -d Use the default raw or non-raw processing profile as set in Preferences > Image Processing > Default Processing Profile -j[1-100] Specify output to be JPEG (default, if -t and -n are not set). Optionally, specify compression 1-100 (default value: 92). -js<1-3> Specify the JPEG chroma subsampling parameter, where: 1 = Best compression: 2x2, 1x1, 1x1 (4:2:0) Chroma halved vertically and horizontally. 2 = Balanced (default): 2x1, 1x1, 1x1 (4:2:2) Chroma halved horizontally. 3 = Best quality: 1x1, 1x1, 1x1 (4:4:4) No chroma subsampling. -b<8|16|16f|32> Specify bit depth per channel. 8 = 8-bit integer. Applies to JPEG, PNG and TIFF. Default for JPEG and PNG. 16 = 16-bit integer. Applies to TIFF and PNG. Default for TIFF. 16f = 16-bit float. Applies to TIFF. 32 = 32-bit float. Applies to TIFF. -t[z] Specify output to be TIFF. Uncompressed by default, or deflate compression with 'z'. -n Specify output to be compressed PNG. Compression is hard-coded to PNG_FILTER_PAETH, Z_RLE. -Y Overwrite output if present. -f Use the custom fast-export processing pipeline. Your pp3 files can be incomplete, RawTherapee will build the final values as follows: 1- A new processing profile is created using neutral values, 2- If the "-d" option is set, the values are overridden by those found in the default raw or non-raw processing profile. 3- If one or more "-p" options are set, the values are overridden by those found in these processing profiles. 4- If the "-s" or "-S" options are set, the values are finally overridden by those found in the sidecar files. The processing profiles are processed in the order specified on the command line.

librsvg

Operation: SVG Files To PNG

OpenBenchmarking.orgSeconds, Fewer Is BetterlibrsvgOperation: SVG Files To PNGUbuntu 19.04918273645SE +/- 0.56, N = 337.621. rsvg-convert version 2.44.10

PlaidML

FP16: No - Mode: Inference - Network: VGG16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: VGG16 - Device: CPUUbuntu 19.0448121620SE +/- 0.21, N = 315.78

PlaidML

FP16: No - Mode: Inference - Network: VGG19 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: VGG19 - Device: CPUUbuntu 19.043691215SE +/- 0.07, N = 313.09

PlaidML

FP16: No - Mode: Inference - Network: IMDB LSTM - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: IMDB LSTM - Device: CPUUbuntu 19.04150300450600750SE +/- 1.69, N = 3716.04

PlaidML

FP16: No - Mode: Inference - Network: Mobilenet - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: Mobilenet - Device: CPUUbuntu 19.043691215SE +/- 0.07, N = 311.74

PlaidML

FP16: No - Mode: Inference - Network: ResNet 50 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: ResNet 50 - Device: CPUUbuntu 19.041.21052.4213.63154.8426.0525SE +/- 0.03, N = 35.38

PlaidML

FP16: No - Mode: Inference - Network: DenseNet 201 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: DenseNet 201 - Device: CPUUbuntu 19.040.52881.05761.58642.11522.644SE +/- 0.00, N = 32.35

PlaidML

FP16: No - Mode: Inference - Network: Inception V3 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: Inception V3 - Device: CPUUbuntu 19.04246810SE +/- 0.05, N = 37.13

PlaidML

FP16: No - Mode: Inference - Network: NASNer Large - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: NASNer Large - Device: CPUUbuntu 19.040.16650.3330.49950.6660.8325SE +/- 0.00, N = 30.74

Blender

Blend File: BMW27 - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: BMW27 - Compute: CPU-OnlyUbuntu 19.0420406080100SE +/- 0.24, N = 376.25

Blender

Blend File: Classroom - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: Classroom - Compute: CPU-OnlyUbuntu 19.0450100150200250SE +/- 0.49, N = 3212.84

Blender

Blend File: Fishy Cat - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: Fishy Cat - Compute: CPU-OnlyUbuntu 19.0420406080100SE +/- 0.25, N = 3111.65

Blender

Blend File: Barbershop - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: Barbershop - Compute: CPU-OnlyUbuntu 19.0470140210280350SE +/- 0.52, N = 3337.51

Blender

Blend File: Pabellon Barcelona - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: Pabellon Barcelona - Compute: CPU-OnlyUbuntu 19.0460120180240300SE +/- 0.44, N = 3252.92

Tesseract OCR

Time To OCR 7 Images

OpenBenchmarking.orgSeconds, Fewer Is BetterTesseract OCR 4.0.0Time To OCR 7 ImagesUbuntu 19.041020304050SE +/- 0.26, N = 342.52


Phoronix Test Suite v10.8.4