AMD EPYC 7601 Ubuntu 19.04

AMD EPYC 7601 32-Core testing with a TYAN B8026T70AE24HR (V1.02.B10 BIOS) and ASPEED on Ubuntu 19.04 via the Phoronix Test Suite.

HTML result view exported from: https://openbenchmarking.org/result/2009077-FI-AMDEPYC7666&grs.

AMD EPYC 7601 Ubuntu 19.04ProcessorMotherboardChipsetMemoryDiskGraphicsMonitorNetworkOSKernelDesktopDisplay ServerDisplay DriverCompilerFile-SystemScreen ResolutionUbuntu 19.04AMD EPYC 7601 32-Core @ 2.20GHz (32 Cores / 64 Threads)TYAN B8026T70AE24HR (V1.02.B10 BIOS)AMD 17h126GB280GB INTEL SSDPE21D280GAASPEEDVE2282 x Broadcom NetXtreme BCM5720 PCIeUbuntu 19.045.5.0-rc7-phx-k10temp6 (x86_64) 20200123GNOME Shell 3.32.2X Server 1.20.4modesetting 1.20.4GCC 8.3.0ext41920x1080OpenBenchmarking.org- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v - Scaling Governor: acpi-cpufreq ondemand - CPU Microcode: 0x8001227- Python 2.7.16 + Python 3.7.3- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional STIBP: disabled RSB filling + tsx_async_abort: Not affected

AMD EPYC 7601 Ubuntu 19.04tesseract-ocr: Time To OCR 7 Imagesblender: Pabellon Barcelona - CPU-Onlyblender: Barbershop - CPU-Onlyblender: Fishy Cat - CPU-Onlyblender: Classroom - CPU-Onlyblender: BMW27 - CPU-Onlyplaidml: No - Inference - NASNer Large - CPUplaidml: No - Inference - Inception V3 - CPUplaidml: No - Inference - DenseNet 201 - CPUplaidml: No - Inference - ResNet 50 - CPUplaidml: No - Inference - Mobilenet - CPUplaidml: No - Inference - IMDB LSTM - CPUplaidml: No - Inference - VGG19 - CPUplaidml: No - Inference - VGG16 - CPUrsvg: SVG Files To PNGrawtherapee: Total Benchmark Timeocrmypdf: Processing 60 Page PDF Documenthugin: Panorama Photo Assistant + Stitching Timegmic: 3D Elevated Function In Rand Colors, 100 Timesgmic: Plotting Isosurface Of A 3D Volume, 1000 Timesgmic: 2D Function Plotting, 1000 Timesastcenc: Exhaustiveastcenc: Thoroughastcenc: Mediumastcenc: Fasttensorflow-lite: Inception ResNet V2tensorflow-lite: Mobilenet Quanttensorflow-lite: Mobilenet Floattensorflow-lite: NASNet Mobiletensorflow-lite: Inception V4tensorflow-lite: SqueezeNetgromacs: Water Benchmarkperf-bench: Syscall Basicperf-bench: Futex Lock-Piperf-bench: Sched Pipeperf-bench: Memset 1MBperf-bench: Memcpy 1MBperf-bench: Futex Hashperf-bench: Epoll Waitbuild-linux-kernel: Time To Compileavifenc: 10avifenc: 8avifenc: 2avifenc: 0svt-av1: Enc Mode 8 - 1080psvt-av1: Enc Mode 4 - 1080psvt-av1: Enc Mode 0 - 1080ponednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Deconvolution Batch deconv_3d - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Deconvolution Batch deconv_3d - f32 - CPUonednn: Deconvolution Batch deconv_1d - f32 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: IP Batch All - u8s8f32 - CPUonednn: IP Batch 1D - u8s8f32 - CPUonednn: IP Batch All - f32 - CPUcompress-zstd: 19compress-zstd: 3namd: ATPase Simulation - 327,506 Atomsonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Deconvolution Batch deconv_1d - u8s8f32 - CPUonednn: IP Batch 1D - f32 - CPUUbuntu 19.0442.516252.92337.51111.65212.8476.250.747.132.355.3811.74716.0413.0915.7837.62464.84231.97673.929115.67226.994335.08991.6211.938.026.65115138064549.959349.2129628130303799445.71.9421497777012426394342.00264214.9960742468188651143.0897.0217.08455.61491.15336.9164.2370.0971.73814111.613442.9874.2995220.55738.936573.7579118.931536.78292.6708270.334461.24652.90.969001.837804.402194.20128OpenBenchmarking.org

Tesseract OCR

Time To OCR 7 Images

OpenBenchmarking.orgSeconds, Fewer Is BetterTesseract OCR 4.0.0Time To OCR 7 ImagesUbuntu 19.041020304050SE +/- 0.26, N = 342.52

Blender

Blend File: Pabellon Barcelona - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: Pabellon Barcelona - Compute: CPU-OnlyUbuntu 19.0460120180240300SE +/- 0.44, N = 3252.92

Blender

Blend File: Barbershop - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: Barbershop - Compute: CPU-OnlyUbuntu 19.0470140210280350SE +/- 0.52, N = 3337.51

Blender

Blend File: Fishy Cat - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: Fishy Cat - Compute: CPU-OnlyUbuntu 19.0420406080100SE +/- 0.25, N = 3111.65

Blender

Blend File: Classroom - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: Classroom - Compute: CPU-OnlyUbuntu 19.0450100150200250SE +/- 0.49, N = 3212.84

Blender

Blend File: BMW27 - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 2.90Blend File: BMW27 - Compute: CPU-OnlyUbuntu 19.0420406080100SE +/- 0.24, N = 376.25

PlaidML

FP16: No - Mode: Inference - Network: NASNer Large - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: NASNer Large - Device: CPUUbuntu 19.040.16650.3330.49950.6660.8325SE +/- 0.00, N = 30.74

PlaidML

FP16: No - Mode: Inference - Network: Inception V3 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: Inception V3 - Device: CPUUbuntu 19.04246810SE +/- 0.05, N = 37.13

PlaidML

FP16: No - Mode: Inference - Network: DenseNet 201 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: DenseNet 201 - Device: CPUUbuntu 19.040.52881.05761.58642.11522.644SE +/- 0.00, N = 32.35

PlaidML

FP16: No - Mode: Inference - Network: ResNet 50 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: ResNet 50 - Device: CPUUbuntu 19.041.21052.4213.63154.8426.0525SE +/- 0.03, N = 35.38

PlaidML

FP16: No - Mode: Inference - Network: Mobilenet - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: Mobilenet - Device: CPUUbuntu 19.043691215SE +/- 0.07, N = 311.74

PlaidML

FP16: No - Mode: Inference - Network: IMDB LSTM - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: IMDB LSTM - Device: CPUUbuntu 19.04150300450600750SE +/- 1.69, N = 3716.04

PlaidML

FP16: No - Mode: Inference - Network: VGG19 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: VGG19 - Device: CPUUbuntu 19.043691215SE +/- 0.07, N = 313.09

PlaidML

FP16: No - Mode: Inference - Network: VGG16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: VGG16 - Device: CPUUbuntu 19.0448121620SE +/- 0.21, N = 315.78

librsvg

Operation: SVG Files To PNG

OpenBenchmarking.orgSeconds, Fewer Is BetterlibrsvgOperation: SVG Files To PNGUbuntu 19.04918273645SE +/- 0.56, N = 337.621. rsvg-convert version 2.44.10

RawTherapee

Total Benchmark Time

OpenBenchmarking.orgSeconds, Fewer Is BetterRawTherapeeTotal Benchmark TimeUbuntu 19.041428425670SE +/- 0.15, N = 364.841. RawTherapee, version 5.5, command line. An advanced, cross-platform program for developing raw photos. Website: http://www.rawtherapee.com/ Documentation: http://rawpedia.rawtherapee.com/ Forum: https://discuss.pixls.us/c/software/rawtherapee Code and bug reports: https://github.com/Beep6581/RawTherapee Symbols: <Chevrons> indicate parameters you can change. [Square brackets] mean the parameter is optional. The pipe symbol | indicates a choice of one or the other. The dash symbol - denotes a range of possible values from one to the other. Usage: rawtherapee-cli -c <dir>|<files> Convert files in batch with default parameters. rawtherapee-cli <other options> -c <dir>|<files> Convert files in batch with your own settings. Options: rawtherapee-cli[-o <output>|-O <output>] [-q] [-a] [-s|-S] [-p <one.pp3> [-p <two.pp3> ...] ] [-d] [ -j[1-100] -js<1-3> | -t[z] -b<8|16|16f|32> | -n -b<8|16> ] [-Y] [-f] -c <input> -c <files> Specify one or more input files or folders. When specifying folders, Rawtherapee will look for image file types which comply with the selected extensions (see also '-a'). -c must be the last option. -o <file>|<dir> Set output file or folder. Saves output file alongside input file if -o is not specified. -O <file>|<dir> Set output file or folder and copy pp3 file into it. Saves output file alongside input file if -O is not specified. -q Quick-start mode. Does not load cached files to speedup start time. -a Process all supported image file types when specifying a folder, even those not currently selected in Preferences > File Browser > Parsed Extensions. -s Use the existing sidecar file to build the processing parameters, e.g. for photo.raw there should be a photo.raw.pp3 file in the same folder. If the sidecar file does not exist, neutral values will be used. -S Like -s but skip if the sidecar file does not exist. -p <file.pp3> Specify processing profile to be used for all conversions. You can specify as many sets of "-p <file.pp3>" options as you like, each will be built on top of the previous one, as explained below. -d Use the default raw or non-raw processing profile as set in Preferences > Image Processing > Default Processing Profile -j[1-100] Specify output to be JPEG (default, if -t and -n are not set). Optionally, specify compression 1-100 (default value: 92). -js<1-3> Specify the JPEG chroma subsampling parameter, where: 1 = Best compression: 2x2, 1x1, 1x1 (4:2:0) Chroma halved vertically and horizontally. 2 = Balanced (default): 2x1, 1x1, 1x1 (4:2:2) Chroma halved horizontally. 3 = Best quality: 1x1, 1x1, 1x1 (4:4:4) No chroma subsampling. -b<8|16|16f|32> Specify bit depth per channel. 8 = 8-bit integer. Applies to JPEG, PNG and TIFF. Default for JPEG and PNG. 16 = 16-bit integer. Applies to TIFF and PNG. Default for TIFF. 16f = 16-bit float. Applies to TIFF. 32 = 32-bit float. Applies to TIFF. -t[z] Specify output to be TIFF. Uncompressed by default, or deflate compression with 'z'. -n Specify output to be compressed PNG. Compression is hard-coded to PNG_FILTER_PAETH, Z_RLE. -Y Overwrite output if present. -f Use the custom fast-export processing pipeline. Your pp3 files can be incomplete, RawTherapee will build the final values as follows: 1- A new processing profile is created using neutral values, 2- If the "-d" option is set, the values are overridden by those found in the default raw or non-raw processing profile. 3- If one or more "-p" options are set, the values are overridden by those found in these processing profiles. 4- If the "-s" or "-S" options are set, the values are finally overridden by those found in the sidecar files. The processing profiles are processed in the order specified on the command line.

OCRMyPDF

Processing 60 Page PDF Document

OpenBenchmarking.orgSeconds, Fewer Is BetterOCRMyPDF 8.0.1+dfsgProcessing 60 Page PDF DocumentUbuntu 19.04714212835SE +/- 0.12, N = 331.98

Hugin

Panorama Photo Assistant + Stitching Time

OpenBenchmarking.orgSeconds, Fewer Is BetterHuginPanorama Photo Assistant + Stitching TimeUbuntu 19.041632486480SE +/- 0.74, N = 1573.93

G'MIC

Test: 3D Elevated Function In Random Colors, 100 Times

OpenBenchmarking.orgSeconds, Fewer Is BetterG'MICTest: 3D Elevated Function In Random Colors, 100 TimesUbuntu 19.04306090120150SE +/- 0.08, N = 3115.671. Version 2.4.5, Copyright (c) 2008-2019, David Tschumperle.

G'MIC

Test: Plotting Isosurface Of A 3D Volume, 1000 Times

OpenBenchmarking.orgSeconds, Fewer Is BetterG'MICTest: Plotting Isosurface Of A 3D Volume, 1000 TimesUbuntu 19.04612182430SE +/- 0.02, N = 326.991. Version 2.4.5, Copyright (c) 2008-2019, David Tschumperle.

G'MIC

Test: 2D Function Plotting, 1000 Times

OpenBenchmarking.orgSeconds, Fewer Is BetterG'MICTest: 2D Function Plotting, 1000 TimesUbuntu 19.0470140210280350SE +/- 0.37, N = 3335.091. Version 2.4.5, Copyright (c) 2008-2019, David Tschumperle.

ASTC Encoder

Preset: Exhaustive

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: ExhaustiveUbuntu 19.0420406080100SE +/- 0.22, N = 391.621. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

ASTC Encoder

Preset: Thorough

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: ThoroughUbuntu 19.043691215SE +/- 0.03, N = 311.931. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

ASTC Encoder

Preset: Medium

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: MediumUbuntu 19.04246810SE +/- 0.04, N = 38.021. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

ASTC Encoder

Preset: Fast

OpenBenchmarking.orgSeconds, Fewer Is BetterASTC Encoder 2.0Preset: FastUbuntu 19.04246810SE +/- 0.05, N = 36.651. (CXX) g++ options: -std=c++14 -fvisibility=hidden -O3 -flto -mfpmath=sse -mavx2 -mpopcnt -lpthread

TensorFlow Lite

Model: Inception ResNet V2

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception ResNet V2Ubuntu 19.04200K400K600K800K1000KSE +/- 5528.66, N = 31151380

TensorFlow Lite

Model: Mobilenet Quant

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet QuantUbuntu 19.0414K28K42K56K70KSE +/- 318.17, N = 364549.9

TensorFlow Lite

Model: Mobilenet Float

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet FloatUbuntu 19.0413K26K39K52K65KSE +/- 235.78, N = 359349.2

TensorFlow Lite

Model: NASNet Mobile

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: NASNet MobileUbuntu 19.0430K60K90K120K150KSE +/- 1752.56, N = 3129628

TensorFlow Lite

Model: Inception V4

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception V4Ubuntu 19.04300K600K900K1200K1500KSE +/- 7397.48, N = 31303037

TensorFlow Lite

Model: SqueezeNet

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: SqueezeNetUbuntu 19.0420K40K60K80K100KSE +/- 242.10, N = 399445.7

GROMACS

Water Benchmark

OpenBenchmarking.orgNs Per Day, More Is BetterGROMACS 2020.1Water BenchmarkUbuntu 19.040.4370.8741.3111.7482.185SE +/- 0.017, N = 31.9421. (CXX) g++ options: -O3 -pthread -lrt -lpthread -lm

perf-bench

Benchmark: Syscall Basic

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Syscall BasicUbuntu 19.043M6M9M12M15MSE +/- 2524.94, N = 3149777701. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

perf-bench

Benchmark: Futex Lock-Pi

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Futex Lock-PiUbuntu 19.04306090120150SE +/- 1.13, N = 151241. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

perf-bench

Benchmark: Sched Pipe

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Sched PipeUbuntu 19.0460K120K180K240K300KSE +/- 2700.16, N = 32639431. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

perf-bench

Benchmark: Memset 1MB

OpenBenchmarking.orgGB/sec, More Is Betterperf-benchBenchmark: Memset 1MBUbuntu 19.041020304050SE +/- 0.22, N = 342.001. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

perf-bench

Benchmark: Memcpy 1MB

OpenBenchmarking.orgGB/sec, More Is Betterperf-benchBenchmark: Memcpy 1MBUbuntu 19.0448121620SE +/- 0.03, N = 315.001. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

perf-bench

Benchmark: Futex Hash

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Futex HashUbuntu 19.04500K1000K1500K2000K2500KSE +/- 548.36, N = 324681881. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

perf-bench

Benchmark: Epoll Wait

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Epoll WaitUbuntu 19.0414002800420056007000SE +/- 106.51, N = 365111. (CC) gcc options: -pthread -shared -Xlinker -export-dynamic -O6 -ggdb3 -funwind-tables -std=gnu99 -lnuma

Timed Linux Kernel Compilation

Time To Compile

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 5.4Time To CompileUbuntu 19.041020304050SE +/- 0.58, N = 543.09

libavif avifenc

Encoder Speed: 10

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 10Ubuntu 19.04246810SE +/- 0.037, N = 37.0211. (CXX) g++ options: -O3 -fPIC

libavif avifenc

Encoder Speed: 8

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 8Ubuntu 19.04246810SE +/- 0.063, N = 37.0841. (CXX) g++ options: -O3 -fPIC

libavif avifenc

Encoder Speed: 2

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 2Ubuntu 19.041224364860SE +/- 0.17, N = 355.611. (CXX) g++ options: -O3 -fPIC

libavif avifenc

Encoder Speed: 0

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.7.3Encoder Speed: 0Ubuntu 19.0420406080100SE +/- 0.37, N = 391.151. (CXX) g++ options: -O3 -fPIC

SVT-AV1

Encoder Mode: Enc Mode 8 - Input: 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8Encoder Mode: Enc Mode 8 - Input: 1080pUbuntu 19.04816243240SE +/- 0.37, N = 336.921. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie

SVT-AV1

Encoder Mode: Enc Mode 4 - Input: 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8Encoder Mode: Enc Mode 4 - Input: 1080pUbuntu 19.040.95331.90662.85993.81324.7665SE +/- 0.057, N = 34.2371. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie

SVT-AV1

Encoder Mode: Enc Mode 0 - Input: 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8Encoder Mode: Enc Mode 0 - Input: 1080pUbuntu 19.040.02180.04360.06540.08720.109SE +/- 0.000, N = 30.0971. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie

oneDNN

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUUbuntu 19.040.39110.78221.17331.56441.9555SE +/- 0.00231, N = 31.73814MIN: 1.621. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUUbuntu 19.0420406080100SE +/- 0.40, N = 3111.61MIN: 109.391. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUUbuntu 19.04100200300400500SE +/- 7.32, N = 3442.99MIN: 421.251. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32 - Engine: CPUUbuntu 19.040.96741.93482.90223.86964.837SE +/- 0.05187, N = 34.29952MIN: 4.061. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUUbuntu 19.04510152025SE +/- 0.02, N = 320.56MIN: 19.441. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch deconv_3d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_3d - Data Type: f32 - Engine: CPUUbuntu 19.04246810SE +/- 0.14257, N = 38.93657MIN: 6.951. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch deconv_1d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_1d - Data Type: f32 - Engine: CPUUbuntu 19.040.84551.6912.53653.3824.2275SE +/- 0.02933, N = 33.75791MIN: 3.441. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUUbuntu 19.04510152025SE +/- 0.02, N = 318.93MIN: 17.831. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: IP Batch All - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch All - Data Type: u8s8f32 - Engine: CPUUbuntu 19.04816243240SE +/- 0.54, N = 1536.78MIN: 32.821. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: IP Batch 1D - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch 1D - Data Type: u8s8f32 - Engine: CPUUbuntu 19.040.60091.20181.80272.40363.0045SE +/- 0.00259, N = 32.67082MIN: 2.591. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: IP Batch All - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch All - Data Type: f32 - Engine: CPUUbuntu 19.041632486480SE +/- 0.57, N = 370.33MIN: 63.691. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

Zstd Compression

Compression Level: 19

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.4.5Compression Level: 19Ubuntu 19.041428425670SE +/- 0.58, N = 361.21. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 3

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.4.5Compression Level: 3Ubuntu 19.0410002000300040005000SE +/- 79.97, N = 34652.91. (CC) gcc options: -O3 -pthread -lz -llzma

NAMD

ATPase Simulation - 327,506 Atoms

OpenBenchmarking.orgdays/ns, Fewer Is BetterNAMD 2.14ATPase Simulation - 327,506 AtomsUbuntu 19.040.2180.4360.6540.8721.09SE +/- 0.00501, N = 30.96900

oneDNN

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUUbuntu 19.040.41350.8271.24051.6542.0675SE +/- 0.06760, N = 151.83780MIN: 11. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32 - Engine: CPUUbuntu 19.040.99051.9812.97153.9624.9525SE +/- 0.08113, N = 154.40219MIN: 3.981. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl

oneDNN

Harness: IP Batch 1D - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 1.5Harness: IP Batch 1D - Data Type: f32 - Engine: CPUUbuntu 19.040.94531.89062.83593.78124.7265SE +/- 0.09463, N = 144.20128MIN: 2.851. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl


Phoronix Test Suite v10.8.5