gh200

ARMv8 Neoverse-V2 testing with a Pegatron JIMBO P4352 (00022432 BIOS) and NVIDIA GH200 144G HBM3e 143GB on Ubuntu 24.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2410122-NE-GH200545357
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
a
October 12
  12 Hours, 26 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


gh200OpenBenchmarking.orgPhoronix Test SuiteARMv8 Neoverse-V2 @ 3.47GHz (72 Cores)Pegatron JIMBO P4352 (00022432 BIOS)1 x 480GB LPDDR5-6400MT/s NVIDIA 699-2G530-0236-RC11000GB CT1000T700SSD3NVIDIA GH200 144G HBM3e 143GB2 x Intel X550Ubuntu 24.046.8.0-45-generic-64k (aarch64)NVIDIAOpenCL 3.0 CUDA 12.6.65GCC 13.2.0ext41920x1200ProcessorMotherboardMemoryDiskGraphicsNetworkOSKernelDisplay DriverOpenCLCompilerFile-SystemScreen ResolutionGh200 BenchmarksSystem Logs- Transparent Huge Pages: madvise- --build=aarch64-linux-gnu --disable-libquadmath --disable-libquadmath-support --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-fix-cortex-a53-843419 --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-backtrace --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-dIwDw0/gcc-13-13.2.0/debian/tmp-nvptx/usr --enable-plugin --enable-shared --enable-threads=posix --host=aarch64-linux-gnu --program-prefix=aarch64-linux-gnu- --target=aarch64-linux-gnu --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-target-system-zlib=auto --without-cuda-driver -v - Scaling Governor: cppc_cpufreq ondemand (Boost: Disabled)- OpenJDK Runtime Environment (build 21.0.4+7-Ubuntu-1ubuntu224.04)- Python 3.12.3- gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + reg_file_data_sampling: Not affected + retbleed: Not affected + spec_rstack_overflow: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of __user pointer sanitization + spectre_v2: Not affected + srbds: Not affected + tsx_async_abort: Not affected

gh200stockfish: Chess Benchmarkxnnpack: QU8MobileNetV3Smallxnnpack: QU8MobileNetV3Largexnnpack: QU8MobileNetV2xnnpack: FP16MobileNetV3Smallxnnpack: FP16MobileNetV3Largexnnpack: FP16MobileNetV2xnnpack: FP32MobileNetV3Smallxnnpack: FP32MobileNetV3Largexnnpack: FP32MobileNetV2stockfish: Chess Benchmarkblender: Barbershop - CPU-Onlylczero: Eigenonnx: ZFNet-512 - CPU - Parallelonnx: ZFNet-512 - CPU - Parallelgraphics-magick: Noise-Gaussianepoch: Conebyte: Whetstone Doublebuild-linux-kernel: defconfigbuild-linux-kernel: allmodconfigbuild-llvm: Unix Makefilespyperformance: gc_collectbyte: Pipebyte: System Callbyte: Dhrystone 2pyperformance: asyncio_tcp_sslgromacs: water_GMX50_barebuild-llvm: Ninjablender: Pabellon Barcelona - CPU-Onlypyperformance: async_tree_iopyperformance: xml_etreepyperformance: python_startuppyperformance: asyncio_websocketsbuild2: Time To Compilemnn: inception-v3mnn: mobilenet-v1-1.0mnn: MobileNetV2_224mnn: SqueezeNetV1.0mnn: resnet-v2-50mnn: squeezenetv1.1mnn: mobilenetV3mnn: nasnetblender: Classroom - CPU-Onlyblender: Fishy Cat - CPU-Onlysimdjson: PartialTweetssimdjson: DistinctUserIDsimdjson: TopTweetx265: Bosphorus 4Konnx: ResNet101_DUC_HDC-12 - CPU - Standardonnx: ResNet101_DUC_HDC-12 - CPU - Standardonnx: ResNet101_DUC_HDC-12 - CPU - Parallelonnx: ResNet101_DUC_HDC-12 - CPU - Parallelonnx: fcn-resnet101-11 - CPU - Standardonnx: fcn-resnet101-11 - CPU - Standardonnx: fcn-resnet101-11 - CPU - Parallelonnx: fcn-resnet101-11 - CPU - Parallelsimdjson: Kostyaonnx: yolov4 - CPU - Parallelonnx: yolov4 - CPU - Parallelonnx: yolov4 - CPU - Standardonnx: yolov4 - CPU - Standardonnx: T5 Encoder - CPU - Standardonnx: T5 Encoder - CPU - Standardonnx: ZFNet-512 - CPU - Standardonnx: ZFNet-512 - CPU - Standardonnx: T5 Encoder - CPU - Parallelonnx: T5 Encoder - CPU - Parallelgraphics-magick: Noise-Gaussiangraphics-magick: Rotategraphics-magick: Sharpenonnx: CaffeNet 12-int8 - CPU - Standardonnx: CaffeNet 12-int8 - CPU - Standardonnx: CaffeNet 12-int8 - CPU - Parallelonnx: CaffeNet 12-int8 - CPU - Parallelgraphics-magick: Resizingonnx: super-resolution-10 - CPU - Parallelonnx: super-resolution-10 - CPU - Parallelonnx: ResNet50 v1-12-int8 - CPU - Parallelonnx: ResNet50 v1-12-int8 - CPU - Parallelonnx: ResNet50 v1-12-int8 - CPU - Standardonnx: ResNet50 v1-12-int8 - CPU - Standardgraphics-magick: Enhancedgraphics-magick: Sharpenonnx: super-resolution-10 - CPU - Standardonnx: super-resolution-10 - CPU - Standardgraphics-magick: Rotategraphics-magick: HWB Color Spacegraphics-magick: Resizinggraphics-magick: Swirlgraphics-magick: Swirlgraphics-magick: Enhancedgraphics-magick: HWB Color Spacegromacs: MPI CPU - water_GMX50_baresimdjson: LargeRandx265: Bosphorus 1080ppyperformance: raytracepyperformance: goblender: BMW27 - CPU-Onlyc-ray: 5K - 16etcpak: Multi-Threaded - ETC2pyperformance: chaospyperformance: json_loadspyperformance: regex_compilepyperformance: django_templatepyperformance: pathlibwarpx: Plasma Accelerationwarpx: Uniform Plasmapyperformance: pickle_pure_pythonpyperformance: nbodypyperformance: floatpyperformance: crypto_pyaesc-ray: 4K - 16compress-7zip: Decompression Ratingcompress-7zip: Compression Ratingcompress-7zip: Decompression Ratingcompress-7zip: Compression Ratingpovray: Trace Timec-ray: 1080p - 16onnx: GPT-2 - CPU - Parallela584967531083148494588112268409451426967168428763381.4536022.145145.2314301188.20721978.066.710285.132276.9261.08202565282.2145868649.34998587529.81.497.156175.030154.4674845.818.751084.78613.6931.7931.5023.39611.3351.8241.1345.00878.3773.024.064.164.148.813298.190.3032222776.330.3601962160.110.4629431715.640.5829023.11172.2255.80711194.3645.146852.56003390.0674.75865210.1179.15193109.2412172091710.7916241262.143.11158321.23744253.625218.64876.19726161.3323.14523317.8683514116.28553159.0683314302826576053596566.0011.1512.6121798.238.0636.208471.19347.417.582.326.315.520.3805382116.9020207620564.556.854.820.3574188193935234205243847757.7865.195OpenBenchmarking.org

Stockfish

OpenBenchmarking.orgNodes Per Second, More Is BetterStockfishChess Benchmarka13M26M39M52M65MSE +/- 959000.15, N = 15584967531. Stockfish 16 by the Stockfish developers (see AUTHORS file)

XNNPACK

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: QU8MobileNetV3Smalla2004006008001000SE +/- 9.82, N = 310831. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: QU8MobileNetV3Largea30060090012001500SE +/- 8.97, N = 314841. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: QU8MobileNetV2a2004006008001000SE +/- 6.69, N = 39451. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: FP16MobileNetV3Smalla2004006008001000SE +/- 20.00, N = 38811. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: FP16MobileNetV3Largea30060090012001500SE +/- 21.31, N = 312261. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: FP16MobileNetV2a2004006008001000SE +/- 15.62, N = 38401. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: FP32MobileNetV3Smalla2004006008001000SE +/- 16.38, N = 39451. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: FP32MobileNetV3Largea30060090012001500SE +/- 6.51, N = 314261. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: FP32MobileNetV2a2004006008001000SE +/- 8.41, N = 39671. (CXX) g++ options: -O3 -lrt -lm

Stockfish

OpenBenchmarking.orgNodes Per Second, More Is BetterStockfish 17Chess Benchmarka40M80M120M160M200MSE +/- 6156005.01, N = 151684287631. (CXX) g++ options: -lgcov -lpthread -fno-exceptions -std=c++17 -fno-peel-loops -fno-tracer -pedantic -O3 -funroll-loops -flto -flto-partition=one -flto=jobserver

Blender

Blender is an open-source 3D creation software project. This test is of Blender's Cycles benchmark with various sample files. GPU computing is supported. This system/blender test profile makes use of the system-supplied Blender. Use pts/blender if wishing to stick to a fixed version of Blender. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.0.2Blend File: Barbershop - Compute: CPU-Onlya80160240320400SE +/- 0.54, N = 3381.45

LeelaChessZero

OpenBenchmarking.orgNodes Per Second, More Is BetterLeelaChessZero 0.31.1Backend: Eigena80160240320400SE +/- 4.26, N = 33601. (CXX) g++ options: -flto -pthread

ONNX Runtime

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: ZFNet-512 - Device: CPU - Executor: Parallela510152025SE +/- 0.25, N = 1522.151. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: ZFNet-512 - Device: CPU - Executor: Parallela1020304050SE +/- 0.50, N = 1545.231. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

GraphicsMagick

This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample high resolution (currently 15400 x 6940) JPEG image. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Noise-Gaussiana70140210280350SE +/- 2.18, N = 153011. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

Epoch

OpenBenchmarking.orgSeconds, Fewer Is BetterEpoch 4.19.4Epoch3D Deck: Conea4080120160200SE +/- 2.18, N = 4188.201. (F9X) gfortran options: -O3 -std=f2003 -Jobj -lsdf -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz

BYTE Unix Benchmark

OpenBenchmarking.orgMWIPS, More Is BetterBYTE Unix Benchmark 5.1.3-gitComputational Test: Whetstone Doublea150K300K450K600K750KSE +/- 19.25, N = 3721978.01. (CC) gcc options: -pedantic -O3 -ffast-math -march=native -mtune=native -lm

Timed Linux Kernel Compilation

This test times how long it takes to build the Linux kernel in a default configuration (defconfig) for the architecture being tested or alternatively an allmodconfig for building all possible kernel modules for the build. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 6.8Build: defconfiga1530456075SE +/- 0.55, N = 1366.71

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 6.8Build: allmodconfiga60120180240300SE +/- 2.59, N = 3285.13

Timed LLVM Compilation

This test times how long it takes to compile/build the LLVM compiler stack. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed LLVM Compilation 16.0Build System: Unix Makefilesa60120180240300SE +/- 0.32, N = 3276.93

PyPerformance

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: gc_collecta0.2430.4860.7290.9721.215SE +/- 0.01, N = 151.08

BYTE Unix Benchmark

OpenBenchmarking.orgLPS, More Is BetterBYTE Unix Benchmark 5.1.3-gitComputational Test: Pipea40M80M120M160M200MSE +/- 32087.94, N = 3202565282.21. (CC) gcc options: -pedantic -O3 -ffast-math -march=native -mtune=native -lm

OpenBenchmarking.orgLPS, More Is BetterBYTE Unix Benchmark 5.1.3-gitComputational Test: System Calla30M60M90M120M150MSE +/- 15202.21, N = 3145868649.31. (CC) gcc options: -pedantic -O3 -ffast-math -march=native -mtune=native -lm

OpenBenchmarking.orgLPS, More Is BetterBYTE Unix Benchmark 5.1.3-gitComputational Test: Dhrystone 2a1100M2200M3300M4400M5500MSE +/- 2591819.88, N = 34998587529.81. (CC) gcc options: -pedantic -O3 -ffast-math -march=native -mtune=native -lm

PyPerformance

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: asyncio_tcp_ssla0.33530.67061.00591.34121.6765SE +/- 0.00, N = 31.49

GROMACS

OpenBenchmarking.orgNs Per Day, More Is BetterGROMACSInput: water_GMX50_barea246810SE +/- 0.004, N = 37.1561. GROMACS version: 2023.3-Ubuntu_2023.3_1ubuntu3

Timed LLVM Compilation

This test times how long it takes to compile/build the LLVM compiler stack. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed LLVM Compilation 16.0Build System: Ninjaa4080120160200SE +/- 1.19, N = 3175.03

Blender

Blender is an open-source 3D creation software project. This test is of Blender's Cycles benchmark with various sample files. GPU computing is supported. This system/blender test profile makes use of the system-supplied Blender. Use pts/blender if wishing to stick to a fixed version of Blender. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.0.2Blend File: Pabellon Barcelona - Compute: CPU-Onlya306090120150SE +/- 0.37, N = 3154.46

PyPerformance

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: async_tree_ioa160320480640800SE +/- 2.96, N = 3748

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: xml_etreea1020304050SE +/- 0.03, N = 345.8

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: python_startupa510152025SE +/- 0.06, N = 318.7

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: asyncio_websocketsa110220330440550SE +/- 0.33, N = 3510

Build2

OpenBenchmarking.orgSeconds, Fewer Is BetterBuild2 0.17Time To Compilea20406080100SE +/- 0.22, N = 384.79

Mobile Neural Network

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: inception-v3a48121620SE +/- 0.02, N = 313.69MIN: 11.51 / MAX: 42.341. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: mobilenet-v1-1.0a0.40340.80681.21021.61362.017SE +/- 0.005, N = 31.793MIN: 1.34 / MAX: 22.051. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: MobileNetV2_224a0.3380.6761.0141.3521.69SE +/- 0.019, N = 31.502MIN: 1.12 / MAX: 13.521. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: SqueezeNetV1.0a0.76411.52822.29233.05643.8205SE +/- 0.027, N = 33.396MIN: 2.14 / MAX: 29.881. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: resnet-v2-50a3691215SE +/- 0.10, N = 311.34MIN: 8.54 / MAX: 42.161. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: squeezenetv1.1a0.41040.82081.23121.64162.052SE +/- 0.044, N = 31.824MIN: 1.17 / MAX: 20.371. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: mobilenetV3a0.25520.51040.76561.02081.276SE +/- 0.009, N = 31.134MIN: 0.69 / MAX: 11.141. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: nasneta1.12682.25363.38044.50725.634SE +/- 0.036, N = 35.008MIN: 4.49 / MAX: 27.911. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

Blender

Blender is an open-source 3D creation software project. This test is of Blender's Cycles benchmark with various sample files. GPU computing is supported. This system/blender test profile makes use of the system-supplied Blender. Use pts/blender if wishing to stick to a fixed version of Blender. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.0.2Blend File: Classroom - Compute: CPU-Onlya20406080100SE +/- 0.08, N = 378.37

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.0.2Blend File: Fishy Cat - Compute: CPU-Onlya1632486480SE +/- 0.44, N = 373.02

simdjson

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 3.10Throughput Test: PartialTweetsa0.91351.8272.74053.6544.5675SE +/- 0.00, N = 34.061. (CXX) g++ options: -O3 -lrt

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 3.10Throughput Test: DistinctUserIDa0.9361.8722.8083.7444.68SE +/- 0.00, N = 34.161. (CXX) g++ options: -O3 -lrt

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 3.10Throughput Test: TopTweeta0.93151.8632.79453.7264.6575SE +/- 0.01, N = 34.141. (CXX) g++ options: -O3 -lrt

x265

OpenBenchmarking.orgFrames Per Second, More Is Betterx265Video Input: Bosphorus 4Ka246810SE +/- 0.03, N = 38.811. x265 [info]: HEVC encoder version 3.5+1-f0c1022b6

ONNX Runtime

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: ResNet101_DUC_HDC-12 - Device: CPU - Executor: Standarda7001400210028003500SE +/- 21.77, N = 33298.191. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: ResNet101_DUC_HDC-12 - Device: CPU - Executor: Standarda0.06820.13640.20460.27280.341SE +/- 0.001997, N = 30.3032221. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: ResNet101_DUC_HDC-12 - Device: CPU - Executor: Parallela6001200180024003000SE +/- 9.74, N = 32776.331. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: ResNet101_DUC_HDC-12 - Device: CPU - Executor: Parallela0.0810.1620.2430.3240.405SE +/- 0.001260, N = 30.3601961. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: fcn-resnet101-11 - Device: CPU - Executor: Standarda5001000150020002500SE +/- 4.77, N = 32160.111. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: fcn-resnet101-11 - Device: CPU - Executor: Standarda0.10420.20840.31260.41680.521SE +/- 0.001019, N = 30.4629431. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: fcn-resnet101-11 - Device: CPU - Executor: Parallela400800120016002000SE +/- 8.64, N = 31715.641. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: fcn-resnet101-11 - Device: CPU - Executor: Parallela0.13120.26240.39360.52480.656SE +/- 0.002945, N = 30.5829021. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

simdjson

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 3.10Throughput Test: Kostyaa0.69981.39962.09942.79923.499SE +/- 0.01, N = 33.111. (CXX) g++ options: -O3 -lrt

ONNX Runtime

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: yolov4 - Device: CPU - Executor: Parallela4080120160200SE +/- 1.46, N = 3172.231. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: yolov4 - Device: CPU - Executor: Parallela1.30662.61323.91985.22646.533SE +/- 0.04903, N = 35.807111. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: yolov4 - Device: CPU - Executor: Standarda4080120160200SE +/- 2.71, N = 3194.361. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: yolov4 - Device: CPU - Executor: Standarda1.1582.3163.4744.6325.79SE +/- 0.07121, N = 35.146851. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: T5 Encoder - Device: CPU - Executor: Standarda0.5761.1521.7282.3042.88SE +/- 0.01750, N = 32.560031. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: T5 Encoder - Device: CPU - Executor: Standarda80160240320400SE +/- 2.61, N = 3390.071. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: ZFNet-512 - Device: CPU - Executor: Standarda1.07072.14143.21214.28285.3535SE +/- 0.05942, N = 34.758651. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: ZFNet-512 - Device: CPU - Executor: Standarda50100150200250SE +/- 2.64, N = 3210.121. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: T5 Encoder - Device: CPU - Executor: Parallela3691215SE +/- 0.06344, N = 39.151931. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: T5 Encoder - Device: CPU - Executor: Parallela20406080100SE +/- 0.76, N = 3109.241. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

GraphicsMagick

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: Noise-Gaussiana50100150200250SE +/- 1.73, N = 32171. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: Rotatea50100150200250SE +/- 0.88, N = 32091. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: Sharpena4080120160200SE +/- 0.33, N = 31711. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

ONNX Runtime

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: CaffeNet 12-int8 - Device: CPU - Executor: Standarda0.17810.35620.53430.71240.8905SE +/- 0.001678, N = 30.7916241. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: CaffeNet 12-int8 - Device: CPU - Executor: Standarda30060090012001500SE +/- 2.68, N = 31262.141. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: CaffeNet 12-int8 - Device: CPU - Executor: Parallela0.70011.40022.10032.80043.5005SE +/- 0.00260, N = 33.111581. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: CaffeNet 12-int8 - Device: CPU - Executor: Parallela70140210280350SE +/- 0.27, N = 3321.241. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

GraphicsMagick

This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample high resolution (currently 15400 x 6940) JPEG image. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Resizinga100200300400500SE +/- 6.06, N = 34421. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

ONNX Runtime

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: super-resolution-10 - Device: CPU - Executor: Parallela1224364860SE +/- 0.29, N = 353.631. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: super-resolution-10 - Device: CPU - Executor: Parallela510152025SE +/- 0.10, N = 318.651. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Parallela246810SE +/- 0.02687, N = 36.197261. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Parallela4080120160200SE +/- 0.70, N = 3161.331. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standarda0.70771.41542.12312.83083.5385SE +/- 0.00610, N = 33.145231. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standarda70140210280350SE +/- 0.61, N = 3317.871. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

GraphicsMagick

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: Enhanceda80160240320400SE +/- 0.33, N = 33511. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Sharpena90180270360450SE +/- 0.58, N = 34111. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

ONNX Runtime

OpenBenchmarking.orgInference Time Cost (ms), Fewer Is BetterONNX Runtime 1.19Model: super-resolution-10 - Device: CPU - Executor: Standarda246810SE +/- 0.04448, N = 36.285531. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: super-resolution-10 - Device: CPU - Executor: Standarda4080120160200SE +/- 1.12, N = 3159.071. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

GraphicsMagick

This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample high resolution (currently 15400 x 6940) JPEG image. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Rotatea70140210280350SE +/- 4.26, N = 33311. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: HWB Color Spacea90180270360450SE +/- 0.67, N = 34301. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: Resizinga60120180240300SE +/- 1.15, N = 32821. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Swirla140280420560700SE +/- 4.26, N = 36571. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: Swirla130260390520650SE +/- 5.51, N = 36051. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Enhanceda80160240320400SE +/- 0.67, N = 33591. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: HWB Color Spacea140280420560700SE +/- 8.82, N = 36561. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

GROMACS

The GROMACS (GROningen MAchine for Chemical Simulations) molecular dynamics package testing with the water_GMX50 data. This test profile allows selecting between CPU and GPU-based GROMACS builds. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNs Per Day, More Is BetterGROMACS 2024Implementation: MPI CPU - Input: water_GMX50_barea246810SE +/- 0.003, N = 36.0011. (CXX) g++ options: -O3 -lm

simdjson

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 3.10Throughput Test: LargeRandoma0.25880.51760.77641.03521.294SE +/- 0.00, N = 31.151. (CXX) g++ options: -O3 -lrt

x265

OpenBenchmarking.orgFrames Per Second, More Is Betterx265Video Input: Bosphorus 1080pa3691215SE +/- 0.18, N = 312.611. x265 [info]: HEVC encoder version 3.5+1-f0c1022b6

PyPerformance

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: raytracea50100150200250SE +/- 0.33, N = 3217

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: goa20406080100SE +/- 0.07, N = 398.2

Blender

Blender is an open-source 3D creation software project. This test is of Blender's Cycles benchmark with various sample files. GPU computing is supported. This system/blender test profile makes use of the system-supplied Blender. Use pts/blender if wishing to stick to a fixed version of Blender. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.0.2Blend File: BMW27 - Compute: CPU-Onlya918273645SE +/- 0.04, N = 338.06

C-Ray

OpenBenchmarking.orgSeconds, Fewer Is BetterC-Ray 2.0Resolution: 5K - Rays Per Pixel: 16a816243240SE +/- 0.02, N = 336.211. (CC) gcc options: -lpthread -lm

Etcpak

OpenBenchmarking.orgMpx/s, More Is BetterEtcpak 2.0Benchmark: Multi-Threaded - Configuration: ETC2a100200300400500SE +/- 2.47, N = 3471.191. (CXX) g++ options: -flto -pthread

PyPerformance

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: chaosa1122334455SE +/- 0.06, N = 347.4

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: json_loadsa48121620SE +/- 0.06, N = 317.5

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: regex_compilea20406080100SE +/- 0.12, N = 382.3

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: django_templatea612182430SE +/- 0.12, N = 326.3

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: pathliba48121620SE +/- 0.03, N = 315.5

WarpX

OpenBenchmarking.orgSeconds, Fewer Is BetterWarpX 24.10Input: Plasma Accelerationa510152025SE +/- 0.03, N = 320.381. (CXX) g++ options: -O3

OpenBenchmarking.orgSeconds, Fewer Is BetterWarpX 24.10Input: Uniform Plasmaa48121620SE +/- 0.18, N = 316.901. (CXX) g++ options: -O3

PyPerformance

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: pickle_pure_pythona4080120160200SE +/- 0.33, N = 3205

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: nbodya1428425670SE +/- 0.09, N = 364.5

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: floata1326395265SE +/- 0.03, N = 356.8

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: crypto_pyaesa1224364860SE +/- 0.03, N = 354.8

C-Ray

OpenBenchmarking.orgSeconds, Fewer Is BetterC-Ray 2.0Resolution: 4K - Rays Per Pixel: 16a510152025SE +/- 0.00, N = 320.361. (CC) gcc options: -lpthread -lm

7-Zip Compression

OpenBenchmarking.orgMIPS, More Is Better7-Zip CompressionTest: Decompression Ratinga90K180K270K360K450KSE +/- 507.84, N = 34188191. 7-Zip 23.01 (arm64) : Copyright (c) 1999-2023 Igor Pavlov : 2023-06-20

OpenBenchmarking.orgMIPS, More Is Better7-Zip CompressionTest: Compression Ratinga80K160K240K320K400KSE +/- 3097.20, N = 33935231. 7-Zip 23.01 (arm64) : Copyright (c) 1999-2023 Igor Pavlov : 2023-06-20

OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 24.05Test: Decompression Ratinga90K180K270K360K450KSE +/- 944.71, N = 34205241. (CXX) g++ options: -lpthread -ldl -O2 -fPIC

OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 24.05Test: Compression Ratinga80K160K240K320K400KSE +/- 4213.31, N = 33847751. (CXX) g++ options: -lpthread -ldl -O2 -fPIC

PostgreSQL

Scaling Factor: 100 - Clients: 800 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1 - Clients: 1000 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 100 - Clients: 1000 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 100 - Clients: 1000 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1 - Clients: 800 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1000 - Clients: 1000 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1000 - Clients: 500 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 100 - Clients: 800 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 100 - Clients: 500 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 100 - Clients: 500 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1 - Clients: 800 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1 - Clients: 500 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1000 - Clients: 800 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1000 - Clients: 1000 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1000 - Clients: 800 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1000 - Clients: 500 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1 - Clients: 500 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1 - Clients: 1000 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

POV-Ray

OpenBenchmarking.orgSeconds, Fewer Is BetterPOV-RayTrace Timea246810SE +/- 0.061, N = 37.7861. POV-Ray 3.7.0.10.unofficial

LeelaChessZero

Backend: BLAS

a: The test quit with a non-zero exit status.

C-Ray

OpenBenchmarking.orgSeconds, Fewer Is BetterC-Ray 2.0Resolution: 1080p - Rays Per Pixel: 16a1.16892.33783.50674.67565.8445SE +/- 0.003, N = 35.1951. (CC) gcc options: -lpthread -lm

Apache Cassandra

Test: Writes

a: The test run did not produce a result.

ONNX Runtime

Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "FasterRCNN-12-int8/FasterRCNN-12-int8.onnx" failed: No such file or directory

Model: GPT-2 - Device: CPU - Executor: Standard

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "GPT2/model.onnx" failed: No such file or directory

GROMACS

The GROMACS (GROningen MAchine for Chemical Simulations) molecular dynamics package testing with the water_GMX50 data. This test profile allows selecting between CPU and GPU-based GROMACS builds. Learn more via the OpenBenchmarking.org test page.

Implementation: NVIDIA CUDA GPU - Input: water_GMX50_bare

a: The test quit with a non-zero exit status. E: ./gromacs: 5: /cuda-build/run-gromacs: not found

ONNX Runtime

Model: bertsquad-12 - Device: CPU - Executor: Standard

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "bertsquad-12/bertsquad-12.onnx" failed: No such file or directory

Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "resnet100/resnet100.onnx" failed: No such file or directory

Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Parallel

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "FasterRCNN-12-int8/FasterRCNN-12-int8.onnx" failed: No such file or directory

Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "resnet100/resnet100.onnx" failed: No such file or directory

Model: bertsquad-12 - Device: CPU - Executor: Parallel

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "bertsquad-12/bertsquad-12.onnx" failed: No such file or directory

Model: GPT-2 - Device: CPU - Executor: Parallel

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "GPT2/model.onnx" failed: No such file or directory

118 Results Shown

Stockfish
XNNPACK:
  QU8MobileNetV3Small
  QU8MobileNetV3Large
  QU8MobileNetV2
  FP16MobileNetV3Small
  FP16MobileNetV3Large
  FP16MobileNetV2
  FP32MobileNetV3Small
  FP32MobileNetV3Large
  FP32MobileNetV2
Stockfish
Blender
LeelaChessZero
ONNX Runtime:
  ZFNet-512 - CPU - Parallel:
    Inference Time Cost (ms)
    Inferences Per Second
GraphicsMagick
Epoch
BYTE Unix Benchmark
Timed Linux Kernel Compilation:
  defconfig
  allmodconfig
Timed LLVM Compilation
PyPerformance
BYTE Unix Benchmark:
  Pipe
  System Call
  Dhrystone 2
PyPerformance
GROMACS
Timed LLVM Compilation
Blender
PyPerformance:
  async_tree_io
  xml_etree
  python_startup
  asyncio_websockets
Build2
Mobile Neural Network:
  inception-v3
  mobilenet-v1-1.0
  MobileNetV2_224
  SqueezeNetV1.0
  resnet-v2-50
  squeezenetv1.1
  mobilenetV3
  nasnet
Blender:
  Classroom - CPU-Only
  Fishy Cat - CPU-Only
simdjson:
  PartialTweets
  DistinctUserID
  TopTweet
x265
ONNX Runtime:
  ResNet101_DUC_HDC-12 - CPU - Standard:
    Inference Time Cost (ms)
    Inferences Per Second
  ResNet101_DUC_HDC-12 - CPU - Parallel:
    Inference Time Cost (ms)
    Inferences Per Second
  fcn-resnet101-11 - CPU - Standard:
    Inference Time Cost (ms)
    Inferences Per Second
  fcn-resnet101-11 - CPU - Parallel:
    Inference Time Cost (ms)
    Inferences Per Second
simdjson
ONNX Runtime:
  yolov4 - CPU - Parallel:
    Inference Time Cost (ms)
    Inferences Per Second
  yolov4 - CPU - Standard:
    Inference Time Cost (ms)
    Inferences Per Second
  T5 Encoder - CPU - Standard:
    Inference Time Cost (ms)
    Inferences Per Second
  ZFNet-512 - CPU - Standard:
    Inference Time Cost (ms)
    Inferences Per Second
  T5 Encoder - CPU - Parallel:
    Inference Time Cost (ms)
    Inferences Per Second
GraphicsMagick:
  Noise-Gaussian
  Rotate
  Sharpen
ONNX Runtime:
  CaffeNet 12-int8 - CPU - Standard:
    Inference Time Cost (ms)
    Inferences Per Second
  CaffeNet 12-int8 - CPU - Parallel:
    Inference Time Cost (ms)
    Inferences Per Second
GraphicsMagick
ONNX Runtime:
  super-resolution-10 - CPU - Parallel:
    Inference Time Cost (ms)
    Inferences Per Second
  ResNet50 v1-12-int8 - CPU - Parallel:
    Inference Time Cost (ms)
    Inferences Per Second
  ResNet50 v1-12-int8 - CPU - Standard:
    Inference Time Cost (ms)
    Inferences Per Second
GraphicsMagick
GraphicsMagick
ONNX Runtime:
  super-resolution-10 - CPU - Standard:
    Inference Time Cost (ms)
    Inferences Per Second
GraphicsMagick
GraphicsMagick:
  HWB Color Space
  Resizing
GraphicsMagick
GraphicsMagick
GraphicsMagick:
  Enhanced
  HWB Color Space
GROMACS
simdjson
x265
PyPerformance:
  raytrace
  go
Blender
C-Ray
Etcpak
PyPerformance:
  chaos
  json_loads
  regex_compile
  django_template
  pathlib
WarpX:
  Plasma Acceleration
  Uniform Plasma
PyPerformance:
  pickle_pure_python
  nbody
  float
  crypto_pyaes
C-Ray
7-Zip Compression:
  Decompression Rating
  Compression Rating
7-Zip Compression:
  Decompression Rating
  Compression Rating
POV-Ray
C-Ray