tr

AMD Ryzen Threadripper PRO 5965WX 24-Cores testing with a ASUS Pro WS WRX80E-SAGE SE WIFI (1201 BIOS) and ASUS NVIDIA NV106 2GB on Ubuntu 23.10 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2311098-NE-TR860765996
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results

Limit displaying results to tests within:

AV1 2 Tests
BLAS (Basic Linear Algebra Sub-Routine) Tests 2 Tests
C++ Boost Tests 3 Tests
Timed Code Compilation 8 Tests
C/C++ Compiler Tests 15 Tests
Compression Tests 3 Tests
CPU Massive 24 Tests
Creator Workloads 18 Tests
Cryptography 2 Tests
Database Test Suite 9 Tests
Encoding 9 Tests
Fortran Tests 6 Tests
Game Development 4 Tests
HPC - High Performance Computing 13 Tests
Java Tests 4 Tests
Common Kernel Benchmarks 5 Tests
LAPACK (Linear Algebra Pack) Tests 2 Tests
Machine Learning 5 Tests
Molecular Dynamics 2 Tests
MPI Benchmarks 4 Tests
Multi-Core 33 Tests
NVIDIA GPU Compute 3 Tests
Intel oneAPI 7 Tests
OpenMPI Tests 9 Tests
Programmer / Developer System Benchmarks 11 Tests
Python Tests 8 Tests
Raytracing 2 Tests
Renderers 3 Tests
Scientific Computing 4 Tests
Server 11 Tests
Server CPU Tests 19 Tests
Video Encoding 9 Tests
Common Workstation Benchmarks 3 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
a
November 07 2023
  20 Hours, 53 Minutes
b
November 09 2023
  19 Hours, 54 Minutes
Invert Hiding All Results Option
  20 Hours, 23 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


trOpenBenchmarking.orgPhoronix Test SuiteAMD Ryzen Threadripper PRO 5965WX 24-Cores @ 3.80GHz (24 Cores / 48 Threads)ASUS Pro WS WRX80E-SAGE SE WIFI (1201 BIOS)AMD Starship/Matisse128GB2048GB SOLIDIGM SSDPFKKW020X7ASUS NVIDIA NV106 2GBAMD Starship/MatisseVA24312 x Intel X550 + Intel Wi-Fi 6 AX200Ubuntu 23.106.5.0-10-generic (x86_64)GNOME Shell 45.0X Server + Waylandnouveau4.3 Mesa 23.2.1-1ubuntu3GCC 13.2.0ext41920x1080ProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerDisplay DriverOpenGLCompilerFile-SystemScreen ResolutionTr BenchmarksSystem Logs- Transparent Huge Pages: madvise- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - NONE / relatime,rw / Block Size: 4096- Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0xa008205- OpenJDK Runtime Environment (build 11.0.20+8-post-Ubuntu-1ubuntu1)- Python 3.11.6- gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Mitigation of safe RET no microcode + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected

a vs. b ComparisonPhoronix Test SuiteBaseline+70.1%+70.1%+140.2%+140.2%+210.3%+210.3%+280.4%+280.4%280.5%94.8%80.5%73.4%63.3%54.6%47.2%46.4%17.4%16.6%16.5%11.3%9.7%9%7.7%7.4%7.1%6.6%6.4%6.2%5.4%5.2%5.2%5.1%5%3.8%3.7%3.6%3.5%3.5%3.4%3.3%3.2%3.2%3.1%2.9%2.9%2.8%2.5%2.5%2.5%2.4%2.3%2.3%2.2%2.2%2.2%2.2%2.1%2%2%CPU-v3-v3 - mobilenet-v3D.B.s - f32 - CPU166.9%100 - 800 - Read Write143.6%100 - 800 - Read Write - Average Latency143.6%100 - 1000 - Read Write - Average Latency120.4%100 - 1000 - Read Write120.4%CPU-v2-v2 - mobilenet-v2CPU - alexnetFile Status - 20 - 100000D.B.s - u8s8f32 - CPUIP Shapes 1D - f32 - CPU57.6%CPU - FastestDetCPU - regnety_400m51.9%CPU - efficientnet-b050.5%CPU - yolov4-tinyCPU - shufflenet-v2IP Shapes 3D - f32 - CPU44.3%IP Shapes 1D - u8s8f32 - CPU31.8%IP Shapes 3D - u8s8f32 - CPU29.4%CPU - googlenet27.6%D.B.s - u8s8f32 - CPU26.3%OpenMP HotSpot3D21%CPU - mnasnet17.6%CPU - squeezenet_ssdCPU - vision_transformerCPU - vgg16Context Switching1 - 500 - Read Write - Average Latency11.2%1 - 500 - Read Write11.1%CPU - blazeface100 - 100 - 800 - 1009.3%EP.DG.C.S.F8 - Compression Speed1 - 1000 - Read Write7.3%1 - 1000 - Read Write - Average Latency7.2%CPU - mobilenetggml-base.en - 2.S.o.t.U6.9%D.RR.N.N.I - f32 - CPUDust 2D tau100.0e.G.B.S - 2406.2%200 - 100 - 800 - 1005.6%Preset 8 - Bosphorus 1080pLiH_ae_MSD1 - 800 - Read Write - Average Latency1 - 800 - Read WriteC.B.S.A - f32 - CPUFused Multiply-Add4.9%500 - 100 - 800 - 4004.5%OpenMP LavaMD4.3%SHA5124.2%Read While Writing4.1%100 - 100 - 800 - 1003.9%blosclz shuffle - 8MB3.9%Bosphorus 1080p - Faster1.R.H.D.F.R.C.CVMAF Optimized - Bosphorus 4KH4_aeCPU - resnet18Bumper Beamclover_bmForkingR.N.N.I - u8s8f32 - CPU2blosclz bitshuffle - 256MB3%Bosphorus 1080p - Very Fast3%C.P.D.T3%800 - 100 - 200 - 4003%2002.9%100 - 100 - 200 - 100144 144 144 - 60500 - 100 - 500 - 4002.8%Preset 4 - Bosphorus 1080p2.8%blosclz bitshuffle - 32MB2.8%Create - 20 - 1000002.8%Preset 12 - Bosphorus 4KMutex2.7%CPU - 32 - GoogLeNet2.6%800 - 100 - 800 - 4002.6%SP.BUnix MakefilesI.a.F.S.I.D.C500 - 100 - 500 - 100blosclz bitshuffle - 64MB2.4%Bosphorus 1080p - Fast200 - 100 - 200 - 1002.3%B.S.o.W2.3%3, Long Mode - Compression Speed2.3%H.E.R.F.I - CPUH.E.R.F.I - CPU100 - 500 - Read Write - Average Latency2.2%100 - 500 - Read Write2.2%O_ae_pyscf_UHFMount St. Helens2.2%blosclz noshuffle - 16MBVMAF Optimized - Bosphorus 1080pD.B.s - f32 - CPU200 - 100 - 800 - 1002%12 - D.S2%19, Long Mode - D.S500 - 100 - 200 - 100NCNNoneDNNPostgreSQLPostgreSQLPostgreSQLPostgreSQLNCNNNCNNApache HadooponeDNNoneDNNNCNNNCNNNCNNNCNNNCNNoneDNNoneDNNoneDNNNCNNoneDNNRodiniaNCNNNCNNNCNNNCNNStress-NGPostgreSQLPostgreSQLNCNNApache IoTDBNAS Parallel BenchmarksStress-NGZstd CompressionPostgreSQLPostgreSQLNCNNWhisper.cpp7-Zip CompressiononeDNNMonte Carlo Simulations of Ionised NebulaeeasyWaveApache IoTDBSVT-AV1QMCPACKPostgreSQLPostgreSQLoneDNNStress-NGApache IoTDBRodiniaOpenSSLRocksDBApache IoTDBC-BloscVVenCClickHouseSVT-VP9QMCPACKNCNNOpenRadiossCloverLeafStress-NGoneDNNSQLiteC-BloscKvazaarOpenRadiossApache IoTDBnginxApache IoTDBHigh Performance Conjugate GradientApache IoTDBSVT-AV1C-BloscApache HadoopSVT-AV1Stress-NGTensorFlowApache IoTDBNAS Parallel BenchmarksTimed LLVM CompilationOpenRadiossApache IoTDBC-BloscVVenCApache IoTDBOpenRadiossZstd CompressionOpenVINOOpenVINOPostgreSQLPostgreSQLQMCPACKSPECFEM3DC-BloscSVT-VP9oneDNNApache IoTDBZstd CompressionZstd CompressionApache IoTDBab

trncnn: CPU-v3-v3 - mobilenet-v3onednn: Deconvolution Batch shapes_3d - f32 - CPUpgbench: 100 - 800 - Read Writepgbench: 100 - 800 - Read Write - Average Latencypgbench: 100 - 1000 - Read Write - Average Latencypgbench: 100 - 1000 - Read Writencnn: CPU-v2-v2 - mobilenet-v2ncnn: CPU - alexnethadoop: File Status - 20 - 100000onednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: IP Shapes 1D - f32 - CPUncnn: CPU - FastestDetncnn: CPU - regnety_400mncnn: CPU - efficientnet-b0ncnn: CPU - yolov4-tinyncnn: CPU - shufflenet-v2onednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUncnn: CPU - googlenetonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUrodinia: OpenMP HotSpot3Dncnn: CPU - mnasnetncnn: CPU - squeezenet_ssdncnn: CPU - vision_transformerncnn: CPU - vgg16stress-ng: Context Switchingpgbench: 1 - 500 - Read Write - Average Latencypgbench: 1 - 500 - Read Writencnn: CPU - blazefaceapache-iotdb: 100 - 100 - 800 - 100npb: EP.Dstress-ng: Glibc C String Functionscompress-zstd: 8 - Compression Speedpgbench: 1 - 1000 - Read Writepgbench: 1 - 1000 - Read Write - Average Latencyncnn: CPU - mobilenetwhisper-cpp: ggml-base.en - 2016 State of the Unioncompress-7zip: Decompression Ratingonednn: Recurrent Neural Network Inference - f32 - CPUmocassin: Dust 2D tau100.0easywave: e2Asean Grid + BengkuluSept2007 Source - 240apache-iotdb: 200 - 100 - 800 - 100svt-av1: Preset 8 - Bosphorus 1080pqmcpack: LiH_ae_MSDpgbench: 1 - 800 - Read Write - Average Latencypgbench: 1 - 800 - Read Writeonednn: Convolution Batch Shapes Auto - f32 - CPUstress-ng: Fused Multiply-Addapache-iotdb: 500 - 100 - 800 - 400rodinia: OpenMP LavaMDopenssl: SHA512rocksdb: Read While Writingapache-iotdb: 100 - 100 - 800 - 100blosc: blosclz shuffle - 8MBvvenc: Bosphorus 1080p - Fasterclickhouse: 100M Rows Hits Dataset, First Run / Cold Cachesvt-vp9: VMAF Optimized - Bosphorus 4Kqmcpack: H4_aencnn: CPU - resnet18openradioss: Bumper Beamcloverleaf: clover_bmstress-ng: Forkingonednn: Recurrent Neural Network Inference - u8s8f32 - CPUsqlite: 2blosc: blosclz bitshuffle - 256MBkvazaar: Bosphorus 1080p - Very Fastopenradioss: Cell Phone Drop Testapache-iotdb: 800 - 100 - 200 - 400nginx: 200apache-iotdb: 100 - 100 - 200 - 100hpcg: 144 144 144 - 60apache-iotdb: 500 - 100 - 500 - 400svt-av1: Preset 4 - Bosphorus 1080pblosc: blosclz bitshuffle - 32MBhadoop: Create - 20 - 100000svt-av1: Preset 12 - Bosphorus 4Kstress-ng: Mutextensorflow: CPU - 32 - GoogLeNetapache-iotdb: 800 - 100 - 800 - 400npb: SP.Bbuild-llvm: Unix Makefilesopenradioss: INIVOL and Fluid Structure Interaction Drop Containerapache-iotdb: 500 - 100 - 500 - 100blosc: blosclz bitshuffle - 64MBvvenc: Bosphorus 1080p - Fastapache-iotdb: 200 - 100 - 200 - 100openradioss: Bird Strike on Windshieldcompress-zstd: 3, Long Mode - Compression Speedopenvino: Handwritten English Recognition FP16-INT8 - CPUopenvino: Handwritten English Recognition FP16-INT8 - CPUpgbench: 100 - 500 - Read Write - Average Latencypgbench: 100 - 500 - Read Writeqmcpack: O_ae_pyscf_UHFspecfem3d: Mount St. Helensblosc: blosclz noshuffle - 16MBsvt-vp9: VMAF Optimized - Bosphorus 1080ponednn: Deconvolution Batch shapes_1d - f32 - CPUapache-iotdb: 200 - 100 - 800 - 100compress-zstd: 12 - Decompression Speedcompress-zstd: 19, Long Mode - Decompression Speedapache-iotdb: 500 - 100 - 200 - 100blosc: blosclz bitshuffle - 8MBonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUapache-iotdb: 800 - 100 - 800 - 100apache-iotdb: 800 - 100 - 200 - 400openradioss: Rubber O-Ring Seal Installationstress-ng: SENDFILEstress-ng: Pipecompress-zstd: 3 - Compression Speedonednn: Recurrent Neural Network Training - u8s8f32 - CPUpgbench: 1 - 500 - Read Only - Average Latencyqmcpack: FeCO6_b3lyp_gmshadoop: Open - 20 - 100000stress-ng: Futexapache-iotdb: 800 - 100 - 800 - 100openvino: Vehicle Detection FP16 - CPUopenvino: Vehicle Detection FP16 - CPUwhisper-cpp: ggml-small.en - 2016 State of the Unionpgbench: 1 - 500 - Read Onlycompress-7zip: Compression Ratingkvazaar: Bosphorus 1080p - Mediumblosc: blosclz shuffle - 16MBqmcpack: simple-H2Onpb: EP.Csvt-av1: Preset 12 - Bosphorus 1080pblosc: blosclz shuffle - 128MBvvenc: Bosphorus 4K - Fastertensorflow: CPU - 16 - GoogLeNetsvt-av1: Preset 13 - Bosphorus 1080pspecfem3d: Layered Halfspacestress-ng: Atomicuvg266: Bosphorus 1080p - Very Fastx265: Bosphorus 4Ksqlite: 4svt-av1: Preset 8 - Bosphorus 4Kkvazaar: Bosphorus 1080p - Ultra Fastopenvino: Handwritten English Recognition FP16 - CPUopenvino: Handwritten English Recognition FP16 - CPUtensorflow: CPU - 64 - GoogLeNeteasywave: e2Asean Grid + BengkuluSept2007 Source - 1200whisper-cpp: ggml-medium.en - 2016 State of the Unionjohn-the-ripper: HMAC-SHA512uvg266: Bosphorus 1080p - Slowkvazaar: Bosphorus 1080p - Slowcompress-zstd: 8, Long Mode - Compression Speedblosc: blosclz noshuffle - 32MBsqlite: 1blosc: blosclz noshuffle - 8MBlibxsmm: 256blosc: blosclz shuffle - 256MBnwchem: C240 Buckyballapache-iotdb: 500 - 100 - 500 - 100uvg266: Bosphorus 4K - Ultra Fastcompress-zstd: 3, Long Mode - Decompression Speedstream: Scaleospray: gravity_spheres_volume/dim_512/scivis/real_timehadoop: Delete - 20 - 100000vvenc: Bosphorus 4K - Fastembree: Pathtracer ISPC - Crownapache-iotdb: 500 - 100 - 200 - 100ospray: particle_volume/pathtracer/real_timesvt-av1: Preset 13 - Bosphorus 4Kstream: Triadapache-iotdb: 200 - 100 - 200 - 100specfem3d: Tomographic Modelapache-iotdb: 100 - 100 - 200 - 100svt-hevc: 7 - Bosphorus 4Kquantlib: Multi-Threadedstream: Copyonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUstress-ng: Floating Pointsqlite: 8rocksdb: Seq Filluvg266: Bosphorus 4K - Very Fastopenvino: Face Detection FP16 - CPUcassandra: Writeshpcg: 160 160 160 - 60rocksdb: Rand Fillrodinia: OpenMP Leukocyteapache-iotdb: 800 - 100 - 800 - 400npb: FT.Copenradioss: Chrysler Neon 1Mapache-iotdb: 200 - 100 - 500 - 100apache-iotdb: 800 - 100 - 500 - 400openvino: Face Detection FP16 - CPUsvt-av1: Preset 4 - Bosphorus 4Kospray-studio: 2 - 1080p - 32 - Path Tracer - CPUpgbench: 100 - 500 - Read Only - Average Latencyblosc: blosclz noshuffle - 128MBospray-studio: 3 - 4K - 32 - Path Tracer - CPUopenvino: Road Segmentation ADAS FP16 - CPUopenvkl: vklBenchmarkCPU ISPCbuild-godot: Time To Compilecompress-zstd: 19, Long Mode - Compression Speedopenvino: Road Segmentation ADAS FP16 - CPUospray-studio: 3 - 1080p - 1 - Path Tracer - CPUtensorflow: CPU - 64 - ResNet-50ospray-studio: 3 - 1080p - 16 - Path Tracer - CPUnpb: CG.Cbuild-linux-kernel: allmodconfigonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUopenvino: Person Vehicle Bike Detection FP16 - CPUopenvkl: vklBenchmarkCPU Scalarstream: Addstress-ng: Matrix Mathpgbench: 1 - 800 - Read Onlyblosc: blosclz shuffle - 64MBqmcpack: Li2_STO_aestress-ng: CPU Cacheopenvino: Person Vehicle Bike Detection FP16 - CPUblender: BMW27 - CPU-Onlycompress-zstd: 8 - Decompression Speedospray-studio: 2 - 4K - 32 - Path Tracer - CPUblosc: blosclz noshuffle - 256MBnpb: BT.Ceasywave: e2Asean Grid + BengkuluSept2007 Source - 2400apache-iotdb: 500 - 100 - 800 - 400npb: MG.Ccompress-zstd: 8, Long Mode - Decompression Speedspecfem3d: Water-layered Halfspaceapache-iotdb: 200 - 100 - 500 - 100blosc: blosclz bitshuffle - 128MBembree: Pathtracer - Asian Dragonpgbench: 100 - 500 - Read Onlyx265: Bosphorus 1080papache-iotdb: 800 - 100 - 500 - 100kvazaar: Bosphorus 1080p - Super Faststress-ng: Semaphoresrodinia: OpenMP CFD Solveruvg266: Bosphorus 1080p - Ultra Fastkvazaar: Bosphorus 4K - Slowncnn: CPU - resnet50openvino: Person Detection FP16 - CPUpgbench: 1 - 800 - Read Only - Average Latencyuvg266: Bosphorus 1080p - Mediumopenvino: Face Detection Retail FP16 - CPUopenvino: Person Detection FP16 - CPUrodinia: OpenMP Streamclusterstress-ng: CPU Stressospray-studio: 2 - 1080p - 16 - Path Tracer - CPUsqlite-speedtest: Timed Time - Size 1,000blender: Pabellon Barcelona - CPU-Onlystress-ng: Mixed Schedulerapache-iotdb: 100 - 100 - 500 - 100ospray: particle_volume/ao/real_timeapache-iotdb: 800 - 100 - 500 - 100ospray-studio: 3 - 4K - 16 - Path Tracer - CPUrocksdb: Update Randtensorflow: CPU - 16 - AlexNetlibxsmm: 32embree: Pathtracer - Asian Dragon Objapache-iotdb: 500 - 100 - 200 - 400openvino: Age Gender Recognition Retail 0013 FP16 - CPUapache-iotdb: 800 - 100 - 200 - 100npb: LU.Capache-iotdb: 100 - 100 - 500 - 100build-nodejs: Time To Compileblender: Fishy Cat - CPU-Onlyclickhouse: 100M Rows Hits Dataset, Third Runcompress-zstd: 19 - Decompression Speedstress-ng: Socket Activitylibxsmm: 128blender: Classroom - CPU-Onlyapache-iotdb: 500 - 100 - 800 - 100openvino: Weld Porosity Detection FP16 - CPUopenvino: Face Detection Retail FP16 - CPUstress-ng: x86_64 RdRandopenvino: Weld Porosity Detection FP16 - CPUopenvino: Machine Translation EN To DE FP16 - CPUsvt-hevc: 7 - Bosphorus 1080pnginx: 1000embree: Pathtracer ISPC - Asian Dragonjohn-the-ripper: WPA PSKospray-studio: 1 - 4K - 32 - Path Tracer - CPUopenvino: Vehicle Detection FP16-INT8 - CPUtensorflow: CPU - 64 - AlexNetospray-studio: 2 - 1080p - 1 - Path Tracer - CPUsvt-vp9: PSNR/SSIM Optimized - Bosphorus 4Kblosc: blosclz noshuffle - 64MBbuild-gcc: Time To Compilebuild-gem5: Time To Compilepgbench: 100 - 800 - Read Only - Average Latencypgbench: 100 - 800 - Read Onlyopenvino: Vehicle Detection FP16-INT8 - CPUopenvino: Machine Translation EN To DE FP16 - CPUospray-studio: 2 - 4K - 16 - Path Tracer - CPUx264: Bosphorus 1080psvt-vp9: Visual Quality Optimized - Bosphorus 4Kspecfem3d: Homogeneous Halfspaceopenradioss: Ford Taurus 10Mkvazaar: Bosphorus 4K - Super Fastkvazaar: Bosphorus 4K - Very Fastospray-studio: 3 - 1080p - 32 - Path Tracer - CPUstress-ng: AVX-512 VNNIospray-studio: 1 - 4K - 16 - Path Tracer - CPUstress-ng: Memory Copyingkvazaar: Bosphorus 4K - Ultra Fastcompress-zstd: 3 - Decompression Speedstress-ng: Mallocembree: Pathtracer - Crownhpcg: 104 104 104 - 60john-the-ripper: MD5apache-iotdb: 800 - 100 - 500 - 400npb: IS.Dnginx: 500openvino: Face Detection Retail FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUpgbench: 100 - 1000 - Read Only - Average Latencyopenvino: Face Detection FP16-INT8 - CPUopenvino: Face Detection Retail FP16-INT8 - CPUapache-iotdb: 500 - 100 - 500 - 400stress-ng: Cryptopgbench: 100 - 1000 - Read Onlyquantlib: Single-Threadedstress-ng: Pthreadbuild-linux-kernel: defconfigstress-ng: Wide Vector Mathstress-ng: Hashrocksdb: Read Rand Write Randtensorflow: CPU - 16 - ResNet-50compress-zstd: 12 - Compression Speeduvg266: Bosphorus 4K - Mediumospray-studio: 1 - 4K - 1 - Path Tracer - CPUlibxsmm: 64openvino: Road Segmentation ADAS FP16-INT8 - CPUospray: gravity_spheres_volume/dim_512/pathtracer/real_timecloverleaf: clover_bm16blender: Barbershop - CPU-Onlyblosc: blosclz shuffle - 32MBjohn-the-ripper: Blowfishopenvino: Road Segmentation ADAS FP16-INT8 - CPUopenvino: Face Detection FP16-INT8 - CPUstress-ng: Vector Floating Pointtensorflow: CPU - 32 - ResNet-50uvg266: Bosphorus 4K - Slowospray-studio: 3 - 4K - 1 - Path Tracer - CPUx264: Bosphorus 4Ksvt-hevc: 1 - Bosphorus 4Kstress-ng: IO_uringhadoop: Rename - 20 - 100000john-the-ripper: bcryptstress-ng: Cloningapache-iotdb: 500 - 100 - 800 - 100openssl: SHA256blosc: blosclz bitshuffle - 16MBclickhouse: 100M Rows Hits Dataset, Second Runapache-iotdb: 800 - 100 - 200 - 100stress-ng: Zlibstress-ng: Matrix 3D Mathopenvino: Person Detection FP32 - CPUmocassin: Gas HII40ospray: particle_volume/scivis/real_timerocksdb: Rand Readstress-ng: Glibc Qsort Data Sortingstress-ng: Vector Shuffleopenssl: RSA4096stress-ng: Vector Mathembree: Pathtracer ISPC - Asian Dragon Objsvt-vp9: PSNR/SSIM Optimized - Bosphorus 1080psvt-hevc: 10 - Bosphorus 1080ponednn: Recurrent Neural Network Training - f32 - CPUospray-studio: 1 - 1080p - 16 - Path Tracer - CPUopenvino: Person Detection FP32 - CPUpgbench: 1 - 1000 - Read Onlyapache-iotdb: 500 - 100 - 200 - 400stress-ng: Pollospray-studio: 1 - 1080p - 32 - Path Tracer - CPUnpb: SP.Csvt-hevc: 1 - Bosphorus 1080pstress-ng: Function Callopenvino: Weld Porosity Detection FP16-INT8 - CPUsvt-vp9: Visual Quality Optimized - Bosphorus 1080popenssl: RSA4096openvino: Weld Porosity Detection FP16-INT8 - CPUsvt-hevc: 10 - Bosphorus 4Kbuild-llvm: Ninjastress-ng: AVL Treeopenssl: ChaCha20-Poly1305ospray: gravity_spheres_volume/dim_512/ao/real_timeospray-studio: 2 - 4K - 1 - Path Tracer - CPUstress-ng: NUMAopenssl: AES-256-GCMcloverleaf: clover_bm64_shortopenssl: AES-128-GCMstress-ng: MMAPstress-ng: System V Message Passingtensorflow: CPU - 32 - AlexNetuvg266: Bosphorus 1080p - Super Faststress-ng: MEMFDopenssl: ChaCha20rocksdb: Rand Fill Syncopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16 - CPUpgbench: 1 - 1000 - Read Only - Average Latencyospray-studio: 1 - 1080p - 1 - Path Tracer - CPUoidn: RTLightmap.hdr.4096x4096 - CPU-Onlyoidn: RT.ldr_alb_nrm.3840x2160 - CPU-Onlyoidn: RT.hdr_alb_nrm.3840x2160 - CPU-Onlyuvg266: Bosphorus 4K - Super Fastkvazaar: Bosphorus 4K - Mediumcompress-zstd: 19 - Compression Speedhpcg: 192 192 192 - 60ab26.13.562424594517.41223.594239016.01174651168.4227310.322412.7724.4317.5733.3923.637.1563716.18940.58617822.282.7746475.78313.1317.17231.5841.844803177.96415.48112032.95142.972633.7830660102.97854.36651504.46718.62147.254091680834811.2109.7174.963103.83126.126125.41103.8717259.7250172143203.62272.9768.2771340889088046177364342717715164.332.371274.4793.6720.8413.61110.9559.4245021.864431.410.085751.6147.346.77114.16137957.75123.3911.7844205.7913.10813355.531279170.9888070540.9360.86282.3635080.84351.297265.4458.531063118.33570.75149.96658.5306.9978.119.86850670257.3127.4700905412068.3390.9215.778660168551819.61488.738.9816072.810.953876.836292962185.49329200.0613012555.282990.17834.310.34166.385747133131000.159963993120.1596.35367.30603146856319646265.5815169.330.8122819.63521.2488198.412.00658.62604.64564.133233251196.84142.3928.4713.51277.101232.33297.7480.5456.93166.513888.524691805400043.8962.16784.810219.18.13812665.56195776.835707925170847.871660.362703.74.35238367246.66830.702147212437175.737173.54570199.12459715030.32415476313379847105.5101118.464896.27757.39738.8920.10856735738.347.5521561311.817552382139.33710311503039511.98576.1985.36901565101579.275.056417950.3876644.517726269.33525162.6029.27172.9135718.892166214846.47520.7844308.92880.4528869855129845.95142733410740.6210.951361420.5913.6148.41755.31529124948.462863.84432.8099176128534749.271796.975.987330693504139578104.534.9525129063387.8253.34206.7343582244.0658.11167.6519.3243.2169.970.5650.245.6870.5336.49470228.21868461.39148.9720266.39138.0110.0576891381289040651074199.65162.231.52774646563123339.736171074571297.9629150108247.94258.35298.051563.818841.57537.7124.7192191074679.252108.596949614.1817.65140.75332.59130670.2932.79385776614962411.61133.621169136.798342.5906.348245.4750.62912726931032.6985.1978446231.86117.6338.87659539831662.8754.1240.83475681350185.39777788525.0166.671624.429708642.7834.887412.0751711238203.771626.47132166.253.9142663.030.79615.973062.067796808967835.6412560243545.3100016.8151.786925723.036322695.54214548319.52248.414.974552315.826.817.47547937.87474.2313381.625277447.22748.8266891.4119.6513.3536867.36.772252032.2734566252924702.2380.924165377886015922.2293.7730.364029.263732.12170.2313.2699.93418145256029737.4119219.26484191210549.0728.4954460.56584.87675.291833470.411447730145.442739015.754112124400.1726.822601.21599.183857413.115185.24298.738279.35827452858404.695534632380.81121391741680107.77132458672400297.9213970811.83118.02155.73397.05123425530680478980.551.020.69111500.511.041.0341.5719.8617.36.869.50781885742.42451.994192338.229.428064525.1582816.26918.2637.126.4422.6916.1410.323421.33990.75877328.443.5049391.70315.4414.63198.6135.915347449.26461.89310832.69156.282871.6633023477.62917.16201613.43417.38157.43421791274521.66103.3445.269109.63132.961119.161049.6417629.258268790290.88285.3471.2451287175173044348234179529614597.433.591284.6397.0820.1313.15107.357.5146465.594295.229.7735581.8143.0448.16117.53134010.32119.912.1205211.6312.74712988.430432175.7257857209.5659.3289.6435972.61342.615259.0757.1610382.518.76272.37153.39644313.976.410.08849565251.7728.07130430812332.3399.4715.4529647039761784.21518.238.231577410.75275.436178523583.94335277.713250944.773044.77694.070.346163.515847953184742.7910130826119.77606.21361.46562144531519339066.6215408.531.2722861.56528.9978078.512.18459.48613.47465.067465554199.69144.4528.8813.70678.2229.07301.9579.4257.7168.751900.335311781800044.4762.9779510351.58.03412827.7611.25705.13526.28023051748.461640.3619554.4039362986.74631.047447738065173.821175.41669457.12434640730.01512884413513355106.53100146.364274.77832.079831.7820.29956203938.77.4821762011.927351901138.97610217485539854.79581.1584.64909107921592.165.016421200.396695.817591569.86529161.3739.2171.61134719.032150414955.35517.0324339.32886.6228669373.7130745.18143715110668.3209.541352360.4513.5248.081766.91519414917.363259.23430.1399120068534538.23178676.446433843507143458056.635.1599128303288.3453.03205.5343835038.7457.776168.6119.4342.96170.890.55750.515.6570.1636.30370596.281858761.072148.220370.2137.3110.1084886933208995650821499.1616331.37654624420823230.576199762671624.5429282060246.83358.61299.371570.718923.68540124.1891813947681.992117.076977389.1617.58140.2331.31130170.0932.67155756114910211.57134.081165136.358369.4903.44244.6890.62712767411035.9785.4678690231.15117.9938.75871056331568.853.9640.71474291354114.89775538549.4766.481619.829792020.9534.790612.0424713164204.311630.76131826.313.942770.70.79416.013069.687777512568002.6912590943536.799776.4451.671923677.846336112.65214994819.48247.9154561315.226.867.46218936.22473.4113404.625234446.46747.6566994.7419.6813.32537667.46.782255307.5534614253274708.6881.034159733510015943.5293.3830.324024.183727.55170.4313.2849.94526145417830738.2119239.65484700.3210762.2528.4666461.01584.237682.641835170.351446506145.562741243.134108824381.5426.8222617.241600.31385.277418.114.99185.36298.923279.52827897831904.693394634380.65121341157650107.74132491461970297.9813973199.87118.04155.71397.08123417875470478980.551.020.69111500.511.041.0341.5719.8617.3OpenBenchmarking.org

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU-v3-v3 - Model: mobilenet-v3ab61218243026.106.86MIN: 4.41 / MAX: 1148.86MIN: 6.31 / MAX: 8.081. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUab36912153.562429.50780MIN: 2.84MIN: 2.411. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 16Scaling Factor: 100 - Clients: 800 - Mode: Read Writeab10K20K30K40K50K45945188571. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 16Scaling Factor: 100 - Clients: 800 - Mode: Read Write - Average Latencyab102030405017.4142.421. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 16Scaling Factor: 100 - Clients: 1000 - Mode: Read Write - Average Latencyab122436486023.5951.991. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 16Scaling Factor: 100 - Clients: 1000 - Mode: Read Writeab9K18K27K36K45K42390192331. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU-v2-v2 - Model: mobilenet-v2ab4812162016.018.22MIN: 4.97 / MAX: 831.66MIN: 4.84 / MAX: 276.841. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: alexnetab4812162017.009.42MIN: 7.67 / MAX: 474.49MIN: 8.15 / MAX: 11.971. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: File Status - Threads: 20 - Files: 100000ab200K400K600K800K1000K465116806452

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUab2468108.422735.15828MIN: 0.98MIN: 1.031. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUab4812162010.3216.27MIN: 1.4MIN: 1.271. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: FastestDetab369121512.778.26MIN: 6.91 / MAX: 368.66MIN: 6.18 / MAX: 156.851. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: regnety_400mab91827364524.4337.10MIN: 17.65 / MAX: 150.31MIN: 17.07 / MAX: 1371.671. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: efficientnet-b0ab61218243017.5726.44MIN: 7.73 / MAX: 723.53MIN: 7.51 / MAX: 925.051. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: yolov4-tinyab81624324033.3922.69MIN: 21.14 / MAX: 535.99MIN: 21.28 / MAX: 38.771. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: shufflenet-v2ab61218243023.6316.14MIN: 6.32 / MAX: 1872.84MIN: 6.2 / MAX: 1027.671. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUab36912157.1563710.32340MIN: 3.52MIN: 3.541. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUab51015202516.1921.34MIN: 0.77MIN: 0.771. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUab0.17070.34140.51210.68280.85350.5861780.758773MIN: 0.4MIN: 0.421. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: googlenetab71421283522.2828.44MIN: 14.13 / MAX: 139.12MIN: 13.51 / MAX: 635.931. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUab0.78861.57722.36583.15443.9432.774643.50493MIN: 1.09MIN: 1.061. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Rodinia

Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes select OpenCL, NVIDIA CUDA and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP HotSpot3Dab2040608010075.7891.701. (CXX) g++ options: -O2 -lOpenCL

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: mnasnetab4812162013.1315.44MIN: 4.46 / MAX: 1049.89MIN: 4.41 / MAX: 1373.161. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: squeezenet_ssdab4812162017.1714.63MIN: 12.44 / MAX: 95.81MIN: 12.69 / MAX: 26.041. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: vision_transformerab50100150200250231.58198.61MIN: 137.7 / MAX: 1126.49MIN: 131.13 / MAX: 589.431. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: vgg16ab102030405041.8435.91MIN: 27.23 / MAX: 449.74MIN: 26.4 / MAX: 273.561. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Stress-NG

Stress-NG is a Linux stress tool developed by Colin Ian King. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: Context Switchingab1.1M2.2M3.3M4.4M5.5M4803177.965347449.261. (CXX) g++ options: -O2 -std=gnu99 -lc

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 16Scaling Factor: 1 - Clients: 500 - Mode: Read Write - Average Latencyab100200300400500415.48461.891. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 16Scaling Factor: 1 - Clients: 500 - Mode: Read Writeab30060090012001500120310831. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: blazefaceab0.66381.32761.99142.65523.3192.952.69MIN: 2.21 / MAX: 21.26MIN: 2.31 / MAX: 8.971. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Latency, Fewer Is BetterApache IoTDB 1.2Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 800 - Client Number: 100ab306090120150142.97156.28MAX: 26023.68MAX: 26493.38

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.Dab60012001800240030002633.782871.661. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.5

Stress-NG

Stress-NG is a Linux stress tool developed by Colin Ian King. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: Glibc C String Functionsab7M14M21M28M35M30660102.9733023477.621. (CXX) g++ options: -O2 -std=gnu99 -lc

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8 - Compression Speedab2004006008001000854.3917.11. (CC) gcc options: -O3 -pthread -lz -llzma

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 16Scaling Factor: 1 - Clients: 1000 - Mode: Read Writeab1402804205607006656201. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 16Scaling Factor: 1 - Clients: 1000 - Mode: Read Write - Average Latencyab300600900120015001504.471613.431. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: mobilenetab51015202518.6217.38MIN: 13.29 / MAX: 43.75MIN: 13.38 / MAX: 76.781. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Whisper.cpp

Whisper.cpp is a port of OpenAI's Whisper model in C/C++. Whisper.cpp is developed by Georgi Gerganov for transcribing WAV audio files to text / speech recognition. Whisper.cpp supports ARM NEON, x86 AVX, and other advanced CPU features. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWhisper.cpp 1.4Model: ggml-base.en - Input: 2016 State of the Unionab306090120150147.25157.431. (CXX) g++ options: -O3 -std=c++11 -fPIC -pthread

7-Zip Compression

This is a test of 7-Zip compression/decompression with its integrated benchmark feature. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 22.01Test: Decompression Ratingab40K80K120K160K200K1680831791271. (CXX) g++ options: -lpthread -ldl -O2 -fPIC

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUab100020003000400050004811.204521.66MIN: 3780.73MIN: 4042.461. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Monte Carlo Simulations of Ionised Nebulae

Mocassin is the Monte Carlo Simulations of Ionised Nebulae. MOCASSIN is a fully 3D or 2D photoionisation and dust radiative transfer code which employs a Monte Carlo approach to the transfer of radiation through media of arbitrary geometry and density distribution. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterMonte Carlo Simulations of Ionised Nebulae 2.02.73.3Input: Dust 2D tau100.0ab20406080100109.72103.341. (F9X) gfortran options: -cpp -Jsource/ -ffree-line-length-0 -lm -std=legacy -O2 -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lz

easyWave

The easyWave software allows simulating tsunami generation and propagation in the context of early warning systems. EasyWave supports making use of OpenMP for CPU multi-threading and there are also GPU ports available but not currently incorporated as part of this test profile. The easyWave tsunami generation software is run with one of the example/reference input files for measuring the CPU execution time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BettereasyWave r34Input: e2Asean Grid + BengkuluSept2007 Source - Time: 240ab1.18552.3713.55654.7425.92754.9635.2691. (CXX) g++ options: -O3 -fopenmp

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Latency, Fewer Is BetterApache IoTDB 1.2Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 800 - Client Number: 100ab20406080100103.83109.63MAX: 24115.48MAX: 24091.01

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 8 - Input: Bosphorus 1080pab306090120150126.13132.961. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

QMCPACK

QMCPACK is a modern high-performance open-source Quantum Monte Carlo (QMC) simulation code making use of MPI for this benchmark of the H20 example code. QMCPACK is an open-source production level many-body ab initio Quantum Monte Carlo code for computing the electronic structure of atoms, molecules, and solids. QMCPACK is supported by the U.S. Department of Energy. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Execution Time - Seconds, Fewer Is BetterQMCPACK 3.17.1Input: LiH_ae_MSDab306090120150125.40119.161. (CXX) g++ options: -fopenmp -foffload=disable -finline-limit=1000 -fstrict-aliasing -funroll-all-loops -ffast-math -march=native -O3 -lm -ldl

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 16Scaling Factor: 1 - Clients: 800 - Mode: Read Write - Average Latencyab20040060080010001103.871049.641. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 16Scaling Factor: 1 - Clients: 800 - Mode: Read Writeab1603204806408007257621. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUab36912159.725019.25820MIN: 2.73MIN: 3.051. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Stress-NG

Stress-NG is a Linux stress tool developed by Colin Ian King. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: Fused Multiply-Addab15M30M45M60M75M72143203.6268790290.881. (CXX) g++ options: -O2 -std=gnu99 -lc

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Latency, Fewer Is BetterApache IoTDB 1.2Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 800 - Client Number: 400ab60120180240300272.97285.34MAX: 27781.33MAX: 27459.7

Rodinia

Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes select OpenCL, NVIDIA CUDA and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP LavaMDab163248648068.2871.251. (CXX) g++ options: -O2 -lOpenCL

OpenSSL

OpenSSL is an open-source toolkit that implements SSL (Secure Sockets Layer) and TLS (Transport Layer Security) protocols. This test profile makes use of the built-in "openssl speed" benchmarking capabilities. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgbyte/s, More Is BetterOpenSSL 3.1Algorithm: SHA512ab3000M6000M9000M12000M15000M13408890880128717517301. (CC) gcc options: -pthread -m64 -O3 -lssl -lcrypto -ldl

RocksDB

This is a benchmark of Meta/Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterRocksDB 8.0Test: Read While Writingab1000K2000K3000K4000K5000K461773644348231. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgpoint/sec, More Is BetterApache IoTDB 1.2Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 800 - Client Number: 100ab9M18M27M36M45M4342717741795296

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz shuffle - Buffer Size: 8MBab3K6K9K12K15K15164.314597.41. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

VVenC

VVenC is the Fraunhofer Versatile Video Encoder as a fast/efficient H.266/VVC encoder. The vvenc encoder makes use of SIMD Everywhere (SIMDe). The vvenc software is published under the Clear BSD License. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.9Video Input: Bosphorus 1080p - Video Preset: Fasterab81624324032.3733.591. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects

ClickHouse

ClickHouse is an open-source, high performance OLAP data management system. This test profile uses ClickHouse's standard benchmark recommendations per https://clickhouse.com/docs/en/operations/performance-test/ / https://github.com/ClickHouse/ClickBench/tree/main/clickhouse with the 100 million rows web analytics dataset. The reported value is the query processing time using the geometric mean of all separate queries performed as an aggregate. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Minute, Geo Mean, More Is BetterClickHouse 22.12.3.5100M Rows Hits Dataset, First Run / Cold Cacheab60120180240300274.47284.63MIN: 21.36 / MAX: 3333.33MIN: 16.6 / MAX: 5000

SVT-VP9

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample YUV input video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-VP9 0.3Tuning: VMAF Optimized - Input: Bosphorus 4Kab2040608010093.6797.081. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm

QMCPACK

QMCPACK is a modern high-performance open-source Quantum Monte Carlo (QMC) simulation code making use of MPI for this benchmark of the H20 example code. QMCPACK is an open-source production level many-body ab initio Quantum Monte Carlo code for computing the electronic structure of atoms, molecules, and solids. QMCPACK is supported by the U.S. Department of Energy. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Execution Time - Seconds, Fewer Is BetterQMCPACK 3.17.1Input: H4_aeab51015202520.8420.131. (CXX) g++ options: -fopenmp -foffload=disable -finline-limit=1000 -fstrict-aliasing -funroll-all-loops -ffast-math -march=native -O3 -lm -ldl

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: resnet18ab369121513.6113.15MIN: 8.3 / MAX: 117.56MIN: 8.09 / MAX: 73.511. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/ and https://github.com/OpenRadioss/ModelExchange/tree/main/Examples. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2023.09.15Model: Bumper Beamab20406080100110.95107.30

CloverLeaf

CloverLeaf is a Lagrangian-Eulerian hydrodynamics benchmark. This test profile currently makes use of CloverLeaf's OpenMP version. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterCloverLeaf 1.3Input: clover_bmab132639526559.4257.511. (F9X) gfortran options: -O3 -march=native -funroll-loops -fopenmp

Stress-NG

Stress-NG is a Linux stress tool developed by Colin Ian King. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: Forkingab10K20K30K40K50K45021.8646465.591. (CXX) g++ options: -O2 -std=gnu99 -lc

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUab100020003000400050004431.404295.22MIN: 3982.7MIN: 3730.181. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

SQLite

This is a simple benchmark of SQLite. At present this test profile just measures the time to perform a pre-defined number of insertions on an indexed database with a variable number of concurrent repetitions -- up to the maximum number of CPU threads available. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSQLite 3.41.2Threads / Copies: 2ab369121510.0809.7731. (CC) gcc options: -O2 -lz -lm

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz bitshuffle - Buffer Size: 256MBab120024003600480060005751.65581.81. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 1080p - Video Preset: Very Fastab306090120150147.30143.041. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/ and https://github.com/OpenRadioss/ModelExchange/tree/main/Examples. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2023.09.15Model: Cell Phone Drop Testab112233445546.7748.16

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Latency, Fewer Is BetterApache IoTDB 1.2Device Count: 800 - Batch Size Per Write: 100 - Sensor Count: 200 - Client Number: 400ab306090120150114.16117.53MAX: 27085.42MAX: 27796.62

nginx

This is a benchmark of the lightweight Nginx HTTP(S) web-server. This Nginx web server benchmark test profile makes use of the wrk program for facilitating the HTTP requests over a fixed period time with a configurable number of concurrent clients/connections. HTTPS with a self-signed OpenSSL certificate is used by this test for local benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is Betternginx 1.23.2Connections: 200ab30K60K90K120K150K137957.75134010.321. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Latency, Fewer Is BetterApache IoTDB 1.2Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 200 - Client Number: 100ab306090120150123.39119.90MAX: 26107.26MAX: 26035.73

High Performance Conjugate Gradient

HPCG is the High Performance Conjugate Gradient and is a new scientific benchmark from Sandia National Lans focused for super-computer testing with modern real-world workloads compared to HPCC. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFLOP/s, More Is BetterHigh Performance Conjugate Gradient 3.1X Y Z: 144 144 144 - RT: 60ab3691215SE +/- 0.14, N = 311.7812.121. (CXX) g++ options: -O3 -ffast-math -ftree-vectorize -lmpi_cxx -lmpi
OpenBenchmarking.orgGFLOP/s, More Is BetterHigh Performance Conjugate Gradient 3.1X Y Z: 144 144 144 - RT: 60ab48121620Min: 11.63 / Avg: 11.78 / Max: 12.071. (CXX) g++ options: -O3 -ffast-math -ftree-vectorize -lmpi_cxx -lmpi

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Latency, Fewer Is BetterApache IoTDB 1.2Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 500 - Client Number: 400ab50100150200250205.79211.63MAX: 27542.34MAX: 26697.43

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 4 - Input: Bosphorus 1080pab369121513.1112.751. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz bitshuffle - Buffer Size: 32MBab3K6K9K12K15K13355.512988.41. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Create - Threads: 20 - Files: 100000ab7K14K21K28K35K3127930432

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 12 - Input: Bosphorus 4Kab4080120160200170.99175.731. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

Stress-NG

Stress-NG is a Linux stress tool developed by Colin Ian King. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: Mutexab2M4M6M8M10M8070540.937857209.561. (CXX) g++ options: -O2 -std=gnu99 -lc

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries if desired for complementary metrics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 32 - Model: GoogLeNetab142842567060.8659.30

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Latency, Fewer Is BetterApache IoTDB 1.2Device Count: 800 - Batch Size Per Write: 100 - Sensor Count: 800 - Client Number: 400ab60120180240300282.36289.64MAX: 27774.74MAX: 27997.68

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: SP.Bab8K16K24K32K40K35080.8435972.611. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.5

Timed LLVM Compilation

This test times how long it takes to compile/build the LLVM compiler stack. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed LLVM Compilation 16.0Build System: Unix Makefilesab80160240320400351.30342.62

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/ and https://github.com/OpenRadioss/ModelExchange/tree/main/Examples. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2023.09.15Model: INIVOL and Fluid Structure Interaction Drop Containerab60120180240300265.44259.07

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Latency, Fewer Is BetterApache IoTDB 1.2Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 500 - Client Number: 100ab132639526558.5357.16MAX: 11473.92MAX: 11495.07

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz bitshuffle - Buffer Size: 64MBab2K4K6K8K10K10631.010382.51. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

VVenC

VVenC is the Fraunhofer Versatile Video Encoder as a fast/efficient H.266/VVC encoder. The vvenc encoder makes use of SIMD Everywhere (SIMDe). The vvenc software is published under the Clear BSD License. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.9Video Input: Bosphorus 1080p - Video Preset: Fastab51015202518.3418.761. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Latency, Fewer Is BetterApache IoTDB 1.2Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 200 - Client Number: 100ab163248648070.7572.37MAX: 24048.84MAX: 24026.47

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/ and https://github.com/OpenRadioss/ModelExchange/tree/main/Examples. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2023.09.15Model: Bird Strike on Windshieldab306090120150149.96153.39

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3, Long Mode - Compression Speedab140280420560700658.5644.01. (CC) gcc options: -O3 -pthread -lz -llzma

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.2.devModel: Handwritten English Recognition FP16-INT8 - Device: CPUab70140210280350306.99313.901. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.2.devModel: Handwritten English Recognition FP16-INT8 - Device: CPUab2040608010078.1176.40MIN: 47.76 / MAX: 391.71MIN: 46.51 / MAX: 315.261. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 16Scaling Factor: 100 - Clients: 500 - Mode: Read Write - Average Latencyab36912159.86810.0881. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 16Scaling Factor: 100 - Clients: 500 - Mode: Read Writeab11K22K33K44K55K50670495651. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

QMCPACK

QMCPACK is a modern high-performance open-source Quantum Monte Carlo (QMC) simulation code making use of MPI for this benchmark of the H20 example code. QMCPACK is an open-source production level many-body ab initio Quantum Monte Carlo code for computing the electronic structure of atoms, molecules, and solids. QMCPACK is supported by the U.S. Department of Energy. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Execution Time - Seconds, Fewer Is BetterQMCPACK 3.17.1Input: O_ae_pyscf_UHFab60120180240300257.31251.771. (CXX) g++ options: -fopenmp -foffload=disable -finline-limit=1000 -fstrict-aliasing -funroll-all-loops -ffast-math -march=native -O3 -lm -ldl

SPECFEM3D

simulates acoustic (fluid), elastic (solid), coupled acoustic/elastic, poroelastic or seismic wave propagation in any type of conforming mesh of hexahedra. This test profile currently relies on CPU-based execution for SPECFEM3D and using a variety of their built-in examples/models for benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Mount St. Helensab71421283527.4728.071. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz noshuffle - Buffer Size: 16MBab3K6K9K12K15K12068.312332.31. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

SVT-VP9

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample YUV input video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-VP9 0.3Tuning: VMAF Optimized - Input: Bosphorus 1080pab90180270360450390.92399.471. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUab4812162015.7815.45MIN: 3.97MIN: 3.921. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgpoint/sec, More Is BetterApache IoTDB 1.2Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 800 - Client Number: 100ab14M28M42M56M70M6601685564703976

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 12 - Decompression Speedab4008001200160020001819.61784.21. (CC) gcc options: -O3 -pthread -lz -llzma

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19, Long Mode - Decompression Speedab300600900120015001488.71518.21. (CC) gcc options: -O3 -pthread -lz -llzma

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Latency, Fewer Is BetterApache IoTDB 1.2Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 200 - Client Number: 100ab91827364538.9838.23MAX: 12726.18MAX: 12732.42

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz bitshuffle - Buffer Size: 8MBab3K6K9K12K15K16072.815774.01. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUab369121510.9510.75MIN: 3.88MIN: 4.011. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Latency, Fewer Is BetterApache IoTDB 1.2Device Count: 800 - Batch Size Per Write: 100 - Sensor Count: 800 - Client Number: 100ab2040608010076.8375.43MAX: 23969.69MAX: 24022.77

OpenBenchmarking.orgpoint/sec, More Is BetterApache IoTDB 1.2Device Count: 800 - Batch Size Per Write: 100 - Sensor Count: 200 - Client Number: 400ab13M26M39M52M65M6292962161785235

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/ and https://github.com/OpenRadioss/ModelExchange/tree/main/Examples. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2023.09.15Model: Rubber O-Ring Seal Installationab2040608010085.4983.94

Stress-NG

Stress-NG is a Linux stress tool developed by Colin Ian King. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: SENDFILEab70K140K210K280K350K329200.06335277.701. (CXX) g++ options: -O2 -std=gnu99 -lc

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: Pipeab3M6M9M12M15M13012555.2813250944.771. (CXX) g++ options: -O2 -std=gnu99 -lc

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3 - Compression Speedab70014002100280035002990.13044.71. (CC) gcc options: -O3 -pthread -lz -llzma

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUab2K4K6K8K10K7834.317694.07MIN: 7257.37MIN: 7140.541. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 16Scaling Factor: 1 - Clients: 500 - Mode: Read Only - Average Latencyab0.07790.15580.23370.31160.38950.3400.3461. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

QMCPACK

QMCPACK is a modern high-performance open-source Quantum Monte Carlo (QMC) simulation code making use of MPI for this benchmark of the H20 example code. QMCPACK is an open-source production level many-body ab initio Quantum Monte Carlo code for computing the electronic structure of atoms, molecules, and solids. QMCPACK is supported by the U.S. Department of Energy. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Execution Time - Seconds, Fewer Is BetterQMCPACK 3.17.1Input: FeCO6_b3lyp_gmsab4080120160200166.38163.511. (CXX) g++ options: -fopenmp -foffload=disable -finline-limit=1000 -fstrict-aliasing -funroll-all-loops -ffast-math -march=native -O3 -lm -ldl

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Open - Threads: 20 - Files: 100000ab130K260K390K520K650K574713584795

Stress-NG

Stress-NG is a Linux stress tool developed by Colin Ian King. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: Futexab700K1400K2100K2800K3500K3131000.153184742.791. (CXX) g++ options: -O2 -std=gnu99 -lc

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgpoint/sec, More Is BetterApache IoTDB 1.2Device Count: 800 - Batch Size Per Write: 100 - Sensor Count: 800 - Client Number: 100ab20M40M60M80M100M99639931101308261

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.2.devModel: Vehicle Detection FP16 - Device: CPUab51015202520.1019.77MIN: 9.28 / MAX: 63.06MIN: 10.86 / MAX: 58.461. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.2.devModel: Vehicle Detection FP16 - Device: CPUab130260390520650596.35606.211. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

Whisper.cpp

Whisper.cpp is a port of OpenAI's Whisper model in C/C++. Whisper.cpp is developed by Georgi Gerganov for transcribing WAV audio files to text / speech recognition. Whisper.cpp supports ARM NEON, x86 AVX, and other advanced CPU features. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWhisper.cpp 1.4Model: ggml-small.en - Input: 2016 State of the Unionab80160240320400367.31361.471. (CXX) g++ options: -O3 -std=c++11 -fPIC -pthread

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 16Scaling Factor: 1 - Clients: 500 - Mode: Read Onlyab300K600K900K1200K1500K146856314453151. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

7-Zip Compression

This is a test of 7-Zip compression/decompression with its integrated benchmark feature. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 22.01Test: Compression Ratingab40K80K120K160K200K1964621933901. (CXX) g++ options: -lpthread -ldl -O2 -fPIC

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 1080p - Video Preset: Mediumab153045607565.5866.621. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz shuffle - Buffer Size: 16MBab3K6K9K12K15K15169.315408.51. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

QMCPACK

QMCPACK is a modern high-performance open-source Quantum Monte Carlo (QMC) simulation code making use of MPI for this benchmark of the H20 example code. QMCPACK is an open-source production level many-body ab initio Quantum Monte Carlo code for computing the electronic structure of atoms, molecules, and solids. QMCPACK is supported by the U.S. Department of Energy. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Execution Time - Seconds, Fewer Is BetterQMCPACK 3.17.1Input: simple-H2Oab71421283530.8131.271. (CXX) g++ options: -fopenmp -foffload=disable -finline-limit=1000 -fstrict-aliasing -funroll-all-loops -ffast-math -march=native -O3 -lm -ldl

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.Cab60012001800240030002819.632861.561. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.5

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 12 - Input: Bosphorus 1080pab110220330440550521.25529.001. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz shuffle - Buffer Size: 128MBab2K4K6K8K10K8198.48078.51. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

VVenC

VVenC is the Fraunhofer Versatile Video Encoder as a fast/efficient H.266/VVC encoder. The vvenc encoder makes use of SIMD Everywhere (SIMDe). The vvenc software is published under the Clear BSD License. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.9Video Input: Bosphorus 4K - Video Preset: Fasterab369121512.0112.181. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries if desired for complementary metrics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 16 - Model: GoogLeNetab132639526558.6259.48

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 13 - Input: Bosphorus 1080pab130260390520650604.65613.471. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

SPECFEM3D

simulates acoustic (fluid), elastic (solid), coupled acoustic/elastic, poroelastic or seismic wave propagation in any type of conforming mesh of hexahedra. This test profile currently relies on CPU-based execution for SPECFEM3D and using a variety of their built-in examples/models for benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Layered Halfspaceab153045607564.1365.071. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz

Stress-NG

Stress-NG is a Linux stress tool developed by Colin Ian King. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: Atomicab4080120160200196.84199.691. (CXX) g++ options: -O2 -std=gnu99 -lc

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 1080p - Video Preset: Very Fastab306090120150142.39144.45

x265

This is a simple test of the x265 encoder run on the CPU with 1080p and 4K options for H.265 video encode performance with x265. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betterx265 3.4Video Input: Bosphorus 4Kab71421283528.4728.881. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

SQLite

This is a simple benchmark of SQLite. At present this test profile just measures the time to perform a pre-defined number of insertions on an indexed database with a variable number of concurrent repetitions -- up to the maximum number of CPU threads available. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSQLite 3.41.2Threads / Copies: 4ab4812162013.5113.711. (CC) gcc options: -O2 -lz -lm

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 8 - Input: Bosphorus 4Kab2040608010077.1078.201. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 1080p - Video Preset: Ultra Fastab50100150200250232.33229.071. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.2.devModel: Handwritten English Recognition FP16 - Device: CPUab70140210280350297.74301.951. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.2.devModel: Handwritten English Recognition FP16 - Device: CPUab2040608010080.5479.42MIN: 53.41 / MAX: 188.39MIN: 49.44 / MAX: 180.711. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries if desired for complementary metrics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 64 - Model: GoogLeNetab132639526556.9357.70

easyWave

The easyWave software allows simulating tsunami generation and propagation in the context of early warning systems. EasyWave supports making use of OpenMP for CPU multi-threading and there are also GPU ports available but not currently incorporated as part of this test profile. The easyWave tsunami generation software is run with one of the example/reference input files for measuring the CPU execution time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BettereasyWave r34Input: e2Asean Grid + BengkuluSept2007 Source - Time: 1200ab4080120160200166.51168.751. (CXX) g++ options: -O3 -fopenmp

Whisper.cpp

Whisper.cpp is a port of OpenAI's Whisper model in C/C++. Whisper.cpp is developed by Georgi Gerganov for transcribing WAV audio files to text / speech recognition. Whisper.cpp supports ARM NEON, x86 AVX, and other advanced CPU features. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWhisper.cpp 1.4Model: ggml-medium.en - Input: 2016 State of the Unionab2004006008001000888.52900.341. (CXX) g++ options: -O3 -std=c++11 -fPIC -pthread

John The Ripper

This is a benchmark of John The Ripper, which is a password cracker. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgReal C/S, More Is BetterJohn The Ripper 2023.03.14Test: HMAC-SHA512ab4M8M12M16M20M18054000178180001. (CC) gcc options: -m64 -lssl -lcrypto -fopenmp -lgmp -lm -lrt -lz -ldl -lcrypt

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 1080p - Video Preset: Slowab102030405043.8944.47

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 1080p - Video Preset: Slowab142842567062.1662.971. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8, Long Mode - Compression Speedab2004006008001000784.8795.01. (CC) gcc options: -O3 -pthread -lz -llzma

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz noshuffle - Buffer Size: 32MBab2K4K6K8K10K10219.110351.51. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

SQLite

This is a simple benchmark of SQLite. At present this test profile just measures the time to perform a pre-defined number of insertions on an indexed database with a variable number of concurrent repetitions -- up to the maximum number of CPU threads available. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSQLite 3.41.2Threads / Copies: 1ab2468108.1388.0341. (CC) gcc options: -O2 -lz -lm

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz noshuffle - Buffer Size: 8MBab3K6K9K12K15K12665.512827.71. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

libxsmm

Libxsmm is an open-source library for specialized dense and sparse matrix operations and deep learning primitives. Libxsmm supports making use of Intel AMX, AVX-512, and other modern CPU instruction set capabilities. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFLOPS/s, More Is Betterlibxsmm 2-1.17-3645M N K: 256ab130260390520650619.0611.21. (CXX) g++ options: -dynamic -Bstatic -static-libgcc -lgomp -lm -lrt -ldl -lquadmath -lstdc++ -pthread -fPIC -std=c++14 -O2 -fopenmp-simd -funroll-loops -ftree-vectorize -fdata-sections -ffunction-sections -fvisibility=hidden -msse4.2

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz shuffle - Buffer Size: 256MBab120024003600480060005776.85705.11. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

NWChem

NWChem is an open-source high performance computational chemistry package. Per NWChem's documentation, "NWChem aims to provide its users with computational chemistry tools that are scalable both in their ability to treat large scientific computational chemistry problems efficiently, and in their use of available parallel computing resources from high-performance parallel supercomputers to conventional workstation clusters." Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterNWChem 7.0.2Input: C240 Buckyballab80016002400320040003570.03526.21. (F9X) gfortran options: -lnwctask -lccsd -lmcscf -lselci -lmp2 -lmoints -lstepper -ldriver -loptim -lnwdft -lgradients -lcphf -lesp -lddscf -ldangchang -lguess -lhessian -lvib -lnwcutil -lrimp2 -lproperty -lsolvation -lnwints -lprepar -lnwmd -lnwpw -lofpw -lpaw -lpspw -lband -lnwpwlib -lcafe -lspace -lanalyze -lqhop -lpfft -ldplot -ldrdy -lvscf -lqmmm -lqmd -letrans -ltce -lbq -lmm -lcons -lperfm -ldntmc -lccca -ldimqm -lga -larmci -lpeigs -l64to32 -lopenblas -lpthread -lrt -llapack -lnwcblas -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz -lcomex -m64 -ffast-math -std=legacy -fdefault-integer-8 -finline-functions -O2

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgpoint/sec, More Is BetterApache IoTDB 1.2Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 500 - Client Number: 100ab20M40M60M80M100M7925170880230517

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 4K - Video Preset: Ultra Fastab112233445547.8748.46

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3, Long Mode - Decompression Speedab4008001200160020001660.31640.31. (CC) gcc options: -O3 -pthread -lz -llzma

Stream

This is a benchmark of Stream, the popular system memory (RAM) benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterStream 2013-01-17Type: Scaleab13K26K39K52K65K62703.761955.01. (CC) gcc options: -mcmodel=medium -O3 -march=native -fopenmp

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: gravity_spheres_volume/dim_512/scivis/real_timeab0.99091.98182.97273.96364.95454.352384.40390

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Delete - Threads: 20 - Files: 100000ab8K16K24K32K40K3672436298

VVenC

VVenC is the Fraunhofer Versatile Video Encoder as a fast/efficient H.266/VVC encoder. The vvenc encoder makes use of SIMD Everywhere (SIMDe). The vvenc software is published under the Clear BSD License. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.9Video Input: Bosphorus 4K - Video Preset: Fastab2468106.6686.7461. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects

Embree

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer ISPC - Model: Crownab71421283530.7031.05MIN: 30.14 / MAX: 31.43MIN: 30.4 / MAX: 32.01

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgpoint/sec, More Is BetterApache IoTDB 1.2Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 200 - Client Number: 100ab10M20M30M40M50M4721243747738065

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: particle_volume/pathtracer/real_timeab4080120160200175.74173.82

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 13 - Input: Bosphorus 4Kab4080120160200173.55175.421. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

Stream

This is a benchmark of Stream, the popular system memory (RAM) benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterStream 2013-01-17Type: Triadab15K30K45K60K75K70199.169457.11. (CC) gcc options: -mcmodel=medium -O3 -march=native -fopenmp

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgpoint/sec, More Is BetterApache IoTDB 1.2Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 200 - Client Number: 100ab5M10M15M20M25M2459715024346407

SPECFEM3D

simulates acoustic (fluid), elastic (solid), coupled acoustic/elastic, poroelastic or seismic wave propagation in any type of conforming mesh of hexahedra. This test profile currently relies on CPU-based execution for SPECFEM3D and using a variety of their built-in examples/models for benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Tomographic Modelab71421283530.3230.021. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgpoint/sec, More Is BetterApache IoTDB 1.2Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 200 - Client Number: 100ab3M6M9M12M15M1337984713513355

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 7 - Input: Bosphorus 4Kab20406080100105.50106.531. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

QuantLib

QuantLib is an open-source library/framework around quantitative finance for modeling, trading and risk management scenarios. QuantLib is written in C++ with Boost and its built-in benchmark used reports the QuantLib Benchmark Index benchmark score. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMFLOPS, More Is BetterQuantLib 1.32Configuration: Multi-Threadedab20K40K60K80K100K101118.4100146.31. (CXX) g++ options: -O3 -march=native -fPIE -pie

Stream

This is a benchmark of Stream, the popular system memory (RAM) benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterStream 2013-01-17Type: Copyab14K28K42K56K70K64896.264274.71. (CC) gcc options: -mcmodel=medium -O3 -march=native -fopenmp

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUab2K4K6K8K10K7757.307832.07MIN: 7171.06MIN: 7093.41. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Stress-NG

Stress-NG is a Linux stress tool developed by Colin Ian King. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: Floating Pointab2K4K6K8K10K9738.899831.781. (CXX) g++ options: -O2 -std=gnu99 -lc

SQLite

This is a simple benchmark of SQLite. At present this test profile just measures the time to perform a pre-defined number of insertions on an indexed database with a variable number of concurrent repetitions -- up to the maximum number of CPU threads available. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSQLite 3.41.2Threads / Copies: 8ab51015202520.1120.301. (CC) gcc options: -O2 -lz -lm

RocksDB

This is a benchmark of Meta/Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterRocksDB 8.0Test: Sequential Fillab120K240K360K480K600K5673575620391. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 4K - Video Preset: Very Fastab91827364538.3438.70

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.2.devModel: Face Detection FP16 - Device: CPUab2468107.557.481. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

Apache Cassandra

This is a benchmark of the Apache Cassandra NoSQL database management system making use of cassandra-stress. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 4.1.3Test: Writesab50K100K150K200K250K215613217620

High Performance Conjugate Gradient

HPCG is the High Performance Conjugate Gradient and is a new scientific benchmark from Sandia National Lans focused for super-computer testing with modern real-world workloads compared to HPCC. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFLOP/s, More Is BetterHigh Performance Conjugate Gradient 3.1X Y Z: 160 160 160 - RT: 60ab3691215SE +/- 0.13, N = 311.8211.931. (CXX) g++ options: -O3 -ffast-math -ftree-vectorize -lmpi_cxx -lmpi
OpenBenchmarking.orgGFLOP/s, More Is BetterHigh Performance Conjugate Gradient 3.1X Y Z: 160 160 160 - RT: 60ab3691215Min: 11.65 / Avg: 11.82 / Max: 12.071. (CXX) g++ options: -O3 -ffast-math -ftree-vectorize -lmpi_cxx -lmpi

RocksDB

This is a benchmark of Meta/Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterRocksDB 8.0Test: Random Fillab110K220K330K440K550K5238215190111. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

Rodinia

Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes select OpenCL, NVIDIA CUDA and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP Leukocyteab91827364539.3438.981. (CXX) g++ options: -O2 -lOpenCL

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgpoint/sec, More Is BetterApache IoTDB 1.2Device Count: 800 - Batch Size Per Write: 100 - Sensor Count: 800 - Client Number: 400ab20M40M60M80M100M103115030102174855

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: FT.Cab9K18K27K36K45K39511.9839854.791. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.5

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/ and https://github.com/OpenRadioss/ModelExchange/tree/main/Examples. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2023.09.15Model: Chrysler Neon 1Mab130260390520650576.19581.15

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Latency, Fewer Is BetterApache IoTDB 1.2Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 500 - Client Number: 100ab2040608010085.3684.64MAX: 24058.19MAX: 24104.59

OpenBenchmarking.orgpoint/sec, More Is BetterApache IoTDB 1.2Device Count: 800 - Batch Size Per Write: 100 - Sensor Count: 500 - Client Number: 400ab20M40M60M80M100M9015651090910792

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.2.devModel: Face Detection FP16 - Device: CPUab300600900120015001579.271592.16MIN: 1364.94 / MAX: 2085.13MIN: 950.29 / MAX: 2096.551. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 4 - Input: Bosphorus 4Kab1.13762.27523.41284.55045.6885.0565.0161. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

OSPRay Studio

Intel OSPRay Studio is an open-source, interactive visualization and ray-tracing software package. OSPRay Studio makes use of Intel OSPRay, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.13Camera: 2 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPUab9K18K27K36K45K4179542120

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 16Scaling Factor: 100 - Clients: 500 - Mode: Read Only - Average Latencyab0.08780.17560.26340.35120.4390.3870.3901. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz noshuffle - Buffer Size: 128MBab140028004200560070006644.56695.81. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

OSPRay Studio

Intel OSPRay Studio is an open-source, interactive visualization and ray-tracing software package. OSPRay Studio makes use of Intel OSPRay, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.13Camera: 3 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPUab40K80K120K160K200K177262175915

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.2.devModel: Road Segmentation ADAS FP16 - Device: CPUab163248648069.3369.86MIN: 19.1 / MAX: 176.01MIN: 20.19 / MAX: 169.391. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

OpenVKL

OpenVKL is the Intel Open Volume Kernel Library that offers high-performance volume computation kernels and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 2.0.0Benchmark: vklBenchmarkCPU ISPCab110220330440550525529MIN: 47 / MAX: 6753MIN: 47 / MAX: 6787

Timed Godot Game Engine Compilation

This test times how long it takes to compile the Godot Game Engine. Godot is a popular, open-source, cross-platform 2D/3D game engine and is built using the SCons build system and targeting the X11 platform. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Godot Game Engine Compilation 4.0Time To Compileab4080120160200162.60161.37

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19, Long Mode - Compression Speedab36912159.279.201. (CC) gcc options: -O3 -pthread -lz -llzma

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.2.devModel: Road Segmentation ADAS FP16 - Device: CPUab4080120160200172.90171.611. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

OSPRay Studio

Intel OSPRay Studio is an open-source, interactive visualization and ray-tracing software package. OSPRay Studio makes use of Intel OSPRay, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.13Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPUab3006009001200150013571347

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries if desired for complementary metrics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 64 - Model: ResNet-50ab51015202518.8919.03

OSPRay Studio

Intel OSPRay Studio is an open-source, interactive visualization and ray-tracing software package. OSPRay Studio makes use of Intel OSPRay, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.13Camera: 3 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPUab5K10K15K20K25K2166221504

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: CG.Cab3K6K9K12K15K14846.4714955.351. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.5

Timed Linux Kernel Compilation

This test times how long it takes to build the Linux kernel in a default configuration (defconfig) for the architecture being tested or alternatively an allmodconfig for building all possible kernel modules for the build. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 6.1Build: allmodconfigab110220330440550520.78517.03

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUab90018002700360045004308.924339.32MIN: 3624.55MIN: 3850.271. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.2.devModel: Person Vehicle Bike Detection FP16 - Device: CPUab2004006008001000880.45886.621. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

OpenVKL

OpenVKL is the Intel Open Volume Kernel Library that offers high-performance volume computation kernels and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 2.0.0Benchmark: vklBenchmarkCPU Scalarab60120180240300288286MIN: 22 / MAX: 5321MIN: 22 / MAX: 5354

Stream

This is a benchmark of Stream, the popular system memory (RAM) benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterStream 2013-01-17Type: Addab15K30K45K60K75K69855.069373.71. (CC) gcc options: -mcmodel=medium -O3 -march=native -fopenmp

Stress-NG

Stress-NG is a Linux stress tool developed by Colin Ian King. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: Matrix Mathab30K60K90K120K150K129845.95130745.181. (CXX) g++ options: -O2 -std=gnu99 -lc

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 16Scaling Factor: 1 - Clients: 800 - Mode: Read Onlyab300K600K900K1200K1500K142733414371511. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz shuffle - Buffer Size: 64MBab2K4K6K8K10K10740.610668.31. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

QMCPACK

QMCPACK is a modern high-performance open-source Quantum Monte Carlo (QMC) simulation code making use of MPI for this benchmark of the H20 example code. QMCPACK is an open-source production level many-body ab initio Quantum Monte Carlo code for computing the electronic structure of atoms, molecules, and solids. QMCPACK is supported by the U.S. Department of Energy. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Execution Time - Seconds, Fewer Is BetterQMCPACK 3.17.1Input: Li2_STO_aeab50100150200250210.95209.541. (CXX) g++ options: -fopenmp -foffload=disable -finline-limit=1000 -fstrict-aliasing -funroll-all-loops -ffast-math -march=native -O3 -lm -ldl

Stress-NG

Stress-NG is a Linux stress tool developed by Colin Ian King. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: CPU Cacheab300K600K900K1200K1500K1361420.591352360.451. (CXX) g++ options: -O2 -std=gnu99 -lc

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.2.devModel: Person Vehicle Bike Detection FP16 - Device: CPUab369121513.6113.52MIN: 7.31 / MAX: 43.72MIN: 7.33 / MAX: 43.311. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.6Blend File: BMW27 - Compute: CPU-Onlyab112233445548.4048.08

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8 - Decompression Speedab4008001200160020001755.31766.91. (CC) gcc options: -O3 -pthread -lz -llzma

OSPRay Studio

Intel OSPRay Studio is an open-source, interactive visualization and ray-tracing software package. OSPRay Studio makes use of Intel OSPRay, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.13Camera: 2 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPUab30K60K90K120K150K152912151941

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz noshuffle - Buffer Size: 256MBab110022003300440055004948.44917.31. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: BT.Cab14K28K42K56K70K62863.8463259.231. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.5

easyWave

The easyWave software allows simulating tsunami generation and propagation in the context of early warning systems. EasyWave supports making use of OpenMP for CPU multi-threading and there are also GPU ports available but not currently incorporated as part of this test profile. The easyWave tsunami generation software is run with one of the example/reference input files for measuring the CPU execution time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BettereasyWave r34Input: e2Asean Grid + BengkuluSept2007 Source - Time: 2400ab90180270360450432.81430.141. (CXX) g++ options: -O3 -fopenmp

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgpoint/sec, More Is BetterApache IoTDB 1.2Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 800 - Client Number: 400ab20M40M60M80M100M9176128591200685

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: MG.Cab7K14K21K28K35K34749.2734538.231. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.5

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8, Long Mode - Decompression Speedab4008001200160020001796.91786.01. (CC) gcc options: -O3 -pthread -lz -llzma

SPECFEM3D

simulates acoustic (fluid), elastic (solid), coupled acoustic/elastic, poroelastic or seismic wave propagation in any type of conforming mesh of hexahedra. This test profile currently relies on CPU-based execution for SPECFEM3D and using a variety of their built-in examples/models for benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Water-layered Halfspaceab2040608010075.9976.451. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgpoint/sec, More Is BetterApache IoTDB 1.2Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 500 - Client Number: 100ab11M22M33M44M55M5041395750714345

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C that focuses on compression of binary data. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.11Test: blosclz bitshuffle - Buffer Size: 128MBab2K4K6K8K10K8104.58056.61. (CC) gcc options: -std=gnu99 -O3 -ldl -lrt -lm

Embree

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer - Model: Asian Dragonab81624324034.9535.16MIN: 34.41 / MAX: 35.71MIN: 34.63 / MAX: 35.8

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 16Scaling Factor: 100 - Clients: 500 - Mode: Read Onlyab300K600K900K1200K1500K129063312830321. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

x265

This is a simple test of the x265 encoder run on the CPU with 1080p and 4K options for H.265 video encode performance with x265. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betterx265 3.4Video Input: Bosphorus 1080pab2040608010087.8288.341. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

Apache IoTDB

Apache IotDB is a time series database and this benchmark is facilitated using the IoT Benchmaark [https://github.com/thulab/iot-benchmark/]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Latency, Fewer Is BetterApache IoTDB 1.2Device Count: 800 - Batch Size Per Write: 100 - Sensor Count: 500 - Client Number: 100ab122436486053.3453.03MAX: 24055.64MAX: 24012.98

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 1080p - Video Preset: Super Fastab50100150200250206.73205.531. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

Stress-NG

Stress-NG is a Linux stress tool developed by Colin Ian King. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: Semaphoresab9M18M27M36M45M43582244.0643835038.741. (CXX) g++ options: -O2 -std=gnu99 -lc

Rodinia

Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes select OpenCL, NVIDIA CUDA and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP CFD Solverab132639526558.1157.781. (CXX) g++ options: -O2 -lOpenCL

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 1080p - Video Preset: Ultra Fastab4080120160200167.65168.61

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 4K - Video Preset: Slowab51015202519.3219.431. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: resnet50ab102030405043.2042.96MIN: 16.39 / MAX: 675.32MIN: 17.49 / MAX: 512.271. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.2.devModel: Person Detection FP16 - Device: CPUab4080120160200169.97170.89MIN: 91.34 / MAX: 276.04MIN: 87.67 / MAX: 290.451. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 16Scaling Factor: 1 - Clients: 800 - Mode: Read Only - Average Latencyab0.1260.2520.3780.5040.630.5600.5571. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 1080p - Video Preset: Mediumab112233445550.2450.51

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.2.devModel: Face Detection Retail FP16 - Device: CPUab1.2782.5563.8345.1126.395.685.65MIN: 2.85 / MAX: 98.82MIN: 2.89 / MAX: 40.951. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.2.devModel: Person Detection FP16 - Device: CPUab163248648070.5370.161. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie

Rodinia

Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes select OpenCL, NVIDIA CUDA and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP Streamclusterab81624324036.4936.301. (CXX) g++ options: -O2 -lOpenCL

Stress-NG

Stress-NG is a Linux stress tool developed by Colin Ian King. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.16.04Test: CPU Stressab15K30K45K60K75K70228.2070596.281. (CXX) g++ options: -O2 -std=gnu99 -lc

OSPRay Studio

Intel OSPRay Studio is an open-source, interactive visualization and ray-tracing software package. OSPRay Studio makes use of Intel OSPRay, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.13Camera: 2 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPUab4K8K12K16K20K1868418587

SQLite Speedtest

This is a benchmark of SQLite's speedtest1 benchmark program with an increased problem size of 1,000. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Better