12400 march

Intel Core i5-12400 testing with a MSI PRO Z690-A WIFI DDR4(MS-7D25) v1.0 (Dasharo coreboot+UEFI v1.0.0 BIOS) and MSI Intel ADL-S GT1 14GB on Ubuntu 22.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2303066-NE-12400MARC37
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts
Allow Limiting Results To Certain Suite(s)

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Toggle/Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
a
March 03 2023
  2 Days, 44 Minutes
b
March 05 2023
  11 Hours, 15 Minutes
c
March 05 2023
  11 Hours, 18 Minutes
Invert Behavior (Only Show Selected Data)
  23 Hours, 46 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


12400 marchOpenBenchmarking.orgPhoronix Test SuiteIntel Core i5-12400 @ 5.60GHz (6 Cores / 12 Threads)MSI PRO Z690-A WIFI DDR4(MS-7D25) v1.0 (Dasharo coreboot+UEFI v1.0.0 BIOS)Intel Device 7aa716GBWestern Digital WD_BLACK SN750 SE 500GBMSI Intel ADL-S GT1 14GB (1450MHz)Realtek ALC897DELL S2409WIntel I225-V + Intel Device 7af0Ubuntu 22.045.15.0-47-generic (x86_64)GNOME Shell 42.2X Server + Wayland4.6 Mesa 22.0.11.2.204GCC 11.2.0GCC 11.3.0ext41920x1080ProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilersFile-SystemScreen Resolution12400 March BenchmarksSystem Logs- Transparent Huge Pages: madvise- a: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - b: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-xKiWfi/gcc-11-11.3.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-xKiWfi/gcc-11-11.3.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - c: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-xKiWfi/gcc-11-11.3.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-xKiWfi/gcc-11-11.3.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x1f - Thermald 2.4.9 - Python 3.10.6- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected

abcResult OverviewPhoronix Test Suite100%105%110%114%119%RocksDBFLAC Audio EncodingPostgreSQLKeyDBEnCodecTimed FFmpeg CompilationVP9 libvpx EncodingNeural Magic DeepSparseSVT-AV1nekRSJPEG XL Decoding libjxlAOM AV1miniBUDEClickHouseStress-NGuvg266Cpuminer-OptJPEG XL libjxlKvazaarCockroachDBEmbreeXmrignginxFFmpegVVenCTimed Linux Kernel CompilationOpenVINOOpenVKLBRL-CADStargate Digital Audio WorkstationOpenRadiossoneDNNBlenderGROMACSTensorFlowOpenEMSlibavif avifencScikit-LearnZstd CompressionUnvanquished

12400 marchblender: Barbershop - CPU-Onlynekrs: TurboPipe Periodicbuild-linux-kernel: allmodconfigclickhouse: 100M Rows Hits Dataset, Third Runclickhouse: 100M Rows Hits Dataset, Second Runclickhouse: 100M Rows Hits Dataset, First Run / Cold Cachetensorflow: CPU - 512 - GoogLeNetopenradioss: INIVOL and Fluid Structure Interaction Drop Containertensorflow: CPU - 256 - GoogLeNetblender: Pabellon Barcelona - CPU-Onlyopenems: pyEMS Couplerbrl-cad: VGR Performance Metricminibude: OpenMP - BM2minibude: OpenMP - BM2blender: Classroom - CPU-Onlytensorflow: CPU - 512 - AlexNetpgbench: 100 - 1000 - Read Write - Average Latencypgbench: 100 - 1000 - Read Writepgbench: 100 - 250 - Read Write - Average Latencypgbench: 100 - 250 - Read Writepgbench: 100 - 500 - Read Write - Average Latencypgbench: 100 - 500 - Read Writepgbench: 100 - 100 - Read Write - Average Latencypgbench: 100 - 100 - Read Writeopenvkl: vklBenchmark ISPCkeydb: HMSET - 100openradioss: Bird Strike on Windshieldpgbench: 1 - 800 - Read Write - Average Latencypgbench: 1 - 800 - Read Writeopenvkl: vklBenchmark Scalarpgbench: 1 - 500 - Read Write - Average Latencypgbench: 1 - 500 - Read Writejpegxl: JPEG - 100jpegxl: PNG - 100xmrig: Monero - 1Mpgbench: 100 - 800 - Read Write - Average Latencypgbench: 100 - 800 - Read Writetensorflow: CPU - 256 - AlexNetffmpeg: libx264 - Uploadffmpeg: libx264 - Uploadcockroach: MoVR - 128blender: Fishy Cat - CPU-Onlyopenradioss: Rubber O-Ring Seal Installationbuild-linux-kernel: defconfigaom-av1: Speed 0 Two-Pass - Bosphorus 4Kcompress-zstd: 19 - Decompression Speedcompress-zstd: 19 - Compression Speedvvenc: Bosphorus 4K - Fastgromacs: MPI CPU - water_GMX50_barerocksdb: Rand Fillkeydb: LPUSH - 100avifenc: 0ffmpeg: libx265 - Video On Demandffmpeg: libx265 - Video On Demandopenvino: Vehicle Detection FP16 - CPUopenvino: Vehicle Detection FP16 - CPUffmpeg: libx265 - Uploadffmpeg: libx265 - Uploadffmpeg: libx265 - Platformffmpeg: libx265 - Platformxmrig: Wownero - 1Mopenradioss: Bumper Beamtensorflow: CPU - 64 - GoogLeNetblender: BMW27 - CPU-Onlyffmpeg: libx264 - Video On Demandffmpeg: libx264 - Video On Demandffmpeg: libx264 - Platformffmpeg: libx264 - Platformpgbench: 1 - 500 - Read Only - Average Latencypgbench: 1 - 500 - Read Onlyencodec: 24 kbpspgbench: 100 - 1000 - Read Only - Average Latencypgbench: 100 - 1000 - Read Onlyjpegxl: JPEG - 80pgbench: 100 - 800 - Read Only - Average Latencypgbench: 100 - 800 - Read Onlykeydb: SADD - 50pgbench: 100 - 500 - Read Only - Average Latencypgbench: 100 - 500 - Read Onlyjpegxl: PNG - 80aom-av1: Speed 4 Two-Pass - Bosphorus 4Kpgbench: 100 - 250 - Read Only - Average Latencypgbench: 100 - 250 - Read Onlyminibude: OpenMP - BM1minibude: OpenMP - BM1openradioss: Cell Phone Drop Testuvg266: Bosphorus 4K - Slowkeydb: LPOP - 50pgbench: 100 - 100 - Read Only - Average Latencypgbench: 100 - 100 - Read Onlyscikit-learn: Sparse Rand Projections, 100 Iterationskeydb: GET - 50pgbench: 1 - 1000 - Read Write - Average Latencypgbench: 1 - 1000 - Read Writeaom-av1: Speed 0 Two-Pass - Bosphorus 1080pstargate: 192000 - 512deepsparse: NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Streamencodec: 3 kbpskeydb: SET - 50pgbench: 1 - 250 - Read Write - Average Latencypgbench: 1 - 250 - Read Writepgbench: 1 - 100 - Read Write - Average Latencypgbench: 1 - 100 - Read Writepgbench: 1 - 800 - Read Only - Average Latencypgbench: 1 - 800 - Read Onlypgbench: 1 - 1000 - Read Only - Average Latencypgbench: 1 - 1000 - Read Onlypgbench: 1 - 100 - Read Only - Average Latencypgbench: 1 - 100 - Read Onlypgbench: 1 - 250 - Read Only - Average Latencypgbench: 1 - 250 - Read Onlyuvg266: Bosphorus 4K - Mediumkeydb: LPOP - 100stargate: 192000 - 1024jpegxl: JPEG - 90jpegxl: PNG - 90deepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Streamvpxenc: Speed 0 - Bosphorus 4Kcockroach: KV, 10% Reads - 512cockroach: KV, 10% Reads - 1024cockroach: KV, 95% Reads - 1024cockroach: KV, 50% Reads - 1024cockroach: KV, 60% Reads - 1024cockroach: KV, 95% Reads - 512cockroach: KV, 50% Reads - 512cockroach: KV, 60% Reads - 512cockroach: KV, 60% Reads - 256cockroach: KV, 50% Reads - 256cockroach: KV, 95% Reads - 256cockroach: KV, 10% Reads - 256cockroach: KV, 50% Reads - 128cockroach: KV, 10% Reads - 128cockroach: KV, 60% Reads - 128cockroach: KV, 95% Reads - 128keydb: HMSET - 50stress-ng: Atomicstress-ng: Context Switchingstress-ng: MMAPstress-ng: Socket Activitystress-ng: Futexstress-ng: IO_uringdeepsparse: CV Detection, YOLOv5s COCO - Synchronous Single-Streamdeepsparse: CV Detection, YOLOv5s COCO - Synchronous Single-Streamvvenc: Bosphorus 4K - Fastercockroach: MoVR - 256cockroach: MoVR - 512cockroach: MoVR - 1024rocksdb: Rand Readencode-flac: WAV To FLACkvazaar: Bosphorus 4K - Slowkvazaar: Bosphorus 4K - Mediumscikit-learn: MNIST Datasettensorflow: CPU - 32 - GoogLeNetnginx: 200nginx: 1000nginx: 500nginx: 100nginx: 20onednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUavifenc: 2stargate: 96000 - 512rocksdb: Update Randonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUstress-ng: MEMFDstress-ng: CPU Cacheembree: Pathtracer - Asian Dragon Objonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUstargate: 96000 - 1024tensorflow: CPU - 64 - AlexNetaom-av1: Speed 6 Two-Pass - Bosphorus 4Kembree: Pathtracer ISPC - Asian Dragon Objembree: Pathtracer - Crownvvenc: Bosphorus 1080p - Fastsvt-av1: Preset 4 - Bosphorus 4Kunvanquished: 1920 x 1080 - Ultraonednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUopenvino: Person Detection FP16 - CPUopenvino: Person Detection FP16 - CPUopenvino: Person Detection FP32 - CPUopenvino: Person Detection FP32 - CPUembree: Pathtracer ISPC - Crownopenvino: Face Detection FP16 - CPUopenvino: Face Detection FP16 - CPUembree: Pathtracer - Asian Dragonstress-ng: Glibc C String Functionskeydb: SET - 100stargate: 44100 - 512compress-zstd: 8 - Decompression Speedcompress-zstd: 8 - Compression Speedrocksdb: Rand Fill Syncstargate: 480000 - 512openvino: Face Detection FP16-INT8 - CPUopenvino: Face Detection FP16-INT8 - CPUcompress-zstd: 3, Long Mode - Decompression Speedcompress-zstd: 3, Long Mode - Compression Speedcompress-zstd: 3 - Decompression Speedcompress-zstd: 3 - Compression Speedcompress-zstd: 12 - Decompression Speedcompress-zstd: 12 - Compression Speedopenvino: Machine Translation EN To DE FP16 - CPUopenvino: Machine Translation EN To DE FP16 - CPUcompress-zstd: 8, Long Mode - Decompression Speedcompress-zstd: 8, Long Mode - Compression Speedcompress-zstd: 19, Long Mode - Decompression Speedcompress-zstd: 19, Long Mode - Compression Speedffmpeg: libx265 - Liveffmpeg: libx265 - Livestargate: 44100 - 1024build-ffmpeg: Time To Compileopenvino: Person Vehicle Bike Detection FP16 - CPUopenvino: Person Vehicle Bike Detection FP16 - CPUopenvino: Weld Porosity Detection FP16 - CPUopenvino: Weld Porosity Detection FP16 - CPUopenvino: Vehicle Detection FP16-INT8 - CPUopenvino: Vehicle Detection FP16-INT8 - CPUrocksdb: Read While Writingopenvino: Weld Porosity Detection FP16-INT8 - CPUopenvino: Weld Porosity Detection FP16-INT8 - CPUstargate: 480000 - 1024openvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16 - CPUopenvino: Age Gender Recognition Retail 0013 FP16 - CPUrocksdb: Read Rand Write Randembree: Pathtracer ISPC - Asian Dragonkeydb: LPUSH - 50jpegxl-decode: 1aom-av1: Speed 4 Two-Pass - Bosphorus 1080pdeepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Streamdeepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Streamdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Streamdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Streamonednn: IP Shapes 1D - f32 - CPUdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Streamonednn: IP Shapes 1D - u8s8f32 - CPUvpxenc: Speed 5 - Bosphorus 4Kvpxenc: Speed 0 - Bosphorus 1080pkeydb: SADD - 100tensorflow: CPU - 16 - GoogLeNetdeepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Asynchronous Multi-Streamdeepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Asynchronous Multi-Streamdeepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Synchronous Single-Streamdeepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Synchronous Single-Streamtensorflow: CPU - 32 - AlexNetunvanquished: 1920 x 1080 - Highdeepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Streamdeepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Streamkeydb: GET - 100uvg266: Bosphorus 4K - Very Fastdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Synchronous Single-Streamdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Synchronous Single-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Synchronous Single-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Synchronous Single-Streamdeepsparse: NLP Text Classification, BERT base uncased SST2 - Synchronous Single-Streamdeepsparse: NLP Text Classification, BERT base uncased SST2 - Synchronous Single-Streamuvg266: Bosphorus 4K - Super Fastopenems: openEMS MSL_NotchFilterdeepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Synchronous Single-Streamdeepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Synchronous Single-Streamcpuminer-opt: LBC, LBRY Creditsdeepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Synchronous Single-Streamdeepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Synchronous Single-Streamffmpeg: libx264 - Liveffmpeg: libx264 - Liveencodec: 6 kbpsdeepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Streamdeepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Streamaom-av1: Speed 6 Realtime - Bosphorus 4Kencodec: 1.5 kbpsdeepsparse: NLP Text Classification, DistilBERT mnli - Synchronous Single-Streamdeepsparse: NLP Text Classification, DistilBERT mnli - Synchronous Single-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Synchronous Single-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Synchronous Single-Streamstress-ng: Pthreadkvazaar: Bosphorus 4K - Very Fastcpuminer-opt: Garlicoinuvg266: Bosphorus 4K - Ultra Fastunvanquished: 1920 x 1080 - Mediumaom-av1: Speed 10 Realtime - Bosphorus 4Konednn: IP Shapes 3D - u8s8f32 - CPUtensorflow: CPU - 16 - AlexNetstress-ng: Zlibstress-ng: System V Message Passingstress-ng: Hashstress-ng: Pollstress-ng: Memory Copyingstress-ng: NUMAcpuminer-opt: Myriad-Groestlstress-ng: Mallocstress-ng: Matrix Mathstress-ng: Function Callstress-ng: Vector Mathcpuminer-opt: Skeincoincpuminer-opt: Blake-2 Sstress-ng: Mutexstress-ng: CPU Stressstress-ng: SENDFILEstress-ng: x86_64 RdRandstress-ng: Glibc Qsort Data Sortingstress-ng: Cryptostress-ng: Semaphoresstress-ng: Forkingcpuminer-opt: Triple SHA-256, Onecoincpuminer-opt: Magicpuminer-opt: Deepcoincpuminer-opt: x25xcpuminer-opt: scryptcpuminer-opt: Ringcoinvvenc: Bosphorus 1080p - Fastercpuminer-opt: Quad SHA-256, Pyritekvazaar: Bosphorus 4K - Super Fastuvg266: Bosphorus 1080p - Slowjpegxl-decode: Allscikit-learn: TSNE MNIST Datasetuvg266: Bosphorus 1080p - Mediumsvt-av1: Preset 8 - Bosphorus 4Kaom-av1: Speed 6 Two-Pass - Bosphorus 1080pavifenc: 10, Losslesssvt-av1: Preset 4 - Bosphorus 1080pkvazaar: Bosphorus 4K - Ultra Fastvpxenc: Speed 5 - Bosphorus 1080pkvazaar: Bosphorus 1080p - Slowkvazaar: Bosphorus 1080p - Mediumaom-av1: Speed 6 Realtime - Bosphorus 1080paom-av1: Speed 8 Realtime - Bosphorus 4Kavifenc: 6, Losslessaom-av1: Speed 10 Realtime - Bosphorus 1080ponednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUaom-av1: Speed 8 Realtime - Bosphorus 1080paom-av1: Speed 9 Realtime - Bosphorus 1080paom-av1: Speed 9 Realtime - Bosphorus 4Konednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUuvg266: Bosphorus 1080p - Very Fastonednn: IP Shapes 3D - f32 - CPUavifenc: 6uvg266: Bosphorus 1080p - Super Fastsvt-av1: Preset 12 - Bosphorus 4Krocksdb: Seq Fillkvazaar: Bosphorus 1080p - Very Fastsvt-av1: Preset 8 - Bosphorus 1080puvg266: Bosphorus 1080p - Ultra Fastsvt-av1: Preset 13 - Bosphorus 4Kkvazaar: Bosphorus 1080p - Super Fastonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUkvazaar: Bosphorus 1080p - Ultra Fastsvt-av1: Preset 12 - Bosphorus 1080psvt-av1: Preset 13 - Bosphorus 1080ponednn: IP Shapes 1D - bf16bf16bf16 - CPUabc1913.71298446555561605.965133.73130.84112.0844.65862.7643.37602.1739.721329319.747243.662493.25117.88246.313407060.8024148113.076447224.5524114120361525.21378.022891.941277681561.2693210.680.692901.8188.7064252107.9113.50187.09176.4242.51236.87117.6530.151516.312.62.6960.852846457976117.59195.06843.25175.1327.97143.2121.20119.1344.40170.615359.8187.4941.05172.0148.80155.2751.90145.951.39335916544.5584.1412415407.933.0312639501345962.361.5733178728.055.120.6343945269.780244.489141.004.111403220.970.185541410142.7291481042.423797.3142630.501.272677129.735123.140138.6631249454.75681.418367258.4273872.6063070183.4582892050.1626177060.5414620714.581301994.001.3986637.797.9465.116746.10095.7625625.026050.840249.930338.732151.843281.732781.235029.434782.830043.849613.517827.616123.09270.520255.752512.6511218.21146.111897936.0774.708703.541579624.881965.7532.733230.57316.065181.6177.4177.44161411017.3246.346.5390.14640.1761096.5841663.1751922.4867993.5262961.454341.924329.654334.9788.2461.9309606299442437.9658.5783.608.71022438.212406.132.07166988.609.119.57068.48757.9772.185100.68.697481.526653384.141.173438.181.159.33662257.171.759.69581220179.931069422.992.5785991732.4202.613512.624567552.797.211619.4869.31592.61567.81758.977.1184.3721.681727.9198.41444.47.00120.1742.022.70512261.46814.44276.7831.87188.1711.47348.4915185158.41712.702.7437840.956331.011.065560.27154665511.05881323318.9750.0513.31371.27078.0798530.83515.64014.09853530.17895.64341.25187112.7312.181148270.2938.8942.337070.8341125.27077.982177.95170.8136.273922.01961283175.5813.25182.98235.4650184.65465.415451.147219.551413.9898.8849.640120.14211843820.326949.1789239.8621.0538.51691.194532.892755.3336.91526.085138.336640.468074.098815.567064.2327128698.0116.451836.5916.99248.258.762.5046260.81717.358216877.601548492.99823762.132301.69244.13112375226456.9945892.7949787.6241733.97478532822074254720.2317781.86177365.8481909.45101.4214276.391189258.3254276.3478157331.496392.10344.30110.871560.0920.1554825020.3821.53217.3326.91424.1127.44131.696.5717.73727.8332.7532.5334.46147.6945.8713.461175.803.597191.20219139.13188.8157.998.938852.0002062.3310.37229.17567.0988.426150279274.7981.45380.0195.56593.4816.592518.2750123.26348.114398.8611918.11306997000001581.695130.95129.64114.0144.7855.1243.6601.7339.861320589.819245.471491.93118.22273.835365268.0723673103.714482123.6424230120360477.28375.592861.265280691536.8553250.690.692943.7217.4623679108.6913.43188.02161.5242.4233.55117.8430.161512.312.32.7420.8518663811010733.94193.62644.31170.9531.54126.7821.11119.62656910744.36170.765362.5187.5340.8169.9751.90145.9551.56146.921.32937630243.5284.1592404268.062.9032755401920122.881.5813162888.195.220.61340792110.18254.507141.914.221585238.250.182549838142.6031373626.384001.5372500.541.272057126.342823.74338.3751253729.88702.475356260.5813842.6543014673.5172843520.166242380.5294721714.641336505.381.3870997.898.0964.231746.69935.7224787.326158.440176.230358.132001.243180.132901.134456.334070.830151.34902417639.615644.69148.719591.851610.2518247.5145.011941869.4650.269385.831723742.381723.1332.288330.96325.893173.6177.3177.44315211219.2976.526.6490.81940.3261654.7742015.7251890.4568835.7864372.014289.214234.914270.7587.0411.9184436245402403.1941.6589.58.71682398.622394.332.04893188.459.159.62968.71898.0852.257100.58.372391.415383363.571.183394.571.169.53952239.721.779.78571273002.921146000.382.5705971733.2203.612392.593242541.217.391618.7866.51596.21588.61761.777.2182.5321.871725.8199.81451.77.06119.6542.212.69743859.35314.31279.331.53190.211.49347.8315311578.32719.912.7228460.946344.821.045661.39156281211.14321252473.6250.8313.31371.96948.0647522.7595.73863.9326526.95955.68881.6407612.0211.571137811.7539.250.842158.9625123.96798.065877.8171.1146.717120.43721382055.3813.57185.08765.4027183.88985.437952.802518.937114.398.7258.906816.9731920020.387449.0317241.9520.8741.17588.782833.777854.9341.02926.566837.635138.429878.043615.371765.0336130862.3817.221908.9917.49248.158.662.4527861.24717.628225295.791549710.05823948.192295.12246.58110905241054.3246046.4849807.0541733.47484103029804273551.9817780.76176973.2481907.97102.5314261.231190334.2752494.8177950327.096428.9367.28111.651583.5820.8594831021.0422.34226.3126.70825.0527.78231.546.4387.80929.1132.3333.2734.52145.5846.213.724177.463.593641.19925139.06171.5358.458.133732.0115464.110.5579.27767.4669.565151136174.7381.39979.8196.5893.5216.612418.4063121.82348.913399.0811921.09306969000001591.758137.07134.75112.4444.62854.2143.37604.2739.951323689.809245.226492.05118.5221.36451841.632600582.83603619.6025102121343680.41376.262990.051268681495.6313340.690.72952.3124.5086425108.8313.45187.76161.5242.75235.09118.7950.161515.112.42.7250.8491102439967773.12195.0744.55170.0326.79149.2121.24118.9044.53170.115388.2188.0640.83169.9851.73146.4552.00145.671.36836548545.3094.3022324458.093.0252645001225640.51.5733178638.225.170.60841146810.172254.306141.14.214794650.184544459142.5011690102.753967.3652520.541.267444125.312223.938140.9031422030.12697.892358252.0033972.5523135033.5212839730.1576380110.5374656194.6213036451.3927467.968.0963.143647.50355.732570926199.140255.330286.731560.143303.932572.934933.334301.129377.949764.217585.4157339128.119566.651080.3506124.09162.651906264.5749.739379.731689762.382856.8932.281830.975.857173.8173.9176.14297599519.6096.456.690.83840.0861396.9342016.8851737.2668707.4964314.454303.624253.174291.9886.4341.9233526641574000.3452.9972.18.71242410.162388.512.0526689.089.219.58378.63018.022.25100.78.44041.40933402.111.153413.871.169.4822256.41.759.75441288214.511049912.882.5753631734.3203.631932.60295549.427.281620.2866.11598.41583.61762.277.3185.4821.541728.6200.41447.27.02118.3642.672.68744959.45314.37278.231.68189.3211.54346.3415325468.35717.72.7171160.946357.311.045687.23156820111.11211317870.3850.4813.31373.23628.0373519.65585.77293.95246520.94945.75860.95994212.2512.691193659.2538.5341.865871.6306123.95478.066677.44170.7135.427822.14911273333.8813.5185.10375.4023183.69245.443750.816219.677314.2498.7748.520720.60621971020.263949.3305235.8021.4240.1290.158533.263655.3138.70625.251839.595739.010176.883115.382564.9894132268.7217.131895.3217.41248.258.532.4353260.69716.148192747.721548182.16823174.052285.43245111005200128.8245978.8449784.1941779.98486702952504254389.2917782.21178115.6481907.39102.7314271.881188758.8454204.6577940326.586368.98368.41111.661586.8620.5924832020.9222.14225.6326.86424.9627.50432.356.4397.7528.8532.3833.2234.5169.9646.0813.727168.563.607051.18228140.47201.2958.389.33672.010564.1310.43429.26867.3668.862146156675.0481.02179.4195.09393.316.589318.2347123.33347.219397.306OpenBenchmarking.org

Blender

Blender is an open-source 3D creation and modeling software project. This test is of Blender's Cycles performance with various sample files. GPU computing via NVIDIA OptiX and NVIDIA CUDA is currently supported as well as HIP for AMD Radeon GPUs and Intel oneAPI for Intel Graphics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.4Blend File: Barbershop - Compute: CPU-Onlyabc400800120016002000SE +/- 2.57, N = 31913.711918.111921.09

nekRS

nekRS is an open-source Navier Stokes solver based on the spectral element method. NekRS supports both CPU and GPU/accelerator support though this test profile is currently configured for CPU execution. NekRS is part of Nek5000 of the Mathematics and Computer Science MCS at Argonne National Laboratory. This nekRS benchmark is primarily relevant to large core count HPC servers and otherwise may be very time consuming. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFLOP/s, More Is BetternekRS 22.0Input: TurboPipe Periodicbca7000M14000M21000M28000M35000MSE +/- 533119138.01, N = 93069970000030696900000298446555561. (CXX) g++ options: -fopenmp -O2 -march=native -mtune=native -ftree-vectorize -lmpi_cxx -lmpi

Timed Linux Kernel Compilation

This test times how long it takes to build the Linux kernel in a default configuration (defconfig) for the architecture being tested or alternatively an allmodconfig for building all possible kernel modules for the build. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 6.1Build: allmodconfigbca30060090012001500SE +/- 2.13, N = 31581.701591.761605.97

ClickHouse

ClickHouse is an open-source, high performance OLAP data management system. This test profile uses ClickHouse's standard benchmark recommendations per https://clickhouse.com/docs/en/operations/performance-test/ / https://github.com/ClickHouse/ClickBench/tree/main/clickhouse with the 100 million rows web analytics dataset. The reported value is the query processing time using the geometric mean of all separate queries performed as an aggregate. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Minute, Geo Mean, More Is BetterClickHouse 22.12.3.5100M Rows Hits Dataset, Third Runcab306090120150SE +/- 0.37, N = 9137.07133.73130.95MIN: 7.03 / MAX: 7500MIN: 6.54 / MAX: 8571.43MIN: 7.08 / MAX: 6000

OpenBenchmarking.orgQueries Per Minute, Geo Mean, More Is BetterClickHouse 22.12.3.5100M Rows Hits Dataset, Second Runcab306090120150SE +/- 0.66, N = 9134.75130.84129.64MIN: 7.04 / MAX: 8571.43MIN: 6.51 / MAX: 8571.43MIN: 7.07 / MAX: 6000

OpenBenchmarking.orgQueries Per Minute, Geo Mean, More Is BetterClickHouse 22.12.3.5100M Rows Hits Dataset, First Run / Cold Cachebca306090120150SE +/- 1.04, N = 9114.01112.44112.08MIN: 6.89 / MAX: 5000MIN: 6.91 / MAX: 5454.55MIN: 6.36 / MAX: 6666.67

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 512 - Model: GoogLeNetbac1020304050SE +/- 0.04, N = 344.7044.6544.62

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: INIVOL and Fluid Structure Interaction Drop Containercba2004006008001000SE +/- 5.89, N = 3854.21855.12862.76

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 256 - Model: GoogLeNetbca1020304050SE +/- 0.04, N = 343.6043.3743.37

Blender

Blender is an open-source 3D creation and modeling software project. This test is of Blender's Cycles performance with various sample files. GPU computing via NVIDIA OptiX and NVIDIA CUDA is currently supported as well as HIP for AMD Radeon GPUs and Intel oneAPI for Intel Graphics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.4Blend File: Pabellon Barcelona - Compute: CPU-Onlybac130260390520650SE +/- 0.41, N = 3601.73602.17604.27

OpenEMS

OpenEMS is a free and open electromagnetic field solver using the FDTD method. This test profile runs OpenEMS and pyEMS benchmark demos. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMCells/s, More Is BetterOpenEMS 0.0.35-86Test: pyEMS Couplercba918273645SE +/- 0.06, N = 339.9539.8639.721. (CXX) g++ options: -O3 -rdynamic -ltinyxml -lcrypto -lcurl -lpthread -lsz -lz -ldl -lm -lexpat

BRL-CAD

BRL-CAD is a cross-platform, open-source solid modeling system with built-in benchmark mode. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgVGR Performance Metric, More Is BetterBRL-CAD 7.34VGR Performance Metricacb30K60K90K120K150K1329311323681320581. (CXX) g++ options: -std=c++14 -pipe -fvisibility=hidden -fno-strict-aliasing -fno-common -fexceptions -ftemplate-depth-128 -m64 -ggdb3 -O3 -fipa-pta -fstrength-reduce -finline-functions -flto -ltcl8.6 -lregex_brl -lz_brl -lnetpbm -ldl -lm -ltk8.6

miniBUDE

MiniBUDE is a mini application for the the core computation of the Bristol University Docking Engine (BUDE). This test profile currently makes use of the OpenMP implementation of miniBUDE for CPU benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBillion Interactions/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM2bca3691215SE +/- 0.016, N = 39.8199.8099.7471. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

OpenBenchmarking.orgGFInst/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM2bca50100150200250SE +/- 0.41, N = 3245.47245.23243.661. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

Blender

Blender is an open-source 3D creation and modeling software project. This test is of Blender's Cycles performance with various sample files. GPU computing via NVIDIA OptiX and NVIDIA CUDA is currently supported as well as HIP for AMD Radeon GPUs and Intel oneAPI for Intel Graphics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.4Blend File: Classroom - Compute: CPU-Onlybca110220330440550SE +/- 0.90, N = 3491.93492.05493.25

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 512 - Model: AlexNetcba306090120150SE +/- 0.47, N = 3118.50118.22117.88

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 1000 - Mode: Read Write - Average Latencycab60120180240300SE +/- 3.72, N = 12221.36246.31273.841. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 1000 - Mode: Read Writecab10002000300040005000SE +/- 59.66, N = 124518407036521. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 250 - Mode: Read Write - Average Latencycab1530456075SE +/- 1.63, N = 1241.6360.8068.071. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 250 - Mode: Read Writecab13002600390052006500SE +/- 122.04, N = 126005414836731. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 500 - Mode: Read Write - Average Latencycba306090120150SE +/- 3.53, N = 1282.83103.71113.081. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 500 - Mode: Read Writecba13002600390052006500SE +/- 146.05, N = 126036482144721. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 100 - Mode: Read Write - Average Latencycba612182430SE +/- 0.75, N = 1219.6023.6424.551. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 100 - Mode: Read Writecba11002200330044005500SE +/- 122.37, N = 125102423041141. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenVKL

OpenVKL is the Intel Open Volume Kernel Library that offers high-performance volume computation kernels and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 1.3.1Benchmark: vklBenchmark ISPCcba306090120150SE +/- 0.67, N = 3121120120MIN: 15 / MAX: 1739MIN: 15 / MAX: 1749MIN: 15 / MAX: 1750

KeyDB

A benchmark of KeyDB as an open-source, multi-threaded fork of the Redis server now developed by Snapchat. The developers self-describe the KeyDB in-memory data structure store as a "faster drop in alternative to Redis". Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterKeyDB 6.3.2Test: HMSET - Parallel Connections: 100abc80K160K240K320K400KSE +/- 3780.33, N = 12361525.21360477.28343680.411. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: Bird Strike on Windshieldbca80160240320400SE +/- 1.35, N = 3375.59376.26378.02

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 800 - Mode: Read Write - Average Latencybac6001200180024003000SE +/- 26.55, N = 122861.272891.942990.051. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 800 - Mode: Read Writebac60120180240300SE +/- 2.63, N = 122802772681. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenVKL

OpenVKL is the Intel Open Volume Kernel Library that offers high-performance volume computation kernels and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 1.3.1Benchmark: vklBenchmark Scalarbca1530456075SE +/- 0.33, N = 3696868MIN: 7 / MAX: 1201MIN: 7 / MAX: 1202MIN: 7 / MAX: 1212

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 500 - Mode: Read Write - Average Latencycba30060090012001500SE +/- 15.49, N = 121495.631536.861561.271. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 500 - Mode: Read Writecba70140210280350SE +/- 3.20, N = 123343253211. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

JPEG XL libjxl

The JPEG XL Image Coding System is designed to provide next-generation JPEG image capabilities with JPEG XL offering better image quality and compression over legacy JPEG. This test profile is currently focused on the multi-threaded JPEG XL image encode performance using the reference libjxl library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterJPEG XL libjxl 0.7Input: JPEG - Quality: 100cba0.15530.31060.46590.62120.7765SE +/- 0.00, N = 30.690.690.681. (CXX) g++ options: -fno-rtti -funwind-tables -O3 -O2 -fPIE -pie -lm -latomic

OpenBenchmarking.orgMP/s, More Is BetterJPEG XL libjxl 0.7Input: PNG - Quality: 100cba0.15750.3150.47250.630.7875SE +/- 0.00, N = 30.700.690.691. (CXX) g++ options: -fno-rtti -funwind-tables -O3 -O2 -fPIE -pie -lm -latomic

Xmrig

Xmrig is an open-source cross-platform CPU/GPU miner for RandomX, KawPow, CryptoNight and AstroBWT. This test profile is setup to measure the Xmlrig CPU mining performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgH/s, More Is BetterXmrig 6.18.1Variant: Monero - Hash Count: 1Mcba6001200180024003000SE +/- 10.62, N = 32952.32943.72901.81. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 800 - Mode: Read Write - Average Latencycab50100150200250SE +/- 3.52, N = 9124.51188.71217.461. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 800 - Mode: Read Writecab14002800420056007000SE +/- 83.04, N = 96425425236791. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 256 - Model: AlexNetcba20406080100SE +/- 0.09, N = 3108.83108.69107.91

FFmpeg

This is a benchmark of the FFmpeg multimedia framework. The FFmpeg test profile is making use of a modified version of vbench from Columbia University's Architecture and Design Lab (ARCADE) [http://arcade.cs.columbia.edu/vbench/] that is a benchmark for video-as-a-service workloads. The test profile offers the options of a range of vbench scenarios based on freely distributable video content and offers the options of using the x264 or x265 video encoders for transcoding. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterFFmpeg 6.0Encoder: libx264 - Scenario: Uploadacb3691215SE +/- 0.02, N = 313.5013.4513.431. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

OpenBenchmarking.orgSeconds, Fewer Is BetterFFmpeg 6.0Encoder: libx264 - Scenario: Uploadacb4080120160200SE +/- 0.26, N = 3187.09187.76188.021. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

CockroachDB

CockroachDB is a cloud-native, distributed SQL database for data intensive applications. This test profile uses a server-less CockroachDB configuration to test various Coackroach workloads on the local host with a single node. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: MoVR - Concurrency: 128acb4080120160200SE +/- 1.32, N = 11176.4161.5161.5

Blender

Blender is an open-source 3D creation and modeling software project. This test is of Blender's Cycles performance with various sample files. GPU computing via NVIDIA OptiX and NVIDIA CUDA is currently supported as well as HIP for AMD Radeon GPUs and Intel oneAPI for Intel Graphics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.4Blend File: Fishy Cat - Compute: CPU-Onlybac50100150200250SE +/- 0.25, N = 3242.40242.51242.75

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: Rubber O-Ring Seal Installationbca50100150200250SE +/- 1.23, N = 3233.55235.09236.87

Timed Linux Kernel Compilation

This test times how long it takes to build the Linux kernel in a default configuration (defconfig) for the architecture being tested or alternatively an allmodconfig for building all possible kernel modules for the build. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 6.1Build: defconfigabc306090120150SE +/- 1.00, N = 8117.65117.84118.80

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google as the AV1 Codec Library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 4Kcba0.0360.0720.1080.1440.18SE +/- 0.00, N = 70.160.160.151. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19 - Decompression Speedacb30060090012001500SE +/- 0.83, N = 151516.31515.11512.31. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19 - Compression Speedacb3691215SE +/- 0.10, N = 1512.612.412.31. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

VVenC

VVenC is the Fraunhofer Versatile Video Encoder as a fast/efficient H.266/VVC encoder. The vvenc encoder makes use of SIMD Everywhere (SIMDe). The vvenc software is published under The Clear BSD License. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.7Video Input: Bosphorus 4K - Video Preset: Fastbca0.6171.2341.8512.4683.085SE +/- 0.016, N = 32.7422.7252.6961. (CXX) g++ options: -O3 -flto -fno-fat-lto-objects -flto=auto

GROMACS

The GROMACS (GROningen MAchine for Chemical Simulations) molecular dynamics package testing with the water_GMX50 data. This test profile allows selecting between CPU and GPU-based GROMACS builds. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNs Per Day, More Is BetterGROMACS 2023Implementation: MPI CPU - Input: water_GMX50_bareabc0.19170.38340.57510.76680.9585SE +/- 0.003, N = 30.8520.8510.8491. (CXX) g++ options: -O3

RocksDB

This is a benchmark of Meta/Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterRocksDB 7.9.2Test: Random Fillcba200K400K600K800K1000KSE +/- 20633.51, N = 1511024398663818464571. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

KeyDB

A benchmark of KeyDB as an open-source, multi-threaded fork of the Redis server now developed by Snapchat. The developers self-describe the KeyDB in-memory data structure store as a "faster drop in alternative to Redis". Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterKeyDB 6.3.2Test: LPUSH - Parallel Connections: 100bac200K400K600K800K1000KSE +/- 10077.27, N = 151010733.94976117.59967773.121. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.11Encoder Speed: 0bac4080120160200SE +/- 0.94, N = 3193.63195.07195.071. (CXX) g++ options: -O3 -fPIC -lm

FFmpeg

This is a benchmark of the FFmpeg multimedia framework. The FFmpeg test profile is making use of a modified version of vbench from Columbia University's Architecture and Design Lab (ARCADE) [http://arcade.cs.columbia.edu/vbench/] that is a benchmark for video-as-a-service workloads. The test profile offers the options of a range of vbench scenarios based on freely distributable video content and offers the options of using the x264 or x265 video encoders for transcoding. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterFFmpeg 6.0Encoder: libx265 - Scenario: Video On Demandcba1020304050SE +/- 0.04, N = 344.5544.3143.251. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

OpenBenchmarking.orgSeconds, Fewer Is BetterFFmpeg 6.0Encoder: libx265 - Scenario: Video On Demandcba4080120160200SE +/- 0.16, N = 3170.03170.95175.131. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Vehicle Detection FP16 - Device: CPUcab714212835SE +/- 0.37, N = 1426.7927.9731.54MIN: 21.9 / MAX: 44.22MIN: 19.8 / MAX: 50.33MIN: 19.58 / MAX: 49.191. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Vehicle Detection FP16 - Device: CPUcab306090120150SE +/- 1.79, N = 14149.21143.21126.781. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

FFmpeg

This is a benchmark of the FFmpeg multimedia framework. The FFmpeg test profile is making use of a modified version of vbench from Columbia University's Architecture and Design Lab (ARCADE) [http://arcade.cs.columbia.edu/vbench/] that is a benchmark for video-as-a-service workloads. The test profile offers the options of a range of vbench scenarios based on freely distributable video content and offers the options of using the x264 or x265 video encoders for transcoding. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterFFmpeg 6.0Encoder: libx265 - Scenario: Uploadcab510152025SE +/- 0.03, N = 321.2421.2021.111. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

OpenBenchmarking.orgSeconds, Fewer Is BetterFFmpeg 6.0Encoder: libx265 - Scenario: Uploadcab306090120150SE +/- 0.17, N = 3118.90119.13119.631. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

OpenBenchmarking.orgFPS, More Is BetterFFmpeg 6.0Encoder: libx265 - Scenario: Platformcab1020304050SE +/- 0.03, N = 344.5344.4044.361. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

OpenBenchmarking.orgSeconds, Fewer Is BetterFFmpeg 6.0Encoder: libx265 - Scenario: Platformcab4080120160200SE +/- 0.12, N = 3170.11170.61170.761. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

Xmrig

Xmrig is an open-source cross-platform CPU/GPU miner for RandomX, KawPow, CryptoNight and AstroBWT. This test profile is setup to measure the Xmlrig CPU mining performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgH/s, More Is BetterXmrig 6.18.1Variant: Wownero - Hash Count: 1Mcba12002400360048006000SE +/- 13.58, N = 35388.25362.55359.81. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: Bumper Beamabc4080120160200SE +/- 0.32, N = 3187.49187.53188.06

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 64 - Model: GoogLeNetacb918273645SE +/- 0.09, N = 341.0540.8340.80

Blender

Blender is an open-source 3D creation and modeling software project. This test is of Blender's Cycles performance with various sample files. GPU computing via NVIDIA OptiX and NVIDIA CUDA is currently supported as well as HIP for AMD Radeon GPUs and Intel oneAPI for Intel Graphics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.4Blend File: BMW27 - Compute: CPU-Onlybca4080120160200SE +/- 1.22, N = 3169.97169.98172.01

FFmpeg

This is a benchmark of the FFmpeg multimedia framework. The FFmpeg test profile is making use of a modified version of vbench from Columbia University's Architecture and Design Lab (ARCADE) [http://arcade.cs.columbia.edu/vbench/] that is a benchmark for video-as-a-service workloads. The test profile offers the options of a range of vbench scenarios based on freely distributable video content and offers the options of using the x264 or x265 video encoders for transcoding. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterFFmpeg 6.0Encoder: libx264 - Scenario: Video On Demandbca1224364860SE +/- 0.66, N = 351.9051.7348.801. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

OpenBenchmarking.orgSeconds, Fewer Is BetterFFmpeg 6.0Encoder: libx264 - Scenario: Video On Demandbca306090120150SE +/- 2.11, N = 3145.95146.45155.271. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

OpenBenchmarking.orgFPS, More Is BetterFFmpeg 6.0Encoder: libx264 - Scenario: Platformcab1224364860SE +/- 0.27, N = 352.0051.9051.561. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

OpenBenchmarking.orgSeconds, Fewer Is BetterFFmpeg 6.0Encoder: libx264 - Scenario: Platformcab306090120150SE +/- 0.75, N = 3145.67145.95146.921. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 500 - Mode: Read Only - Average Latencybca0.31340.62680.94021.25361.567SE +/- 0.016, N = 41.3291.3681.3931. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 500 - Mode: Read Onlybca80K160K240K320K400KSE +/- 3941.00, N = 43763023654853591651. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

EnCodec

EnCodec is a Facebook/Meta developed AI means of compressing audio files using High Fidelity Neural Audio Compression. EnCodec is designed to provide codec compression at 6 kbps using their novel AI-powered compression technique. The test profile uses a lengthy JFK speech as the audio input for benchmarking and the performance measurement is measuring the time to encode the EnCodec file from WAV. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterEnCodec 0.1.1Target Bandwidth: 24 kbpsbac1020304050SE +/- 0.52, N = 1543.5344.5645.31

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 1000 - Mode: Read Only - Average Latencyabc0.9681.9362.9043.8724.84SE +/- 0.034, N = 34.1414.1594.3021. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 1000 - Mode: Read Onlyabc50K100K150K200K250KSE +/- 1961.46, N = 32415402404262324451. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

JPEG XL libjxl

The JPEG XL Image Coding System is designed to provide next-generation JPEG image capabilities with JPEG XL offering better image quality and compression over legacy JPEG. This test profile is currently focused on the multi-threaded JPEG XL image encode performance using the reference libjxl library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterJPEG XL libjxl 0.7Input: JPEG - Quality: 80cba246810SE +/- 0.01, N = 38.098.067.931. (CXX) g++ options: -fno-rtti -funwind-tables -O3 -O2 -fPIE -pie -lm -latomic

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 800 - Mode: Read Only - Average Latencybca0.6821.3642.0462.7283.41SE +/- 0.027, N = 32.9033.0253.0311. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 800 - Mode: Read Onlybca60K120K180K240K300KSE +/- 2360.08, N = 32755402645002639501. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

KeyDB

A benchmark of KeyDB as an open-source, multi-threaded fork of the Redis server now developed by Snapchat. The developers self-describe the KeyDB in-memory data structure store as a "faster drop in alternative to Redis". Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterKeyDB 6.3.2Test: SADD - Parallel Connections: 50bac400K800K1200K1600K2000KSE +/- 44518.65, N = 151920122.881345962.361225640.501. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 500 - Mode: Read Only - Average Latencyacb0.35570.71141.06711.42281.7785SE +/- 0.012, N = 31.5731.5731.5811. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 500 - Mode: Read Onlyacb70K140K210K280K350KSE +/- 2326.54, N = 33178723178633162881. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

JPEG XL libjxl

The JPEG XL Image Coding System is designed to provide next-generation JPEG image capabilities with JPEG XL offering better image quality and compression over legacy JPEG. This test profile is currently focused on the multi-threaded JPEG XL image encode performance using the reference libjxl library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterJPEG XL libjxl 0.7Input: PNG - Quality: 80cba246810SE +/- 0.00, N = 38.228.198.051. (CXX) g++ options: -fno-rtti -funwind-tables -O3 -O2 -fPIE -pie -lm -latomic

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google as the AV1 Codec Library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 4Kbca1.17452.3493.52354.6985.8725SE +/- 0.00, N = 35.225.175.121. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 250 - Mode: Read Only - Average Latencycba0.14270.28540.42810.57080.7135SE +/- 0.005, N = 30.6080.6130.6341. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 250 - Mode: Read Onlycba90K180K270K360K450KSE +/- 2783.61, N = 34114684079213945261. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

miniBUDE

MiniBUDE is a mini application for the the core computation of the Bristol University Docking Engine (BUDE). This test profile currently makes use of the OpenMP implementation of miniBUDE for CPU benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgBillion Interactions/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM1bca3691215SE +/- 0.087, N = 710.18010.1729.7801. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

OpenBenchmarking.orgGFInst/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM1bca60120180240300SE +/- 2.17, N = 7254.51254.31244.491. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2022.10.13Model: Cell Phone Drop Testacb306090120150SE +/- 0.15, N = 3141.00141.10141.91

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 4K - Video Preset: Slowbca0.94951.8992.84853.7984.7475SE +/- 0.05, N = 34.224.204.11

KeyDB

A benchmark of KeyDB as an open-source, multi-threaded fork of the Redis server now developed by Snapchat. The developers self-describe the KeyDB in-memory data structure store as a "faster drop in alternative to Redis". Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterKeyDB 6.3.2Test: LPOP - Parallel Connections: 50bca300K600K900K1200K1500KSE +/- 22923.50, N = 151585238.251479465.001403220.971. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 100 - Mode: Read Only - Average Latencybca0.04160.08320.12480.16640.208SE +/- 0.001, N = 30.1820.1840.1851. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 100 - Clients: 100 - Mode: Read Onlybca120K240K360K480K600KSE +/- 1426.41, N = 35498385444595414101. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

Scikit-Learn

Scikit-learn is a Python module for machine learning Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.1.3Benchmark: Sparse Random Projections, 100 Iterationscba306090120150SE +/- 0.14, N = 3142.50142.60142.73

KeyDB

A benchmark of KeyDB as an open-source, multi-threaded fork of the Redis server now developed by Snapchat. The developers self-describe the KeyDB in-memory data structure store as a "faster drop in alternative to Redis". Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterKeyDB 6.3.2Test: GET - Parallel Connections: 50cab400K800K1200K1600K2000KSE +/- 63048.39, N = 151690102.751481042.421373626.381. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 1000 - Mode: Read Write - Average Latencyacb9001800270036004500SE +/- 35.63, N = 33797.313967.374001.541. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 1000 - Mode: Read Writeacb60120180240300SE +/- 2.49, N = 32632522501. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google as the AV1 Codec Library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 1080pcba0.12150.2430.36450.4860.6075SE +/- 0.00, N = 150.540.540.501. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

Stargate Digital Audio Workstation

Stargate is an open-source, cross-platform digital audio workstation (DAW) software package with "a unique and carefully curated experience" with scalability from old systems up through modern multi-core systems. Stargate is GPLv3 licensed and makes use of Qt5 (PyQt5) for its user-interface. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRender Ratio, More Is BetterStargate Digital Audio Workstation 22.11.5Sample Rate: 192000 - Buffer Size: 512abc0.28640.57280.85921.14561.432SE +/- 0.003098, N = 31.2726771.2720571.2674441. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions

Neural Magic DeepSparse

This is a benchmark of Neural Magic's DeepSparse using its built-in deepsparse.benchmark utility and various models from their SparseZoo (https://sparsezoo.neuralmagic.com/). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Streamcba306090120150SE +/- 1.18, N = 12125.31126.34129.74

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Streamcba612182430SE +/- 0.20, N = 1223.9423.7423.14

EnCodec

EnCodec is a Facebook/Meta developed AI means of compressing audio files using High Fidelity Neural Audio Compression. EnCodec is designed to provide codec compression at 6 kbps using their novel AI-powered compression technique. The test profile uses a lengthy JFK speech as the audio input for benchmarking and the performance measurement is measuring the time to encode the EnCodec file from WAV. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterEnCodec 0.1.1Target Bandwidth: 3 kbpsbac918273645SE +/- 0.60, N = 1538.3838.6640.90

KeyDB

A benchmark of KeyDB as an open-source, multi-threaded fork of the Redis server now developed by Snapchat. The developers self-describe the KeyDB in-memory data structure store as a "faster drop in alternative to Redis". Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterKeyDB 6.3.2Test: SET - Parallel Connections: 50cba300K600K900K1200K1500KSE +/- 40517.71, N = 121422030.121253729.881249454.751. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD

PostgreSQL

This is a benchmark of PostgreSQL using the integrated pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 250 - Mode: Read Write - Average Latencyacb150300450600750SE +/- 4.83, N = 3681.42697.89702.481. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 250 - Mode: Read Writeacb80160240320400SE +/- 2.61, N = 33673583561. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 100 - Mode: Read Write - Average Latencycab60120180240300SE +/- 1.27, N = 3252.00258.43260.581. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 100 - Mode: Read Writecab90180270360450SE +/- 1.90, N = 33973873841. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 800 - Mode: Read Only - Average Latencycab0.59721.19441.79162.38882.986SE +/- 0.012, N = 32.5522.6062.6541. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 800 - Mode: Read Onlycab70K140K210K280K350KSE +/- 1379.07, N = 33135033070183014671. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 1000 - Mode: Read Only - Average Latencyabc0.79221.58442.37663.16883.961SE +/- 0.034, N = 33.4583.5173.5211. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 1000 - Mode: Read Onlyabc60K120K180K240K300KSE +/- 2823.16, N = 32892052843522839731. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 100 - Mode: Read Only - Average Latencycba0.03650.0730.10950.1460.1825SE +/- 0.001, N = 30.1570.1600.1621. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 100 - Mode: Read Onlycba140K280K420K560K700KSE +/- 4488.67, N = 36380116242386177061. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 250 - Mode: Read Only - Average Latencybca0.12170.24340.36510.48680.6085SE +/- 0.008, N = 30.5290.5370.5411. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL 15Scaling Factor: 1 - Clients: 250 - Mode: Read Onlybca100K200K300K400K500KSE +/- 6603.78, N = 34721714656194620711. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 4K - Video Preset: Mediumbca1.0442.0883.1324.1765.22SE +/- 0.01, N = 34.644.624.58

KeyDB

A benchmark of KeyDB as an open-source, multi-threaded fork of the Redis server now developed by Snapchat. The developers self-describe the KeyDB in-memory data structure store as a "faster drop in alternative to Redis". Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterKeyDB 6.3.2Test: LPOP - Parallel Connections: 100bca300K600K900K1200K1500KSE +/- 37181.44, N = 121336505.381303645.001301994.001. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD

Stargate Digital Audio Workstation

Stargate is an open-source, cross-platform digital audio workstation (DAW) software package with "a unique and carefully curated experience" with scalability from old systems up through modern multi-core systems. Stargate is GPLv3 licensed and makes use of Qt5 (PyQt5) for its user-interface. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRender Ratio, More Is BetterStargate Digital Audio Workstation 22.11.5Sample Rate: 192000 - Buffer Size: 1024acb0.31470.62940.94411.25881.5735SE +/- 0.005117, N = 31.3986631.3927461.3870991. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions

JPEG XL libjxl

The JPEG XL Image Coding System is designed to provide next-generation JPEG image capabilities with JPEG XL offering better image quality and compression over legacy JPEG. This test profile is currently focused on the multi-threaded JPEG XL image encode performance using the reference libjxl library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterJPEG XL libjxl 0.7Input: JPEG - Quality: 90cba246810SE +/- 0.01, N = 37.967.897.791. (CXX) g++ options: -fno-rtti -funwind-tables -O3 -O2 -fPIE -pie -lm -latomic

OpenBenchmarking.orgMP/s, More Is BetterJPEG XL libjxl 0.7Input: PNG - Quality: 90cba246810SE +/- 0.01, N = 38.098.097.941. (CXX) g++ options: -fno-rtti -funwind-tables -O3 -O2 -fPIE -pie -lm -latomic

Neural Magic DeepSparse

This is a benchmark of Neural Magic's DeepSparse using its built-in deepsparse.benchmark utility and various models from their SparseZoo (https://sparsezoo.neuralmagic.com/). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Streamcba1530456075SE +/- 0.58, N = 1263.1464.2365.12

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Streamcba1122334455SE +/- 0.38, N = 1247.5046.7046.10

VP9 libvpx Encoding

This is a standard video encoding performance test of Google's libvpx library and the vpxenc command for the VP9 video format. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVP9 libvpx Encoding 1.13Speed: Speed 0 - Input: Bosphorus 4Kacb1.2962.5923.8885.1846.48SE +/- 0.02, N = 35.765.735.721. (CXX) g++ options: -m64 -lm -lpthread -O3 -fPIC -U_FORTIFY_SOURCE -std=gnu++11

CockroachDB

CockroachDB is a cloud-native, distributed SQL database for data intensive applications. This test profile uses a server-less CockroachDB configuration to test various Coackroach workloads on the local host with a single node. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 10% Reads - Concurrency: 512cab6K12K18K24K30KSE +/- 297.75, N = 325709.025625.024787.3

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 10% Reads - Concurrency: 1024cba6K12K18K24K30KSE +/- 369.95, N = 326199.126158.426050.8

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 95% Reads - Concurrency: 1024cab9K18K27K36K45KSE +/- 52.10, N = 340255.340249.940176.2

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 50% Reads - Concurrency: 1024bac7K14K21K28K35KSE +/- 65.26, N = 330358.130338.730286.7

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 60% Reads - Concurrency: 1024abc7K14K21K28K35KSE +/- 47.74, N = 332151.832001.231560.1

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 95% Reads - Concurrency: 512cab9K18K27K36K45KSE +/- 83.04, N = 343303.943281.743180.1

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 50% Reads - Concurrency: 512bac7K14K21K28K35KSE +/- 338.62, N = 332901.132781.232572.9

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 60% Reads - Concurrency: 512acb8K16K24K32K40KSE +/- 342.69, N = 335029.434933.334456.3

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 60% Reads - Concurrency: 256acb7K14K21K28K35KSE +/- 405.43, N = 334782.834301.134070.8

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 50% Reads - Concurrency: 256bac6K12K18K24K30KSE +/- 375.67, N = 330151.330043.829377.9

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 95% Reads - Concurrency: 256cab11K22K33K44K55KSE +/- 168.75, N = 349764.249613.549024.0

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 10% Reads - Concurrency: 256abc4K8K12K16K20KSE +/- 126.23, N = 317827.617639.617585.4

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 50% Reads - Concurrency: 128acb3K6K9K12K15KSE +/- 199.13, N = 316123.015733.015644.6

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 10% Reads - Concurrency: 128abc2K4K6K8K10KSE +/- 11.01, N = 39270.59148.79128.1

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 60% Reads - Concurrency: 128abc4K8K12K16K20KSE +/- 104.08, N = 320255.719591.819566.6

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: KV, 95% Reads - Concurrency: 128abc11K22K33K44K55KSE +/- 345.86, N = 352512.651610.251080.3

KeyDB

A benchmark of KeyDB as an open-source, multi-threaded fork of the Redis server now developed by Snapchat. The developers self-describe the KeyDB in-memory data structure store as a "faster drop in alternative to Redis". Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterKeyDB 6.3.2Test: HMSET - Parallel Connections: 50bac110K220K330K440K550KSE +/- 2335.08, N = 3518247.50511218.21506124.091. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Atomiccab4080120160200SE +/- 3.12, N = 15162.65146.11145.011. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Context Switchingbca400K800K1200K1600K2000KSE +/- 48908.21, N = 151941869.461906264.571897936.071. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: MMAPabc20406080100SE +/- 2.50, N = 1574.7050.2649.731. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Socket Activitybca2K4K6K8K10KSE +/- 146.72, N = 159385.839379.738703.541. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Futexbca400K800K1200K1600K2000KSE +/- 41062.68, N = 151723742.381689762.381579624.881. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: IO_uringcab6001200180024003000SE +/- 36.56, N = 152856.891965.751723.131. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

Neural Magic DeepSparse

This is a benchmark of Neural Magic's DeepSparse using its built-in deepsparse.benchmark utility and various models from their SparseZoo (https://sparsezoo.neuralmagic.com/). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: CV Detection, YOLOv5s COCO - Scenario: Synchronous Single-Streamcba816243240SE +/- 0.33, N = 1232.2832.2932.73

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: CV Detection, YOLOv5s COCO - Scenario: Synchronous Single-Streamcba714212835SE +/- 0.28, N = 1230.9730.9630.57

VVenC

VVenC is the Fraunhofer Versatile Video Encoder as a fast/efficient H.266/VVC encoder. The vvenc encoder makes use of SIMD Everywhere (SIMDe). The vvenc software is published under The Clear BSD License. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.7Video Input: Bosphorus 4K - Video Preset: Fasterabc246810SE +/- 0.017, N = 36.0655.8935.8571. (CXX) g++ options: -O3 -flto -fno-fat-lto-objects -flto=auto

CockroachDB

CockroachDB is a cloud-native, distributed SQL database for data intensive applications. This test profile uses a server-less CockroachDB configuration to test various Coackroach workloads on the local host with a single node. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: MoVR - Concurrency: 256acb4080120160200SE +/- 0.00, N = 3181.6173.8173.6

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: MoVR - Concurrency: 512abc4080120160200SE +/- 0.00, N = 3177.4177.3173.9

OpenBenchmarking.orgops/s, More Is BetterCockroachDB 22.2Workload: MoVR - Concurrency: 1024bac4080120160200SE +/- 0.06, N = 3177.4177.4176.1

RocksDB

This is a benchmark of Meta/Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterRocksDB 7.9.2Test: Random Readbca9M18M27M36M45MSE +/- 414703.47, N = 64315211242975995416141101. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

FLAC Audio Encoding

This test times how long it takes to encode a sample WAV file to FLAC audio format ten times using the --best preset settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterFLAC Audio Encoding 1.4WAV To FLACabc510152025SE +/- 0.15, N = 2517.3219.3019.611. (CXX) g++ options: -O3 -fvisibility=hidden -logg -lm

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 4K - Video Preset: Slowbca246810SE +/- 0.03, N = 36.526.456.341. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 4K - Video Preset: Mediumbca246810SE +/- 0.02, N = 36.646.606.531. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

Scikit-Learn

Scikit-learn is a Python module for machine learning Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.1.3Benchmark: MNIST Datasetabc20406080100SE +/- 0.23, N = 390.1590.8290.84

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 32 - Model: GoogLeNetbac918273645SE +/- 0.11, N = 340.3240.1740.08

nginx

This is a benchmark of the lightweight Nginx HTTP(S) web-server. This Nginx web server benchmark test profile makes use of the wrk program for facilitating the HTTP requests over a fixed period time with a configurable number of concurrent clients/connections. HTTPS with a self-signed OpenSSL certificate is used by this test for local benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is Betternginx 1.23.2Connections: 200bca13K26K39K52K65KSE +/- 149.19, N = 361654.7761396.9361096.581. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

OpenBenchmarking.orgRequests Per Second, More Is Betternginx 1.23.2Connections: 1000cba9K18K27K36K45KSE +/- 203.43, N = 342016.8842015.7241663.171. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

OpenBenchmarking.orgRequests Per Second, More Is Betternginx 1.23.2Connections: 500abc11K22K33K44K55KSE +/- 61.89, N = 351922.4851890.4551737.261. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

OpenBenchmarking.orgRequests Per Second, More Is Betternginx 1.23.2Connections: 100bca15K30K45K60K75KSE +/- 335.89, N = 368835.7868707.4967993.521. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

OpenBenchmarking.orgRequests Per Second, More Is Betternginx 1.23.2Connections: 20bca14K28K42K56K70KSE +/- 804.71, N = 364372.0164314.4562961.451. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUbca9001800270036004500SE +/- 2.56, N = 34289.214303.624341.92MIN: 4045.19MIN: 4079.69MIN: 4124.681. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUbca9001800270036004500SE +/- 15.35, N = 34234.914253.174329.65MIN: 4027.35MIN: 4043.25MIN: 4088.591. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUbca9001800270036004500SE +/- 5.86, N = 34270.754291.984334.97MIN: 4043.43MIN: 4067.51MIN: 4122.671. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.11Encoder Speed: 2cba20406080100SE +/- 0.11, N = 386.4387.0488.251. (CXX) g++ options: -O3 -fPIC -lm

Stargate Digital Audio Workstation

Stargate is an open-source, cross-platform digital audio workstation (DAW) software package with "a unique and carefully curated experience" with scalability from old systems up through modern multi-core systems. Stargate is GPLv3 licensed and makes use of Qt5 (PyQt5) for its user-interface. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRender Ratio, More Is BetterStargate Digital Audio Workstation 22.11.5Sample Rate: 96000 - Buffer Size: 512acb0.43450.8691.30351.7382.1725SE +/- 0.000948, N = 31.9309601.9233521.9184431. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions

RocksDB

This is a benchmark of Meta/Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterRocksDB 7.9.2Test: Update Randomcab140K280K420K560K700KSE +/- 6561.15, N = 56641576299446245401. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUbac9001800270036004500SE +/- 4.94, N = 32403.192437.964000.34MIN: 2251.07MIN: 2279.1MIN: 3992.831. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: MEMFDacb1326395265SE +/- 3.98, N = 1258.5752.9941.651. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: CPU Cachebac20406080100SE +/- 2.33, N = 1289.5083.6072.101. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.0Binary: Pathtracer - Model: Asian Dragon Objbca246810SE +/- 0.0103, N = 38.71688.71248.7102MIN: 8.66 / MAX: 8.83MIN: 8.66 / MAX: 8.86MIN: 8.64 / MAX: 8.85

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUbca5001000150020002500SE +/- 3.98, N = 32398.622410.162438.21MIN: 2251.99MIN: 2254.61MIN: 2279.81. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUcba5001000150020002500SE +/- 22.77, N = 32388.512394.332406.13MIN: 2252.37MIN: 2261.91MIN: 2255.861. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Stargate Digital Audio Workstation

Stargate is an open-source, cross-platform digital audio workstation (DAW) software package with "a unique and carefully curated experience" with scalability from old systems up through modern multi-core systems. Stargate is GPLv3 licensed and makes use of Qt5 (PyQt5) for its user-interface. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRender Ratio, More Is BetterStargate Digital Audio Workstation 22.11.5Sample Rate: 96000 - Buffer Size: 1024acb0.46610.93221.39831.86442.3305SE +/- 0.001510, N = 32.0716692.0526602.0489311. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 64 - Model: AlexNetcab20406080100SE +/- 0.49, N = 389.0888.6088.45

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google as the AV1 Codec Library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 4Kcba3691215SE +/- 0.03, N = 39.219.159.111. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.0Binary: Pathtracer ISPC - Model: Asian Dragon Objbca3691215SE +/- 0.0123, N = 39.62969.58379.5706MIN: 9.56 / MAX: 9.84MIN: 9.5 / MAX: 9.8MIN: 9.47 / MAX: 9.79

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.0Binary: Pathtracer - Model: Crownbca246810SE +/- 0.0433, N = 38.71898.63018.4875MIN: 8.66 / MAX: 8.9MIN: 8.56 / MAX: 8.78MIN: 8.32 / MAX: 8.82

VVenC

VVenC is the Fraunhofer Versatile Video Encoder as a fast/efficient H.266/VVC encoder. The vvenc encoder makes use of SIMD Everywhere (SIMDe). The vvenc software is published under The Clear BSD License. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.7Video Input: Bosphorus 1080p - Video Preset: Fastbca246810SE +/- 0.014, N = 38.0858.0207.9771. (CXX) g++ options: -O3 -flto -fno-fat-lto-objects -flto=auto

SVT-AV1

This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 4 - Input: Bosphorus 4Kbca0.50781.01561.52342.03122.539SE +/- 0.007, N = 32.2572.2502.1851. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

Unvanquished

Unvanquished is a modern fork of the Tremulous first person shooter. Unvanquished is powered by the Daemon engine, a combination of the ioquake3 (id Tech 3) engine with the graphically-beautiful XreaL engine. Unvanquished supports a modern OpenGL 3 renderer and other advanced graphics features for this open-source, cross-platform shooter game. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterUnvanquished 0.54Resolution: 1920 x 1080 - Effects Quality: Ultracab20406080100SE +/- 0.03, N = 3100.7100.6100.5

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUbca246810SE +/- 0.17183, N = 158.372398.440408.69748MIN: 5.85MIN: 5.8MIN: 5.761. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUcba0.34350.6871.03051.3741.7175SE +/- 0.01138, N = 151.409301.415381.52665MIN: 1.39MIN: 1.39MIN: 1.391. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Person Detection FP16 - Device: CPUbac7001400210028003500SE +/- 16.56, N = 33363.573384.143402.11MIN: 3197.76 / MAX: 3701.88MIN: 3150.48 / MAX: 4797.77MIN: 3157.9 / MAX: 5910.281. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Person Detection FP16 - Device: CPUbac0.26550.5310.79651.0621.3275SE +/- 0.01, N = 31.181.171.151. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Person Detection FP32 - Device: CPUbca7001400210028003500SE +/- 29.12, N = 33394.573413.873438.18MIN: 3178.76 / MAX: 3780.74MIN: 3197.21 / MAX: 4433.43MIN: 3203.29 / MAX: 4311.691. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Person Detection FP32 - Device: CPUcba0.2610.5220.7831.0441.305SE +/- 0.01, N = 31.161.161.151. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.0Binary: Pathtracer ISPC - Model: Crownbca3691215SE +/- 0.0598, N = 39.53959.48209.3366MIN: 9.33 / MAX: 9.85MIN: 9.26 / MAX: 9.81MIN: 9.13 / MAX: 9.88

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Face Detection FP16 - Device: CPUbca5001000150020002500SE +/- 12.95, N = 32239.722256.402257.17MIN: 2149.54 / MAX: 2500.68MIN: 2134.57 / MAX: 2467.9MIN: 2132.88 / MAX: 3216.451. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Face Detection FP16 - Device: CPUbca0.39830.79661.19491.59321.9915SE +/- 0.02, N = 31.771.751.751. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.0Binary: Pathtracer - Model: Asian Dragonbca3691215SE +/- 0.0065, N = 39.78579.75449.6958MIN: 9.73 / MAX: 9.93MIN: 9.7 / MAX: 9.87MIN: 9.6 / MAX: 9.92

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Glibc C String Functionscba300K600K900K1200K1500KSE +/- 9938.23, N = 91288214.511273002.921220179.931. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

KeyDB

A benchmark of KeyDB as an open-source, multi-threaded fork of the Redis server now developed by Snapchat. The developers self-describe the KeyDB in-memory data structure store as a "faster drop in alternative to Redis". Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterKeyDB 6.3.2Test: SET - Parallel Connections: 100bac200K400K600K800K1000KSE +/- 12223.15, N = 41146000.381069422.991049912.881. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD

Stargate Digital Audio Workstation

Stargate is an open-source, cross-platform digital audio workstation (DAW) software package with "a unique and carefully curated experience" with scalability from old systems up through modern multi-core systems. Stargate is GPLv3 licensed and makes use of Qt5 (PyQt5) for its user-interface. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRender Ratio, More Is BetterStargate Digital Audio Workstation 22.11.5Sample Rate: 44100 - Buffer Size: 512acb0.58021.16041.74062.32082.901SE +/- 0.005177, N = 32.5785992.5753632.5705971. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8 - Decompression Speedcba400800120016002000SE +/- 0.97, N = 31734.31733.21732.41. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8 - Compression Speedcba4080120160200SE +/- 0.20, N = 3203.6203.6202.61. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

RocksDB

This is a benchmark of Meta/Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterRocksDB 7.9.2Test: Random Fill Synccab7001400210028003500SE +/- 15.17, N = 33193135112391. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

Stargate Digital Audio Workstation

Stargate is an open-source, cross-platform digital audio workstation (DAW) software package with "a unique and carefully curated experience" with scalability from old systems up through modern multi-core systems. Stargate is GPLv3 licensed and makes use of Qt5 (PyQt5) for its user-interface. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRender Ratio, More Is BetterStargate Digital Audio Workstation 22.11.5Sample Rate: 480000 - Buffer Size: 512acb0.59051.1811.77152.3622.9525SE +/- 0.002482, N = 32.6245672.6029502.5932421. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Face Detection FP16-INT8 - Device: CPUbca120240360480600SE +/- 2.23, N = 3541.21549.42552.79MIN: 503.72 / MAX: 633.37MIN: 502.92 / MAX: 645.96MIN: 501.54 / MAX: 634.831. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Face Detection FP16-INT8 - Device: CPUbca246810SE +/- 0.04, N = 37.397.287.211. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3, Long Mode - Decompression Speedcab30060090012001500SE +/- 0.35, N = 31620.21619.41618.71. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3, Long Mode - Compression Speedabc2004006008001000SE +/- 6.47, N = 3869.3866.5866.11. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3 - Decompression Speedcba30060090012001500SE +/- 0.86, N = 31598.41596.21592.61. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3 - Compression Speedbca30060090012001500SE +/- 1.71, N = 31588.61583.61567.81. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 12 - Decompression Speedcba400800120016002000SE +/- 3.74, N = 31762.21761.71758.91. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 12 - Compression Speedcba20406080100SE +/- 0.23, N = 377.377.277.11. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Machine Translation EN To DE FP16 - Device: CPUbac4080120160200SE +/- 0.73, N = 3182.53184.37185.48MIN: 165.34 / MAX: 243.94MIN: 164.97 / MAX: 226.63MIN: 166.33 / MAX: 257.371. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Machine Translation EN To DE FP16 - Device: CPUbac510152025SE +/- 0.09, N = 321.8721.6821.541. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8, Long Mode - Decompression Speedcab400800120016002000SE +/- 1.39, N = 31728.61727.91725.81. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8, Long Mode - Compression Speedcba4080120160200SE +/- 0.15, N = 3200.4199.8198.41. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19, Long Mode - Decompression Speedbca30060090012001500SE +/- 1.17, N = 31451.71447.21444.41. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19, Long Mode - Compression Speedbca246810SE +/- 0.01, N = 37.067.027.001. (CC) gcc options: -O3 -pthread -lz -llzma -llz4

FFmpeg

This is a benchmark of the FFmpeg multimedia framework. The FFmpeg test profile is making use of a modified version of vbench from Columbia University's Architecture and Design Lab (ARCADE) [http://arcade.cs.columbia.edu/vbench/] that is a benchmark for video-as-a-service workloads. The test profile offers the options of a range of vbench scenarios based on freely distributable video content and offers the options of using the x264 or x265 video encoders for transcoding. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterFFmpeg 6.0Encoder: libx265 - Scenario: Liveabc306090120150SE +/- 0.43, N = 3120.17119.65118.361. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

OpenBenchmarking.orgSeconds, Fewer Is BetterFFmpeg 6.0Encoder: libx265 - Scenario: Liveabc1020304050SE +/- 0.15, N = 342.0242.2142.671. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

Stargate Digital Audio Workstation

Stargate is an open-source, cross-platform digital audio workstation (DAW) software package with "a unique and carefully curated experience" with scalability from old systems up through modern multi-core systems. Stargate is GPLv3 licensed and makes use of Qt5 (PyQt5) for its user-interface. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRender Ratio, More Is BetterStargate Digital Audio Workstation 22.11.5Sample Rate: 44100 - Buffer Size: 1024abc0.60871.21741.82612.43483.0435SE +/- 0.000534, N = 32.7051222.6974382.6874491. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions

Timed FFmpeg Compilation

This test times how long it takes to build the FFmpeg multimedia library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed FFmpeg Compilation 6.0Time To Compilebca1428425670SE +/- 0.68, N = 359.3559.4561.47

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Person Vehicle Bike Detection FP16 - Device: CPUbca48121620SE +/- 0.07, N = 314.3114.3714.44MIN: 12.84 / MAX: 26.05MIN: 12.83 / MAX: 26.08MIN: 12.82 / MAX: 32.761. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Person Vehicle Bike Detection FP16 - Device: CPUbca60120180240300SE +/- 1.32, N = 3279.30278.20276.781. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Weld Porosity Detection FP16 - Device: CPUbca714212835SE +/- 0.15, N = 331.5331.6831.87MIN: 28.51 / MAX: 38.58MIN: 28.34 / MAX: 39.14MIN: 28.47 / MAX: 39.411. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Weld Porosity Detection FP16 - Device: CPUbca4080120160200SE +/- 0.88, N = 3190.20189.32188.171. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Vehicle Detection FP16-INT8 - Device: CPUabc3691215SE +/- 0.05, N = 311.4711.4911.54MIN: 10.23 / MAX: 55.04MIN: 10.24 / MAX: 17.75MIN: 10.23 / MAX: 58.551. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Vehicle Detection FP16-INT8 - Device: CPUabc80160240320400SE +/- 1.63, N = 3348.49347.83346.341. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

RocksDB

This is a benchmark of Meta/Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterRocksDB 7.9.2Test: Read While Writingcba300K600K900K1200K1500KSE +/- 18612.51, N = 31532546153115715185151. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Weld Porosity Detection FP16-INT8 - Device: CPUbca246810SE +/- 0.04, N = 38.328.358.41MIN: 7.49 / MAX: 14.34MIN: 7.49 / MAX: 38.94MIN: 7.49 / MAX: 16.851. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Weld Porosity Detection FP16-INT8 - Device: CPUbca160320480640800SE +/- 3.68, N = 3719.91717.70712.701. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

Stargate Digital Audio Workstation

Stargate is an open-source, cross-platform digital audio workstation (DAW) software package with "a unique and carefully curated experience" with scalability from old systems up through modern multi-core systems. Stargate is GPLv3 licensed and makes use of Qt5 (PyQt5) for its user-interface. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRender Ratio, More Is BetterStargate Digital Audio Workstation 22.11.5Sample Rate: 480000 - Buffer Size: 1024abc0.61741.23481.85222.46963.087SE +/- 0.003205, N = 32.7437842.7228462.7171161. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUbca0.21380.42760.64140.85521.069SE +/- 0.00, N = 30.940.940.95MIN: 0.85 / MAX: 6.61MIN: 0.84 / MAX: 7.1MIN: 0.85 / MAX: 7.321. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUcba14002800420056007000SE +/- 19.43, N = 36357.316344.826331.011. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.3Model: Age Gender Recognition Retail 0013 FP16 - Device: CPUbca0.23850.4770.71550.9541.1925SE +/- 0.01, N = 31.041.041.06MIN: 0.92 / MAX: 7.17MIN: 0.92 / MAX: 4.88MIN: 0.92 / MAX: 68.141. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.3Model: Age Gender Recognition Retail 0013 FP16 - Device: CPUcba12002400360048006000SE +/- 27.53, N = 35687.235661.395560.271. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared

RocksDB

This is a benchmark of Meta/Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterRocksDB 7.9.2Test: Read Random Write Randomcba300K600K900K1200K1500KSE +/- 6061.26, N = 31568201156281215466551. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.0Binary: Pathtracer ISPC - Model: Asian Dragonbca3691215SE +/- 0.03, N = 311.1411.1111.06MIN: 11.05 / MAX: 11.33MIN: 11.03 / MAX: 11.35MIN: 10.93 / MAX: 11.31

KeyDB

A benchmark of KeyDB as an open-source, multi-threaded fork of the Redis server now developed by Snapchat. The developers self-describe the KeyDB in-memory data structure store as a "faster drop in alternative to Redis". Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterKeyDB 6.3.2Test: LPUSH - Parallel Connections: 50acb300K600K900K1200K1500KSE +/- 14149.76, N = 41323318.971317870.381252473.621. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD

JPEG XL Decoding libjxl

The JPEG XL Image Coding System is designed to provide next-generation JPEG image capabilities with JPEG XL offering better image quality and compression over legacy JPEG. This test profile is suited for JPEG XL decode performance testing to PNG output file, the pts/jpexl test is for encode performance. The JPEG XL encoding/decoding is done using the libjxl codebase. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterJPEG XL Decoding libjxl 0.7CPU Threads: 1bca1122334455SE +/- 0.34, N = 350.8350.4850.05

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google as the AV1 Codec Library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 1080pcba3691215SE +/- 0.13, N = 313.3113.3113.311. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

Neural Magic DeepSparse

This is a benchmark of Neural Magic's DeepSparse using its built-in deepsparse.benchmark utility and various models from their SparseZoo (https://sparsezoo.neuralmagic.com/). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Streamabc80160240320400SE +/- 0.80, N = 3371.27371.97373.24

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Streamabc246810SE +/- 0.0174, N = 38.07988.06478.0373

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Streamcba110220330440550SE +/- 1.89, N = 3519.66522.76530.84

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Streamcba1.29892.59783.89675.19566.4945SE +/- 0.0204, N = 35.77295.73865.6401

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUbca0.92221.84442.76663.68884.611SE +/- 0.02631, N = 153.932603.952464.09853MIN: 3.81MIN: 3.82MIN: 3.81. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Neural Magic DeepSparse

This is a benchmark of Neural Magic's DeepSparse using its built-in deepsparse.benchmark utility and various models from their SparseZoo (https://sparsezoo.neuralmagic.com/). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Streamcba110220330440550SE +/- 1.42, N = 3520.95526.96530.18

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Streamcba1.29572.59143.88715.18286.4785SE +/- 0.0208, N = 35.75865.68885.6434

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUcab0.36920.73841.10761.47681.846SE +/- 0.105801, N = 150.9599421.2518711.640760MIN: 0.91MIN: 0.91MIN: 1.591. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

VP9 libvpx Encoding

This is a standard video encoding performance test of Google's libvpx library and the vpxenc command for the VP9 video format. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVP9 libvpx Encoding 1.13Speed: Speed 5 - Input: Bosphorus 4Kacb3691215SE +/- 0.07, N = 312.7312.2512.021. (CXX) g++ options: -m64 -lm -lpthread -O3 -fPIC -U_FORTIFY_SOURCE -std=gnu++11

OpenBenchmarking.orgFrames Per Second, More Is BetterVP9 libvpx Encoding 1.13Speed: Speed 0 - Input: Bosphorus 1080pcab3691215SE +/- 0.13, N = 312.6912.1811.571. (CXX) g++ options: -m64 -lm -lpthread -O3 -fPIC -U_FORTIFY_SOURCE -std=gnu++11

KeyDB

A benchmark of KeyDB as an open-source, multi-threaded fork of the Redis server now developed by Snapchat. The developers self-describe the KeyDB in-memory data structure store as a "faster drop in alternative to Redis". Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterKeyDB 6.3.2Test: SADD - Parallel Connections: 100cab300K600K900K1200K1500KSE +/- 8944.31, N = 31193659.251148270.291137811.751. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 16 - Model: GoogLeNetbac918273645SE +/- 0.03, N = 339.2038.8938.53

Neural Magic DeepSparse

This is a benchmark of Neural Magic's DeepSparse using its built-in deepsparse.benchmark utility and various models from their SparseZoo (https://sparsezoo.neuralmagic.com/). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Streamcab1122334455SE +/- 0.04, N = 341.8742.3450.84

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Streamcab1632486480SE +/- 0.07, N = 371.6370.8358.96

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Synchronous Single-Streamcba306090120150SE +/- 0.41, N = 3123.95123.97125.27

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Synchronous Single-Streamcba246810SE +/- 0.0263, N = 38.06668.06587.9821

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 32 - Model: AlexNetabc20406080100SE +/- 0.36, N = 377.9577.8077.44

Unvanquished

Unvanquished is a modern fork of the Tremulous first person shooter. Unvanquished is powered by the Daemon engine, a combination of the ioquake3 (id Tech 3) engine with the graphically-beautiful XreaL engine. Unvanquished supports a modern OpenGL 3 renderer and other advanced graphics features for this open-source, cross-platform shooter game. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterUnvanquished 0.54Resolution: 1920 x 1080 - Effects Quality: Highbac4080120160200SE +/- 0.20, N = 3171.1170.8170.7

Neural Magic DeepSparse

This is a benchmark of Neural Magic's DeepSparse using its built-in deepsparse.benchmark utility and various models from their SparseZoo (https://sparsezoo.neuralmagic.com/). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Streamcab306090120150SE +/- 1.91, N = 3135.43136.27146.72

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Streamcab510152025SE +/- 0.30, N = 322.1522.0220.44

KeyDB

A benchmark of KeyDB as an open-source, multi-threaded fork of the Redis server now developed by Snapchat. The developers self-describe the KeyDB in-memory data structure store as a "faster drop in alternative to Redis". Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterKeyDB 6.3.2Test: GET - Parallel Connections: 100bac300K600K900K1200K1500KSE +/- 15206.41, N = 31382055.381283175.581273333.881. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 4K - Video Preset: Very Fastbca3691215SE +/- 0.08, N = 313.5713.5013.25

Neural Magic DeepSparse

This is a benchmark of Neural Magic's DeepSparse using its built-in deepsparse.benchmark utility and various models from their SparseZoo (https://sparsezoo.neuralmagic.com/). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-Streamabc4080120160200SE +/- 0.52, N = 3182.98185.09185.10

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-Streamabc1.22962.45923.68884.91846.148SE +/- 0.0157, N = 35.46505.40275.4023

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-Streamcba4080120160200SE +/- 0.43, N = 3183.69183.89184.65

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-Streamcba1.22482.44963.67444.89926.124SE +/- 0.0127, N = 35.44375.43795.4154

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Synchronous Single-Streamcab1224364860SE +/- 0.32, N = 350.8251.1552.80

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Synchronous Single-Streamcab510152025SE +/- 0.12, N = 319.6819.5518.94

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 4K - Video Preset: Super Fastbca48121620SE +/- 0.08, N = 314.3014.2413.98

OpenEMS

OpenEMS is a free and open electromagnetic field solver using the FDTD method. This test profile runs OpenEMS and pyEMS benchmark demos. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMCells/s, More Is BetterOpenEMS 0.0.35-86Test: openEMS MSL_NotchFilteracb20406080100SE +/- 0.06, N = 398.8898.7798.721. (CXX) g++ options: -O3 -rdynamic -ltinyxml -lcrypto -lcurl -lpthread -lsz -lz -ldl -lm -lexpat

Neural Magic DeepSparse

This is a benchmark of Neural Magic's DeepSparse using its built-in deepsparse.benchmark utility and various models from their SparseZoo (https://sparsezoo.neuralmagic.com/). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Synchronous Single-Streamcab1326395265SE +/- 0.15, N = 348.5249.6458.91

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Synchronous Single-Streamcab510152025SE +/- 0.06, N = 320.6120.1416.97

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: LBC, LBRY Creditscba4K8K12K16K20KSE +/- 186.26, N = 51971019200184381. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

Neural Magic DeepSparse

This is a benchmark of Neural Magic's DeepSparse using its built-in deepsparse.benchmark utility and various models from their SparseZoo (https://sparsezoo.neuralmagic.com/). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Synchronous Single-Streamcab510152025SE +/- 0.08, N = 320.2620.3320.39

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Synchronous Single-Streamcab1122334455SE +/- 0.19, N = 349.3349.1849.03

FFmpeg

This is a benchmark of the FFmpeg multimedia framework. The FFmpeg test profile is making use of a modified version of vbench from Columbia University's Architecture and Design Lab (ARCADE) [http://arcade.cs.columbia.edu/vbench/] that is a benchmark for video-as-a-service workloads. The test profile offers the options of a range of vbench scenarios based on freely distributable video content and offers the options of using the x264 or x265 video encoders for transcoding. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterFFmpeg 6.0Encoder: libx264 - Scenario: Livebac50100150200250SE +/- 0.14, N = 3241.95239.86235.801. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

OpenBenchmarking.orgSeconds, Fewer Is BetterFFmpeg 6.0Encoder: libx264 - Scenario: Livebac510152025SE +/- 0.01, N = 320.8721.0521.421. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma

EnCodec

EnCodec is a Facebook/Meta developed AI means of compressing audio files using High Fidelity Neural Audio Compression. EnCodec is designed to provide codec compression at 6 kbps using their novel AI-powered compression technique. The test profile uses a lengthy JFK speech as the audio input for benchmarking and the performance measurement is measuring the time to encode the EnCodec file from WAV. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterEnCodec 0.1.1Target Bandwidth: 6 kbpsacb918273645SE +/- 0.47, N = 338.5240.1241.18

Neural Magic DeepSparse

This is a benchmark of Neural Magic's DeepSparse using its built-in deepsparse.benchmark utility and various models from their SparseZoo (https://sparsezoo.neuralmagic.com/). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Streambca20406080100SE +/- 0.97, N = 388.7890.1691.19

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Streambca816243240SE +/- 0.35, N = 333.7833.2632.89

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google as the AV1 Codec Library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 6 Realtime - Input: Bosphorus 4Kacb1224364860SE +/- 0.37, N = 1455.3355.3154.931. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

EnCodec

EnCodec is a Facebook/Meta developed AI means of compressing audio files using High Fidelity Neural Audio Compression. EnCodec is designed to provide codec compression at 6 kbps using their novel AI-powered compression technique. The test profile uses a lengthy JFK speech as the audio input for benchmarking and the performance measurement is measuring the time to encode the EnCodec file from WAV. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterEnCodec 0.1.1Target Bandwidth: 1.5 kbpsacb918273645SE +/- 0.20, N = 336.9238.7141.03

Neural Magic DeepSparse

This is a benchmark of Neural Magic's DeepSparse using its built-in deepsparse.benchmark utility and various models from their SparseZoo (https://sparsezoo.neuralmagic.com/). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-Streamcab612182430SE +/- 0.23, N = 325.2526.0926.57

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-Streamcab918273645SE +/- 0.34, N = 339.6038.3437.64

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Streambca918273645SE +/- 0.13, N = 338.4339.0140.47

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Streambca20406080100SE +/- 0.23, N = 378.0476.8874.10

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.3.2Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-Streambca48121620SE +/- 0.17, N = 315.3715.3815.57

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.3.2Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-Streambca1530456075SE +/- 0.70, N = 365.0364.9964.23

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Pthreadcba30K60K90K120K150KSE +/- 1404.99, N = 4132268.72130862.38128698.011. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 4K - Video Preset: Very Fastbca48121620SE +/- 0.12, N = 317.2217.1316.451. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Garlicoinbca400800120016002000SE +/- 21.45, N = 41908.991895.321836.591. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 4K - Video Preset: Ultra Fastbca48121620SE +/- 0.04, N = 317.4917.4116.99

Unvanquished

Unvanquished is a modern fork of the Tremulous first person shooter. Unvanquished is powered by the Daemon engine, a combination of the ioquake3 (id Tech 3) engine with the graphically-beautiful XreaL engine. Unvanquished supports a modern OpenGL 3 renderer and other advanced graphics features for this open-source, cross-platform shooter game. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterUnvanquished 0.54Resolution: 1920 x 1080 - Effects Quality: Mediumcab50100150200250SE +/- 0.09, N = 3248.2248.2248.1

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google as the AV1 Codec Library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 10 Realtime - Input: Bosphorus 4Kabc1326395265SE +/- 0.42, N = 1258.7658.6658.531. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUcba0.56351.1271.69052.2542.8175SE +/- 0.02403, N = 152.435322.452782.50462MIN: 2.39MIN: 2.41MIN: 2.341. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 16 - Model: AlexNetbac1428425670SE +/- 0.10, N = 361.2460.8160.69

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Zlibbac150300450600750SE +/- 0.29, N = 3717.62717.35716.141. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: System V Message Passingbac2M4M6M8M10MSE +/- 10695.76, N = 38225295.798216877.608192747.721. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Hashbac300K600K900K1200K1500KSE +/- 854.18, N = 31549710.051548492.991548182.161. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Pollbac200K400K600K800K1000KSE +/- 379.65, N = 3823948.19823762.13823174.051. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Memory Copyingabc5001000150020002500SE +/- 14.39, N = 32301.692295.122285.431. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: NUMAbca50100150200250SE +/- 0.70, N = 3246.58245.00244.131. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Myriad-Groestlacb2K4K6K8K10KSE +/- 68.39, N = 31123711100110901. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Mallocbac1.1M2.2M3.3M4.4M5.5MSE +/- 19639.79, N = 35241054.325226456.995200128.821. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Matrix Mathbca10K20K30K40K50KSE +/- 69.96, N = 346046.4845978.8445892.791. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Function Callbac11K22K33K44K55KSE +/- 1.32, N = 349807.0549787.6249784.191. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Vector Mathcab9K18K27K36K45KSE +/- 3.69, N = 341779.9841733.9741733.471. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Skeincoincba10K20K30K40K50KSE +/- 360.94, N = 34867048410478531. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Blake-2 Sbca60K120K180K240K300KSE +/- 791.71, N = 33029802952502822071. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Mutexbac900K1800K2700K3600K4500KSE +/- 2569.53, N = 34273551.984254720.234254389.291. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: CPU Stresscab4K8K12K16K20KSE +/- 32.25, N = 317782.2117781.8617780.761. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: SENDFILEcab40K80K120K160K200KSE +/- 184.17, N = 3178115.64177365.84176973.241. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: x86_64 RdRandabc20K40K60K80K100KSE +/- 0.57, N = 381909.4581907.9781907.391. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Glibc Qsort Data Sortingcba20406080100SE +/- 0.60, N = 3102.73102.53101.421. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Cryptoacb3K6K9K12K15KSE +/- 2.58, N = 314276.3914271.8814261.231. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Semaphoresbac300K600K900K1200K1500KSE +/- 593.53, N = 31190334.271189258.321188758.841. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.15.04Test: Forkingacb12K24K36K48K60KSE +/- 302.19, N = 354276.3454204.6552494.811. (CC) gcc options: -std=gnu99 -O2 -lm -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lGLESv2 -ljpeg -lrt -lsctp -lz -pthread

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Triple SHA-256, Onecoinabc20K40K60K80K100KSE +/- 108.37, N = 37815777950779401. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Magiabc70140210280350SE +/- 2.70, N = 3331.49327.09326.581. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Deepcoinbac14002800420056007000SE +/- 12.35, N = 36428.906392.106368.981. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: x25xcba80160240320400SE +/- 0.59, N = 3368.41367.28344.301. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: scryptcba20406080100SE +/- 0.41, N = 3111.66111.65110.871. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Ringcoincba30060090012001500SE +/- 14.19, N = 31586.861583.581560.091. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

VVenC

VVenC is the Fraunhofer Versatile Video Encoder as a fast/efficient H.266/VVC encoder. The vvenc encoder makes use of SIMD Everywhere (SIMDe). The vvenc software is published under The Clear BSD License. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.7Video Input: Bosphorus 1080p - Video Preset: Fasterbca510152025SE +/- 0.07, N = 320.8620.5920.161. (CXX) g++ options: -O3 -flto -fno-fat-lto-objects -flto=auto

Cpuminer-Opt

Cpuminer-Opt is a fork of cpuminer-multi that carries a wide range of CPU performance optimizations for measuring the potential cryptocurrency mining performance of the CPU/processor with a wide variety of cryptocurrencies. The benchmark reports the hash speed for the CPU mining performance for the selected cryptocurrency. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.20.3Algorithm: Quad SHA-256, Pyritecba10K20K30K40K50KSE +/- 70.24, N = 34832048310482501. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 4K - Video Preset: Super Fastbca510152025SE +/- 0.11, N = 321.0420.9220.381. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 1080p - Video Preset: Slowbca510152025SE +/- 0.02, N = 322.3422.1421.53

JPEG XL Decoding libjxl

The JPEG XL Image Coding System is designed to provide next-generation JPEG image capabilities with JPEG XL offering better image quality and compression over legacy JPEG. This test profile is suited for JPEG XL decode performance testing to PNG output file, the pts/jpexl test is for encode performance. The JPEG XL encoding/decoding is done using the libjxl codebase. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterJPEG XL Decoding libjxl 0.7CPU Threads: Allbca50100150200250SE +/- 0.07, N = 3226.31225.63217.33

Scikit-Learn

Scikit-learn is a Python module for machine learning Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.1.3Benchmark: TSNE MNIST Datasetbca612182430SE +/- 0.22, N = 326.7126.8626.91

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 1080p - Video Preset: Mediumbca612182430SE +/- 0.14, N = 325.0524.9624.11

SVT-AV1

This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 8 - Input: Bosphorus 4Kbca714212835SE +/- 0.22, N = 327.7827.5027.441. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google as the AV1 Codec Library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 1080pcab816243240SE +/- 0.42, N = 332.3531.6931.541. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.11Encoder Speed: 10, Losslessbca246810SE +/- 0.185, N = 156.4386.4396.5711. (CXX) g++ options: -O3 -fPIC -lm

SVT-AV1

This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 4 - Input: Bosphorus 1080pbca246810SE +/- 0.018, N = 37.8097.7507.7371. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 4K - Video Preset: Ultra Fastbca714212835SE +/- 0.21, N = 329.1128.8527.831. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

VP9 libvpx Encoding

This is a standard video encoding performance test of Google's libvpx library and the vpxenc command for the VP9 video format. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVP9 libvpx Encoding 1.13Speed: Speed 5 - Input: Bosphorus 1080pacb816243240SE +/- 0.02, N = 332.7532.3832.331. (CXX) g++ options: -m64 -lm -lpthread -O3 -fPIC -U_FORTIFY_SOURCE -std=gnu++11

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 1080p - Video Preset: Slowbca816243240SE +/- 0.14, N = 333.2733.2232.531. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 1080p - Video Preset: Mediumbca816243240SE +/- 0.14, N = 334.5234.5034.461. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google as the AV1 Codec Library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 6 Realtime - Input: Bosphorus 1080pcab4080120160200SE +/- 2.71, N = 15169.96147.69145.581. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 8 Realtime - Input: Bosphorus 4Kbca1020304050SE +/- 0.05, N = 346.2046.0845.871. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.11Encoder Speed: 6, Losslessabc48121620SE +/- 0.14, N = 313.4613.7213.731. (CXX) g++ options: -O3 -fPIC -lm

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google as the AV1 Codec Library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 10 Realtime - Input: Bosphorus 1080pbac4080120160200SE +/- 3.61, N = 15177.46175.80168.561. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUbac0.81161.62322.43483.24644.058SE +/- 0.00059, N = 33.593643.597193.60705MIN: 3.53MIN: 3.53MIN: 3.541. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUcba0.27050.5410.81151.0821.3525SE +/- 0.00770, N = 31.182281.199251.20219MIN: 1.12MIN: 1.13MIN: 1.131. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google as the AV1 Codec Library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 8 Realtime - Input: Bosphorus 1080pcab306090120150SE +/- 1.01, N = 11140.47139.13139.061. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 9 Realtime - Input: Bosphorus 1080pcab4080120160200SE +/- 3.99, N = 15201.29188.81171.531. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.6Encoder Mode: Speed 9 Realtime - Input: Bosphorus 4Kbca1326395265SE +/- 0.04, N = 358.4558.3857.991. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUbac3691215SE +/- 0.66035, N = 158.133738.938859.33670MIN: 8.1MIN: 7.53MIN: 9.281. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUacb0.45260.90521.35781.81042.263SE +/- 0.01841, N = 152.000202.010502.01154MIN: 1.92MIN: 2MIN: 21. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 1080p - Video Preset: Very Fastcba1428425670SE +/- 0.17, N = 364.1364.1062.33

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUacb3691215SE +/- 0.00, N = 310.3710.4310.56MIN: 10.1MIN: 10.17MIN: 10.291. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.11Encoder Speed: 6acb3691215SE +/- 0.071, N = 39.1759.2689.2771. (CXX) g++ options: -O3 -fPIC -lm

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 1080p - Video Preset: Super Fastbca1530456075SE +/- 0.36, N = 367.4667.3667.09

SVT-AV1

This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 12 - Input: Bosphorus 4Kabc20406080100SE +/- 0.17, N = 388.4369.5768.861. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

RocksDB

This is a benchmark of Meta/Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterRocksDB 7.9.2Test: Sequential Fillbac300K600K900K1200K1500KSE +/- 3413.48, N = 31511361150279214615661. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 1080p - Video Preset: Very Fastcab20406080100SE +/- 0.07, N = 375.0474.7974.731. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

SVT-AV1

This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 8 - Input: Bosphorus 1080pabc20406080100SE +/- 0.56, N = 381.4581.4081.021. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

uvg266

uvg266 is an open-source VVC/H.266 (Versatile Video Coding) encoder based on Kvazaar as part of the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betteruvg266 0.4.1Video Input: Bosphorus 1080p - Video Preset: Ultra Fastabc20406080100SE +/- 0.06, N = 380.0179.8179.41

SVT-AV1

This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 13 - Input: Bosphorus 4Kbac20406080100SE +/- 0.27, N = 396.5895.5795.091. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 1080p - Video Preset: Super Fastbac20406080100SE +/- 0.19, N = 393.5293.4893.301. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUcab48121620SE +/- 0.00, N = 316.5916.5916.61MIN: 16.43MIN: 16.45MIN: 16.471. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.0Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUcab510152025SE +/- 0.03, N = 318.2318.2818.41MIN: 17.93MIN: 17.89MIN: 18.021. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Kvazaar

This is a test of Kvazaar as a CPU-based H.265/HEVC video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterKvazaar 2.2Video Input: Bosphorus 1080p - Video Preset: Ultra Fastcab306090120150SE +/- 0.14, N = 3123.33123.26121.821. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt

SVT-AV1

This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 12 - Input: Bosphorus 1080pbac80160240320400SE +/- 1.08, N = 3348.91348.11347.221. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.4Encoder Mode: Preset 13 - Input: Bosphorus 1080pbac90180270360450SE +/- 0.84, N = 3399.08398.86397.311. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

nginx

This is a benchmark of the lightweight Nginx HTTP(S) web-server. This Nginx web server benchmark test profile makes use of the wrk program for facilitating the HTTP requests over a fixed period time with a configurable number of concurrent clients/connections. HTTPS with a self-signed OpenSSL certificate is used by this test for local benchmarking. Learn more via the OpenBenchmarking.org test page.

Connections: 1

a: The test quit with a non-zero exit status.

b: The test quit with a non-zero exit status.

c: The test quit with a non-zero exit status.

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU

a: The test run did not produce a result.

b: The test run did not produce a result.

c: The test run did not produce a result.

Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU

a: The test run did not produce a result.

b: The test run did not produce a result.

c: The test run did not produce a result.

Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU

a: The test run did not produce a result.

b: The test run did not produce a result.

c: The test run did not produce a result.

Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU

a: The test run did not produce a result.

b: The test run did not produce a result.

c: The test run did not produce a result.

Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU

a: The test run did not produce a result.

b: The test run did not produce a result.

c: The test run did not produce a result.

Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU

a: The test run did not produce a result.

b: The test run did not produce a result.

c: The test run did not produce a result.

359 Results Shown

Blender
nekRS
Timed Linux Kernel Compilation
ClickHouse:
  100M Rows Hits Dataset, Third Run
  100M Rows Hits Dataset, Second Run
  100M Rows Hits Dataset, First Run / Cold Cache
TensorFlow
OpenRadioss
TensorFlow
Blender
OpenEMS
BRL-CAD
miniBUDE:
  OpenMP - BM2:
    Billion Interactions/s
    GFInst/s
Blender
TensorFlow
PostgreSQL:
  100 - 1000 - Read Write - Average Latency
  100 - 1000 - Read Write
  100 - 250 - Read Write - Average Latency
  100 - 250 - Read Write
  100 - 500 - Read Write - Average Latency
  100 - 500 - Read Write
  100 - 100 - Read Write - Average Latency
  100 - 100 - Read Write
OpenVKL
KeyDB
OpenRadioss
PostgreSQL:
  1 - 800 - Read Write - Average Latency
  1 - 800 - Read Write
OpenVKL
PostgreSQL:
  1 - 500 - Read Write - Average Latency
  1 - 500 - Read Write
JPEG XL libjxl:
  JPEG - 100
  PNG - 100
Xmrig
PostgreSQL:
  100 - 800 - Read Write - Average Latency
  100 - 800 - Read Write
TensorFlow
FFmpeg:
  libx264 - Upload:
    FPS
    Seconds
CockroachDB
Blender
OpenRadioss
Timed Linux Kernel Compilation
AOM AV1
Zstd Compression:
  19 - Decompression Speed
  19 - Compression Speed
VVenC
GROMACS
RocksDB
KeyDB
libavif avifenc
FFmpeg:
  libx265 - Video On Demand:
    FPS
    Seconds
OpenVINO:
  Vehicle Detection FP16 - CPU:
    ms
    FPS
FFmpeg:
  libx265 - Upload:
    FPS
    Seconds
  libx265 - Platform:
    FPS
    Seconds
Xmrig
OpenRadioss
TensorFlow
Blender
FFmpeg:
  libx264 - Video On Demand:
    FPS
    Seconds
  libx264 - Platform:
    FPS
    Seconds
PostgreSQL:
  1 - 500 - Read Only - Average Latency
  1 - 500 - Read Only
EnCodec
PostgreSQL:
  100 - 1000 - Read Only - Average Latency
  100 - 1000 - Read Only
JPEG XL libjxl
PostgreSQL:
  100 - 800 - Read Only - Average Latency
  100 - 800 - Read Only
KeyDB
PostgreSQL:
  100 - 500 - Read Only - Average Latency
  100 - 500 - Read Only
JPEG XL libjxl
AOM AV1
PostgreSQL:
  100 - 250 - Read Only - Average Latency
  100 - 250 - Read Only
miniBUDE:
  OpenMP - BM1:
    Billion Interactions/s
    GFInst/s
OpenRadioss
uvg266
KeyDB
PostgreSQL:
  100 - 100 - Read Only - Average Latency
  100 - 100 - Read Only
Scikit-Learn
KeyDB
PostgreSQL:
  1 - 1000 - Read Write - Average Latency
  1 - 1000 - Read Write
AOM AV1
Stargate Digital Audio Workstation
Neural Magic DeepSparse:
  NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Stream:
    ms/batch
    items/sec
EnCodec
KeyDB
PostgreSQL:
  1 - 250 - Read Write - Average Latency
  1 - 250 - Read Write
  1 - 100 - Read Write - Average Latency
  1 - 100 - Read Write
  1 - 800 - Read Only - Average Latency
  1 - 800 - Read Only
  1 - 1000 - Read Only - Average Latency
  1 - 1000 - Read Only
  1 - 100 - Read Only - Average Latency
  1 - 100 - Read Only
  1 - 250 - Read Only - Average Latency
  1 - 250 - Read Only
uvg266
KeyDB
Stargate Digital Audio Workstation
JPEG XL libjxl:
  JPEG - 90
  PNG - 90
Neural Magic DeepSparse:
  NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Stream:
    ms/batch
    items/sec
VP9 libvpx Encoding
CockroachDB:
  KV, 10% Reads - 512
  KV, 10% Reads - 1024
  KV, 95% Reads - 1024
  KV, 50% Reads - 1024
  KV, 60% Reads - 1024
  KV, 95% Reads - 512
  KV, 50% Reads - 512
  KV, 60% Reads - 512
  KV, 60% Reads - 256
  KV, 50% Reads - 256
  KV, 95% Reads - 256
  KV, 10% Reads - 256
  KV, 50% Reads - 128
  KV, 10% Reads - 128
  KV, 60% Reads - 128
  KV, 95% Reads - 128
KeyDB
Stress-NG:
  Atomic
  Context Switching
  MMAP
  Socket Activity
  Futex
  IO_uring
Neural Magic DeepSparse:
  CV Detection, YOLOv5s COCO - Synchronous Single-Stream:
    ms/batch
    items/sec
VVenC
CockroachDB:
  MoVR - 256
  MoVR - 512
  MoVR - 1024
RocksDB
FLAC Audio Encoding
Kvazaar:
  Bosphorus 4K - Slow
  Bosphorus 4K - Medium
Scikit-Learn
TensorFlow
nginx:
  200
  1000
  500
  100
  20
oneDNN:
  Recurrent Neural Network Training - u8s8f32 - CPU
  Recurrent Neural Network Training - f32 - CPU
  Recurrent Neural Network Training - bf16bf16bf16 - CPU
libavif avifenc
Stargate Digital Audio Workstation
RocksDB
oneDNN
Stress-NG:
  MEMFD
  CPU Cache
Embree
oneDNN:
  Recurrent Neural Network Inference - f32 - CPU
  Recurrent Neural Network Inference - u8s8f32 - CPU
Stargate Digital Audio Workstation
TensorFlow
AOM AV1
Embree:
  Pathtracer ISPC - Asian Dragon Obj
  Pathtracer - Crown
VVenC
SVT-AV1
Unvanquished
oneDNN:
  Deconvolution Batch shapes_1d - f32 - CPU
  Deconvolution Batch shapes_1d - u8s8f32 - CPU
OpenVINO:
  Person Detection FP16 - CPU:
    ms
    FPS
  Person Detection FP32 - CPU:
    ms
    FPS
Embree
OpenVINO:
  Face Detection FP16 - CPU:
    ms
    FPS
Embree
Stress-NG
KeyDB
Stargate Digital Audio Workstation
Zstd Compression:
  8 - Decompression Speed
  8 - Compression Speed
RocksDB
Stargate Digital Audio Workstation
OpenVINO:
  Face Detection FP16-INT8 - CPU:
    ms
    FPS
Zstd Compression:
  3, Long Mode - Decompression Speed
  3, Long Mode - Compression Speed
  3 - Decompression Speed
  3 - Compression Speed
  12 - Decompression Speed
  12 - Compression Speed
OpenVINO:
  Machine Translation EN To DE FP16 - CPU:
    ms
    FPS
Zstd Compression:
  8, Long Mode - Decompression Speed
  8, Long Mode - Compression Speed
  19, Long Mode - Decompression Speed
  19, Long Mode - Compression Speed
FFmpeg:
  libx265 - Live:
    FPS
    Seconds
Stargate Digital Audio Workstation
Timed FFmpeg Compilation
OpenVINO:
  Person Vehicle Bike Detection FP16 - CPU:
    ms
    FPS
  Weld Porosity Detection FP16 - CPU:
    ms
    FPS
  Vehicle Detection FP16-INT8 - CPU:
    ms
    FPS
RocksDB
OpenVINO:
  Weld Porosity Detection FP16-INT8 - CPU:
    ms
    FPS
Stargate Digital Audio Workstation
OpenVINO:
  Age Gender Recognition Retail 0013 FP16-INT8 - CPU:
    ms
    FPS
  Age Gender Recognition Retail 0013 FP16 - CPU:
    ms
    FPS
RocksDB
Embree
KeyDB
JPEG XL Decoding libjxl
AOM AV1
Neural Magic DeepSparse:
  CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Stream:
    ms/batch
    items/sec
oneDNN
Neural Magic DeepSparse:
  NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Stream:
    ms/batch
    items/sec
oneDNN
VP9 libvpx Encoding:
  Speed 5 - Bosphorus 4K
  Speed 0 - Bosphorus 1080p
KeyDB
TensorFlow
Neural Magic DeepSparse:
  NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  CV Segmentation, 90% Pruned YOLACT Pruned - Synchronous Single-Stream:
    ms/batch
    items/sec
TensorFlow
Unvanquished
Neural Magic DeepSparse:
  NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Stream:
    ms/batch
    items/sec
KeyDB
uvg266
Neural Magic DeepSparse:
  NLP Document Classification, oBERT base uncased on IMDB - Synchronous Single-Stream:
    ms/batch
    items/sec
  NLP Token Classification, BERT base uncased conll2003 - Synchronous Single-Stream:
    ms/batch
    items/sec
  NLP Text Classification, BERT base uncased SST2 - Synchronous Single-Stream:
    ms/batch
    items/sec
uvg266
OpenEMS
Neural Magic DeepSparse:
  NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Synchronous Single-Stream:
    ms/batch
    items/sec
Cpuminer-Opt
Neural Magic DeepSparse:
  NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Synchronous Single-Stream:
    ms/batch
    items/sec
FFmpeg:
  libx264 - Live:
    FPS
    Seconds
EnCodec
Neural Magic DeepSparse:
  CV Detection, YOLOv5s COCO - Asynchronous Multi-Stream:
    ms/batch
    items/sec
AOM AV1
EnCodec
Neural Magic DeepSparse:
  NLP Text Classification, DistilBERT mnli - Synchronous Single-Stream:
    ms/batch
    items/sec
  CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  CV Classification, ResNet-50 ImageNet - Synchronous Single-Stream:
    ms/batch
    items/sec
Stress-NG
Kvazaar
Cpuminer-Opt
uvg266
Unvanquished
AOM AV1
oneDNN
TensorFlow
Stress-NG:
  Zlib
  System V Message Passing
  Hash
  Poll
  Memory Copying
  NUMA
Cpuminer-Opt
Stress-NG:
  Malloc
  Matrix Math
  Function Call
  Vector Math
Cpuminer-Opt:
  Skeincoin
  Blake-2 S
Stress-NG:
  Mutex
  CPU Stress
  SENDFILE
  x86_64 RdRand
  Glibc Qsort Data Sorting
  Crypto
  Semaphores
  Forking
Cpuminer-Opt:
  Triple SHA-256, Onecoin
  Magi
  Deepcoin
  x25x
  scrypt
  Ringcoin
VVenC
Cpuminer-Opt
Kvazaar
uvg266
JPEG XL Decoding libjxl
Scikit-Learn
uvg266
SVT-AV1
AOM AV1
libavif avifenc
SVT-AV1
Kvazaar
VP9 libvpx Encoding
Kvazaar:
  Bosphorus 1080p - Slow
  Bosphorus 1080p - Medium
AOM AV1:
  Speed 6 Realtime - Bosphorus 1080p
  Speed 8 Realtime - Bosphorus 4K
libavif avifenc
AOM AV1
oneDNN:
  Matrix Multiply Batch Shapes Transformer - f32 - CPU
  Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU
AOM AV1:
  Speed 8 Realtime - Bosphorus 1080p
  Speed 9 Realtime - Bosphorus 1080p
  Speed 9 Realtime - Bosphorus 4K
oneDNN:
  Deconvolution Batch shapes_3d - f32 - CPU
  Deconvolution Batch shapes_3d - u8s8f32 - CPU
uvg266
oneDNN
libavif avifenc
uvg266
SVT-AV1
RocksDB
Kvazaar
SVT-AV1
uvg266
SVT-AV1
Kvazaar
oneDNN:
  Convolution Batch Shapes Auto - f32 - CPU
  Convolution Batch Shapes Auto - u8s8f32 - CPU
Kvazaar
SVT-AV1:
  Preset 12 - Bosphorus 1080p
  Preset 13 - Bosphorus 1080p