Benchmarks for a future article.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 2403025-NE-2402297NE31 Linux Distros Emerald Rapids - Phoronix Test Suite Linux Distros Emerald Rapids Benchmarks for a future article.
HTML result view exported from: https://openbenchmarking.org/result/2403025-NE-2402297NE31&sor&gru .
Linux Distros Emerald Rapids Processor Motherboard Chipset Memory Disk Graphics Network OS Kernel Compiler File-System Screen Resolution Desktop Display Server Ubuntu Linux 23.10 CentOS Stream 9 Fedora Server 39 2 x INTEL XEON PLATINUM 8592+ @ 3.90GHz (128 Cores / 256 Threads) Quanta Cloud QuantaGrid D54Q-2U S6Q-MB-MPS (3B05.TEL4P1 BIOS) Intel Device 1bce 1008GB 3201GB Micron_7450_MTFDKCB3T2TFS ASPEED 2 x Intel X710 for 10GBASE-T Ubuntu 23.10 6.5.0-17-generic (x86_64) GCC 13.2.0 ext4 1024x768 3201GB Micron_7450_MTFDKCB3T2TFS + 257GB Flash Drive CentOS Stream 9 5.14.0-419.el9.x86_64 (x86_64) GNOME Shell 40.10 X Server GCC 11.4.1 20231218 xfs 1920x1200 Fedora Linux 39 6.7.6-200.fc39.x86_64 (x86_64) GCC 13.2.1 20231205 1024x768 OpenBenchmarking.org Kernel Details - Ubuntu Linux 23.10: Transparent Huge Pages: madvise - CentOS Stream 9: Transparent Huge Pages: always - Fedora Server 39: Transparent Huge Pages: madvise Compiler Details - Ubuntu Linux 23.10: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - CentOS Stream 9: --build=x86_64-redhat-linux --disable-libunwind-exceptions --enable-__cxa_atexit --enable-bootstrap --enable-cet --enable-checking=release --enable-gnu-indirect-function --enable-gnu-unique-object --enable-host-bind-now --enable-host-pie --enable-initfini-array --enable-languages=c,c++,fortran,lto --enable-link-serialization=1 --enable-multilib --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --mandir=/usr/share/man --with-arch_32=x86-64 --with-arch_64=x86-64-v2 --with-build-config=bootstrap-lto --with-gcc-major-version-only --with-linker-hash-style=gnu --with-tune=generic --without-cuda-driver --without-isl - Fedora Server 39: --build=x86_64-redhat-linux --disable-libunwind-exceptions --enable-__cxa_atexit --enable-bootstrap --enable-cet --enable-checking=release --enable-gnu-indirect-function --enable-gnu-unique-object --enable-initfini-array --enable-languages=c,c++,fortran,objc,obj-c++,ada,go,d,m2,lto --enable-libstdcxx-backtrace --enable-link-serialization=1 --enable-multilib --enable-offload-defaulted --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --mandir=/usr/share/man --with-arch_32=i686 --with-build-config=bootstrap-lto --with-gcc-major-version-only --with-libstdcxx-zoneinfo=/usr/share/zoneinfo --with-linker-hash-style=gnu --with-tune=generic --without-cuda-driver Processor Details - Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x21000161 Python Details - Ubuntu Linux 23.10: Python 3.11.6 - CentOS Stream 9: Python 3.9.18 - Fedora Server 39: Python 3.12.2 Security Details - Ubuntu Linux 23.10: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Not affected - CentOS Stream 9: SELinux + gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Not affected - Fedora Server 39: SELinux + gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Not affected
Linux Distros Emerald Rapids graph500: 26 graph500: 26 quicksilver: CTS2 quicksilver: CORAL2 P1 quicksilver: CORAL2 P2 openvino: Person Detection FP16 - CPU openvino: Face Detection FP16-INT8 - CPU openvino: Vehicle Detection FP16-INT8 - CPU openvino: Face Detection Retail FP16-INT8 - CPU openvino: Road Segmentation ADAS FP16-INT8 - CPU openvino: Machine Translation EN To DE FP16 - CPU openvino: Weld Porosity Detection FP16-INT8 - CPU openvino: Person Vehicle Bike Detection FP16 - CPU openvino: Handwritten English Recognition FP16-INT8 - CPU openvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPU embree: Pathtracer ISPC - Crown embree: Pathtracer ISPC - Asian Dragon svt-av1: Preset 4 - Bosphorus 4K svt-av1: Preset 8 - Bosphorus 4K svt-av1: Preset 12 - Bosphorus 4K svt-av1: Preset 13 - Bosphorus 4K vvenc: Bosphorus 4K - Fast vvenc: Bosphorus 4K - Faster hpcg: 104 104 104 - 60 hpcg: 144 144 144 - 60 mt-dgemm: Sustained Floating-Point Rate oidn: RT.hdr_alb_nrm.3840x2160 - CPU-Only oidn: RT.ldr_alb_nrm.3840x2160 - CPU-Only oidn: RTLightmap.hdr.4096x4096 - CPU-Only tensorflow: CPU - 512 - ResNet-50 openvkl: vklBenchmarkCPU ISPC ospray: particle_volume/ao/real_time ospray: particle_volume/scivis/real_time ospray: particle_volume/pathtracer/real_time ospray: gravity_spheres_volume/dim_512/ao/real_time ospray: gravity_spheres_volume/dim_512/scivis/real_time ospray: gravity_spheres_volume/dim_512/pathtracer/real_time deepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Stream deepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Asynchronous Multi-Stream deepsparse: ResNet-50, Baseline - Asynchronous Multi-Stream deepsparse: ResNet-50, Sparse INT8 - Asynchronous Multi-Stream deepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Stream deepsparse: BERT-Large, NLP Question Answering - Asynchronous Multi-Stream deepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Stream deepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Stream deepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Stream deepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Stream deepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Stream deepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Stream compress-7zip: Compression Rating compress-7zip: Decompression Rating lczero: BLAS lczero: Eigen gromacs: MPI CPU - water_GMX50_bare namd: ATPase with 327,506 Atoms namd: STMV with 1,066,628 Atoms lammps: 20k Atoms lammps: Rhodopsin Protein rocksdb: Rand Read rocksdb: Update Rand rocksdb: Read While Writing rocksdb: Read Rand Write Rand speedb: Rand Read memcached: 1:10 memcached: 1:100 clickhouse: 100M Rows Hits Dataset, First Run / Cold Cache clickhouse: 100M Rows Hits Dataset, Second Run clickhouse: 100M Rows Hits Dataset, Third Run redis: GET - 500 redis: SET - 500 graph500: 26 graph500: 26 pgbench: 100 - 1000 - Read Only pgbench: 100 - 1000 - Read Write pgbench: 100 - 1000 - Read Only - Average Latency pgbench: 100 - 1000 - Read Write - Average Latency openvino: Person Detection FP16 - CPU openvino: Face Detection FP16-INT8 - CPU openvino: Vehicle Detection FP16-INT8 - CPU openvino: Face Detection Retail FP16-INT8 - CPU openvino: Road Segmentation ADAS FP16-INT8 - CPU openvino: Machine Translation EN To DE FP16 - CPU openvino: Weld Porosity Detection FP16-INT8 - CPU openvino: Person Vehicle Bike Detection FP16 - CPU openvino: Handwritten English Recognition FP16-INT8 - CPU openvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPU ospray-studio: 1 - 4K - 1 - Path Tracer - CPU ospray-studio: 2 - 4K - 1 - Path Tracer - CPU ospray-studio: 3 - 4K - 1 - Path Tracer - CPU ospray-studio: 1 - 4K - 16 - Path Tracer - CPU ospray-studio: 1 - 4K - 32 - Path Tracer - CPU ospray-studio: 2 - 4K - 16 - Path Tracer - CPU ospray-studio: 2 - 4K - 32 - Path Tracer - CPU ospray-studio: 3 - 4K - 16 - Path Tracer - CPU ospray-studio: 3 - 4K - 32 - Path Tracer - CPU ospray-studio: 1 - 1080p - 1 - Path Tracer - CPU ospray-studio: 2 - 1080p - 1 - Path Tracer - CPU ospray-studio: 3 - 1080p - 1 - Path Tracer - CPU ospray-studio: 1 - 1080p - 16 - Path Tracer - CPU ospray-studio: 1 - 1080p - 32 - Path Tracer - CPU ospray-studio: 2 - 1080p - 16 - Path Tracer - CPU ospray-studio: 2 - 1080p - 32 - Path Tracer - CPU ospray-studio: 3 - 1080p - 16 - Path Tracer - CPU ospray-studio: 3 - 1080p - 32 - Path Tracer - CPU deepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Stream deepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Asynchronous Multi-Stream deepsparse: ResNet-50, Baseline - Asynchronous Multi-Stream deepsparse: ResNet-50, Sparse INT8 - Asynchronous Multi-Stream deepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Stream deepsparse: BERT-Large, NLP Question Answering - Asynchronous Multi-Stream deepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Stream deepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Stream deepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Stream deepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Stream deepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Stream deepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Stream cloverleaf: clover_bm16 cloverleaf: clover_bm64_short nwchem: C240 Buckyball incompact3d: X3D-benchmarking input.i3d easywave: e2Asean Grid + BengkuluSept2007 Source - 2400 y-cruncher: 1B y-cruncher: 5B rawtherapee: Total Benchmark Time gpaw: Carbon Nanotube Ubuntu Linux 23.10 CentOS Stream 9 Fedora Server 39 1289580000 1226550000 7680222 5931500 6744778 375.75 519.83 8901.61 24570.10 2377.90 1143.25 43516.98 10033.95 3217.84 61020.16 100.3399 144.9195 2.307 21.317 66.857 66.430 2.785 5.611 71.1371 69.3362 37.971018 4.53 4.54 2.18 73.12 2821 20.6546 20.7725 82.9700 28.6088 28.9162 32.3212 132.6789 4547.5395 1837.1113 11099.6440 829.5063 155.8178 1826.0447 853.2079 1223.1639 176.5691 1866.2259 133.2063 464813 455056 15 380 9.630 4.70711 1.49004 56.683 37.518 599822046 125560 9775989 955461 600703384 3370014.38 3477137.28 205.20 214.31 215.62 1541713.80 2134872.89 664991000 483534000 679317 39542 1.489 25.616 85.06 245.98 14.37 5.20 53.79 27.93 2.85 12.74 39.76 0.80 914 911 1079 12086 32279 12243 33174 14508 37285 234 233 276 3718 6448 3694 6324 4353 7321 478.7631 14.0573 34.7603 5.7515 77.0180 409.0089 34.9884 74.8838 52.2774 361.2877 34.2545 477.1362 279.21 32.10 1750.3 169.178579 177.754 7.207 29.821 129.079 38.637 1499600000 1464120000 7398333 5800833 6381000 328.37 528.79 8908.74 24295.71 2412.51 617.45 40187.34 10113.47 3196.47 60429.54 105.6194 177.7960 2.358 21.560 69.868 69.718 2.823 5.808 71.1332 69.1728 63.635840 5.04 4.90 2.51 76.68 2815 20.5021 20.4427 83.5652 29.7626 29.5104 33.2938 132.9101 4462.5255 1863.3429 11022.3020 827.9914 152.5678 1866.5153 853.3374 1223.6470 173.9515 1810.9508 133.2818 751070 434200 5.939 3.20328 0.50251 57.836 39.795 3337339.10 2615816.73 183.62 200.18 201.05 1821521.39 1978937.18 640794000 466398000 97.34 241.67 14.36 5.25 53.02 52.09 3.02 12.65 40.02 0.83 478.0315 14.3182 34.2906 5.7833 77.1671 416.6816 34.2262 74.8038 52.2360 366.3831 35.2846 476.5785 314.49 25.01 184.192332 205.005 7.913 32.850 1253250000 1195510000 7367500 5835818 6759600 340.97 529.14 8862.07 24296.54 2383.22 978.70 42419.89 10082.93 3255.16 56624.98 142.1675 179.5403 2.346 21.161 71.518 70.874 2.834 5.804 72.4575 70.3805 37.720076 4.68 4.70 2.25 2821 20.4442 20.5419 82.1380 28.6637 28.430 31.6999 496138 428601 7.180 5.40706 1.92977 57.677 39.009 652854125 123092 8567006 1184837 626261826 5054518.76 8983240.52 188.71 202.14 206.83 1838671.43 2125843.97 663412000 493031000 93.74 241.70 14.43 5.25 53.67 32.67 2.86 12.69 39.30 0.83 902 897 1058 12187 28035 12262 27524 14422 34698 235 233 276 3683 6314 3655 6303 4377 7424 255.39 31.63 168.671819 156.237 8.270 33.734 118.228 194.508 OpenBenchmarking.org
Graph500 Scale: 26 OpenBenchmarking.org bfs max_TEPS, More Is Better Graph500 3.0 Scale: 26 CentOS Stream 9 Ubuntu Linux 23.10 Fedora Server 39 300M 600M 900M 1200M 1500M 1499600000 1289580000 1253250000 1. (CC) gcc options: -fcommon -O3 -lpthread -lm -lmpi
Graph500 Scale: 26 OpenBenchmarking.org bfs median_TEPS, More Is Better Graph500 3.0 Scale: 26 CentOS Stream 9 Ubuntu Linux 23.10 Fedora Server 39 300M 600M 900M 1200M 1500M 1464120000 1226550000 1195510000 1. (CC) gcc options: -fcommon -O3 -lpthread -lm -lmpi
Quicksilver Input: CTS2 OpenBenchmarking.org Figure Of Merit, More Is Better Quicksilver 20230818 Input: CTS2 Ubuntu Linux 23.10 CentOS Stream 9 Fedora Server 39 1.6M 3.2M 4.8M 6.4M 8M SE +/- 80526.35, N = 9 SE +/- 121572.66, N = 6 SE +/- 90618.71, N = 4 7680222 7398333 7367500 1. (CXX) g++ options: -fopenmp -O3 -march=native
Quicksilver Input: CORAL2 P1 OpenBenchmarking.org Figure Of Merit, More Is Better Quicksilver 20230818 Input: CORAL2 P1 Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 1.3M 2.6M 3.9M 5.2M 6.5M SE +/- 65941.38, N = 12 SE +/- 42699.12, N = 11 SE +/- 105910.76, N = 12 5931500 5835818 5800833 1. (CXX) g++ options: -fopenmp -O3 -march=native
Quicksilver Input: CORAL2 P2 OpenBenchmarking.org Figure Of Merit, More Is Better Quicksilver 20230818 Input: CORAL2 P2 Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 1.4M 2.8M 4.2M 5.6M 7M SE +/- 74346.22, N = 5 SE +/- 81512.06, N = 9 SE +/- 51868.42, N = 3 6759600 6744778 6381000 1. (CXX) g++ options: -fopenmp -O3 -march=native
OpenVINO Model: Person Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Person Detection FP16 - Device: CPU Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 80 160 240 320 400 SE +/- 0.48, N = 3 SE +/- 0.63, N = 3 SE +/- 2.69, N = 3 375.75 340.97 328.37 -pie -pie -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Face Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Face Detection FP16-INT8 - Device: CPU Fedora Server 39 CentOS Stream 9 Ubuntu Linux 23.10 110 220 330 440 550 SE +/- 6.06, N = 4 SE +/- 1.20, N = 3 SE +/- 5.90, N = 4 529.14 528.79 519.83 -pie -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF -pie 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Vehicle Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Vehicle Detection FP16-INT8 - Device: CPU CentOS Stream 9 Ubuntu Linux 23.10 Fedora Server 39 2K 4K 6K 8K 10K SE +/- 7.04, N = 3 SE +/- 20.48, N = 3 SE +/- 9.74, N = 3 8908.74 8901.61 8862.07 -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF -pie -pie 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Face Detection Retail FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Face Detection Retail FP16-INT8 - Device: CPU Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 5K 10K 15K 20K 25K SE +/- 208.25, N = 3 SE +/- 130.19, N = 3 SE +/- 36.89, N = 3 24570.10 24296.54 24295.71 -pie -pie -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Road Segmentation ADAS FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Road Segmentation ADAS FP16-INT8 - Device: CPU CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 500 1000 1500 2000 2500 SE +/- 10.50, N = 3 SE +/- 4.46, N = 3 SE +/- 15.67, N = 3 2412.51 2383.22 2377.90 -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF -pie -pie 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Machine Translation EN To DE FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Machine Translation EN To DE FP16 - Device: CPU Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 200 400 600 800 1000 SE +/- 4.60, N = 3 SE +/- 8.04, N = 9 SE +/- 13.25, N = 15 1143.25 978.70 617.45 -pie -pie -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Weld Porosity Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Weld Porosity Detection FP16-INT8 - Device: CPU Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 9K 18K 27K 36K 45K SE +/- 43.54, N = 3 SE +/- 194.22, N = 3 SE +/- 347.22, N = 3 43516.98 42419.89 40187.34 -pie -pie -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Person Vehicle Bike Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Person Vehicle Bike Detection FP16 - Device: CPU CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 2K 4K 6K 8K 10K SE +/- 104.96, N = 5 SE +/- 124.93, N = 4 SE +/- 140.94, N = 3 10113.47 10082.93 10033.95 -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF -pie -pie 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Handwritten English Recognition FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Handwritten English Recognition FP16-INT8 - Device: CPU Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 700 1400 2100 2800 3500 SE +/- 4.78, N = 3 SE +/- 1.65, N = 3 SE +/- 1.03, N = 3 3255.16 3217.84 3196.47 -pie -pie -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2023.2.dev Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU Ubuntu Linux 23.10 CentOS Stream 9 Fedora Server 39 13K 26K 39K 52K 65K SE +/- 827.51, N = 15 SE +/- 478.67, N = 3 SE +/- 271.02, N = 3 61020.16 60429.54 56624.98 -pie -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF -pie 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
Embree Binary: Pathtracer ISPC - Model: Crown OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Crown Fedora Server 39 CentOS Stream 9 Ubuntu Linux 23.10 30 60 90 120 150 SE +/- 1.32, N = 15 SE +/- 1.24, N = 15 SE +/- 0.25, N = 3 142.17 105.62 100.34 MIN: 115.81 / MAX: 164.58 MIN: 95.01 / MAX: 132.75 MIN: 95.92 / MAX: 109.46
Embree Binary: Pathtracer ISPC - Model: Asian Dragon OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Asian Dragon Fedora Server 39 CentOS Stream 9 Ubuntu Linux 23.10 40 80 120 160 200 SE +/- 0.73, N = 3 SE +/- 3.90, N = 15 SE +/- 0.50, N = 3 179.54 177.80 144.92 MIN: 145.33 / MAX: 207.47 MIN: 142.88 / MAX: 210.57 MIN: 137.85 / MAX: 155.04
SVT-AV1 Encoder Mode: Preset 4 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.8 Encoder Mode: Preset 4 - Input: Bosphorus 4K CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 0.5306 1.0612 1.5918 2.1224 2.653 SE +/- 0.009, N = 3 SE +/- 0.008, N = 3 SE +/- 0.007, N = 3 2.358 2.346 2.307 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 8 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.8 Encoder Mode: Preset 8 - Input: Bosphorus 4K CentOS Stream 9 Ubuntu Linux 23.10 Fedora Server 39 5 10 15 20 25 SE +/- 0.21, N = 3 SE +/- 0.19, N = 3 SE +/- 0.10, N = 3 21.56 21.32 21.16 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 12 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.8 Encoder Mode: Preset 12 - Input: Bosphorus 4K Fedora Server 39 CentOS Stream 9 Ubuntu Linux 23.10 16 32 48 64 80 SE +/- 0.99, N = 3 SE +/- 0.39, N = 3 SE +/- 0.48, N = 15 71.52 69.87 66.86 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 13 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.8 Encoder Mode: Preset 13 - Input: Bosphorus 4K Fedora Server 39 CentOS Stream 9 Ubuntu Linux 23.10 16 32 48 64 80 SE +/- 0.53, N = 3 SE +/- 0.99, N = 3 SE +/- 0.74, N = 5 70.87 69.72 66.43 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
VVenC Video Input: Bosphorus 4K - Video Preset: Fast OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.11 Video Input: Bosphorus 4K - Video Preset: Fast Fedora Server 39 CentOS Stream 9 Ubuntu Linux 23.10 0.6377 1.2754 1.9131 2.5508 3.1885 SE +/- 0.017, N = 3 SE +/- 0.026, N = 3 SE +/- 0.035, N = 3 2.834 2.823 2.785 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
VVenC Video Input: Bosphorus 4K - Video Preset: Faster OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.11 Video Input: Bosphorus 4K - Video Preset: Faster CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 1.3068 2.6136 3.9204 5.2272 6.534 SE +/- 0.013, N = 3 SE +/- 0.035, N = 3 SE +/- 0.061, N = 4 5.808 5.804 5.611 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
High Performance Conjugate Gradient X Y Z: 104 104 104 - RT: 60 OpenBenchmarking.org GFLOP/s, More Is Better High Performance Conjugate Gradient 3.1 X Y Z: 104 104 104 - RT: 60 Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 16 32 48 64 80 SE +/- 0.14, N = 3 SE +/- 0.19, N = 3 SE +/- 0.12, N = 3 72.46 71.14 71.13 1. (CXX) g++ options: -O3 -ffast-math -ftree-vectorize -lmpi_cxx -lmpi
High Performance Conjugate Gradient X Y Z: 144 144 144 - RT: 60 OpenBenchmarking.org GFLOP/s, More Is Better High Performance Conjugate Gradient 3.1 X Y Z: 144 144 144 - RT: 60 Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 16 32 48 64 80 SE +/- 0.27, N = 3 SE +/- 0.52, N = 3 SE +/- 0.83, N = 4 70.38 69.34 69.17 1. (CXX) g++ options: -O3 -ffast-math -ftree-vectorize -lmpi_cxx -lmpi
ACES DGEMM Sustained Floating-Point Rate OpenBenchmarking.org GFLOP/s, More Is Better ACES DGEMM 1.0 Sustained Floating-Point Rate CentOS Stream 9 Ubuntu Linux 23.10 Fedora Server 39 14 28 42 56 70 SE +/- 0.51, N = 3 SE +/- 0.09, N = 3 SE +/- 0.04, N = 3 63.64 37.97 37.72 1. (CC) gcc options: -O3 -march=native -fopenmp
Intel Open Image Denoise Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.2 Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Only CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 1.134 2.268 3.402 4.536 5.67 SE +/- 0.07, N = 3 SE +/- 0.06, N = 3 SE +/- 0.02, N = 3 5.04 4.68 4.53
Intel Open Image Denoise Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.2 Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Only CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 1.1025 2.205 3.3075 4.41 5.5125 SE +/- 0.03, N = 3 SE +/- 0.03, N = 15 SE +/- 0.06, N = 3 4.90 4.70 4.54
Intel Open Image Denoise Run: RTLightmap.hdr.4096x4096 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.2 Run: RTLightmap.hdr.4096x4096 - Device: CPU-Only CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 0.5648 1.1296 1.6944 2.2592 2.824 SE +/- 0.03, N = 4 SE +/- 0.01, N = 15 SE +/- 0.01, N = 3 2.51 2.25 2.18
TensorFlow Device: CPU - Batch Size: 512 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 512 - Model: ResNet-50 CentOS Stream 9 Ubuntu Linux 23.10 20 40 60 80 100 SE +/- 0.83, N = 4 SE +/- 0.84, N = 3 76.68 73.12
OpenVKL Benchmark: vklBenchmarkCPU ISPC OpenBenchmarking.org Items / Sec, More Is Better OpenVKL 2.0.0 Benchmark: vklBenchmarkCPU ISPC Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 600 1200 1800 2400 3000 SE +/- 22.66, N = 3 SE +/- 20.99, N = 3 SE +/- 33.05, N = 3 2821 2821 2815 MIN: 178 / MAX: 36193 MIN: 175 / MAX: 34505 MIN: 186 / MAX: 35204
OSPRay Benchmark: particle_volume/ao/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: particle_volume/ao/real_time Ubuntu Linux 23.10 CentOS Stream 9 Fedora Server 39 5 10 15 20 25 SE +/- 0.04, N = 3 SE +/- 0.01, N = 3 SE +/- 0.06, N = 3 20.65 20.50 20.44
OSPRay Benchmark: particle_volume/scivis/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: particle_volume/scivis/real_time Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 5 10 15 20 25 SE +/- 0.08, N = 3 SE +/- 0.02, N = 3 SE +/- 0.02, N = 3 20.77 20.54 20.44
OSPRay Benchmark: particle_volume/pathtracer/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: particle_volume/pathtracer/real_time CentOS Stream 9 Ubuntu Linux 23.10 Fedora Server 39 20 40 60 80 100 SE +/- 0.23, N = 3 SE +/- 0.35, N = 3 SE +/- 0.58, N = 3 83.57 82.97 82.14
OSPRay Benchmark: gravity_spheres_volume/dim_512/ao/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: gravity_spheres_volume/dim_512/ao/real_time CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 7 14 21 28 35 SE +/- 0.26, N = 15 SE +/- 0.10, N = 3 SE +/- 0.05, N = 3 29.76 28.66 28.61
OSPRay Benchmark: gravity_spheres_volume/dim_512/scivis/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: gravity_spheres_volume/dim_512/scivis/real_time CentOS Stream 9 Ubuntu Linux 23.10 Fedora Server 39 7 14 21 28 35 SE +/- 0.17, N = 3 SE +/- 0.11, N = 3 SE +/- 0.17, N = 3 29.51 28.92 28.43
OSPRay Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 3.1 Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_time CentOS Stream 9 Ubuntu Linux 23.10 Fedora Server 39 8 16 24 32 40 SE +/- 0.48, N = 15 SE +/- 0.05, N = 3 SE +/- 0.12, N = 3 33.29 32.32 31.70
Neural Magic DeepSparse Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream CentOS Stream 9 Ubuntu Linux 23.10 30 60 90 120 150 SE +/- 0.30, N = 3 SE +/- 0.29, N = 3 132.91 132.68
Neural Magic DeepSparse Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream Ubuntu Linux 23.10 CentOS Stream 9 1000 2000 3000 4000 5000 SE +/- 40.60, N = 7 SE +/- 40.12, N = 7 4547.54 4462.53
Neural Magic DeepSparse Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream CentOS Stream 9 Ubuntu Linux 23.10 400 800 1200 1600 2000 SE +/- 4.86, N = 3 SE +/- 7.36, N = 3 1863.34 1837.11
Neural Magic DeepSparse Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream Ubuntu Linux 23.10 CentOS Stream 9 2K 4K 6K 8K 10K SE +/- 102.73, N = 7 SE +/- 43.63, N = 3 11099.64 11022.30
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream Ubuntu Linux 23.10 CentOS Stream 9 200 400 600 800 1000 SE +/- 5.31, N = 3 SE +/- 4.43, N = 3 829.51 827.99
Neural Magic DeepSparse Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Stream Ubuntu Linux 23.10 CentOS Stream 9 30 60 90 120 150 SE +/- 0.87, N = 3 SE +/- 0.32, N = 3 155.82 152.57
Neural Magic DeepSparse Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream CentOS Stream 9 Ubuntu Linux 23.10 400 800 1200 1600 2000 SE +/- 3.56, N = 3 SE +/- 16.26, N = 3 1866.52 1826.04
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream CentOS Stream 9 Ubuntu Linux 23.10 200 400 600 800 1000 SE +/- 8.06, N = 3 SE +/- 6.05, N = 3 853.34 853.21
Neural Magic DeepSparse Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream CentOS Stream 9 Ubuntu Linux 23.10 300 600 900 1200 1500 SE +/- 15.64, N = 3 SE +/- 14.85, N = 3 1223.65 1223.16
Neural Magic DeepSparse Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream Ubuntu Linux 23.10 CentOS Stream 9 40 80 120 160 200 SE +/- 1.16, N = 3 SE +/- 0.21, N = 3 176.57 173.95
Neural Magic DeepSparse Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream Ubuntu Linux 23.10 CentOS Stream 9 400 800 1200 1600 2000 SE +/- 20.52, N = 3 SE +/- 16.06, N = 3 1866.23 1810.95
Neural Magic DeepSparse Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.6 Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream CentOS Stream 9 Ubuntu Linux 23.10 30 60 90 120 150 SE +/- 0.09, N = 3 SE +/- 0.24, N = 3 133.28 133.21
7-Zip Compression Test: Compression Rating OpenBenchmarking.org MIPS, More Is Better 7-Zip Compression 22.01 Test: Compression Rating CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 160K 320K 480K 640K 800K SE +/- 5583.39, N = 15 SE +/- 4508.27, N = 7 SE +/- 4291.24, N = 3 751070 496138 464813 1. (CXX) g++ options: -lpthread -ldl -O2 -fPIC
7-Zip Compression Test: Decompression Rating OpenBenchmarking.org MIPS, More Is Better 7-Zip Compression 22.01 Test: Decompression Rating Ubuntu Linux 23.10 CentOS Stream 9 Fedora Server 39 100K 200K 300K 400K 500K SE +/- 39319.74, N = 3 SE +/- 2615.74, N = 15 SE +/- 14994.49, N = 7 455056 434200 428601 1. (CXX) g++ options: -lpthread -ldl -O2 -fPIC
LeelaChessZero Backend: BLAS OpenBenchmarking.org Nodes Per Second, More Is Better LeelaChessZero 0.30 Backend: BLAS Ubuntu Linux 23.10 4 8 12 16 20 SE +/- 1.18, N = 9 15 1. (CXX) g++ options: -flto -pthread
LeelaChessZero Backend: Eigen OpenBenchmarking.org Nodes Per Second, More Is Better LeelaChessZero 0.30 Backend: Eigen Ubuntu Linux 23.10 80 160 240 320 400 SE +/- 4.52, N = 4 380 1. (CXX) g++ options: -flto -pthread
GROMACS Implementation: MPI CPU - Input: water_GMX50_bare OpenBenchmarking.org Ns Per Day, More Is Better GROMACS 2024 Implementation: MPI CPU - Input: water_GMX50_bare Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 3 6 9 12 15 SE +/- 1.403, N = 7 SE +/- 0.850, N = 12 SE +/- 0.323, N = 12 9.630 7.180 5.939 1. (CXX) g++ options: -O3 -lm
NAMD Input: ATPase with 327,506 Atoms OpenBenchmarking.org ns/day, More Is Better NAMD 3.0b6 Input: ATPase with 327,506 Atoms Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 1.2166 2.4332 3.6498 4.8664 6.083 SE +/- 0.33952, N = 15 SE +/- 0.33618, N = 15 SE +/- 0.30866, N = 12 5.40706 4.70711 3.20328
NAMD Input: STMV with 1,066,628 Atoms OpenBenchmarking.org ns/day, More Is Better NAMD 3.0b6 Input: STMV with 1,066,628 Atoms Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 0.4342 0.8684 1.3026 1.7368 2.171 SE +/- 0.05361, N = 15 SE +/- 0.14829, N = 15 SE +/- 0.04889, N = 12 1.92977 1.49004 0.50251
LAMMPS Molecular Dynamics Simulator Model: 20k Atoms OpenBenchmarking.org ns/day, More Is Better LAMMPS Molecular Dynamics Simulator 23Jun2022 Model: 20k Atoms CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 13 26 39 52 65 SE +/- 0.46, N = 3 SE +/- 0.40, N = 3 SE +/- 0.42, N = 3 57.84 57.68 56.68 1. (CXX) g++ options: -O3 -lm -ldl
LAMMPS Molecular Dynamics Simulator Model: Rhodopsin Protein OpenBenchmarking.org ns/day, More Is Better LAMMPS Molecular Dynamics Simulator 23Jun2022 Model: Rhodopsin Protein CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 9 18 27 36 45 SE +/- 0.76, N = 15 SE +/- 0.95, N = 15 SE +/- 0.93, N = 15 39.80 39.01 37.52 1. (CXX) g++ options: -O3 -lm -ldl
RocksDB Test: Random Read OpenBenchmarking.org Op/s, More Is Better RocksDB 8.0 Test: Random Read Fedora Server 39 Ubuntu Linux 23.10 140M 280M 420M 560M 700M SE +/- 3865049.53, N = 3 SE +/- 7527805.83, N = 15 652854125 599822046 -lpthread 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti
RocksDB Test: Update Random OpenBenchmarking.org Op/s, More Is Better RocksDB 8.0 Test: Update Random Ubuntu Linux 23.10 Fedora Server 39 30K 60K 90K 120K 150K SE +/- 440.29, N = 3 SE +/- 1187.74, N = 3 125560 123092 -lpthread 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti
RocksDB Test: Read While Writing OpenBenchmarking.org Op/s, More Is Better RocksDB 8.0 Test: Read While Writing Ubuntu Linux 23.10 Fedora Server 39 2M 4M 6M 8M 10M SE +/- 81611.31, N = 8 SE +/- 133209.05, N = 14 9775989 8567006 -lpthread 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti
RocksDB Test: Read Random Write Random OpenBenchmarking.org Op/s, More Is Better RocksDB 8.0 Test: Read Random Write Random Fedora Server 39 Ubuntu Linux 23.10 300K 600K 900K 1200K 1500K SE +/- 13512.70, N = 15 SE +/- 5328.60, N = 3 1184837 955461 -lpthread 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti
Speedb Test: Random Read OpenBenchmarking.org Op/s, More Is Better Speedb 2.7 Test: Random Read Fedora Server 39 Ubuntu Linux 23.10 130M 260M 390M 520M 650M SE +/- 8524314.28, N = 3 SE +/- 5282775.06, N = 8 626261826 600703384 -lpthread 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti
Memcached Set To Get Ratio: 1:10 OpenBenchmarking.org Ops/sec, More Is Better Memcached 1.6.19 Set To Get Ratio: 1:10 Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 1.1M 2.2M 3.3M 4.4M 5.5M SE +/- 65254.44, N = 3 SE +/- 36897.17, N = 5 SE +/- 43085.73, N = 13 5054518.76 3370014.38 3337339.10 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Memcached Set To Get Ratio: 1:100 OpenBenchmarking.org Ops/sec, More Is Better Memcached 1.6.19 Set To Get Ratio: 1:100 Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 2M 4M 6M 8M 10M SE +/- 380886.14, N = 12 SE +/- 20632.12, N = 3 SE +/- 20448.57, N = 10 8983240.52 3477137.28 2615816.73 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
ClickHouse 100M Rows Hits Dataset, First Run / Cold Cache OpenBenchmarking.org Queries Per Minute, Geo Mean, More Is Better ClickHouse 22.12.3.5 100M Rows Hits Dataset, First Run / Cold Cache Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 50 100 150 200 250 SE +/- 2.70, N = 3 SE +/- 2.54, N = 3 SE +/- 0.84, N = 3 205.20 188.71 183.62 MIN: 35.4 / MAX: 1818.18 MIN: 30.08 / MAX: 1428.57 MIN: 34.56 / MAX: 1714.29
ClickHouse 100M Rows Hits Dataset, Second Run OpenBenchmarking.org Queries Per Minute, Geo Mean, More Is Better ClickHouse 22.12.3.5 100M Rows Hits Dataset, Second Run Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 50 100 150 200 250 SE +/- 1.67, N = 3 SE +/- 1.90, N = 3 SE +/- 3.30, N = 3 214.31 202.14 200.18 MIN: 36.39 / MAX: 1764.71 MIN: 36.41 / MAX: 1463.41 MIN: 37.41 / MAX: 2000
ClickHouse 100M Rows Hits Dataset, Third Run OpenBenchmarking.org Queries Per Minute, Geo Mean, More Is Better ClickHouse 22.12.3.5 100M Rows Hits Dataset, Third Run Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 50 100 150 200 250 SE +/- 3.81, N = 3 SE +/- 3.01, N = 3 SE +/- 2.16, N = 3 215.62 206.83 201.05 MIN: 37.43 / MAX: 1714.29 MIN: 35.91 / MAX: 1935.48 MIN: 36.08 / MAX: 1500
Redis Test: GET - Parallel Connections: 500 OpenBenchmarking.org Requests Per Second, More Is Better Redis 7.0.4 Test: GET - Parallel Connections: 500 Fedora Server 39 CentOS Stream 9 Ubuntu Linux 23.10 400K 800K 1200K 1600K 2000K SE +/- 30474.05, N = 15 SE +/- 19161.91, N = 15 SE +/- 15103.72, N = 15 1838671.43 1821521.39 1541713.80 1. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
Redis Test: SET - Parallel Connections: 500 OpenBenchmarking.org Requests Per Second, More Is Better Redis 7.0.4 Test: SET - Parallel Connections: 500 Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 500K 1000K 1500K 2000K 2500K SE +/- 48178.71, N = 15 SE +/- 73166.31, N = 15 SE +/- 47416.25, N = 15 2134872.89 2125843.97 1978937.18 1. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3
Graph500 Scale: 26 OpenBenchmarking.org sssp max_TEPS, More Is Better Graph500 3.0 Scale: 26 Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 140M 280M 420M 560M 700M 664991000 663412000 640794000 1. (CC) gcc options: -fcommon -O3 -lpthread -lm -lmpi
Graph500 Scale: 26 OpenBenchmarking.org sssp median_TEPS, More Is Better Graph500 3.0 Scale: 26 Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 110M 220M 330M 440M 550M 493031000 483534000 466398000 1. (CC) gcc options: -fcommon -O3 -lpthread -lm -lmpi
PostgreSQL Scaling Factor: 100 - Clients: 1000 - Mode: Read Only OpenBenchmarking.org TPS, More Is Better PostgreSQL 16 Scaling Factor: 100 - Clients: 1000 - Mode: Read Only Ubuntu Linux 23.10 150K 300K 450K 600K 750K SE +/- 22056.84, N = 12 679317 1. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm
PostgreSQL Scaling Factor: 100 - Clients: 1000 - Mode: Read Write OpenBenchmarking.org TPS, More Is Better PostgreSQL 16 Scaling Factor: 100 - Clients: 1000 - Mode: Read Write Ubuntu Linux 23.10 8K 16K 24K 32K 40K SE +/- 1662.84, N = 9 39542 1. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm
PostgreSQL Scaling Factor: 100 - Clients: 1000 - Mode: Read Only - Average Latency OpenBenchmarking.org ms, Fewer Is Better PostgreSQL 16 Scaling Factor: 100 - Clients: 1000 - Mode: Read Only - Average Latency Ubuntu Linux 23.10 0.335 0.67 1.005 1.34 1.675 SE +/- 0.047, N = 12 1.489 1. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm
PostgreSQL Scaling Factor: 100 - Clients: 1000 - Mode: Read Write - Average Latency OpenBenchmarking.org ms, Fewer Is Better PostgreSQL 16 Scaling Factor: 100 - Clients: 1000 - Mode: Read Write - Average Latency Ubuntu Linux 23.10 6 12 18 24 30 SE +/- 0.97, N = 9 25.62 1. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lm
OpenVINO Model: Person Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Person Detection FP16 - Device: CPU Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 20 40 60 80 100 SE +/- 0.11, N = 3 SE +/- 0.18, N = 3 SE +/- 0.81, N = 3 85.06 93.74 97.34 -pie - MIN: 39.38 / MAX: 260.96 -pie - MIN: 34.75 / MAX: 283.24 -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF - MIN: 33.82 / MAX: 325.84 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Face Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Face Detection FP16-INT8 - Device: CPU CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 50 100 150 200 250 SE +/- 0.55, N = 3 SE +/- 2.84, N = 4 SE +/- 2.84, N = 4 241.67 241.70 245.98 -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF - MIN: 160.04 / MAX: 556.9 -pie - MIN: 163.92 / MAX: 559.83 -pie - MIN: 176.44 / MAX: 390.61 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Vehicle Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Vehicle Detection FP16-INT8 - Device: CPU CentOS Stream 9 Ubuntu Linux 23.10 Fedora Server 39 4 8 12 16 20 SE +/- 0.01, N = 3 SE +/- 0.03, N = 3 SE +/- 0.02, N = 3 14.36 14.37 14.43 -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF - MIN: 12.55 / MAX: 76.83 -pie - MIN: 12.05 / MAX: 72.49 -pie - MIN: 12.18 / MAX: 71.66 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Face Detection Retail FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Face Detection Retail FP16-INT8 - Device: CPU Ubuntu Linux 23.10 CentOS Stream 9 Fedora Server 39 1.1813 2.3626 3.5439 4.7252 5.9065 SE +/- 0.05, N = 3 SE +/- 0.01, N = 3 SE +/- 0.03, N = 3 5.20 5.25 5.25 -pie - MIN: 4.59 / MAX: 31.78 -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF - MIN: 4.69 / MAX: 58.72 -pie - MIN: 4.62 / MAX: 31.26 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Road Segmentation ADAS FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Road Segmentation ADAS FP16-INT8 - Device: CPU CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 12 24 36 48 60 SE +/- 0.23, N = 3 SE +/- 0.10, N = 3 SE +/- 0.36, N = 3 53.02 53.67 53.79 -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF - MIN: 42.98 / MAX: 178.63 -pie - MIN: 43.29 / MAX: 161.25 -pie - MIN: 44.62 / MAX: 198.12 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Machine Translation EN To DE FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Machine Translation EN To DE FP16 - Device: CPU Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 12 24 36 48 60 SE +/- 0.11, N = 3 SE +/- 0.27, N = 9 SE +/- 1.14, N = 15 27.93 32.67 52.09 -pie - MIN: 19.99 / MAX: 261.46 -pie - MIN: 19.33 / MAX: 484.58 -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF - MIN: 17.25 / MAX: 578.83 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Weld Porosity Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Weld Porosity Detection FP16-INT8 - Device: CPU Ubuntu Linux 23.10 Fedora Server 39 CentOS Stream 9 0.6795 1.359 2.0385 2.718 3.3975 SE +/- 0.00, N = 3 SE +/- 0.01, N = 3 SE +/- 0.02, N = 3 2.85 2.86 3.02 -pie - MIN: 2.08 / MAX: 32.04 -pie - MIN: 1.94 / MAX: 41.9 -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF - MIN: 2.07 / MAX: 29.73 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Person Vehicle Bike Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Person Vehicle Bike Detection FP16 - Device: CPU CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 3 6 9 12 15 SE +/- 0.14, N = 5 SE +/- 0.16, N = 4 SE +/- 0.18, N = 3 12.65 12.69 12.74 -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF - MIN: 10.9 / MAX: 74.36 -pie - MIN: 10.9 / MAX: 75.69 -pie - MIN: 10.88 / MAX: 73.62 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Handwritten English Recognition FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Handwritten English Recognition FP16-INT8 - Device: CPU Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 9 18 27 36 45 SE +/- 0.06, N = 3 SE +/- 0.02, N = 3 SE +/- 0.01, N = 3 39.30 39.76 40.02 -pie - MIN: 33.26 / MAX: 101.52 -pie - MIN: 37.84 / MAX: 104.98 -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF - MIN: 33.6 / MAX: 102.02 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OpenVINO Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2023.2.dev Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU Ubuntu Linux 23.10 CentOS Stream 9 Fedora Server 39 0.1868 0.3736 0.5604 0.7472 0.934 SE +/- 0.00, N = 15 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 0.80 0.83 0.83 -pie - MIN: 0.2 / MAX: 28.24 -isystem -std=c++11 -fPIC -fvisibility=hidden -MD -MT -MF - MIN: 0.21 / MAX: 20.36 -pie - MIN: 0.2 / MAX: 31.65 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv
OSPRay Studio Camera: 1 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Fedora Server 39 Ubuntu Linux 23.10 200 400 600 800 1000 SE +/- 9.82, N = 4 SE +/- 3.48, N = 3 902 914
OSPRay Studio Camera: 2 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Fedora Server 39 Ubuntu Linux 23.10 200 400 600 800 1000 SE +/- 2.03, N = 3 SE +/- 3.53, N = 3 897 911
OSPRay Studio Camera: 3 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Fedora Server 39 Ubuntu Linux 23.10 200 400 600 800 1000 SE +/- 2.96, N = 3 SE +/- 2.40, N = 3 1058 1079
OSPRay Studio Camera: 1 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Ubuntu Linux 23.10 Fedora Server 39 3K 6K 9K 12K 15K SE +/- 3.51, N = 3 SE +/- 22.15, N = 3 12086 12187
OSPRay Studio Camera: 1 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Fedora Server 39 Ubuntu Linux 23.10 7K 14K 21K 28K 35K SE +/- 1257.60, N = 15 SE +/- 481.26, N = 15 28035 32279
OSPRay Studio Camera: 2 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Ubuntu Linux 23.10 Fedora Server 39 3K 6K 9K 12K 15K SE +/- 25.12, N = 3 SE +/- 36.03, N = 3 12243 12262
OSPRay Studio Camera: 2 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Fedora Server 39 Ubuntu Linux 23.10 7K 14K 21K 28K 35K SE +/- 1255.46, N = 15 SE +/- 600.94, N = 15 27524 33174
OSPRay Studio Camera: 3 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Fedora Server 39 Ubuntu Linux 23.10 3K 6K 9K 12K 15K SE +/- 37.95, N = 3 SE +/- 177.62, N = 4 14422 14508
OSPRay Studio Camera: 3 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Fedora Server 39 Ubuntu Linux 23.10 8K 16K 24K 32K 40K SE +/- 119.27, N = 3 SE +/- 564.37, N = 15 34698 37285
OSPRay Studio Camera: 1 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Ubuntu Linux 23.10 Fedora Server 39 50 100 150 200 250 SE +/- 0.33, N = 3 SE +/- 0.88, N = 3 234 235
OSPRay Studio Camera: 2 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Ubuntu Linux 23.10 Fedora Server 39 50 100 150 200 250 SE +/- 0.33, N = 3 SE +/- 0.33, N = 3 233 233
OSPRay Studio Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer - Acceleration: CPU Ubuntu Linux 23.10 Fedora Server 39 60 120 180 240 300 SE +/- 0.00, N = 3 SE +/- 1.45, N = 3 276 276
OSPRay Studio Camera: 1 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Fedora Server 39 Ubuntu Linux 23.10 800 1600 2400 3200 4000 SE +/- 16.22, N = 3 SE +/- 3.21, N = 3 3683 3718
OSPRay Studio Camera: 1 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Fedora Server 39 Ubuntu Linux 23.10 1400 2800 4200 5600 7000 SE +/- 13.35, N = 3 SE +/- 77.79, N = 4 6314 6448
OSPRay Studio Camera: 2 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Fedora Server 39 Ubuntu Linux 23.10 800 1600 2400 3200 4000 SE +/- 43.59, N = 3 SE +/- 7.45, N = 3 3655 3694
OSPRay Studio Camera: 2 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Fedora Server 39 Ubuntu Linux 23.10 1400 2800 4200 5600 7000 SE +/- 28.01, N = 3 SE +/- 60.98, N = 6 6303 6324
OSPRay Studio Camera: 3 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer - Acceleration: CPU Ubuntu Linux 23.10 Fedora Server 39 900 1800 2700 3600 4500 SE +/- 47.59, N = 4 SE +/- 21.55, N = 3 4353 4377
OSPRay Studio Camera: 3 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU OpenBenchmarking.org ms, Fewer Is Better OSPRay Studio 1.0 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer - Acceleration: CPU Ubuntu Linux 23.10 Fedora Server 39 1600 3200 4800 6400 8000 SE +/- 31.90, N = 3 SE +/- 91.12, N = 4 7321 7424
Neural Magic DeepSparse Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream CentOS Stream 9 Ubuntu Linux 23.10 100 200 300 400 500 SE +/- 0.62, N = 3 SE +/- 0.91, N = 3 478.03 478.76
Neural Magic DeepSparse Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream Ubuntu Linux 23.10 CentOS Stream 9 4 8 12 16 20 SE +/- 0.13, N = 7 SE +/- 0.13, N = 7 14.06 14.32
Neural Magic DeepSparse Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream CentOS Stream 9 Ubuntu Linux 23.10 8 16 24 32 40 SE +/- 0.09, N = 3 SE +/- 0.12, N = 3 34.29 34.76
Neural Magic DeepSparse Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream Ubuntu Linux 23.10 CentOS Stream 9 1.3012 2.6024 3.9036 5.2048 6.506 SE +/- 0.0557, N = 7 SE +/- 0.0226, N = 3 5.7515 5.7833
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream Ubuntu Linux 23.10 CentOS Stream 9 20 40 60 80 100 SE +/- 0.46, N = 3 SE +/- 0.41, N = 3 77.02 77.17
Neural Magic DeepSparse Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Stream Ubuntu Linux 23.10 CentOS Stream 9 90 180 270 360 450 SE +/- 2.51, N = 3 SE +/- 0.86, N = 3 409.01 416.68
Neural Magic DeepSparse Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream CentOS Stream 9 Ubuntu Linux 23.10 8 16 24 32 40 SE +/- 0.07, N = 3 SE +/- 0.31, N = 3 34.23 34.99
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream CentOS Stream 9 Ubuntu Linux 23.10 20 40 60 80 100 SE +/- 0.70, N = 3 SE +/- 0.50, N = 3 74.80 74.88
Neural Magic DeepSparse Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream CentOS Stream 9 Ubuntu Linux 23.10 12 24 36 48 60 SE +/- 0.70, N = 3 SE +/- 0.63, N = 3 52.24 52.28
Neural Magic DeepSparse Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream Ubuntu Linux 23.10 CentOS Stream 9 80 160 240 320 400 SE +/- 2.37, N = 3 SE +/- 0.14, N = 3 361.29 366.38
Neural Magic DeepSparse Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream Ubuntu Linux 23.10 CentOS Stream 9 8 16 24 32 40 SE +/- 0.38, N = 3 SE +/- 0.32, N = 3 34.25 35.28
Neural Magic DeepSparse Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.6 Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream CentOS Stream 9 Ubuntu Linux 23.10 100 200 300 400 500 SE +/- 0.34, N = 3 SE +/- 0.78, N = 3 476.58 477.14
CloverLeaf Input: clover_bm16 OpenBenchmarking.org Seconds, Fewer Is Better CloverLeaf 1.3 Input: clover_bm16 Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 70 140 210 280 350 SE +/- 1.01, N = 3 SE +/- 3.60, N = 3 SE +/- 1.51, N = 3 255.39 279.21 314.49 1. (F9X) gfortran options: -O3 -march=native -funroll-loops -fopenmp
CloverLeaf Input: clover_bm64_short OpenBenchmarking.org Seconds, Fewer Is Better CloverLeaf 1.3 Input: clover_bm64_short CentOS Stream 9 Fedora Server 39 Ubuntu Linux 23.10 7 14 21 28 35 SE +/- 0.12, N = 3 SE +/- 0.40, N = 3 SE +/- 0.23, N = 11 25.01 31.63 32.10 1. (F9X) gfortran options: -O3 -march=native -funroll-loops -fopenmp
NWChem Input: C240 Buckyball OpenBenchmarking.org Seconds, Fewer Is Better NWChem 7.0.2 Input: C240 Buckyball Ubuntu Linux 23.10 400 800 1200 1600 2000 1750.3 1. (F9X) gfortran options: -lnwctask -lccsd -lmcscf -lselci -lmp2 -lmoints -lstepper -ldriver -loptim -lnwdft -lgradients -lcphf -lesp -lddscf -ldangchang -lguess -lhessian -lvib -lnwcutil -lrimp2 -lproperty -lsolvation -lnwints -lprepar -lnwmd -lnwpw -lofpw -lpaw -lpspw -lband -lnwpwlib -lcafe -lspace -lanalyze -lqhop -lpfft -ldplot -ldrdy -lvscf -lqmmm -lqmd -letrans -ltce -lbq -lmm -lcons -lperfm -ldntmc -lccca -ldimqm -lga -larmci -lpeigs -l64to32 -lopenblas -lpthread -lrt -llapack -lnwcblas -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz -lcomex -m64 -ffast-math -std=legacy -fdefault-integer-8 -finline-functions -O2
Xcompact3d Incompact3d Input: X3D-benchmarking input.i3d OpenBenchmarking.org Seconds, Fewer Is Better Xcompact3d Incompact3d 2021-03-11 Input: X3D-benchmarking input.i3d Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 40 80 120 160 200 SE +/- 0.67, N = 3 SE +/- 0.44, N = 3 SE +/- 0.56, N = 3 168.67 169.18 184.19 -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 1. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -lmpi_usempif08 -lmpi_mpifh -lmpi
easyWave Input: e2Asean Grid + BengkuluSept2007 Source - Time: 2400 OpenBenchmarking.org Seconds, Fewer Is Better easyWave r34 Input: e2Asean Grid + BengkuluSept2007 Source - Time: 2400 Fedora Server 39 Ubuntu Linux 23.10 CentOS Stream 9 40 80 120 160 200 SE +/- 15.88, N = 12 SE +/- 8.98, N = 12 SE +/- 10.40, N = 12 156.24 177.75 205.01 1. (CXX) g++ options: -O3 -fopenmp
Y-Cruncher Pi Digits To Calculate: 1B OpenBenchmarking.org Seconds, Fewer Is Better Y-Cruncher 0.8.3 Pi Digits To Calculate: 1B Ubuntu Linux 23.10 CentOS Stream 9 Fedora Server 39 2 4 6 8 10 SE +/- 0.037, N = 3 SE +/- 0.045, N = 3 SE +/- 0.088, N = 5 7.207 7.913 8.270
Y-Cruncher Pi Digits To Calculate: 5B OpenBenchmarking.org Seconds, Fewer Is Better Y-Cruncher 0.8.3 Pi Digits To Calculate: 5B Ubuntu Linux 23.10 CentOS Stream 9 Fedora Server 39 8 16 24 32 40 SE +/- 0.07, N = 3 SE +/- 0.17, N = 3 SE +/- 0.43, N = 3 29.82 32.85 33.73
RawTherapee Total Benchmark Time OpenBenchmarking.org Seconds, Fewer Is Better RawTherapee Total Benchmark Time Fedora Server 39 Ubuntu Linux 23.10 30 60 90 120 150 SE +/- 0.45, N = 3 SE +/- 0.65, N = 3 118.23 129.08 1. Fedora Server 39: RawTherapee, version 5.10, command line. 2. Ubuntu Linux 23.10: RawTherapee, version 5.9, command line.
GPAW Input: Carbon Nanotube OpenBenchmarking.org Seconds, Fewer Is Better GPAW 23.6 Input: Carbon Nanotube Ubuntu Linux 23.10 Fedora Server 39 40 80 120 160 200 SE +/- 0.35, N = 15 SE +/- 0.94, N = 3 38.64 194.51 -fwrapv -O2 -fno-strict-overflow -fcf-protection -fexceptions -fPIC -UNDEBUG -std=c99 1. (CC) gcc options: -shared -lxc -lblas -lmpi
Phoronix Test Suite v10.8.4