dddxxx Tests for a future article. Intel Core i7-8565U testing with a Dell 0KTW76 (1.17.0 BIOS) and Intel UHD 620 WHL GT2 15GB on Ubuntu 22.04 via the Phoronix Test Suite.
HTML result view exported from: https://openbenchmarking.org/result/2308061-NE-DDDXXX46317&sor .
dddxxx Processor Motherboard Chipset Memory Disk Graphics Audio Network OS Kernel Desktop Display Server OpenGL OpenCL Vulkan Compiler File-System Screen Resolution a b Intel Core i7-8565U @ 4.60GHz (4 Cores / 8 Threads) Dell 0KTW76 (1.17.0 BIOS) Intel Cannon Point-LP 16GB SK hynix PC401 NVMe 256GB Intel UHD 620 WHL GT2 15GB (1150MHz) Realtek ALC3271 Qualcomm Atheros QCA6174 802.11ac Ubuntu 22.04 5.19.0-rc6-phx-retbleed (x86_64) GNOME Shell 42.2 X Server + Wayland 4.6 Mesa 22.0.1 OpenCL 3.0 1.3.204 GCC 11.3.0 ext4 1920x1080 GCC 11.4.0 OpenBenchmarking.org Kernel Details - Transparent Huge Pages: madvise Compiler Details - a: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-xKiWfi/gcc-11-11.3.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-xKiWfi/gcc-11-11.3.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - b: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-XeT9lY/gcc-11-11.4.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-XeT9lY/gcc-11-11.4.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Disk Details - NONE / errors=remount-ro,relatime,rw / Block Size: 4096 Processor Details - Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0xf0 - Thermald 2.4.9 Java Details - a: OpenJDK Runtime Environment (build 11.0.18+10-post-Ubuntu-0ubuntu122.04) - b: OpenJDK Runtime Environment (build 11.0.20+8-post-Ubuntu-1ubuntu122.04) Python Details - a: Python 3.10.6 - b: Python 3.10.12 Security Details - itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Not affected + mds: Mitigation of Clear buffers; SMT vulnerable + meltdown: Not affected + mmio_stale_data: Mitigation of Clear buffers; SMT vulnerable + retbleed: Mitigation of IBRS + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of IBRS IBPB: conditional RSB filling + srbds: Mitigation of Microcode + tsx_async_abort: Not affected
dddxxx sqlite: 1 sqlite: 2 sqlite: 4 vkpeak: fp32-scalar xonotic: 1920 x 1080 - Low xonotic: 1920 x 1080 - High xonotic: 1920 x 1080 - Ultra xonotic: 1920 x 1080 - Ultimate quantlib: libxsmm: 128 libxsmm: 32 libxsmm: 64 z3: 1.smt2 z3: 2.smt2 dav1d: Chimera 1080p dav1d: Summer Nature 4K dav1d: Summer Nature 1080p dav1d: Chimera 1080p 10-bit embree: Pathtracer - Crown embree: Pathtracer ISPC - Crown embree: Pathtracer - Asian Dragon embree: Pathtracer - Asian Dragon Obj embree: Pathtracer ISPC - Asian Dragon embree: Pathtracer ISPC - Asian Dragon Obj svt-av1: Preset 4 - Bosphorus 4K svt-av1: Preset 8 - Bosphorus 4K svt-av1: Preset 12 - Bosphorus 4K svt-av1: Preset 13 - Bosphorus 4K svt-av1: Preset 4 - Bosphorus 1080p svt-av1: Preset 8 - Bosphorus 1080p svt-av1: Preset 12 - Bosphorus 1080p svt-av1: Preset 13 - Bosphorus 1080p vvenc: Bosphorus 4K - Fast vvenc: Bosphorus 4K - Faster vvenc: Bosphorus 1080p - Fast vvenc: Bosphorus 1080p - Faster oidn: RT.hdr_alb_nrm.3840x2160 - CPU-Only oidn: RT.ldr_alb_nrm.3840x2160 - CPU-Only oidn: RTLightmap.hdr.4096x4096 - CPU-Only ospray: particle_volume/ao/real_time ospray: particle_volume/scivis/real_time ospray: particle_volume/pathtracer/real_time ospray: gravity_spheres_volume/dim_512/ao/real_time ospray: gravity_spheres_volume/dim_512/scivis/real_time ospray: gravity_spheres_volume/dim_512/pathtracer/real_time build-godot: Time To Compile build-llvm: Ninja build-llvm: Unix Makefiles build2: Time To Compile encode-opus: WAV To Opus Encode liquid-dsp: 1 - 256 - 32 liquid-dsp: 1 - 256 - 57 liquid-dsp: 2 - 256 - 32 liquid-dsp: 2 - 256 - 57 liquid-dsp: 4 - 256 - 32 liquid-dsp: 4 - 256 - 57 liquid-dsp: 8 - 256 - 32 liquid-dsp: 8 - 256 - 57 liquid-dsp: 1 - 256 - 512 liquid-dsp: 2 - 256 - 512 liquid-dsp: 4 - 256 - 512 liquid-dsp: 8 - 256 - 512 apache-iotdb: 100 - 1 - 200 apache-iotdb: 100 - 1 - 200 apache-iotdb: 100 - 1 - 500 apache-iotdb: 100 - 1 - 500 apache-iotdb: 200 - 1 - 200 apache-iotdb: 200 - 1 - 200 apache-iotdb: 200 - 1 - 500 apache-iotdb: 200 - 1 - 500 apache-iotdb: 500 - 1 - 200 apache-iotdb: 500 - 1 - 200 apache-iotdb: 500 - 1 - 500 apache-iotdb: 500 - 1 - 500 apache-iotdb: 100 - 100 - 200 apache-iotdb: 100 - 100 - 200 apache-iotdb: 100 - 100 - 500 apache-iotdb: 100 - 100 - 500 apache-iotdb: 200 - 100 - 200 apache-iotdb: 200 - 100 - 200 apache-iotdb: 200 - 100 - 500 apache-iotdb: 200 - 100 - 500 apache-iotdb: 500 - 100 - 200 apache-iotdb: 500 - 100 - 200 apache-iotdb: 500 - 100 - 500 apache-iotdb: 500 - 100 - 500 memcached: 1:5 memcached: 1:10 memcached: 1:100 deepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Stream deepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Stream deepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Asynchronous Multi-Stream deepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Asynchronous Multi-Stream deepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Asynchronous Multi-Stream deepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Asynchronous Multi-Stream deepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Stream deepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Stream deepsparse: ResNet-50, Baseline - Asynchronous Multi-Stream deepsparse: ResNet-50, Baseline - Asynchronous Multi-Stream deepsparse: ResNet-50, Sparse INT8 - Asynchronous Multi-Stream deepsparse: ResNet-50, Sparse INT8 - Asynchronous Multi-Stream deepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Stream deepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Stream deepsparse: BERT-Large, NLP Question Answering - Asynchronous Multi-Stream deepsparse: BERT-Large, NLP Question Answering - Asynchronous Multi-Stream deepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Stream deepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Stream deepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Stream deepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Stream deepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Stream deepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Stream deepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Stream deepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Stream deepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Stream deepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Stream deepsparse: NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Stream deepsparse: NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Stream deepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Stream deepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Stream memtier-benchmark: Redis - 50 - 1:5 memtier-benchmark: Redis - 100 - 1:5 memtier-benchmark: Redis - 50 - 1:10 memtier-benchmark: Redis - 500 - 1:5 memtier-benchmark: Redis - 100 - 1:10 memtier-benchmark: Redis - 500 - 1:10 stress-ng: Hash stress-ng: MMAP stress-ng: NUMA stress-ng: Pipe stress-ng: Poll stress-ng: Zlib stress-ng: Futex stress-ng: MEMFD stress-ng: Mutex stress-ng: Atomic stress-ng: Crypto stress-ng: Malloc stress-ng: Cloning stress-ng: Forking stress-ng: Pthread stress-ng: AVL Tree stress-ng: IO_uring stress-ng: SENDFILE stress-ng: CPU Cache stress-ng: CPU Stress stress-ng: Semaphores stress-ng: Matrix Math stress-ng: Vector Math stress-ng: Function Call stress-ng: x86_64 RdRand stress-ng: Floating Point stress-ng: Matrix 3D Math stress-ng: Memory Copying stress-ng: Vector Shuffle stress-ng: Socket Activity stress-ng: Wide Vector Math stress-ng: Context Switching stress-ng: Fused Multiply-Add stress-ng: Vector Floating Point stress-ng: Glibc C String Functions stress-ng: Glibc Qsort Data Sorting stress-ng: System V Message Passing ncnn: CPU - mobilenet ncnn: CPU-v2-v2 - mobilenet-v2 ncnn: CPU-v3-v3 - mobilenet-v3 ncnn: CPU - shufflenet-v2 ncnn: CPU - mnasnet ncnn: CPU - efficientnet-b0 ncnn: CPU - blazeface ncnn: CPU - googlenet ncnn: CPU - vgg16 ncnn: CPU - resnet18 ncnn: CPU - alexnet ncnn: CPU - resnet50 ncnn: CPU - yolov4-tiny ncnn: CPU - squeezenet_ssd ncnn: CPU - regnety_400m ncnn: CPU - vision_transformer ncnn: CPU - FastestDet ncnn: Vulkan GPU - mobilenet ncnn: Vulkan GPU-v2-v2 - mobilenet-v2 ncnn: Vulkan GPU-v3-v3 - mobilenet-v3 ncnn: Vulkan GPU - shufflenet-v2 ncnn: Vulkan GPU - mnasnet ncnn: Vulkan GPU - efficientnet-b0 ncnn: Vulkan GPU - blazeface ncnn: Vulkan GPU - googlenet ncnn: Vulkan GPU - vgg16 ncnn: Vulkan GPU - resnet18 ncnn: Vulkan GPU - alexnet ncnn: Vulkan GPU - resnet50 ncnn: Vulkan GPU - yolov4-tiny ncnn: Vulkan GPU - squeezenet_ssd ncnn: Vulkan GPU - regnety_400m ncnn: Vulkan GPU - vision_transformer ncnn: Vulkan GPU - FastestDet cassandra: Writes a b 30.327 123.465 125.830 268.33 205.6062326 91.1145989 77.9338736 59.1708882 2379.8 79.9 47.7 90.9 43.973 130.941 244.12 67.01 273.35 191.30 3.5218 3.7369 4.2183 3.8229 4.7883 4.1670 1.092 9.490 35.497 34.049 4.260 35.739 154.597 202.852 1.180 2.693 3.905 9.882 0.12 0.12 0.06 1.49334 1.37399 52.5904 0.745146 0.622182 1.06164 1372.604 2892.088 2967.528 555.936 35.396 46516500 43429000 85550500 73795500 140280000 119590000 177945000 140345000 7721800 15179500 25830500 35719000 529599 22.62 996242.81 34.81 816344.2 17.34 1333775.59 29.65 1036429.93 15.75 611157.73 75.4 15017343.49 107.84 13467917.31 342.21 12613348.93 144.53 8609112.97 550.47 8900311.03 212.43 4948660.19 984.7 525368.39 502050.43 496309.76 2.3527 855.6769 65.4250 30.5398 28.9185 69.5593 8.3009 244.5617 30.5673 65.9517 196.7843 10.2465 14.6720 137.9489 2.6677 764.2676 29.6585 68.0491 12.5441 162.1647 17.5124 116.2683 2.9815 669.1809 21.2985 93.8482 9.4553 211.9830 2.0309 979.1797 1039311.12 1016703.27 1109812.44 890374.23 1028641.19 914462.02 523699.70 20.86 53.95 1422109.84 284300.80 273.05 486078.68 56.59 604156.72 224.51 4590.53 375933.99 669.41 9551.79 34974.26 16.41 170492.16 36779.86 929178.43 7310.05 3218725.51 17318.7 12666.19 2058.70 3267.31 743.83 383.49 1078.73 2431.22 2495.57 147314.85 705297.96 2978232.73 7289.32 2485856.07 69.13 2652462.62 25.93 6.69 5.14 4.29 6.04 9.78 0.91 16.23 97.42 12.61 11.68 36.02 38.37 16.55 10.03 233.96 5.39 28.15 6.67 5.11 3.66 5.23 9.22 0.94 16.29 97.70 12.73 11.67 33.18 39.39 15.97 9.82 240.03 5.66 26422 88.585 170.014 173.045 268.33 205.4486260 90.9080095 77.5195887 58.9499391 2404.8 83.7 48.0 90.8 43.419 131.405 249.62 66.38 310.94 190.87 3.4542 3.6962 4.1998 3.8503 4.7636 3.9923 0.948 9.436 28.978 30.975 4.773 31.014 157.692 206.769 1.170 2.672 3.790 9.868 0.12 0.12 0.06 1.3692 1.34467 50.0829 0.734783 0.709986 1.012953 1384.766 2908.878 2929.911 569.609 35.426 46825000 43427500 85664000 73620500 139950000 112205000 185540000 132830000 7710850 15023500 25812500 36523000 524660.18 22.93 1009840.28 34.48 812170.86 16.68 1298988.31 29.75 1040683.13 15.11 637041.62 70.46 15526243.84 104.43 12006039.14 387.15 12464073 143.24 9394851.37 485.93 9246224.42 198.41 4629287.79 1026.93 518308.20 486807.49 485922.63 2.3352 864.9932 65.3564 30.5664 28.4640 70.8213 8.2832 245.6430 29.7603 68.0003 195.4858 10.336 14.6069 139.0475 3.2461 614.6793 28.9622 69.6181 12.9502 154.2586 20.7264 98.7143 3.2320 628.3639 34.3231 58.2368 11.1719 179.0772 2.3883 835.8517 1398883.01 1372033.18 1455171.86 1248610.70 1377833.48 1253523.69 638650.51 28.46 60.49 1533355.11 318466.50 335.91 624014.17 43.09 694562.47 249.99 5684.53 434156.10 715.58 12140.60 42229.61 19.3 182936.68 43752.37 1153192.75 7866.4 3241408.21 17116.79 12444.33 2164.20 3134.04 762.86 396.03 1088.75 2431.01 2508.63 148388.79 733269.79 2737483.30 7877.70 2360589.10 74.13 2519916.10 27.68 6.15 4.31 3.02 4.49 9.43 0.93 16.19 98.75 12.62 11.64 34.33 39.31 15.80 10.12 232.17 5.55 27.38 6.67 5.11 3.67 5.26 9.34 0.96 16.19 96.26 12.56 11.69 33.60 39.17 15.48 10.41 235.51 5.47 26438 OpenBenchmarking.org
SQLite Threads / Copies: 1 OpenBenchmarking.org Seconds, Fewer Is Better SQLite 3.41.2 Threads / Copies: 1 a b 20 40 60 80 100 SE +/- 0.86, N = 2 SE +/- 0.19, N = 2 30.33 88.59 1. (CC) gcc options: -O2 -lreadline -ltermcap -lz -lm
SQLite Threads / Copies: 2 OpenBenchmarking.org Seconds, Fewer Is Better SQLite 3.41.2 Threads / Copies: 2 a b 40 80 120 160 200 SE +/- 24.93, N = 2 SE +/- 1.42, N = 2 123.47 170.01 1. (CC) gcc options: -O2 -lreadline -ltermcap -lz -lm
SQLite Threads / Copies: 4 OpenBenchmarking.org Seconds, Fewer Is Better SQLite 3.41.2 Threads / Copies: 4 a b 40 80 120 160 200 SE +/- 17.49, N = 2 SE +/- 0.50, N = 2 125.83 173.05 1. (CC) gcc options: -O2 -lreadline -ltermcap -lz -lm
vkpeak fp32-scalar OpenBenchmarking.org GFLOPS, More Is Better vkpeak 20230730 fp32-scalar b a 60 120 180 240 300 SE +/- 0.01, N = 2 SE +/- 0.00, N = 2 268.33 268.33
Xonotic Resolution: 1920 x 1080 - Effects Quality: Low OpenBenchmarking.org Frames Per Second, More Is Better Xonotic 0.8.6 Resolution: 1920 x 1080 - Effects Quality: Low a b 50 100 150 200 250 SE +/- 0.72, N = 2 SE +/- 0.60, N = 2 205.61 205.45 MIN: 99 / MAX: 336 MIN: 95 / MAX: 334
Xonotic Resolution: 1920 x 1080 - Effects Quality: High OpenBenchmarking.org Frames Per Second, More Is Better Xonotic 0.8.6 Resolution: 1920 x 1080 - Effects Quality: High a b 20 40 60 80 100 SE +/- 0.04, N = 2 SE +/- 0.32, N = 2 91.11 90.91 MIN: 39 / MAX: 130 MIN: 38 / MAX: 130
Xonotic Resolution: 1920 x 1080 - Effects Quality: Ultra OpenBenchmarking.org Frames Per Second, More Is Better Xonotic 0.8.6 Resolution: 1920 x 1080 - Effects Quality: Ultra a b 20 40 60 80 100 SE +/- 0.16, N = 2 SE +/- 0.41, N = 2 77.93 77.52 MIN: 34 / MAX: 120 MIN: 32 / MAX: 120
Xonotic Resolution: 1920 x 1080 - Effects Quality: Ultimate OpenBenchmarking.org Frames Per Second, More Is Better Xonotic 0.8.6 Resolution: 1920 x 1080 - Effects Quality: Ultimate a b 13 26 39 52 65 SE +/- 0.07, N = 2 SE +/- 0.06, N = 2 59.17 58.95 MIN: 24 / MAX: 94 MIN: 24 / MAX: 94
QuantLib OpenBenchmarking.org MFLOPS, More Is Better QuantLib 1.30 b a 500 1000 1500 2000 2500 SE +/- 234.25, N = 2 SE +/- 210.20, N = 2 2404.8 2379.8 1. (CXX) g++ options: -O3 -march=native -fPIE -pie
libxsmm M N K: 128 OpenBenchmarking.org GFLOPS/s, More Is Better libxsmm 2-1.17-3645 M N K: 128 b a 20 40 60 80 100 SE +/- 4.10, N = 2 SE +/- 4.60, N = 2 83.7 79.9 1. (CXX) g++ options: -dynamic -Bstatic -static-libgcc -lgomp -lm -lrt -ldl -lquadmath -lstdc++ -pthread -fPIC -std=c++14 -O2 -fopenmp-simd -funroll-loops -ftree-vectorize -fdata-sections -ffunction-sections -fvisibility=hidden -march=core-avx2
libxsmm M N K: 32 OpenBenchmarking.org GFLOPS/s, More Is Better libxsmm 2-1.17-3645 M N K: 32 b a 11 22 33 44 55 SE +/- 0.55, N = 2 SE +/- 0.35, N = 2 48.0 47.7 1. (CXX) g++ options: -dynamic -Bstatic -static-libgcc -lgomp -lm -lrt -ldl -lquadmath -lstdc++ -pthread -fPIC -std=c++14 -O2 -fopenmp-simd -funroll-loops -ftree-vectorize -fdata-sections -ffunction-sections -fvisibility=hidden -march=core-avx2
libxsmm M N K: 64 OpenBenchmarking.org GFLOPS/s, More Is Better libxsmm 2-1.17-3645 M N K: 64 a b 20 40 60 80 100 SE +/- 4.45, N = 2 SE +/- 4.80, N = 2 90.9 90.8 1. (CXX) g++ options: -dynamic -Bstatic -static-libgcc -lgomp -lm -lrt -ldl -lquadmath -lstdc++ -pthread -fPIC -std=c++14 -O2 -fopenmp-simd -funroll-loops -ftree-vectorize -fdata-sections -ffunction-sections -fvisibility=hidden -march=core-avx2
Z3 Theorem Prover SMT File: 1.smt2 OpenBenchmarking.org Seconds, Fewer Is Better Z3 Theorem Prover 4.12.1 SMT File: 1.smt2 b a 10 20 30 40 50 SE +/- 0.18, N = 2 SE +/- 0.05, N = 2 43.42 43.97 1. (CXX) g++ options: -lpthread -std=c++17 -fvisibility=hidden -mfpmath=sse -msse -msse2 -O3 -fPIC
Z3 Theorem Prover SMT File: 2.smt2 OpenBenchmarking.org Seconds, Fewer Is Better Z3 Theorem Prover 4.12.1 SMT File: 2.smt2 a b 30 60 90 120 150 SE +/- 0.11, N = 2 SE +/- 0.05, N = 2 130.94 131.41 1. (CXX) g++ options: -lpthread -std=c++17 -fvisibility=hidden -mfpmath=sse -msse -msse2 -O3 -fPIC
dav1d Video Input: Chimera 1080p OpenBenchmarking.org FPS, More Is Better dav1d 1.2.1 Video Input: Chimera 1080p b a 50 100 150 200 250 SE +/- 1.95, N = 2 SE +/- 2.76, N = 2 249.62 244.12 1. (CC) gcc options: -pthread -lm
dav1d Video Input: Summer Nature 4K OpenBenchmarking.org FPS, More Is Better dav1d 1.2.1 Video Input: Summer Nature 4K a b 15 30 45 60 75 SE +/- 0.31, N = 2 SE +/- 2.35, N = 2 67.01 66.38 1. (CC) gcc options: -pthread -lm
dav1d Video Input: Summer Nature 1080p OpenBenchmarking.org FPS, More Is Better dav1d 1.2.1 Video Input: Summer Nature 1080p b a 70 140 210 280 350 SE +/- 2.55, N = 2 SE +/- 4.15, N = 2 310.94 273.35 1. (CC) gcc options: -pthread -lm
dav1d Video Input: Chimera 1080p 10-bit OpenBenchmarking.org FPS, More Is Better dav1d 1.2.1 Video Input: Chimera 1080p 10-bit a b 40 80 120 160 200 SE +/- 19.41, N = 2 SE +/- 12.81, N = 2 191.30 190.87 1. (CC) gcc options: -pthread -lm
Embree Binary: Pathtracer - Model: Crown OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.1 Binary: Pathtracer - Model: Crown a b 0.7924 1.5848 2.3772 3.1696 3.962 SE +/- 0.0068, N = 2 SE +/- 0.0668, N = 2 3.5218 3.4542 MIN: 2.81 / MAX: 4.3 MIN: 2.82 / MAX: 4.4
Embree Binary: Pathtracer ISPC - Model: Crown OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.1 Binary: Pathtracer ISPC - Model: Crown a b 0.8408 1.6816 2.5224 3.3632 4.204 SE +/- 0.0660, N = 2 SE +/- 0.1040, N = 2 3.7369 3.6962 MIN: 3.08 / MAX: 4.7 MIN: 3.06 / MAX: 4.78
Embree Binary: Pathtracer - Model: Asian Dragon OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.1 Binary: Pathtracer - Model: Asian Dragon a b 0.9491 1.8982 2.8473 3.7964 4.7455 SE +/- 0.0414, N = 2 SE +/- 0.0239, N = 2 4.2183 4.1998 MIN: 3.54 / MAX: 5.31 MIN: 3.54 / MAX: 5.43
Embree Binary: Pathtracer - Model: Asian Dragon Obj OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.1 Binary: Pathtracer - Model: Asian Dragon Obj b a 0.8663 1.7326 2.5989 3.4652 4.3315 SE +/- 0.0122, N = 2 SE +/- 0.0521, N = 2 3.8503 3.8229 MIN: 3.22 / MAX: 4.83 MIN: 3.19 / MAX: 4.76
Embree Binary: Pathtracer ISPC - Model: Asian Dragon OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.1 Binary: Pathtracer ISPC - Model: Asian Dragon a b 1.0774 2.1548 3.2322 4.3096 5.387 SE +/- 0.0697, N = 2 SE +/- 0.0364, N = 2 4.7883 4.7636 MIN: 4.05 / MAX: 5.89 MIN: 4.05 / MAX: 5.98
Embree Binary: Pathtracer ISPC - Model: Asian Dragon Obj OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.1 Binary: Pathtracer ISPC - Model: Asian Dragon Obj a b 0.9376 1.8752 2.8128 3.7504 4.688 SE +/- 0.0272, N = 2 SE +/- 0.1882, N = 2 4.1670 3.9923 MIN: 3.5 / MAX: 5.07 MIN: 3.51 / MAX: 5.03
SVT-AV1 Encoder Mode: Preset 4 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.6 Encoder Mode: Preset 4 - Input: Bosphorus 4K a b 0.2457 0.4914 0.7371 0.9828 1.2285 SE +/- 0.023, N = 2 SE +/- 0.003, N = 2 1.092 0.948 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 8 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.6 Encoder Mode: Preset 8 - Input: Bosphorus 4K a b 3 6 9 12 15 SE +/- 0.307, N = 2 SE +/- 1.160, N = 2 9.490 9.436 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 12 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.6 Encoder Mode: Preset 12 - Input: Bosphorus 4K a b 8 16 24 32 40 SE +/- 0.91, N = 2 SE +/- 0.81, N = 2 35.50 28.98 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 13 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.6 Encoder Mode: Preset 13 - Input: Bosphorus 4K a b 8 16 24 32 40 SE +/- 2.10, N = 2 SE +/- 2.71, N = 2 34.05 30.98 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 4 - Input: Bosphorus 1080p OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.6 Encoder Mode: Preset 4 - Input: Bosphorus 1080p b a 1.0739 2.1478 3.2217 4.2956 5.3695 SE +/- 0.109, N = 2 SE +/- 0.495, N = 2 4.773 4.260 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 8 - Input: Bosphorus 1080p OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.6 Encoder Mode: Preset 8 - Input: Bosphorus 1080p a b 8 16 24 32 40 SE +/- 0.87, N = 2 SE +/- 2.58, N = 2 35.74 31.01 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 12 - Input: Bosphorus 1080p OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.6 Encoder Mode: Preset 12 - Input: Bosphorus 1080p b a 30 60 90 120 150 SE +/- 0.70, N = 2 SE +/- 0.24, N = 2 157.69 154.60 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 13 - Input: Bosphorus 1080p OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.6 Encoder Mode: Preset 13 - Input: Bosphorus 1080p b a 50 100 150 200 250 SE +/- 0.84, N = 2 SE +/- 0.66, N = 2 206.77 202.85 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
VVenC Video Input: Bosphorus 4K - Video Preset: Fast OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.9 Video Input: Bosphorus 4K - Video Preset: Fast a b 0.2655 0.531 0.7965 1.062 1.3275 SE +/- 0.013, N = 2 SE +/- 0.013, N = 2 1.180 1.170 1. (CXX) g++ options: -O3 -flto -fno-fat-lto-objects -flto=auto
VVenC Video Input: Bosphorus 4K - Video Preset: Faster OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.9 Video Input: Bosphorus 4K - Video Preset: Faster a b 0.6059 1.2118 1.8177 2.4236 3.0295 SE +/- 0.052, N = 2 SE +/- 0.045, N = 2 2.693 2.672 1. (CXX) g++ options: -O3 -flto -fno-fat-lto-objects -flto=auto
VVenC Video Input: Bosphorus 1080p - Video Preset: Fast OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.9 Video Input: Bosphorus 1080p - Video Preset: Fast a b 0.8786 1.7572 2.6358 3.5144 4.393 SE +/- 0.015, N = 2 SE +/- 0.042, N = 2 3.905 3.790 1. (CXX) g++ options: -O3 -flto -fno-fat-lto-objects -flto=auto
VVenC Video Input: Bosphorus 1080p - Video Preset: Faster OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.9 Video Input: Bosphorus 1080p - Video Preset: Faster a b 3 6 9 12 15 SE +/- 0.090, N = 2 SE +/- 0.603, N = 2 9.882 9.868 1. (CXX) g++ options: -O3 -flto -fno-fat-lto-objects -flto=auto
Intel Open Image Denoise Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.0 Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Only b a 0.027 0.054 0.081 0.108 0.135 SE +/- 0.00, N = 2 SE +/- 0.00, N = 2 0.12 0.12
Intel Open Image Denoise Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.0 Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Only b a 0.027 0.054 0.081 0.108 0.135 SE +/- 0.00, N = 2 SE +/- 0.00, N = 2 0.12 0.12
Intel Open Image Denoise Run: RTLightmap.hdr.4096x4096 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.0 Run: RTLightmap.hdr.4096x4096 - Device: CPU-Only b a 0.0135 0.027 0.0405 0.054 0.0675 SE +/- 0.00, N = 2 SE +/- 0.00, N = 2 0.06 0.06
OSPRay Benchmark: particle_volume/ao/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 2.12 Benchmark: particle_volume/ao/real_time a b 0.336 0.672 1.008 1.344 1.68 SE +/- 0.09686, N = 2 SE +/- 0.02805, N = 2 1.49334 1.36920
OSPRay Benchmark: particle_volume/scivis/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 2.12 Benchmark: particle_volume/scivis/real_time a b 0.3091 0.6182 0.9273 1.2364 1.5455 SE +/- 0.00084, N = 2 SE +/- 0.00287, N = 2 1.37399 1.34467
OSPRay Benchmark: particle_volume/pathtracer/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 2.12 Benchmark: particle_volume/pathtracer/real_time a b 12 24 36 48 60 SE +/- 0.59, N = 2 SE +/- 0.02, N = 2 52.59 50.08
OSPRay Benchmark: gravity_spheres_volume/dim_512/ao/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 2.12 Benchmark: gravity_spheres_volume/dim_512/ao/real_time a b 0.1677 0.3354 0.5031 0.6708 0.8385 SE +/- 0.009807, N = 2 SE +/- 0.008961, N = 2 0.745146 0.734783
OSPRay Benchmark: gravity_spheres_volume/dim_512/scivis/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 2.12 Benchmark: gravity_spheres_volume/dim_512/scivis/real_time b a 0.1597 0.3194 0.4791 0.6388 0.7985 SE +/- 0.025213, N = 2 SE +/- 0.014296, N = 2 0.709986 0.622182
OSPRay Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_time OpenBenchmarking.org Items Per Second, More Is Better OSPRay 2.12 Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_time a b 0.2389 0.4778 0.7167 0.9556 1.1945 SE +/- 0.011835, N = 2 SE +/- 0.016867, N = 2 1.061640 1.012953
Timed Godot Game Engine Compilation Time To Compile OpenBenchmarking.org Seconds, Fewer Is Better Timed Godot Game Engine Compilation 4.0 Time To Compile a b 300 600 900 1200 1500 SE +/- 0.29, N = 2 SE +/- 12.96, N = 2 1372.60 1384.77
Timed LLVM Compilation Build System: Ninja OpenBenchmarking.org Seconds, Fewer Is Better Timed LLVM Compilation 16.0 Build System: Ninja a b 600 1200 1800 2400 3000 SE +/- 4.75, N = 2 SE +/- 4.95, N = 2 2892.09 2908.88
Timed LLVM Compilation Build System: Unix Makefiles OpenBenchmarking.org Seconds, Fewer Is Better Timed LLVM Compilation 16.0 Build System: Unix Makefiles b a 600 1200 1800 2400 3000 SE +/- 7.61, N = 2 SE +/- 45.61, N = 2 2929.91 2967.53
Build2 Time To Compile OpenBenchmarking.org Seconds, Fewer Is Better Build2 0.15 Time To Compile a b 120 240 360 480 600 SE +/- 1.14, N = 2 SE +/- 4.22, N = 2 555.94 569.61
Opus Codec Encoding WAV To Opus Encode OpenBenchmarking.org Seconds, Fewer Is Better Opus Codec Encoding 1.4 WAV To Opus Encode a b 8 16 24 32 40 SE +/- 0.10, N = 2 SE +/- 0.06, N = 2 35.40 35.43 1. (CXX) g++ options: -O3 -fvisibility=hidden -logg -lm
Liquid-DSP Threads: 1 - Buffer Length: 256 - Filter Length: 32 OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 1.6 Threads: 1 - Buffer Length: 256 - Filter Length: 32 b a 10M 20M 30M 40M 50M SE +/- 77000.00, N = 2 SE +/- 451500.00, N = 2 46825000 46516500 1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid
Liquid-DSP Threads: 1 - Buffer Length: 256 - Filter Length: 57 OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 1.6 Threads: 1 - Buffer Length: 256 - Filter Length: 57 a b 9M 18M 27M 36M 45M SE +/- 225000.00, N = 2 SE +/- 154500.00, N = 2 43429000 43427500 1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid
Liquid-DSP Threads: 2 - Buffer Length: 256 - Filter Length: 32 OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 1.6 Threads: 2 - Buffer Length: 256 - Filter Length: 32 b a 20M 40M 60M 80M 100M SE +/- 1161000.00, N = 2 SE +/- 873500.00, N = 2 85664000 85550500 1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid
Liquid-DSP Threads: 2 - Buffer Length: 256 - Filter Length: 57 OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 1.6 Threads: 2 - Buffer Length: 256 - Filter Length: 57 a b 16M 32M 48M 64M 80M SE +/- 128500.00, N = 2 SE +/- 620500.00, N = 2 73795500 73620500 1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid
Liquid-DSP Threads: 4 - Buffer Length: 256 - Filter Length: 32 OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 1.6 Threads: 4 - Buffer Length: 256 - Filter Length: 32 a b 30M 60M 90M 120M 150M SE +/- 2650000.00, N = 2 SE +/- 2610000.00, N = 2 140280000 139950000 1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid
Liquid-DSP Threads: 4 - Buffer Length: 256 - Filter Length: 57 OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 1.6 Threads: 4 - Buffer Length: 256 - Filter Length: 57 a b 30M 60M 90M 120M 150M SE +/- 6360000.00, N = 2 SE +/- 3395000.00, N = 2 119590000 112205000 1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid
Liquid-DSP Threads: 8 - Buffer Length: 256 - Filter Length: 32 OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 1.6 Threads: 8 - Buffer Length: 256 - Filter Length: 32 b a 40M 80M 120M 160M 200M SE +/- 5480000.00, N = 2 SE +/- 11475000.00, N = 2 185540000 177945000 1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid
Liquid-DSP Threads: 8 - Buffer Length: 256 - Filter Length: 57 OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 1.6 Threads: 8 - Buffer Length: 256 - Filter Length: 57 a b 30M 60M 90M 120M 150M SE +/- 5615000.00, N = 2 SE +/- 7270000.00, N = 2 140345000 132830000 1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid
Liquid-DSP Threads: 1 - Buffer Length: 256 - Filter Length: 512 OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 1.6 Threads: 1 - Buffer Length: 256 - Filter Length: 512 a b 1.7M 3.4M 5.1M 6.8M 8.5M SE +/- 8100.00, N = 2 SE +/- 20850.00, N = 2 7721800 7710850 1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid
Liquid-DSP Threads: 2 - Buffer Length: 256 - Filter Length: 512 OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 1.6 Threads: 2 - Buffer Length: 256 - Filter Length: 512 a b 3M 6M 9M 12M 15M SE +/- 158500.00, N = 2 SE +/- 149500.00, N = 2 15179500 15023500 1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid
Liquid-DSP Threads: 4 - Buffer Length: 256 - Filter Length: 512 OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 1.6 Threads: 4 - Buffer Length: 256 - Filter Length: 512 a b 6M 12M 18M 24M 30M SE +/- 466500.00, N = 2 SE +/- 478500.00, N = 2 25830500 25812500 1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid
Liquid-DSP Threads: 8 - Buffer Length: 256 - Filter Length: 512 OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 1.6 Threads: 8 - Buffer Length: 256 - Filter Length: 512 b a 8M 16M 24M 32M 40M SE +/- 2570000.00, N = 2 SE +/- 3468000.00, N = 2 36523000 35719000 1. (CC) gcc options: -O3 -pthread -lm -lc -lliquid
Apache IoTDB Device Count: 100 - Batch Size Per Write: 1 - Sensor Count: 200 OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 1 - Sensor Count: 200 a b 110K 220K 330K 440K 550K 529599.00 524660.18
Apache IoTDB Device Count: 100 - Batch Size Per Write: 1 - Sensor Count: 200 OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 1 - Sensor Count: 200 b a 5 10 15 20 25 22.93 22.62 MAX: 1216.89 MAX: 1174.88
Apache IoTDB Device Count: 100 - Batch Size Per Write: 1 - Sensor Count: 500 OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 1 - Sensor Count: 500 b a 200K 400K 600K 800K 1000K 1009840.28 996242.81
Apache IoTDB Device Count: 100 - Batch Size Per Write: 1 - Sensor Count: 500 OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 1 - Sensor Count: 500 a b 8 16 24 32 40 34.81 34.48 MAX: 1426.89 MAX: 1495.07
Apache IoTDB Device Count: 200 - Batch Size Per Write: 1 - Sensor Count: 200 OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 200 - Batch Size Per Write: 1 - Sensor Count: 200 a b 200K 400K 600K 800K 1000K 816344.20 812170.86
Apache IoTDB Device Count: 200 - Batch Size Per Write: 1 - Sensor Count: 200 OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 200 - Batch Size Per Write: 1 - Sensor Count: 200 a b 4 8 12 16 20 17.34 16.68 MAX: 866.08 MAX: 1251.48
Apache IoTDB Device Count: 200 - Batch Size Per Write: 1 - Sensor Count: 500 OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 200 - Batch Size Per Write: 1 - Sensor Count: 500 a b 300K 600K 900K 1200K 1500K 1333775.59 1298988.31
Apache IoTDB Device Count: 200 - Batch Size Per Write: 1 - Sensor Count: 500 OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 200 - Batch Size Per Write: 1 - Sensor Count: 500 b a 7 14 21 28 35 29.75 29.65 MAX: 1271.26 MAX: 1193.78
Apache IoTDB Device Count: 500 - Batch Size Per Write: 1 - Sensor Count: 200 OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 500 - Batch Size Per Write: 1 - Sensor Count: 200 b a 200K 400K 600K 800K 1000K 1040683.13 1036429.93
Apache IoTDB Device Count: 500 - Batch Size Per Write: 1 - Sensor Count: 200 OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 500 - Batch Size Per Write: 1 - Sensor Count: 200 a b 4 8 12 16 20 15.75 15.11 MAX: 1425.54 MAX: 1458.64
Apache IoTDB Device Count: 500 - Batch Size Per Write: 1 - Sensor Count: 500 OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 500 - Batch Size Per Write: 1 - Sensor Count: 500 b a 140K 280K 420K 560K 700K 637041.62 611157.73
Apache IoTDB Device Count: 500 - Batch Size Per Write: 1 - Sensor Count: 500 OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 500 - Batch Size Per Write: 1 - Sensor Count: 500 a b 20 40 60 80 100 75.40 70.46 MAX: 1528.52 MAX: 1429.13
Apache IoTDB Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 200 OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 200 b a 3M 6M 9M 12M 15M 15526243.84 15017343.49
Apache IoTDB Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 200 OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 200 a b 20 40 60 80 100 107.84 104.43 MAX: 1809.87 MAX: 2007.51
Apache IoTDB Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 500 OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 500 a b 3M 6M 9M 12M 15M 13467917.31 12006039.14
Apache IoTDB Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 500 OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 500 b a 80 160 240 320 400 387.15 342.21 MAX: 2248.18 MAX: 2349.14
Apache IoTDB Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 200 OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 200 a b 3M 6M 9M 12M 15M 12613348.93 12464073.00
Apache IoTDB Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 200 OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 200 a b 30 60 90 120 150 144.53 143.24 MAX: 1992.27 MAX: 2073.51
Apache IoTDB Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 500 OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 500 b a 2M 4M 6M 8M 10M 9394851.37 8609112.97
Apache IoTDB Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 500 OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 500 a b 120 240 360 480 600 550.47 485.93 MAX: 3020.53 MAX: 2859.52
Apache IoTDB Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 200 OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 200 b a 2M 4M 6M 8M 10M 9246224.42 8900311.03
Apache IoTDB Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 200 OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 200 a b 50 100 150 200 250 212.43 198.41 MAX: 2080.14 MAX: 2427.81
Apache IoTDB Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 500 OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 500 a b 1.1M 2.2M 3.3M 4.4M 5.5M 4948660.19 4629287.79
Apache IoTDB Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 500 OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 500 - Batch Size Per Write: 100 - Sensor Count: 500 b a 200 400 600 800 1000 1026.93 984.70 MAX: 6211.26 MAX: 4944.15
Memcached Set To Get Ratio: 1:5 OpenBenchmarking.org Ops/sec, More Is Better Memcached 1.6.19 Set To Get Ratio: 1:5 a b 110K 220K 330K 440K 550K SE +/- 12286.93, N = 2 SE +/- 13966.62, N = 2 525368.39 518308.20 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Memcached Set To Get Ratio: 1:10 OpenBenchmarking.org Ops/sec, More Is Better Memcached 1.6.19 Set To Get Ratio: 1:10 a b 110K 220K 330K 440K 550K SE +/- 9547.84, N = 2 SE +/- 4490.17, N = 2 502050.43 486807.49 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Memcached Set To Get Ratio: 1:100 OpenBenchmarking.org Ops/sec, More Is Better Memcached 1.6.19 Set To Get Ratio: 1:100 a b 110K 220K 330K 440K 550K SE +/- 10780.90, N = 2 SE +/- 14781.65, N = 2 496309.76 485922.63 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Neural Magic DeepSparse Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream a b 0.5294 1.0588 1.5882 2.1176 2.647 SE +/- 0.2053, N = 2 SE +/- 0.2323, N = 2 2.3527 2.3352
Neural Magic DeepSparse Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream a b 200 400 600 800 1000 SE +/- 75.66, N = 2 SE +/- 86.04, N = 2 855.68 864.99
Neural Magic DeepSparse Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream a b 15 30 45 60 75 SE +/- 0.37, N = 2 SE +/- 0.29, N = 2 65.43 65.36
Neural Magic DeepSparse Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream a b 7 14 21 28 35 SE +/- 0.17, N = 2 SE +/- 0.13, N = 2 30.54 30.57
Neural Magic DeepSparse Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Stream a b 7 14 21 28 35 SE +/- 2.30, N = 2 SE +/- 2.67, N = 2 28.92 28.46
Neural Magic DeepSparse Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Stream a b 16 32 48 64 80 SE +/- 5.54, N = 2 SE +/- 6.64, N = 2 69.56 70.82
Neural Magic DeepSparse Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Stream a b 2 4 6 8 10 SE +/- 1.0816, N = 2 SE +/- 1.0939, N = 2 8.3009 8.2832
Neural Magic DeepSparse Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Stream a b 50 100 150 200 250 SE +/- 31.56, N = 2 SE +/- 32.45, N = 2 244.56 245.64
Neural Magic DeepSparse Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream a b 7 14 21 28 35 SE +/- 2.83, N = 2 SE +/- 3.30, N = 2 30.57 29.76
Neural Magic DeepSparse Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream a b 15 30 45 60 75 SE +/- 6.09, N = 2 SE +/- 7.55, N = 2 65.95 68.00
Neural Magic DeepSparse Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream a b 40 80 120 160 200 SE +/- 20.06, N = 2 SE +/- 21.75, N = 2 196.78 195.49
Neural Magic DeepSparse Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream a b 3 6 9 12 15 SE +/- 1.04, N = 2 SE +/- 1.15, N = 2 10.25 10.34
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream a b 4 8 12 16 20 SE +/- 1.61, N = 2 SE +/- 1.82, N = 2 14.67 14.61
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream a b 30 60 90 120 150 SE +/- 15.16, N = 2 SE +/- 17.32, N = 2 137.95 139.05
Neural Magic DeepSparse Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Stream b a 0.7304 1.4608 2.1912 2.9216 3.652 SE +/- 0.0188, N = 2 SE +/- 0.3971, N = 2 3.2461 2.6677
Neural Magic DeepSparse Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Stream b a 160 320 480 640 800 SE +/- 2.10, N = 2 SE +/- 113.73, N = 2 614.68 764.27
Neural Magic DeepSparse Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream a b 7 14 21 28 35 SE +/- 2.87, N = 2 SE +/- 2.67, N = 2 29.66 28.96
Neural Magic DeepSparse Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream a b 15 30 45 60 75 SE +/- 6.60, N = 2 SE +/- 6.42, N = 2 68.05 69.62
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream b a 3 6 9 12 15 SE +/- 0.10, N = 2 SE +/- 1.68, N = 2 12.95 12.54
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream b a 40 80 120 160 200 SE +/- 1.05, N = 2 SE +/- 21.74, N = 2 154.26 162.16
Neural Magic DeepSparse Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream b a 5 10 15 20 25 SE +/- 3.12, N = 2 SE +/- 2.36, N = 2 20.73 17.51
Neural Magic DeepSparse Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream b a 30 60 90 120 150 SE +/- 14.87, N = 2 SE +/- 15.66, N = 2 98.71 116.27
Neural Magic DeepSparse Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream b a 0.7272 1.4544 2.1816 2.9088 3.636 SE +/- 0.4116, N = 2 SE +/- 0.0495, N = 2 3.2320 2.9815
Neural Magic DeepSparse Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream b a 140 280 420 560 700 SE +/- 79.47, N = 2 SE +/- 12.84, N = 2 628.36 669.18
Neural Magic DeepSparse Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream b a 8 16 24 32 40 SE +/- 0.13, N = 2 SE +/- 0.07, N = 2 34.32 21.30
Neural Magic DeepSparse Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream b a 20 40 60 80 100 SE +/- 0.22, N = 2 SE +/- 0.32, N = 2 58.24 93.85
Neural Magic DeepSparse Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Stream b a 3 6 9 12 15 SE +/- 0.2364, N = 2 SE +/- 0.6045, N = 2 11.1719 9.4553
Neural Magic DeepSparse Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Stream b a 50 100 150 200 250 SE +/- 3.79, N = 2 SE +/- 13.47, N = 2 179.08 211.98
Neural Magic DeepSparse Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.5 Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream b a 0.5374 1.0748 1.6122 2.1496 2.687 SE +/- 0.0327, N = 2 SE +/- 0.0085, N = 2 2.3883 2.0309
Neural Magic DeepSparse Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.5 Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream b a 200 400 600 800 1000 SE +/- 13.19, N = 2 SE +/- 0.29, N = 2 835.85 979.18
Redis 7.0.12 + memtier_benchmark Protocol: Redis - Clients: 50 - Set To Get Ratio: 1:5 OpenBenchmarking.org Ops/sec, More Is Better Redis 7.0.12 + memtier_benchmark 2.0 Protocol: Redis - Clients: 50 - Set To Get Ratio: 1:5 b a 300K 600K 900K 1200K 1500K SE +/- 4729.88, N = 2 SE +/- 43465.56, N = 2 1398883.01 1039311.12 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Redis 7.0.12 + memtier_benchmark Protocol: Redis - Clients: 100 - Set To Get Ratio: 1:5 OpenBenchmarking.org Ops/sec, More Is Better Redis 7.0.12 + memtier_benchmark 2.0 Protocol: Redis - Clients: 100 - Set To Get Ratio: 1:5 b a 300K 600K 900K 1200K 1500K SE +/- 39391.55, N = 2 SE +/- 44860.53, N = 2 1372033.18 1016703.27 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Redis 7.0.12 + memtier_benchmark Protocol: Redis - Clients: 50 - Set To Get Ratio: 1:10 OpenBenchmarking.org Ops/sec, More Is Better Redis 7.0.12 + memtier_benchmark 2.0 Protocol: Redis - Clients: 50 - Set To Get Ratio: 1:10 b a 300K 600K 900K 1200K 1500K SE +/- 18012.65, N = 2 SE +/- 19181.49, N = 2 1455171.86 1109812.44 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Redis 7.0.12 + memtier_benchmark Protocol: Redis - Clients: 500 - Set To Get Ratio: 1:5 OpenBenchmarking.org Ops/sec, More Is Better Redis 7.0.12 + memtier_benchmark 2.0 Protocol: Redis - Clients: 500 - Set To Get Ratio: 1:5 b a 300K 600K 900K 1200K 1500K SE +/- 25673.44, N = 2 SE +/- 8374.42, N = 2 1248610.70 890374.23 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Redis 7.0.12 + memtier_benchmark Protocol: Redis - Clients: 100 - Set To Get Ratio: 1:10 OpenBenchmarking.org Ops/sec, More Is Better Redis 7.0.12 + memtier_benchmark 2.0 Protocol: Redis - Clients: 100 - Set To Get Ratio: 1:10 b a 300K 600K 900K 1200K 1500K SE +/- 8107.22, N = 2 SE +/- 568.99, N = 2 1377833.48 1028641.19 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Redis 7.0.12 + memtier_benchmark Protocol: Redis - Clients: 500 - Set To Get Ratio: 1:10 OpenBenchmarking.org Ops/sec, More Is Better Redis 7.0.12 + memtier_benchmark 2.0 Protocol: Redis - Clients: 500 - Set To Get Ratio: 1:10 b a 300K 600K 900K 1200K 1500K SE +/- 63299.47, N = 2 SE +/- 45279.14, N = 2 1253523.69 914462.02 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Stress-NG Test: Hash OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Hash b a 140K 280K 420K 560K 700K SE +/- 38690.57, N = 2 SE +/- 38337.11, N = 2 638650.51 523699.70 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: MMAP OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: MMAP b a 7 14 21 28 35 SE +/- 3.70, N = 2 SE +/- 3.19, N = 2 28.46 20.86 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: NUMA OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: NUMA b a 14 28 42 56 70 SE +/- 5.27, N = 2 SE +/- 5.63, N = 2 60.49 53.95 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Pipe OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Pipe b a 300K 600K 900K 1200K 1500K SE +/- 131648.17, N = 2 SE +/- 30380.48, N = 2 1533355.11 1422109.84 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Poll OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Poll b a 70K 140K 210K 280K 350K SE +/- 23847.48, N = 2 SE +/- 20478.34, N = 2 318466.50 284300.80 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Zlib OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Zlib b a 70 140 210 280 350 SE +/- 21.49, N = 2 SE +/- 31.94, N = 2 335.91 273.05 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Futex OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Futex b a 130K 260K 390K 520K 650K SE +/- 57507.32, N = 2 SE +/- 36419.89, N = 2 624014.17 486078.68 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: MEMFD OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: MEMFD a b 13 26 39 52 65 SE +/- 0.79, N = 2 SE +/- 0.72, N = 2 56.59 43.09 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Mutex OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Mutex b a 150K 300K 450K 600K 750K SE +/- 20730.70, N = 2 SE +/- 76155.40, N = 2 694562.47 604156.72 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Atomic OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Atomic b a 50 100 150 200 250 SE +/- 11.52, N = 2 SE +/- 10.78, N = 2 249.99 224.51 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Crypto OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Crypto b a 1200 2400 3600 4800 6000 SE +/- 96.10, N = 2 SE +/- 47.72, N = 2 5684.53 4590.53 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Malloc OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Malloc b a 90K 180K 270K 360K 450K SE +/- 46761.51, N = 2 SE +/- 39637.02, N = 2 434156.10 375933.99 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Cloning OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Cloning b a 150 300 450 600 750 SE +/- 2.47, N = 2 SE +/- 28.62, N = 2 715.58 669.41 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Forking OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Forking b a 3K 6K 9K 12K 15K SE +/- 1330.94, N = 2 SE +/- 301.28, N = 2 12140.60 9551.79 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Pthread OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Pthread b a 9K 18K 27K 36K 45K SE +/- 4375.87, N = 2 SE +/- 4732.28, N = 2 42229.61 34974.26 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: AVL Tree OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: AVL Tree b a 5 10 15 20 25 SE +/- 0.15, N = 2 SE +/- 0.05, N = 2 19.30 16.41 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: IO_uring OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: IO_uring b a 40K 80K 120K 160K 200K SE +/- 499.46, N = 2 SE +/- 2497.26, N = 2 182936.68 170492.16 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: SENDFILE OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: SENDFILE b a 9K 18K 27K 36K 45K SE +/- 1151.61, N = 2 SE +/- 1669.87, N = 2 43752.37 36779.86 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: CPU Cache OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: CPU Cache b a 200K 400K 600K 800K 1000K SE +/- 138960.51, N = 2 SE +/- 244998.38, N = 2 1153192.75 929178.43 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: CPU Stress OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: CPU Stress b a 2K 4K 6K 8K 10K SE +/- 158.65, N = 2 SE +/- 567.82, N = 2 7866.40 7310.05 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Semaphores OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Semaphores b a 700K 1400K 2100K 2800K 3500K SE +/- 70160.30, N = 2 SE +/- 287315.12, N = 2 3241408.21 3218725.51 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Matrix Math OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Matrix Math a b 4K 8K 12K 16K 20K SE +/- 1025.15, N = 2 SE +/- 854.63, N = 2 17318.70 17116.79 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Vector Math OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Vector Math a b 3K 6K 9K 12K 15K SE +/- 425.64, N = 2 SE +/- 913.67, N = 2 12666.19 12444.33 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Function Call OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Function Call b a 500 1000 1500 2000 2500 SE +/- 40.41, N = 2 SE +/- 96.37, N = 2 2164.20 2058.70 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: x86_64 RdRand OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: x86_64 RdRand a b 700 1400 2100 2800 3500 SE +/- 13.37, N = 2 SE +/- 68.45, N = 2 3267.31 3134.04 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Floating Point OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Floating Point b a 160 320 480 640 800 SE +/- 25.67, N = 2 SE +/- 25.00, N = 2 762.86 743.83 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Matrix 3D Math OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Matrix 3D Math b a 90 180 270 360 450 SE +/- 6.17, N = 2 SE +/- 5.36, N = 2 396.03 383.49 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Memory Copying OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Memory Copying b a 200 400 600 800 1000 SE +/- 40.05, N = 2 SE +/- 40.86, N = 2 1088.75 1078.73 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Vector Shuffle OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Vector Shuffle a b 500 1000 1500 2000 2500 SE +/- 33.39, N = 2 SE +/- 17.07, N = 2 2431.22 2431.01 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Socket Activity OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Socket Activity b a 500 1000 1500 2000 2500 SE +/- 271.08, N = 2 SE +/- 249.85, N = 2 2508.63 2495.57 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Wide Vector Math OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Wide Vector Math b a 30K 60K 90K 120K 150K SE +/- 3329.63, N = 2 SE +/- 2993.76, N = 2 148388.79 147314.85 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Context Switching OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Context Switching b a 160K 320K 480K 640K 800K SE +/- 16234.88, N = 2 SE +/- 53162.51, N = 2 733269.79 705297.96 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Fused Multiply-Add OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Fused Multiply-Add a b 600K 1200K 1800K 2400K 3000K SE +/- 63230.65, N = 2 SE +/- 262560.86, N = 2 2978232.73 2737483.30 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Vector Floating Point OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Vector Floating Point b a 2K 4K 6K 8K 10K SE +/- 310.83, N = 2 SE +/- 364.11, N = 2 7877.70 7289.32 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Glibc C String Functions OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Glibc C String Functions a b 500K 1000K 1500K 2000K 2500K SE +/- 92139.37, N = 2 SE +/- 158119.57, N = 2 2485856.07 2360589.10 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: Glibc Qsort Data Sorting OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: Glibc Qsort Data Sorting b a 16 32 48 64 80 SE +/- 3.74, N = 2 SE +/- 5.90, N = 2 74.13 69.13 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
Stress-NG Test: System V Message Passing OpenBenchmarking.org Bogo Ops/s, More Is Better Stress-NG 0.15.10 Test: System V Message Passing a b 600K 1200K 1800K 2400K 3000K SE +/- 119752.37, N = 2 SE +/- 200089.21, N = 2 2652462.62 2519916.10 1. (CXX) g++ options: -lm -laio -lapparmor -latomic -lc -lcrypt -ldl -lEGL -lgbm -lGLESv2 -ljpeg -lpthread -lrt -lsctp -lz
NCNN Target: CPU - Model: mobilenet OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: mobilenet a b 7 14 21 28 35 SE +/- 0.19, N = 2 SE +/- 0.03, N = 2 25.93 27.68 MIN: 23.82 / MAX: 45.07 MIN: 24.92 / MAX: 47.55 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU-v2-v2 - Model: mobilenet-v2 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU-v2-v2 - Model: mobilenet-v2 b a 2 4 6 8 10 SE +/- 0.07, N = 2 SE +/- 0.70, N = 2 6.15 6.69 MIN: 5.65 / MAX: 26.38 MIN: 5.6 / MAX: 28.58 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU-v3-v3 - Model: mobilenet-v3 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU-v3-v3 - Model: mobilenet-v3 b a 1.1565 2.313 3.4695 4.626 5.7825 SE +/- 0.04, N = 2 SE +/- 0.80, N = 2 4.31 5.14 MIN: 4.05 / MAX: 20.17 MIN: 4.05 / MAX: 26.81 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: shufflenet-v2 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: shufflenet-v2 b a 0.9653 1.9306 2.8959 3.8612 4.8265 SE +/- 0.01, N = 2 SE +/- 0.02, N = 2 3.02 4.29 MIN: 2.73 / MAX: 18.89 MIN: 3.99 / MAX: 24.93 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: mnasnet OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: mnasnet b a 2 4 6 8 10 SE +/- 0.02, N = 2 SE +/- 0.02, N = 2 4.49 6.04 MIN: 4.03 / MAX: 20.38 MIN: 5.41 / MAX: 26.6 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: efficientnet-b0 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: efficientnet-b0 b a 3 6 9 12 15 SE +/- 0.02, N = 2 SE +/- 0.65, N = 2 9.43 9.78 MIN: 8.62 / MAX: 26.5 MIN: 8.38 / MAX: 32.56 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: blazeface OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: blazeface a b 0.2093 0.4186 0.6279 0.8372 1.0465 SE +/- 0.01, N = 2 SE +/- 0.00, N = 2 0.91 0.93 MIN: 0.78 / MAX: 3.03 MIN: 0.8 / MAX: 3.09 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: googlenet OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: googlenet b a 4 8 12 16 20 SE +/- 0.10, N = 2 SE +/- 0.08, N = 2 16.19 16.23 MIN: 14.97 / MAX: 32.25 MIN: 15.01 / MAX: 32.85 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: vgg16 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: vgg16 a b 20 40 60 80 100 SE +/- 0.26, N = 2 SE +/- 2.34, N = 2 97.42 98.75 MIN: 94.57 / MAX: 115.91 MIN: 90.66 / MAX: 1242.52 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: resnet18 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: resnet18 a b 3 6 9 12 15 SE +/- 0.04, N = 2 SE +/- 0.11, N = 2 12.61 12.62 MIN: 11.89 / MAX: 29.06 MIN: 11.83 / MAX: 28.8 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: alexnet OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: alexnet b a 3 6 9 12 15 SE +/- 0.07, N = 2 SE +/- 0.06, N = 2 11.64 11.68 MIN: 10.99 / MAX: 27.75 MIN: 10.92 / MAX: 27.91 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: resnet50 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: resnet50 b a 8 16 24 32 40 SE +/- 1.96, N = 2 SE +/- 0.05, N = 2 34.33 36.02 MIN: 30.77 / MAX: 59.68 MIN: 31.18 / MAX: 61.6 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: yolov4-tiny OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: yolov4-tiny a b 9 18 27 36 45 SE +/- 0.08, N = 2 SE +/- 0.83, N = 2 38.37 39.31 MIN: 36.54 / MAX: 54.65 MIN: 36.44 / MAX: 59.39 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: squeezenet_ssd OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: squeezenet_ssd b a 4 8 12 16 20 SE +/- 0.37, N = 2 SE +/- 0.30, N = 2 15.80 16.55 MIN: 13.81 / MAX: 36.58 MIN: 15.48 / MAX: 37.44 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: regnety_400m OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: regnety_400m a b 3 6 9 12 15 SE +/- 0.14, N = 2 SE +/- 0.04, N = 2 10.03 10.12 MIN: 8.93 / MAX: 25.51 MIN: 9.48 / MAX: 25.78 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: vision_transformer OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: vision_transformer b a 50 100 150 200 250 SE +/- 5.04, N = 2 SE +/- 0.17, N = 2 232.17 233.96 MIN: 187.55 / MAX: 291.61 MIN: 196.52 / MAX: 292.9 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: CPU - Model: FastestDet OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: FastestDet a b 1.2488 2.4976 3.7464 4.9952 6.244 SE +/- 0.04, N = 2 SE +/- 0.16, N = 2 5.39 5.55 MIN: 5.09 / MAX: 24.31 MIN: 5.23 / MAX: 25.23 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: mobilenet OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: mobilenet b a 7 14 21 28 35 SE +/- 0.57, N = 2 SE +/- 0.19, N = 2 27.38 28.15 MIN: 24.23 / MAX: 47.57 MIN: 26.87 / MAX: 48.42 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU-v2-v2 - Model: mobilenet-v2 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU-v2-v2 - Model: mobilenet-v2 a b 2 4 6 8 10 SE +/- 0.67, N = 2 SE +/- 0.65, N = 2 6.67 6.67 MIN: 5.54 / MAX: 27.83 MIN: 5.58 / MAX: 27.94 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU-v3-v3 - Model: mobilenet-v3 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU-v3-v3 - Model: mobilenet-v3 a b 1.1498 2.2996 3.4494 4.5992 5.749 SE +/- 0.77, N = 2 SE +/- 0.80, N = 2 5.11 5.11 MIN: 4.07 / MAX: 20.14 MIN: 4.05 / MAX: 26.7 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: shufflenet-v2 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: shufflenet-v2 a b 0.8258 1.6516 2.4774 3.3032 4.129 SE +/- 0.63, N = 2 SE +/- 0.62, N = 2 3.66 3.67 MIN: 2.74 / MAX: 16.28 MIN: 2.75 / MAX: 19.08 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: mnasnet OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: mnasnet a b 1.1835 2.367 3.5505 4.734 5.9175 SE +/- 0.81, N = 2 SE +/- 0.78, N = 2 5.23 5.26 MIN: 4.05 / MAX: 26.68 MIN: 4.06 / MAX: 22.06 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: efficientnet-b0 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: efficientnet-b0 a b 3 6 9 12 15 SE +/- 0.07, N = 2 SE +/- 0.02, N = 2 9.22 9.34 MIN: 8.52 / MAX: 24.96 MIN: 8.61 / MAX: 25.62 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: blazeface OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: blazeface a b 0.216 0.432 0.648 0.864 1.08 SE +/- 0.00, N = 2 SE +/- 0.01, N = 2 0.94 0.96 MIN: 0.86 / MAX: 3.07 MIN: 0.8 / MAX: 3.11 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: googlenet OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: googlenet b a 4 8 12 16 20 SE +/- 0.01, N = 2 SE +/- 0.00, N = 2 16.19 16.29 MIN: 14.91 / MAX: 32.9 MIN: 14.94 / MAX: 33.61 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: vgg16 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: vgg16 b a 20 40 60 80 100 SE +/- 0.11, N = 2 SE +/- 0.23, N = 2 96.26 97.70 MIN: 93.43 / MAX: 114.68 MIN: 94.39 / MAX: 117.31 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: resnet18 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: resnet18 b a 3 6 9 12 15 SE +/- 0.04, N = 2 SE +/- 0.04, N = 2 12.56 12.73 MIN: 11.83 / MAX: 29.03 MIN: 11.94 / MAX: 28.78 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: alexnet OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: alexnet a b 3 6 9 12 15 SE +/- 0.03, N = 2 SE +/- 0.04, N = 2 11.67 11.69 MIN: 11.06 / MAX: 27.75 MIN: 10.95 / MAX: 28.31 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: resnet50 OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: resnet50 a b 8 16 24 32 40 SE +/- 0.55, N = 2 SE +/- 1.36, N = 2 33.18 33.60 MIN: 31.12 / MAX: 58.5 MIN: 30.7 / MAX: 59.9 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: yolov4-tiny OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: yolov4-tiny b a 9 18 27 36 45 SE +/- 1.11, N = 2 SE +/- 1.02, N = 2 39.17 39.39 MIN: 36.39 / MAX: 59.83 MIN: 36.4 / MAX: 60.27 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: squeezenet_ssd OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: squeezenet_ssd b a 4 8 12 16 20 SE +/- 0.84, N = 2 SE +/- 0.47, N = 2 15.48 15.97 MIN: 13.66 / MAX: 36.39 MIN: 13.93 / MAX: 35.68 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: regnety_400m OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: regnety_400m a b 3 6 9 12 15 SE +/- 0.01, N = 2 SE +/- 0.07, N = 2 9.82 10.41 MIN: 9.26 / MAX: 25.33 MIN: 9.74 / MAX: 26.42 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: vision_transformer OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: vision_transformer b a 50 100 150 200 250 SE +/- 3.18, N = 2 SE +/- 3.20, N = 2 235.51 240.03 MIN: 187.15 / MAX: 291.28 MIN: 196.33 / MAX: 300.62 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN Target: Vulkan GPU - Model: FastestDet OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: FastestDet b a 1.2735 2.547 3.8205 5.094 6.3675 SE +/- 0.04, N = 2 SE +/- 0.20, N = 2 5.47 5.66 MIN: 5.21 / MAX: 24.93 MIN: 5.23 / MAX: 24.86 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
Apache Cassandra Test: Writes OpenBenchmarking.org Op/s, More Is Better Apache Cassandra 4.1.3 Test: Writes b a 6K 12K 18K 24K 30K SE +/- 71.50, N = 2 SE +/- 149.00, N = 2 26438 26422
Phoronix Test Suite v10.8.5