Intel Core i7-1165G7 testing with a Dell 0GG9PT (3.15.0 BIOS) and Intel Xe TGL GT2 15GB on Ubuntu 23.04 via the Phoronix Test Suite.
a Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0xa6 - Thermald 2.5.2Java Notes: OpenJDK Runtime Environment (build 11.0.19+7-post-Ubuntu-0ubuntu123.04)Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Not affected
b c Processor: Intel Core i7-1165G7 @ 4.70GHz (4 Cores / 8 Threads), Motherboard: Dell 0GG9PT (3.15.0 BIOS), Chipset: Intel Tiger Lake-LP, Memory: 16GB, Disk: Kioxia KBG40ZNS256G NVMe 256GB, Graphics: Intel Xe TGL GT2 15GB (1300MHz), Audio: Realtek ALC289, Network: Intel Wi-Fi 6 AX201
OS: Ubuntu 23.04, Kernel: 6.2.0-24-generic (x86_64), Desktop: GNOME Shell 44.0, Display Server: X Server + Wayland, OpenGL: 4.6 Mesa 23.0.2, Compiler: GCC 12.2.0, File-System: ext4, Screen Resolution: 1920x1200
OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 500 a b c 160 320 480 640 800 284.41 370.36 757.34 MAX: 2551.38 MAX: 7948.19 MAX: 21117.8
OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 500 b a c 110 220 330 440 550 200.97 211.96 499.98 MAX: 1387.04 MAX: 1593.95 MAX: 7223.32
OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 200 - Batch Size Per Write: 100 - Sensor Count: 200 b a c 40 80 120 160 200 85.64 91.67 164.99 MAX: 1651.27 MAX: 7883.93 MAX: 11718.16
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: blazeface a c b 0.297 0.594 0.891 1.188 1.485 SE +/- 0.03, N = 2 SE +/- 0.03, N = 2 SE +/- 0.02, N = 2 1.32 0.96 0.94 MIN: 1.2 / MAX: 4.08 MIN: 0.9 / MAX: 5.52 MIN: 0.9 / MAX: 1.09 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: regnety_400m a b c 3 6 9 12 15 SE +/- 0.10, N = 2 SE +/- 0.03, N = 2 SE +/- 0.01, N = 2 11.76 8.61 8.39 MIN: 11.26 / MAX: 22.16 MIN: 8.21 / MAX: 17.64 MIN: 8.14 / MAX: 18.42 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: efficientnet-b0 a b c 2 4 6 8 10 SE +/- 0.29, N = 2 SE +/- 0.04, N = 2 SE +/- 0.05, N = 2 8.85 6.92 6.64 MIN: 6.6 / MAX: 21.28 MIN: 6.5 / MAX: 16.04 MIN: 6.38 / MAX: 17.34 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: regnety_400m a c b 3 6 9 12 15 SE +/- 0.30, N = 2 SE +/- 1.13, N = 2 SE +/- 0.01, N = 2 11.19 9.62 8.57 MIN: 10.38 / MAX: 21.88 MIN: 8.23 / MAX: 23.19 MIN: 8.18 / MAX: 21.08 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
Apache IoTDB OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 200 c b a 5M 10M 15M 20M 25M 17612257.13 21253793.53 22588608.67
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: googlenet a b c 4 8 12 16 20 SE +/- 0.07, N = 2 SE +/- 0.20, N = 2 SE +/- 0.27, N = 2 15.82 12.78 12.42 MIN: 15.12 / MAX: 28.62 MIN: 11.9 / MAX: 22.46 MIN: 11.68 / MAX: 23.37 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: blazeface a c b 0.2835 0.567 0.8505 1.134 1.4175 SE +/- 0.02, N = 2 SE +/- 0.14, N = 2 SE +/- 0.05, N = 2 1.26 1.08 0.99 MIN: 1.16 / MAX: 4.12 MIN: 0.91 / MAX: 3.75 MIN: 0.91 / MAX: 3.77 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: resnet18 a b c 3 6 9 12 15 SE +/- 0.04, N = 2 SE +/- 0.11, N = 2 SE +/- 0.50, N = 2 11.17 9.25 9.19 MIN: 10.66 / MAX: 21.33 MIN: 8.51 / MAX: 24.74 MIN: 8.38 / MAX: 20.09 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: mnasnet a b c 1.0238 2.0476 3.0714 4.0952 5.119 SE +/- 0.63, N = 2 SE +/- 0.07, N = 2 SE +/- 0.01, N = 2 4.55 3.86 3.83 MIN: 3.81 / MAX: 15.97 MIN: 3.69 / MAX: 12.48 MIN: 3.68 / MAX: 12.77 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: alexnet a b c 2 4 6 8 10 SE +/- 0.04, N = 2 SE +/- 0.09, N = 2 SE +/- 0.25, N = 2 8.61 7.43 7.30 MIN: 8.27 / MAX: 20.05 MIN: 6.9 / MAX: 16.46 MIN: 6.79 / MAX: 17.09 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: shufflenet-v2 a c b 0.9068 1.8136 2.7204 3.6272 4.534 SE +/- 0.57, N = 2 SE +/- 0.02, N = 2 SE +/- 0.01, N = 2 4.03 3.47 3.47 MIN: 3.35 / MAX: 10.89 MIN: 3.36 / MAX: 13.07 MIN: 3.32 / MAX: 12.41 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: googlenet a b c 4 8 12 16 20 SE +/- 0.02, N = 2 SE +/- 1.53, N = 2 SE +/- 1.79, N = 2 15.86 14.09 14.01 MIN: 15.16 / MAX: 26.57 MIN: 12.03 / MAX: 25.74 MIN: 11.83 / MAX: 25.65 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
Apache IoTDB OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 100 - Sensor Count: 200 a b c 20 40 60 80 100 68.48 74.08 76.87 MAX: 1696.56 MAX: 1547.17 MAX: 9168.34
Dragonflydb Dragonfly is an open-source database server that is a "modern Redis replacement" that aims to be the fastest memory store while being compliant with the Redis and Memcached protocols. For benchmarking Dragonfly, Memtier_benchmark is used as a NoSQL Redis/Memcache traffic generation plus benchmarking tool developed by Redis Labs. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Ops/sec, More Is Better Dragonflydb 1.6.2 Clients Per Thread: 20 - Set To Get Ratio: 1:100 a b c 400K 800K 1200K 1600K 2000K SE +/- 66621.90, N = 2 SE +/- 71249.46, N = 2 SE +/- 168860.22, N = 2 1504800.26 1532061.63 1655250.86 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: vision_transformer a b c 50 100 150 200 250 SE +/- 8.01, N = 2 SE +/- 0.73, N = 2 SE +/- 0.68, N = 2 207.46 193.90 188.79 MIN: 169.78 / MAX: 244.36 MIN: 167.64 / MAX: 231.56 MIN: 169.1 / MAX: 243.33 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: resnet18 a c b 3 6 9 12 15 SE +/- 0.02, N = 2 SE +/- 0.82, N = 2 SE +/- 1.05, N = 2 11.18 10.40 10.22 MIN: 10.62 / MAX: 21.34 MIN: 8.57 / MAX: 21.1 MIN: 8.49 / MAX: 20.91 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
VVenC VVenC is the Fraunhofer Versatile Video Encoder as a fast/efficient H.266/VVC encoder. The vvenc encoder makes use of SIMD Everywhere (SIMDe). The vvenc software is published under the Clear BSD License. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.9 Video Input: Bosphorus 1080p - Video Preset: Faster b a c 4 8 12 16 20 SE +/- 0.11, N = 2 SE +/- 0.08, N = 2 SE +/- 0.79, N = 2 12.64 12.80 13.80 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: vgg16 a b c 13 26 39 52 65 SE +/- 0.11, N = 2 SE +/- 1.96, N = 2 SE +/- 2.37, N = 2 59.09 54.25 54.13 MIN: 57.54 / MAX: 76.29 MIN: 50.32 / MAX: 71.7 MIN: 48.55 / MAX: 72.36 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: resnet50 a c b 7 14 21 28 35 SE +/- 0.18, N = 2 SE +/- 2.20, N = 2 SE +/- 2.23, N = 2 28.74 26.57 26.57 MIN: 27.89 / MAX: 40 MIN: 23.37 / MAX: 38.7 MIN: 23.28 / MAX: 38.91 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: vision_transformer b c a 40 80 120 160 200 SE +/- 3.85, N = 2 SE +/- 5.48, N = 2 SE +/- 4.42, N = 2 204.36 200.23 189.40 MIN: 168.8 / MAX: 231.88 MIN: 169.16 / MAX: 234.64 MIN: 169.61 / MAX: 225.69 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: squeezenet_ssd a b c 3 6 9 12 15 SE +/- 0.03, N = 2 SE +/- 0.86, N = 2 SE +/- 0.97, N = 2 12.95 12.23 12.05 MIN: 12.67 / MAX: 23.3 MIN: 10.77 / MAX: 31.43 MIN: 10.72 / MAX: 23.89 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: alexnet a c b 2 4 6 8 10 SE +/- 0.04, N = 2 SE +/- 0.55, N = 2 SE +/- 0.52, N = 2 8.68 8.15 8.08 MIN: 8.25 / MAX: 19.5 MIN: 7.01 / MAX: 17.98 MIN: 7.01 / MAX: 17.55 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
Apache IoTDB OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 200 - Batch Size Per Write: 1 - Sensor Count: 200 a b c 3 6 9 12 15 12.00 12.52 12.78 MAX: 771.24 MAX: 794.23 MAX: 800.68
Dragonflydb Dragonfly is an open-source database server that is a "modern Redis replacement" that aims to be the fastest memory store while being compliant with the Redis and Memcached protocols. For benchmarking Dragonfly, Memtier_benchmark is used as a NoSQL Redis/Memcache traffic generation plus benchmarking tool developed by Redis Labs. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Ops/sec, More Is Better Dragonflydb 1.6.2 Clients Per Thread: 10 - Set To Get Ratio: 1:5 b a c 300K 600K 900K 1200K 1500K SE +/- 115490.54, N = 2 SE +/- 124523.27, N = 2 SE +/- 156954.29, N = 2 1310891.48 1331051.90 1394989.58 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
OpenBenchmarking.org Ops/sec, More Is Better Dragonflydb 1.6.2 Clients Per Thread: 20 - Set To Get Ratio: 1:10 b a c 300K 600K 900K 1200K 1500K SE +/- 144264.17, N = 2 SE +/- 93795.63, N = 2 SE +/- 153108.99, N = 2 1532531.22 1553942.49 1620928.16 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
VVenC VVenC is the Fraunhofer Versatile Video Encoder as a fast/efficient H.266/VVC encoder. The vvenc encoder makes use of SIMD Everywhere (SIMDe). The vvenc software is published under the Clear BSD License. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.9 Video Input: Bosphorus 4K - Video Preset: Faster a b c 0.8669 1.7338 2.6007 3.4676 4.3345 SE +/- 0.014, N = 2 SE +/- 0.014, N = 2 SE +/- 0.081, N = 2 3.680 3.781 3.853 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
Apache IoTDB OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 1 - Sensor Count: 200 c b a 4 8 12 16 20 16.03 16.41 16.78 MAX: 1027.73 MAX: 1017.33 MAX: 906.87
OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 200 - Batch Size Per Write: 1 - Sensor Count: 500 a c b 6 12 18 24 30 22.08 22.63 23.10 MAX: 863.43 MAX: 909.27 MAX: 906.47
Dragonflydb Dragonfly is an open-source database server that is a "modern Redis replacement" that aims to be the fastest memory store while being compliant with the Redis and Memcached protocols. For benchmarking Dragonfly, Memtier_benchmark is used as a NoSQL Redis/Memcache traffic generation plus benchmarking tool developed by Redis Labs. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Ops/sec, More Is Better Dragonflydb 1.6.2 Clients Per Thread: 10 - Set To Get Ratio: 1:10 b a c 300K 600K 900K 1200K 1500K SE +/- 90004.70, N = 2 SE +/- 119195.59, N = 2 SE +/- 143882.01, N = 2 1275557.39 1283252.40 1325104.13 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
OpenBenchmarking.org Ops/sec, More Is Better Dragonflydb 1.6.2 Clients Per Thread: 20 - Set To Get Ratio: 1:5 c b a 300K 600K 900K 1200K 1500K SE +/- 157442.25, N = 2 SE +/- 146499.45, N = 2 SE +/- 138897.05, N = 2 1561730.49 1572424.89 1620397.17 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
OpenBenchmarking.org Ops/sec, More Is Better Dragonflydb 1.6.2 Clients Per Thread: 50 - Set To Get Ratio: 1:5 a b c 300K 600K 900K 1200K 1500K SE +/- 35029.57, N = 2 SE +/- 81765.99, N = 2 SE +/- 148775.44, N = 2 1509161.50 1546553.32 1559859.23 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
OpenBenchmarking.org Ops/sec, More Is Better Dragonflydb 1.6.2 Clients Per Thread: 50 - Set To Get Ratio: 1:100 b c a 300K 600K 900K 1200K 1500K SE +/- 98659.37, N = 2 SE +/- 65233.84, N = 2 SE +/- 113274.82, N = 2 1538729.61 1548401.54 1590238.43 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
VkFFT OpenBenchmarking.org Benchmark Score, More Is Better VkFFT 1.2.31 Test: FFT + iFFT C2C multidimensional in single precision b a c 1100 2200 3300 4400 5500 SE +/- 6.50, N = 2 SE +/- 17.50, N = 2 SE +/- 9.00, N = 2 4937 4944 5087 1. (CXX) g++ options: -O3
Dragonflydb Dragonfly is an open-source database server that is a "modern Redis replacement" that aims to be the fastest memory store while being compliant with the Redis and Memcached protocols. For benchmarking Dragonfly, Memtier_benchmark is used as a NoSQL Redis/Memcache traffic generation plus benchmarking tool developed by Redis Labs. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Ops/sec, More Is Better Dragonflydb 1.6.2 Clients Per Thread: 10 - Set To Get Ratio: 1:100 a c b 300K 600K 900K 1200K 1500K SE +/- 117619.34, N = 2 SE +/- 141978.48, N = 2 SE +/- 95110.31, N = 2 1267626.25 1297415.25 1305698.48 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Apache IoTDB OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 200 - Batch Size Per Write: 1 - Sensor Count: 500 b c a 400K 800K 1200K 1600K 2000K 1651562.36 1659495.29 1696504.02
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: efficientnet-b0 a b c 2 4 6 8 10 SE +/- 0.02, N = 2 SE +/- 0.02, N = 2 SE +/- 0.15, N = 2 6.99 6.98 6.81 MIN: 6.55 / MAX: 16.31 MIN: 6.57 / MAX: 17.79 MIN: 6.42 / MAX: 17.77 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
VVenC VVenC is the Fraunhofer Versatile Video Encoder as a fast/efficient H.266/VVC encoder. The vvenc encoder makes use of SIMD Everywhere (SIMDe). The vvenc software is published under the Clear BSD License. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.9 Video Input: Bosphorus 1080p - Video Preset: Fast b a c 1.181 2.362 3.543 4.724 5.905 SE +/- 0.008, N = 2 SE +/- 0.000, N = 2 SE +/- 0.115, N = 2 5.124 5.140 5.249 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: FastestDet c a b 0.9068 1.8136 2.7204 3.6272 4.534 SE +/- 0.10, N = 2 SE +/- 0.09, N = 2 SE +/- 0.00, N = 2 4.03 3.97 3.94 MIN: 3.74 / MAX: 10.51 MIN: 3.7 / MAX: 15.04 MIN: 3.73 / MAX: 12.78 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU-v3-v3 - Model: mobilenet-v3 b a c 0.8078 1.6156 2.4234 3.2312 4.039 SE +/- 0.04, N = 2 SE +/- 0.03, N = 2 SE +/- 0.00, N = 2 3.59 3.58 3.51 MIN: 3.37 / MAX: 14.17 MIN: 3.35 / MAX: 13.78 MIN: 3.32 / MAX: 11.83 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: FastestDet a b c 0.8933 1.7866 2.6799 3.5732 4.4665 SE +/- 0.00, N = 2 SE +/- 0.00, N = 2 SE +/- 0.03, N = 2 3.97 3.91 3.89 MIN: 3.76 / MAX: 14.59 MIN: 3.71 / MAX: 13.7 MIN: 3.67 / MAX: 13.73 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: mobilenet a c b 5 10 15 20 25 SE +/- 0.19, N = 2 SE +/- 0.03, N = 2 SE +/- 0.03, N = 2 21.07 20.69 20.65 MIN: 20.27 / MAX: 31.83 MIN: 20.24 / MAX: 31.31 MIN: 20.22 / MAX: 31.66 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: vgg16 a c b 13 26 39 52 65 SE +/- 0.15, N = 2 SE +/- 0.65, N = 2 SE +/- 1.40, N = 2 59.03 58.28 57.88 MIN: 57.6 / MAX: 79.49 MIN: 52.14 / MAX: 69.93 MIN: 52 / MAX: 70.84 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
Dragonflydb Dragonfly is an open-source database server that is a "modern Redis replacement" that aims to be the fastest memory store while being compliant with the Redis and Memcached protocols. For benchmarking Dragonfly, Memtier_benchmark is used as a NoSQL Redis/Memcache traffic generation plus benchmarking tool developed by Redis Labs. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Ops/sec, More Is Better Dragonflydb 1.6.2 Clients Per Thread: 50 - Set To Get Ratio: 1:10 b a c 300K 600K 900K 1200K 1500K SE +/- 111709.66, N = 2 SE +/- 99951.62, N = 2 SE +/- 128692.32, N = 2 1545294.47 1559408.51 1574780.45 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
VkFFT OpenBenchmarking.org Benchmark Score, More Is Better VkFFT 1.2.31 Test: FFT + iFFT R2C / C2R a b c 1200 2400 3600 4800 6000 SE +/- 3.50, N = 2 SE +/- 30.50, N = 2 SE +/- 69.00, N = 2 5585 5589 5688 1. (CXX) g++ options: -O3
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU-v3-v3 - Model: mobilenet-v3 a b c 0.81 1.62 2.43 3.24 4.05 SE +/- 0.03, N = 2 SE +/- 0.02, N = 2 SE +/- 0.01, N = 2 3.60 3.59 3.54 MIN: 3.38 / MAX: 12.52 MIN: 3.4 / MAX: 12.21 MIN: 3.33 / MAX: 11.68 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: yolov4-tiny a b c 7 14 21 28 35 SE +/- 0.26, N = 2 SE +/- 0.15, N = 2 SE +/- 0.05, N = 2 29.60 29.23 29.11 MIN: 28.6 / MAX: 44.5 MIN: 28.34 / MAX: 41.02 MIN: 28.38 / MAX: 40.39 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
Apache IoTDB OpenBenchmarking.org Average Latency, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 1 - Sensor Count: 500 c a b 6 12 18 24 30 24.73 24.98 25.14 MAX: 998.23 MAX: 1064.56 MAX: 1026.24
VVenC VVenC is the Fraunhofer Versatile Video Encoder as a fast/efficient H.266/VVC encoder. The vvenc encoder makes use of SIMD Everywhere (SIMDe). The vvenc software is published under the Clear BSD License. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better VVenC 1.9 Video Input: Bosphorus 4K - Video Preset: Fast b a c 0.3773 0.7546 1.1319 1.5092 1.8865 SE +/- 0.028, N = 2 SE +/- 0.037, N = 2 SE +/- 0.036, N = 2 1.650 1.654 1.677 1. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: mnasnet a b c 0.8843 1.7686 2.6529 3.5372 4.4215 SE +/- 0.01, N = 2 SE +/- 0.03, N = 2 SE +/- 0.04, N = 2 3.93 3.89 3.87 MIN: 3.73 / MAX: 14.41 MIN: 3.74 / MAX: 12.97 MIN: 3.71 / MAX: 12.28 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU-v2-v2 - Model: mobilenet-v2 b a c 1.035 2.07 3.105 4.14 5.175 SE +/- 0.02, N = 2 SE +/- 0.04, N = 2 SE +/- 0.02, N = 2 4.60 4.60 4.53 MIN: 4.34 / MAX: 14.83 MIN: 4.38 / MAX: 12.37 MIN: 4.3 / MAX: 15.17 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
Apache IoTDB OpenBenchmarking.org point/sec, More Is Better Apache IoTDB 1.1.2 Device Count: 100 - Batch Size Per Write: 1 - Sensor Count: 500 b a c 300K 600K 900K 1200K 1500K 1247796.00 1265418.91 1266283.94
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: yolov4-tiny a c b 7 14 21 28 35 SE +/- 0.42, N = 2 SE +/- 0.03, N = 2 SE +/- 0.03, N = 2 29.44 29.25 29.03 MIN: 28.43 / MAX: 46.03 MIN: 28.42 / MAX: 40.16 MIN: 28.45 / MAX: 40.14 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU-v2-v2 - Model: mobilenet-v2 a b c 1.0485 2.097 3.1455 4.194 5.2425 SE +/- 0.00, N = 2 SE +/- 0.03, N = 2 SE +/- 0.04, N = 2 4.66 4.63 4.60 MIN: 4.46 / MAX: 14.54 MIN: 4.44 / MAX: 14.09 MIN: 4.38 / MAX: 14 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: resnet50 a c b 7 14 21 28 35 SE +/- 0.19, N = 2 SE +/- 0.15, N = 2 SE +/- 0.12, N = 2 28.84 28.78 28.71 MIN: 27.92 / MAX: 41.94 MIN: 27.96 / MAX: 39.21 MIN: 27.9 / MAX: 39.22 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: squeezenet_ssd a b c 3 6 9 12 15 SE +/- 0.04, N = 2 SE +/- 0.01, N = 2 SE +/- 0.03, N = 2 13.05 13.01 13.00 MIN: 12.7 / MAX: 23.56 MIN: 12.67 / MAX: 27.85 MIN: 12.68 / MAX: 23.98 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: Vulkan GPU - Model: shufflenet-v2 a c b 0.7875 1.575 2.3625 3.15 3.9375 SE +/- 0.01, N = 2 SE +/- 0.02, N = 2 SE +/- 0.01, N = 2 3.50 3.49 3.49 MIN: 3.35 / MAX: 12.33 MIN: 3.35 / MAX: 11.83 MIN: 3.33 / MAX: 13.62 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
BRL-CAD BRL-CAD is a cross-platform, open-source solid modeling system with built-in benchmark mode. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org VGR Performance Metric, More Is Better BRL-CAD 7.36 VGR Performance Metric c b a 11K 22K 33K 44K 55K SE +/- 34.00, N = 2 SE +/- 147.00, N = 2 SE +/- 16.50, N = 2 51880 51967 52008 1. (CXX) g++ options: -std=c++14 -pipe -fvisibility=hidden -fno-strict-aliasing -fno-common -fexceptions -ftemplate-depth-128 -m64 -ggdb3 -O3 -fipa-pta -fstrength-reduce -finline-functions -flto -ltcl8.6 -lregex_brl -lz_brl -lnetpbm -ldl -lm -ltk8.6
VkFFT OpenBenchmarking.org Benchmark Score, More Is Better VkFFT 1.2.31 Test: FFT + iFFT C2C Bluestein in single precision a b c 200 400 600 800 1000 SE +/- 0.50, N = 2 SE +/- 1.00, N = 2 SE +/- 1.00, N = 2 1033 1034 1035 1. (CXX) g++ options: -O3
OpenBenchmarking.org Benchmark Score, More Is Better VkFFT 1.2.31 Test: FFT + iFFT C2C 1D batched in single precision c b a 1600 3200 4800 6400 8000 SE +/- 3.50, N = 2 SE +/- 4.00, N = 2 SE +/- 14.00, N = 2 7478 7482 7486 1. (CXX) g++ options: -O3
OpenBenchmarking.org Benchmark Score, More Is Better VkFFT 1.2.31 Test: FFT + iFFT C2C 1D batched in half precision b c a 3K 6K 9K 12K 15K SE +/- 1.50, N = 2 SE +/- 4.00, N = 2 SE +/- 5.00, N = 2 14232 14241 14246 1. (CXX) g++ options: -O3
OpenBenchmarking.org Benchmark Score, More Is Better VkFFT 1.2.31 Test: FFT + iFFT C2C 1D batched in single precision, no reshuffling a b c 2K 4K 6K 8K 10K SE +/- 4.00, N = 2 SE +/- 0.50, N = 2 SE +/- 5.00, N = 2 8176 8176 8183 1. (CXX) g++ options: -O3
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20230517 Target: CPU - Model: mobilenet c b a 5 10 15 20 25 SE +/- 0.07, N = 2 SE +/- 0.04, N = 2 SE +/- 0.05, N = 2 20.74 20.74 20.74 MIN: 20.26 / MAX: 32.28 MIN: 20.23 / MAX: 31.9 MIN: 20.17 / MAX: 31.42 1. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread
VkFFT Test: FFT + iFFT C2C Bluestein benchmark in double precision
a: The test quit with a non-zero exit status.
b: The test quit with a non-zero exit status.
c: The test quit with a non-zero exit status.
Test: FFT + iFFT C2C 1D batched in double precision
a: The test quit with a non-zero exit status.
b: The test quit with a non-zero exit status.
c: The test quit with a non-zero exit status.
a Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0xa6 - Thermald 2.5.2Java Notes: OpenJDK Runtime Environment (build 11.0.19+7-post-Ubuntu-0ubuntu123.04)Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 3 August 2023 14:30 by user phoronix.
b Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0xa6 - Thermald 2.5.2Java Notes: OpenJDK Runtime Environment (build 11.0.19+7-post-Ubuntu-0ubuntu123.04)Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 3 August 2023 18:07 by user phoronix.
c Processor: Intel Core i7-1165G7 @ 4.70GHz (4 Cores / 8 Threads), Motherboard: Dell 0GG9PT (3.15.0 BIOS), Chipset: Intel Tiger Lake-LP, Memory: 16GB, Disk: Kioxia KBG40ZNS256G NVMe 256GB, Graphics: Intel Xe TGL GT2 15GB (1300MHz), Audio: Realtek ALC289, Network: Intel Wi-Fi 6 AX201
OS: Ubuntu 23.04, Kernel: 6.2.0-24-generic (x86_64), Desktop: GNOME Shell 44.0, Display Server: X Server + Wayland, OpenGL: 4.6 Mesa 23.0.2, Compiler: GCC 12.2.0, File-System: ext4, Screen Resolution: 1920x1200
Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0xa6 - Thermald 2.5.2Java Notes: OpenJDK Runtime Environment (build 11.0.19+7-post-Ubuntu-0ubuntu123.04)Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 3 August 2023 21:36 by user phoronix.