AMD Ryzen 7 PRO 6850U testing with a LENOVO 21CM0001US (R22ET51W 1.21 BIOS) and AMD Radeon 680M 1GB on Ubuntu 22.10 via the Phoronix Test Suite.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 2301235-NE-DSLKEY60028 ds lkey - Phoronix Test Suite ds lkey AMD Ryzen 7 PRO 6850U testing with a LENOVO 21CM0001US (R22ET51W 1.21 BIOS) and AMD Radeon 680M 1GB on Ubuntu 22.10 via the Phoronix Test Suite.
HTML result view exported from: https://openbenchmarking.org/result/2301235-NE-DSLKEY60028&sor .
ds lkey Processor Motherboard Chipset Memory Disk Graphics Audio Network OS Kernel Desktop Display Server OpenGL Vulkan Compiler File-System Screen Resolution a b c AMD Ryzen 7 PRO 6850U @ 4.77GHz (8 Cores / 16 Threads) LENOVO 21CM0001US (R22ET51W 1.21 BIOS) AMD Device 14b5 16GB 512GB Micron MTFDKBA512TFK AMD Radeon 680M 1GB (2200/400MHz) AMD Rembrandt Radeon HD Audio Qualcomm QCNFA765 Ubuntu 22.10 6.1.0-060100rc2daily20221028-generic (x86_64) GNOME Shell 43.0 X Server + Wayland 4.6 Mesa 22.2.1 (LLVM 15.0.2 DRM 3.49) 1.3.224 GCC 12.2.0 ext4 1920x1200 OpenBenchmarking.org Kernel Details - Transparent Huge Pages: madvise Compiler Details - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-12-U8K4Qv/gcc-12-12.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-12-U8K4Qv/gcc-12-12.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details - Scaling Governor: amd-pstate schedutil (Boost: Enabled) - Platform Profile: performance - CPU Microcode: 0xa404102 - ACPI Profile: performance Python Details - Python 3.10.7 Security Details - itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected
ds lkey keydb: GET - 50 keydb: SET - 50 keydb: GET - 100 keydb: GET - 500 keydb: GET - 900 keydb: LPOP - 50 keydb: SADD - 50 keydb: SET - 100 keydb: SET - 500 keydb: SET - 900 keydb: HMSET - 50 keydb: LPOP - 100 keydb: LPOP - 500 keydb: LPOP - 900 keydb: LPUSH - 50 keydb: SADD - 100 keydb: SADD - 500 keydb: SADD - 900 keydb: HMSET - 100 keydb: HMSET - 500 keydb: HMSET - 900 keydb: LPUSH - 100 keydb: LPUSH - 500 keydb: LPUSH - 900 deepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Stream deepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Stream deepsparse: NLP Document Classification, oBERT base uncased on IMDB - Synchronous Single-Stream deepsparse: NLP Document Classification, oBERT base uncased on IMDB - Synchronous Single-Stream deepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Asynchronous Multi-Stream deepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Asynchronous Multi-Stream deepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Synchronous Single-Stream deepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Synchronous Single-Stream deepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Stream deepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Stream deepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Synchronous Single-Stream deepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Synchronous Single-Stream deepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Stream deepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Stream deepsparse: CV Detection, YOLOv5s COCO - Synchronous Single-Stream deepsparse: CV Detection, YOLOv5s COCO - Synchronous Single-Stream deepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Stream deepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Stream deepsparse: CV Classification, ResNet-50 ImageNet - Synchronous Single-Stream deepsparse: CV Classification, ResNet-50 ImageNet - Synchronous Single-Stream deepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Stream deepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Stream deepsparse: NLP Text Classification, DistilBERT mnli - Synchronous Single-Stream deepsparse: NLP Text Classification, DistilBERT mnli - Synchronous Single-Stream deepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Stream deepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Stream deepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Synchronous Single-Stream deepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Synchronous Single-Stream deepsparse: NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Stream deepsparse: NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Stream deepsparse: NLP Text Classification, BERT base uncased SST2 - Synchronous Single-Stream deepsparse: NLP Text Classification, BERT base uncased SST2 - Synchronous Single-Stream deepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Stream deepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Stream deepsparse: NLP Token Classification, BERT base uncased conll2003 - Synchronous Single-Stream deepsparse: NLP Token Classification, BERT base uncased conll2003 - Synchronous Single-Stream a b c 400374.75 391239.34 317397.84 312160.53 308276.59 399629.16 395707.34 316035.62 302518.75 297450.84 301397.28 326756.81 319443.91 313709.75 377803.31 314192.72 309906.47 307442.56 220859.77 205129.89 201398.5 318623.53 277528 270349.19 5.1659 774.2867 5.0356 198.5696 64.9385 61.5757 51.437 19.4233 20.0433 199.16 18.1037 55.2168 28.502 140.2665 26.1404 38.2318 62.169 64.3196 56.3184 17.7369 45.5971 87.704 41.2481 24.228 7.0942 563.8116 7.0091 142.6454 22.1905 180.2356 20.2841 49.2832 5.1498 776.6661 5.0161 199.3423 397769.31 389462.69 318068.16 310491.5 305945.12 370255.16 395848.34 323218.75 297642.09 294012.12 301688.25 288963.91 279348.31 279925.44 378526.91 319358.22 311701.25 304371.41 214634.66 204631.22 203707.48 290675.72 285523.38 269025.47 5.1708 773.5511 5.0409 198.3615 64.8899 61.6214 51.8801 19.2558 20.1989 197.8215 18.1337 55.1247 28.4848 140.3583 26.1645 38.1981 62.0323 64.4599 56.3495 17.7286 45.6629 87.5767 41.2819 24.2071 7.1392 560.2575 7.0162 142.4995 22.2688 179.6011 20.3081 49.2237 5.1574 773.8573 5.0184 199.2496 396253.03 389289.88 315931.81 312734.53 306020.03 368172.25 396227.91 306421.38 300516.88 294821.75 299625.47 307917.81 281265.47 289251.41 378091.84 314827.75 310927.88 305295.03 219591.98 203269.38 202206.47 299638.03 270480.81 267054.06 5.1593 775.2753 5.0291 198.8241 64.5712 61.9249 51.8537 19.2663 20.2206 197.7918 18.1931 54.9457 28.4099 140.7331 26.1412 38.2315 62.3265 64.1565 56.2898 17.7464 45.7318 87.446 41.1178 24.3051 7.1222 561.5936 7.017 142.4844 22.1912 180.2307 20.2112 49.4602 5.169 773.8149 5.0327 198.6827 OpenBenchmarking.org
KeyDB Test: GET - Parallel Connections: 50 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: GET - Parallel Connections: 50 a b c 90K 180K 270K 360K 450K 400374.75 397769.31 396253.03 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: SET - Parallel Connections: 50 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: SET - Parallel Connections: 50 a b c 80K 160K 240K 320K 400K 391239.34 389462.69 389289.88 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: GET - Parallel Connections: 100 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: GET - Parallel Connections: 100 b a c 70K 140K 210K 280K 350K 318068.16 317397.84 315931.81 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: GET - Parallel Connections: 500 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: GET - Parallel Connections: 500 c a b 70K 140K 210K 280K 350K 312734.53 312160.53 310491.50 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: GET - Parallel Connections: 900 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: GET - Parallel Connections: 900 a c b 70K 140K 210K 280K 350K 308276.59 306020.03 305945.12 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: LPOP - Parallel Connections: 50 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: LPOP - Parallel Connections: 50 a b c 90K 180K 270K 360K 450K 399629.16 370255.16 368172.25 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: SADD - Parallel Connections: 50 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: SADD - Parallel Connections: 50 c b a 80K 160K 240K 320K 400K 396227.91 395848.34 395707.34 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: SET - Parallel Connections: 100 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: SET - Parallel Connections: 100 b a c 70K 140K 210K 280K 350K 323218.75 316035.62 306421.38 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: SET - Parallel Connections: 500 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: SET - Parallel Connections: 500 a c b 60K 120K 180K 240K 300K 302518.75 300516.88 297642.09 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: SET - Parallel Connections: 900 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: SET - Parallel Connections: 900 a c b 60K 120K 180K 240K 300K 297450.84 294821.75 294012.12 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: HMSET - Parallel Connections: 50 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: HMSET - Parallel Connections: 50 b a c 60K 120K 180K 240K 300K 301688.25 301397.28 299625.47 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: LPOP - Parallel Connections: 100 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: LPOP - Parallel Connections: 100 a c b 70K 140K 210K 280K 350K 326756.81 307917.81 288963.91 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: LPOP - Parallel Connections: 500 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: LPOP - Parallel Connections: 500 a c b 70K 140K 210K 280K 350K 319443.91 281265.47 279348.31 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: LPOP - Parallel Connections: 900 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: LPOP - Parallel Connections: 900 a c b 70K 140K 210K 280K 350K 313709.75 289251.41 279925.44 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: LPUSH - Parallel Connections: 50 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: LPUSH - Parallel Connections: 50 b c a 80K 160K 240K 320K 400K 378526.91 378091.84 377803.31 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: SADD - Parallel Connections: 100 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: SADD - Parallel Connections: 100 b c a 70K 140K 210K 280K 350K 319358.22 314827.75 314192.72 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: SADD - Parallel Connections: 500 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: SADD - Parallel Connections: 500 b c a 70K 140K 210K 280K 350K 311701.25 310927.88 309906.47 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: SADD - Parallel Connections: 900 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: SADD - Parallel Connections: 900 a c b 70K 140K 210K 280K 350K 307442.56 305295.03 304371.41 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: HMSET - Parallel Connections: 100 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: HMSET - Parallel Connections: 100 a c b 50K 100K 150K 200K 250K 220859.77 219591.98 214634.66 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: HMSET - Parallel Connections: 500 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: HMSET - Parallel Connections: 500 a b c 40K 80K 120K 160K 200K 205129.89 204631.22 203269.38 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: HMSET - Parallel Connections: 900 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: HMSET - Parallel Connections: 900 b c a 40K 80K 120K 160K 200K 203707.48 202206.47 201398.50 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: LPUSH - Parallel Connections: 100 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: LPUSH - Parallel Connections: 100 a c b 70K 140K 210K 280K 350K 318623.53 299638.03 290675.72 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: LPUSH - Parallel Connections: 500 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: LPUSH - Parallel Connections: 500 b a c 60K 120K 180K 240K 300K 285523.38 277528.00 270480.81 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
KeyDB Test: LPUSH - Parallel Connections: 900 OpenBenchmarking.org Requests Per Second, More Is Better KeyDB 6.3.2 Test: LPUSH - Parallel Connections: 900 a b c 60K 120K 180K 240K 300K 270349.19 269025.47 267054.06 1. (CXX) g++ options: -ggdb -rdynamic -lm -lz -lcrypto -lbz2 -lzstd -llz4 -lsnappy -latomic -ldl -pthread -lrt -luuid -lcurl -lsystemd -lssl -std=c++14 -pedantic -fno-rtti -O2 -flto -MMD
Neural Magic DeepSparse Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream b a c 1.1634 2.3268 3.4902 4.6536 5.817 5.1708 5.1659 5.1593
Neural Magic DeepSparse Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream b a c 200 400 600 800 1000 773.55 774.29 775.28
Neural Magic DeepSparse Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-Stream b a c 1.1342 2.2684 3.4026 4.5368 5.671 5.0409 5.0356 5.0291
Neural Magic DeepSparse Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-Stream b a c 40 80 120 160 200 198.36 198.57 198.82
Neural Magic DeepSparse Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Stream a b c 14 28 42 56 70 64.94 64.89 64.57
Neural Magic DeepSparse Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Stream a b c 14 28 42 56 70 61.58 61.62 61.92
Neural Magic DeepSparse Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Synchronous Single-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Synchronous Single-Stream b c a 12 24 36 48 60 51.88 51.85 51.44
Neural Magic DeepSparse Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Synchronous Single-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Synchronous Single-Stream b c a 5 10 15 20 25 19.26 19.27 19.42
Neural Magic DeepSparse Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Stream c b a 5 10 15 20 25 20.22 20.20 20.04
Neural Magic DeepSparse Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Stream c b a 40 80 120 160 200 197.79 197.82 199.16
Neural Magic DeepSparse Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Synchronous Single-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Synchronous Single-Stream c b a 4 8 12 16 20 18.19 18.13 18.10
Neural Magic DeepSparse Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Synchronous Single-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Synchronous Single-Stream c b a 12 24 36 48 60 54.95 55.12 55.22
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream a b c 7 14 21 28 35 28.50 28.48 28.41
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Stream a b c 30 60 90 120 150 140.27 140.36 140.73
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO - Scenario: Synchronous Single-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: CV Detection, YOLOv5s COCO - Scenario: Synchronous Single-Stream b c a 6 12 18 24 30 26.16 26.14 26.14
Neural Magic DeepSparse Model: CV Detection, YOLOv5s COCO - Scenario: Synchronous Single-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: CV Detection, YOLOv5s COCO - Scenario: Synchronous Single-Stream b c a 9 18 27 36 45 38.20 38.23 38.23
Neural Magic DeepSparse Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream c a b 14 28 42 56 70 62.33 62.17 62.03
Neural Magic DeepSparse Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream c a b 14 28 42 56 70 64.16 64.32 64.46
Neural Magic DeepSparse Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-Stream b a c 13 26 39 52 65 56.35 56.32 56.29
Neural Magic DeepSparse Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-Stream b a c 4 8 12 16 20 17.73 17.74 17.75
Neural Magic DeepSparse Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream c b a 10 20 30 40 50 45.73 45.66 45.60
Neural Magic DeepSparse Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream c b a 20 40 60 80 100 87.45 87.58 87.70
Neural Magic DeepSparse Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-Stream b a c 9 18 27 36 45 41.28 41.25 41.12
Neural Magic DeepSparse Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-Stream b a c 6 12 18 24 30 24.21 24.23 24.31
Neural Magic DeepSparse Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream b c a 2 4 6 8 10 7.1392 7.1222 7.0942
Neural Magic DeepSparse Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream b c a 120 240 360 480 600 560.26 561.59 563.81
Neural Magic DeepSparse Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Synchronous Single-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Synchronous Single-Stream c b a 2 4 6 8 10 7.0170 7.0162 7.0091
Neural Magic DeepSparse Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Synchronous Single-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Synchronous Single-Stream c b a 30 60 90 120 150 142.48 142.50 142.65
Neural Magic DeepSparse Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Stream b c a 5 10 15 20 25 22.27 22.19 22.19
Neural Magic DeepSparse Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Stream b c a 40 80 120 160 200 179.60 180.23 180.24
Neural Magic DeepSparse Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Synchronous Single-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Synchronous Single-Stream b a c 5 10 15 20 25 20.31 20.28 20.21
Neural Magic DeepSparse Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Synchronous Single-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Synchronous Single-Stream b a c 11 22 33 44 55 49.22 49.28 49.46
Neural Magic DeepSparse Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream c b a 1.163 2.326 3.489 4.652 5.815 5.1690 5.1574 5.1498
Neural Magic DeepSparse Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream c b a 200 400 600 800 1000 773.81 773.86 776.67
Neural Magic DeepSparse Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-Stream OpenBenchmarking.org items/sec, More Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-Stream c b a 1.1324 2.2648 3.3972 4.5296 5.662 5.0327 5.0184 5.0161
Neural Magic DeepSparse Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-Stream OpenBenchmarking.org ms/batch, Fewer Is Better Neural Magic DeepSparse 1.3.2 Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-Stream c b a 40 80 120 160 200 198.68 199.25 199.34
Phoronix Test Suite v10.8.5