3950X Sep

AMD Ryzen 9 3950X 16-Core testing with a ASUS ROG CROSSHAIR VII HERO (WI-FI) (3103 BIOS) and Sapphire AMD Radeon RX 470 4GB on Ubuntu 22.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2209044-NE-3950XSEP507
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results

Limit displaying results to tests within:

Timed Code Compilation 5 Tests
C/C++ Compiler Tests 3 Tests
Compression Tests 2 Tests
CPU Massive 3 Tests
Creator Workloads 3 Tests
HPC - High Performance Computing 2 Tests
Machine Learning 2 Tests
Multi-Core 9 Tests
Programmer / Developer System Benchmarks 6 Tests
Python Tests 2 Tests
Server CPU Tests 2 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
A
September 03 2022
  1 Hour, 56 Minutes
B
September 03 2022
  2 Hours, 15 Minutes
C
September 03 2022
  2 Hours, 17 Minutes
D
September 03 2022
  2 Hours, 18 Minutes
E
September 03 2022
  13 Hours, 13 Minutes
Invert Hiding All Results Option
  4 Hours, 24 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


3950X SepOpenBenchmarking.orgPhoronix Test SuiteAMD Ryzen 9 3950X 16-Core @ 3.50GHz (16 Cores)ASUS ROG CROSSHAIR VII HERO (WI-FI) (3103 BIOS)AMD Starship/Matisse16GBSamsung SSD 970 EVO 250GBSapphire AMD Radeon RX 470 4GB (1260/1750MHz)AMD Ellesmere HDMI AudioDELL S2409WIntel I211 + Realtek RTL8822BE 802.11a/b/g/n/acUbuntu 22.045.19.0-051900daily20220813-generic (x86_64)GNOME Shell 42.2X Server + Wayland4.6 Mesa 22.0.5 (LLVM 13.0.1 DRM 3.48)1.3.204GCC 11.2.0ext41920x1080ProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilerFile-SystemScreen Resolution3950X Sep BenchmarksSystem Logs- Transparent Huge Pages: madvise- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - NONE / errors=remount-ro,relatime,rw / Block Size: 4096- Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0x8701021 - Python 3.10.4- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Mitigation of IBPB + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: always-on STIBP: forced RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected

ABCDEResult OverviewPhoronix Test Suite100%116%132%148%etcdNatronMobile Neural Network7-Zip CompressionTimed PHP CompilationTimed Erlang/OTP CompilationGraphicsMagickC-BloscUnpacking The Linux KernelOpenVINOTimed Wasmer CompilationTimed CPython CompilationTimed Node.js Compilation

3950X Sepetcd: RANGE - 50 - 100 - Average Latencyetcd: RANGE - 50 - 100etcd: PUT - 500 - 100 - Average Latencyetcd: PUT - 500 - 100etcd: RANGE - 100 - 100 - Average Latencyetcd: PUT - 100 - 100 - Average Latencyetcd: RANGE - 100 - 100etcd: PUT - 50 - 100 - Average Latencyetcd: PUT - 100 - 100etcd: PUT - 50 - 100etcd: RANGE - 500 - 100etcd: RANGE - 500 - 100 - Average Latencyetcd: RANGE - 500 - 1000 - Average Latencyetcd: RANGE - 500 - 1000etcd: PUT - 500 - 1000 - Average Latencyetcd: PUT - 500 - 1000etcd: PUT - 50 - 1000 - Average Latencymnn: SqueezeNetV1.0openvino: Vehicle Detection FP16 - CPUmnn: squeezenetv1.1openvino: Vehicle Detection FP16 - CPUmnn: MobileNetV2_224compress-7zip: Compression Ratingmnn: mobilenetV3mnn: nasnetopenvino: Person Detection FP16 - CPUopenvino: Machine Translation EN To DE FP16 - CPUopenvino: Machine Translation EN To DE FP16 - CPUgraphics-magick: Rotateopenvino: Face Detection FP16 - CPUopenvino: Person Detection FP32 - CPUopenvino: Person Detection FP16 - CPUmnn: resnet-v2-50graphics-magick: Resizingopenvino: Person Detection FP32 - CPUgraphics-magick: HWB Color Spaceopenvino: Age Gender Recognition Retail 0013 FP16 - CPUblosc: blosclz shufflemnn: inception-v3build-erlang: Time To Compileopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUbuild-php: Time To Compileopenvino: Face Detection FP16 - CPUgraphics-magick: Swirlbuild-python: Released Build, PGO + LTO Optimizedgraphics-magick: Sharpenopenvino: Age Gender Recognition Retail 0013 FP16 - CPUblosc: blosclz bitshuffleopenvino: Vehicle Detection FP16-INT8 - CPUopenvino: Vehicle Detection FP16-INT8 - CPUunpack-linux: linux-5.19.tar.xzmnn: mobilenet-v1-1.0openvino: Weld Porosity Detection FP16-INT8 - CPUopenvino: Weld Porosity Detection FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUgraphics-magick: Enhancedopenvino: Face Detection FP16-INT8 - CPUbuild-wasmer: Time To Compileopenvino: Face Detection FP16-INT8 - CPUopenvino: Weld Porosity Detection FP16 - CPUopenvino: Weld Porosity Detection FP16 - CPUcompress-7zip: Decompression Ratinggraphics-magick: Noise-Gaussianbuild-nodejs: Time To Compilebuild-python: Defaultopenvino: Person Vehicle Bike Detection FP16 - CPUopenvino: Person Vehicle Bike Detection FP16 - CPUnatron: Spaceshipetcd: RANGE - 100 - 1000 - Average Latencyetcd: RANGE - 100 - 1000etcd: RANGE - 50 - 1000 - Average Latencyetcd: RANGE - 50 - 1000etcd: PUT - 100 - 1000 - Average Latencyetcd: PUT - 100 - 1000etcd: PUT - 50 - 1000ABCDE3.330099.58133.329940.61613.73.326866.57593.329893.434230093.385818991.52395.314.865917.241214.766273.0768115.421351.333.37911.373.592842052.01312.9782.37110.0136.324681258.341659.891668.8520.17716882.3972910973.6714794.427.847111.9391.0160.3783.16659272.5782511.4410091.19.13437.737.6982.698570.8128.0115767.85364695.5257.3525.74321.7812.4175100290497.1918.361504.467.923.311.784607.34371190098.199511.685435.511190351.73066.216178.2496.116446.37866.2616165.5907616572.280116614.853516096.47416.214.865709.454114.865970.609911.15.566343.853.37411.613.571823502.00613.0612.38110.9236.044631255.921637.811661.5320.06416742.4272211065.9614622.127.608112.06160.6273.16657272.3582511.4310039.29.14397.6822.692572.4127.9215826.05363693.7857.3175.76321.1812.4374832289497.08118.414503.247.933.411.784592.90721189816.411911.784546.061289723.13166.315954.55576.315901.95266.26.316010.01466.315915.464515974.855215859.12486.314.865751.575214.865829.2852115.401344.153.27911.63.542832911.97112.7982.39110.8436.064611271.041641.561663.2620.32116842.4172211106.2414763.627.76112.666160.573.14658272.8032511.4310074.19.09439.427.6532.7572.8427.9115841.7364692.657.325.77321.3812.4275045289495.67418.379504.037.923.311.684946.78541189932.694411.684943.4389853.1686.415613.65856.415561.35946.46.415511.56426.415687.981615721.313215521.42516.414.865750.654614.865753.482311.15.541344.263.3711.63.508840962.01413.0572.41110.5536.154621274.451643.311645.120.06916732.4272311031.8414684.227.825113.131160.493.13657272.4892511.4410053.49.1438.997.6752.69572.1927.9415834.93363693.8357.6185.76321.9912.474994289496.20418.366504.367.923.411.784795.25441190539.598811.784669.789289674.36127.413528.95826.615135.46517.36.513662.18316.415439.577715591.583512552.82918.020.547997.950618.852164.511413.75.380340.863.31211.713.506834731.97212.8102.39111.7935.754611269.031661.551654.2920.03216662.3972011013.0814626.627.527112.5941.0160.0393.14653274.5852491.4410021.89.15436.567.6482.683569.2528.0815751.27362696.4057.3535.74320.4112.4674898289496.36618.387503.157.943.318.155692.008417.557913.176816.062573.154672969.6471OpenBenchmarking.org

etcd

Etcd is a distributed, reliable key-value store intended for critical data of a distributed system. Etcd is written in Golang and part of the Cloud Native Computing Foundation (CNCF) and used by Kubernetes, Rook, CoreDNS, and other open-source software. This test profile uses Etcd's built-in benchmark to stress the PUT and RANGE performance of a single node / local system. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: RANGE - Connections: 50 - Clients: 100 - Average LatencyEDCBA246810SE +/- 0.07, N = 77.46.46.36.23.3
OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: RANGE - Connections: 50 - Clients: 100 - Average LatencyEDCBA3691215Min: 7.3 / Avg: 7.39 / Max: 7.8

OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: RANGE - Connections: 50 - Clients: 100EDCBA6K12K18K24K30KSE +/- 119.86, N = 713528.9615613.6615954.5616178.2530099.58
OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: RANGE - Connections: 50 - Clients: 100EDCBA5K10K15K20K25KMin: 12831.17 / Avg: 13528.96 / Max: 13765.73

OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: PUT - Connections: 500 - Clients: 100 - Average LatencyEDCBA246810SE +/- 0.07, N = 66.66.46.36.13.3
OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: PUT - Connections: 500 - Clients: 100 - Average LatencyEDCBA3691215Min: 6.5 / Avg: 6.6 / Max: 6.9

OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: PUT - Connections: 500 - Clients: 100EDCBA6K12K18K24K30KSE +/- 148.13, N = 615135.4715561.3615901.9516446.3829940.62
OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: PUT - Connections: 500 - Clients: 100EDCBA5K10K15K20K25KMin: 14450.03 / Avg: 15135.47 / Max: 15402.58

OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: RANGE - Connections: 100 - Clients: 100 - Average LatencyEDCBA246810SE +/- 0.11, N = 97.36.46.26.23.7
OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: RANGE - Connections: 100 - Clients: 100 - Average LatencyEDCBA3691215Min: 6.7 / Avg: 7.31 / Max: 7.9

OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: PUT - Connections: 100 - Clients: 100 - Average LatencyEDCBA246810SE +/- 0.00, N = 36.56.46.36.03.3
OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: PUT - Connections: 100 - Clients: 100 - Average LatencyEDCBA3691215Min: 6.5 / Avg: 6.5 / Max: 6.5

OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: RANGE - Connections: 100 - Clients: 100EDCBA6K12K18K24K30KSE +/- 215.62, N = 913662.1815511.5616010.0116165.5926866.58
OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: RANGE - Connections: 100 - Clients: 100EDCBA5K10K15K20K25KMin: 12593.49 / Avg: 13662.18 / Max: 14914.31

OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: PUT - Connections: 50 - Clients: 100 - Average LatencyEDCBA246810SE +/- 0.00, N = 36.46.46.36.03.3
OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: PUT - Connections: 50 - Clients: 100 - Average LatencyEDCBA3691215Min: 6.4 / Avg: 6.4 / Max: 6.4

OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: PUT - Connections: 100 - Clients: 100EDCBA6K12K18K24K30KSE +/- 18.08, N = 315439.5815687.9815915.4616572.2829893.43
OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: PUT - Connections: 100 - Clients: 100EDCBA5K10K15K20K25KMin: 15415.88 / Avg: 15439.58 / Max: 15475.07

OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: PUT - Connections: 50 - Clients: 100EDCBA6K12K18K24K30KSE +/- 9.73, N = 315591.5815721.3115974.8616614.8530093.39
OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: PUT - Connections: 50 - Clients: 100EDCBA5K10K15K20K25KMin: 15579.68 / Avg: 15591.58 / Max: 15610.87

OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: RANGE - Connections: 500 - Clients: 100EDCBA4K8K12K16K20KSE +/- 112.68, N = 712552.8315521.4315859.1216096.4718991.52
OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: RANGE - Connections: 500 - Clients: 100EDCBA3K6K9K12K15KMin: 11941.77 / Avg: 12552.83 / Max: 12902.03

OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: RANGE - Connections: 500 - Clients: 100 - Average LatencyEDCBA246810SE +/- 0.08, N = 78.06.46.36.25.3
OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: RANGE - Connections: 500 - Clients: 100 - Average LatencyEDCBA3691215Min: 7.7 / Avg: 7.97 / Max: 8.4

OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: RANGE - Connections: 500 - Clients: 1000 - Average LatencyEDCBA510152025SE +/- 0.14, N = 1520.514.814.814.814.8
OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: RANGE - Connections: 500 - Clients: 1000 - Average LatencyEDCBA510152025Min: 18.7 / Avg: 20.45 / Max: 21

OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: RANGE - Connections: 500 - Clients: 1000EBDCA14K28K42K56K70KSE +/- 337.08, N = 1547997.9565709.4565750.6565751.5865917.24
OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: RANGE - Connections: 500 - Clients: 1000EBDCA11K22K33K44K55KMin: 46750.7 / Avg: 47997.95 / Max: 52291.17

OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: PUT - Connections: 500 - Clients: 1000 - Average LatencyEDCBA510152025SE +/- 0.21, N = 1518.814.814.814.814.7
OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: PUT - Connections: 500 - Clients: 1000 - Average LatencyEDCBA510152025Min: 16.3 / Avg: 18.81 / Max: 19.6

OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: PUT - Connections: 500 - Clients: 1000EDCBA14K28K42K56K70KSE +/- 625.45, N = 1552164.5165753.4865829.2965970.6166273.08
OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: PUT - Connections: 500 - Clients: 1000EDCBA11K22K33K44K55KMin: 49978.83 / Avg: 52164.51 / Max: 60081.22

OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: PUT - Connections: 50 - Clients: 1000 - Average LatencyEDBCA48121620SE +/- 0.21, N = 1513.711.111.111.011.0
OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: PUT - Connections: 50 - Clients: 1000 - Average LatencyEDBCA48121620Min: 11 / Avg: 13.66 / Max: 14.5

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: SqueezeNetV1.0BDACE1.25242.50483.75725.00966.262SE +/- 0.019, N = 35.5665.5415.4215.4015.380MIN: 5.51 / MAX: 6.67MIN: 5.49 / MAX: 6.76MIN: 5.35 / MAX: 21.72MIN: 5.35 / MAX: 6.22MIN: 5.3 / MAX: 6.521. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: SqueezeNetV1.0BDACE246810Min: 5.35 / Avg: 5.38 / Max: 5.421. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16 - Device: CPUEBCDA80160240320400SE +/- 2.43, N = 3340.86343.85344.15344.26351.331. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16 - Device: CPUEBCDA60120180240300Min: 336.05 / Avg: 340.86 / Max: 343.851. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: squeezenetv1.1ABDEC0.76031.52062.28093.04123.8015SE +/- 0.008, N = 33.3793.3743.3703.3123.279MIN: 3.34 / MAX: 4.37MIN: 3.33 / MAX: 4.48MIN: 3.32 / MAX: 4.47MIN: 3.26 / MAX: 4.42MIN: 3.24 / MAX: 4.291. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: squeezenetv1.1ABDEC246810Min: 3.3 / Avg: 3.31 / Max: 3.331. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16 - Device: CPUEBDCA3691215SE +/- 0.08, N = 311.7111.6111.6011.6011.37MIN: 10.26 / MAX: 21.14MIN: 10.6 / MAX: 28.3MIN: 10.66 / MAX: 22.12MIN: 10.53 / MAX: 23.08MIN: 10.52 / MAX: 21.671. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16 - Device: CPUEBDCA3691215Min: 11.61 / Avg: 11.71 / Max: 11.881. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: MobileNetV2_224ABCDE0.80821.61642.42463.23284.041SE +/- 0.022, N = 33.5923.5713.5423.5083.506MIN: 3.55 / MAX: 6.84MIN: 3.53 / MAX: 4.69MIN: 3.5 / MAX: 5.07MIN: 3.47 / MAX: 4.6MIN: 3.44 / MAX: 9.591. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: MobileNetV2_224ABCDE246810Min: 3.48 / Avg: 3.51 / Max: 3.551. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

7-Zip Compression

This is a test of 7-Zip compression/decompression with its integrated benchmark feature. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 22.01Test: Compression RatingBCEDA20K40K60K80K100KSE +/- 30.72, N = 382350832918347384096842051. (CXX) g++ options: -lpthread -ldl -O2 -fPIC
OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 22.01Test: Compression RatingBCEDA15K30K45K60K75KMin: 83439 / Avg: 83472.67 / Max: 835341. (CXX) g++ options: -lpthread -ldl -O2 -fPIC

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: mobilenetV3DABEC0.45320.90641.35961.81282.266SE +/- 0.009, N = 32.0142.0132.0061.9721.971MIN: 1.99 / MAX: 2.25MIN: 1.99 / MAX: 3.11MIN: 1.98 / MAX: 2.75MIN: 1.93 / MAX: 3.06MIN: 1.94 / MAX: 3.071. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: mobilenetV3DABEC246810Min: 1.96 / Avg: 1.97 / Max: 1.981. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: nasnetBDAEC3691215SE +/- 0.02, N = 313.0613.0612.9812.8112.80MIN: 12.97 / MAX: 19.07MIN: 12.94 / MAX: 19.32MIN: 12.88 / MAX: 19.14MIN: 12.68 / MAX: 30.81MIN: 12.68 / MAX: 19.171. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: nasnetBDAEC48121620Min: 12.79 / Avg: 12.81 / Max: 12.851. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Person Detection FP16 - Device: CPUABCED0.54231.08461.62692.16922.7115SE +/- 0.01, N = 32.372.382.392.392.411. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Person Detection FP16 - Device: CPUABCED246810Min: 2.37 / Avg: 2.39 / Max: 2.411. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Machine Translation EN To DE FP16 - Device: CPUEBCDA306090120150SE +/- 0.33, N = 3111.79110.92110.84110.55110.01MIN: 86.68 / MAX: 133.95MIN: 92.03 / MAX: 133.28MIN: 84.72 / MAX: 133.04MIN: 89.56 / MAX: 138.28MIN: 85.11 / MAX: 135.511. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Machine Translation EN To DE FP16 - Device: CPUEBCDA20406080100Min: 111.33 / Avg: 111.79 / Max: 112.421. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Machine Translation EN To DE FP16 - Device: CPUEBCDA816243240SE +/- 0.10, N = 335.7536.0436.0636.1536.321. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Machine Translation EN To DE FP16 - Device: CPUEBCDA816243240Min: 35.55 / Avg: 35.75 / Max: 35.91. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

GraphicsMagick

This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: RotateCEDBA100200300400500SE +/- 0.33, N = 34614614624634681. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: RotateCEDBA80160240320400Min: 460 / Avg: 460.67 / Max: 4611. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Face Detection FP16 - Device: CPUDCEAB30060090012001500SE +/- 1.95, N = 31274.451271.041269.031258.341255.92MIN: 1226.87 / MAX: 1355.68MIN: 1241.13 / MAX: 1380.49MIN: 1210.75 / MAX: 1378.29MIN: 1201.93 / MAX: 1341.23MIN: 1206.49 / MAX: 1356.421. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Face Detection FP16 - Device: CPUDCEAB2004006008001000Min: 1265.13 / Avg: 1269.03 / Max: 1271.021. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Person Detection FP32 - Device: CPUEADCB400800120016002000SE +/- 11.54, N = 31661.551659.891643.311641.561637.81MIN: 1464.72 / MAX: 1861.45MIN: 1461.24 / MAX: 1823.91MIN: 1495.03 / MAX: 1843.12MIN: 1477.93 / MAX: 1797.33MIN: 1490.45 / MAX: 1850.311. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Person Detection FP32 - Device: CPUEADCB30060090012001500Min: 1639.5 / Avg: 1661.55 / Max: 1678.491. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Person Detection FP16 - Device: CPUACBED400800120016002000SE +/- 9.27, N = 31668.851663.261661.531654.291645.10MIN: 1447.91 / MAX: 1780.07MIN: 1482.48 / MAX: 1805.04MIN: 1480.92 / MAX: 1803.33MIN: 1481.86 / MAX: 1868.97MIN: 1489.68 / MAX: 1839.491. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Person Detection FP16 - Device: CPUACBED30060090012001500Min: 1642.23 / Avg: 1654.29 / Max: 1672.511. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: resnet-v2-50CADBE510152025SE +/- 0.07, N = 320.3220.1820.0720.0620.03MIN: 20.04 / MAX: 36.83MIN: 19.89 / MAX: 26.06MIN: 19.84 / MAX: 34.96MIN: 19.82 / MAX: 36.51MIN: 19.6 / MAX: 34.61. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: resnet-v2-50CADBE510152025Min: 19.9 / Avg: 20.03 / Max: 20.121. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

GraphicsMagick

This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: ResizingEDBCA400800120016002000SE +/- 3.67, N = 3166616731674168416881. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: ResizingEDBCA30060090012001500Min: 1659 / Avg: 1666.33 / Max: 16701. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Person Detection FP32 - Device: CPUAECBD0.54451.0891.63352.1782.7225SE +/- 0.02, N = 32.392.392.412.422.421. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Person Detection FP32 - Device: CPUAECBD246810Min: 2.36 / Avg: 2.39 / Max: 2.431. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

GraphicsMagick

This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: HWB Color SpaceEBCDA160320480640800SE +/- 0.58, N = 37207227227237291. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: HWB Color SpaceEBCDA130260390520650Min: 719 / Avg: 720 / Max: 7211. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16 - Device: CPUAEDBC2K4K6K8K10KSE +/- 33.81, N = 310973.6711013.0811031.8411065.9611106.241. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16 - Device: CPUAEDBC2K4K6K8K10KMin: 10965.56 / Avg: 11013.08 / Max: 11078.51. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.3Test: blosclz shuffleBEDCA3K6K9K12K15KSE +/- 39.75, N = 314622.114626.614684.214763.614794.41. (CC) gcc options: -std=gnu99 -O3 -lrt -lm
OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.3Test: blosclz shuffleBEDCA3K6K9K12K15KMin: 14558 / Avg: 14626.6 / Max: 14695.71. (CC) gcc options: -std=gnu99 -O3 -lrt -lm

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: inception-v3ADCBE714212835SE +/- 0.09, N = 327.8527.8327.7627.6127.53MIN: 27.42 / MAX: 33.72MIN: 27.33 / MAX: 33.14MIN: 27.3 / MAX: 72.79MIN: 27.19 / MAX: 34.67MIN: 27.08 / MAX: 47.121. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: inception-v3ADCBE612182430Min: 27.4 / Avg: 27.53 / Max: 27.71. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

Timed Erlang/OTP Compilation

This test times how long it takes to compile Erlang/OTP. Erlang is a programming language and run-time for massively scalable soft real-time systems with high availability requirements. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Erlang/OTP Compilation 25.0Time To CompileDCEBA306090120150SE +/- 0.20, N = 3113.13112.67112.59112.06111.94
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Erlang/OTP Compilation 25.0Time To CompileDCEBA20406080100Min: 112.19 / Avg: 112.59 / Max: 112.81

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUEADCB0.22730.45460.68190.90921.1365SE +/- 0.00, N = 31.011.011.001.001.00MIN: 0.95 / MAX: 17.36MIN: 0.95 / MAX: 7.42MIN: 0.95 / MAX: 5.25MIN: 0.95 / MAX: 7.16MIN: 0.95 / MAX: 7.421. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUEADCB246810Min: 1.01 / Avg: 1.01 / Max: 1.011. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Timed PHP Compilation

This test times how long it takes to build PHP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed PHP Compilation 8.1.9Time To CompileBCDAE1428425670SE +/- 0.22, N = 360.6360.5760.4960.3860.04
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed PHP Compilation 8.1.9Time To CompileBCDAE1224364860Min: 59.79 / Avg: 60.04 / Max: 60.49

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Face Detection FP16 - Device: CPUDCEAB0.7111.4222.1332.8443.555SE +/- 0.00, N = 33.133.143.143.163.161. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Face Detection FP16 - Device: CPUDCEAB246810Min: 3.13 / Avg: 3.14 / Max: 3.141. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

GraphicsMagick

This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: SwirlEBDCA140280420560700SE +/- 1.20, N = 36536576576586591. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: SwirlEBDCA120240360480600Min: 651 / Avg: 652.67 / Max: 6551. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread

Timed CPython Compilation

This test times how long it takes to build the reference Python implementation, CPython, with optimizations and LTO enabled for a release build. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed CPython Compilation 3.10.6Build Configuration: Released Build, PGO + LTO OptimizedECADB60120180240300274.59272.80272.58272.49272.36

GraphicsMagick

This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: SharpenEABCD50100150200250SE +/- 0.67, N = 32492512512512511. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: SharpenEABCD50100150200250Min: 248 / Avg: 248.67 / Max: 2501. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16 - Device: CPUEDACB0.3240.6480.9721.2961.62SE +/- 0.00, N = 31.441.441.441.431.43MIN: 1.26 / MAX: 17.8MIN: 1.27 / MAX: 18.07MIN: 1.23 / MAX: 17.9MIN: 1.25 / MAX: 17.38MIN: 1.27 / MAX: 18.121. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16 - Device: CPUEDACB246810Min: 1.43 / Avg: 1.44 / Max: 1.441. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

C-Blosc

C-Blosc (c-blosc2) simple, compressed, fast and persistent data store library for C. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.3Test: blosclz bitshuffleEBDCA2K4K6K8K10KSE +/- 4.68, N = 310021.810039.210053.410074.110091.11. (CC) gcc options: -std=gnu99 -O3 -lrt -lm
OpenBenchmarking.orgMB/s, More Is BetterC-Blosc 2.3Test: blosclz bitshuffleEBDCA2K4K6K8K10KMin: 10014.3 / Avg: 10021.77 / Max: 10030.41. (CC) gcc options: -std=gnu99 -O3 -lrt -lm

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16-INT8 - Device: CPUEADBC3691215SE +/- 0.01, N = 39.159.139.109.109.09MIN: 8.87 / MAX: 26.46MIN: 8.85 / MAX: 16.16MIN: 8.85 / MAX: 25.66MIN: 8.83 / MAX: 15.74MIN: 8.83 / MAX: 25.561. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16-INT8 - Device: CPUEADBC3691215Min: 9.14 / Avg: 9.15 / Max: 9.161. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16-INT8 - Device: CPUEADBC100200300400500SE +/- 0.31, N = 3436.56437.73438.99439.00439.421. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16-INT8 - Device: CPUEADBC80160240320400Min: 436.01 / Avg: 436.56 / Max: 437.081. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Unpacking The Linux Kernel

This test measures how long it takes to extract the .tar.xz Linux kernel source tree package. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterUnpacking The Linux Kernel 5.19linux-5.19.tar.xzABDCE246810SE +/- 0.009, N = 47.6987.6827.6757.6537.648
OpenBenchmarking.orgSeconds, Fewer Is BetterUnpacking The Linux Kernel 5.19linux-5.19.tar.xzABDCE3691215Min: 7.63 / Avg: 7.65 / Max: 7.67

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: mobilenet-v1-1.0CABDE0.60751.2151.82252.433.0375SE +/- 0.024, N = 32.7002.6982.6922.6902.683MIN: 2.66 / MAX: 3.68MIN: 2.66 / MAX: 3.73MIN: 2.65 / MAX: 3.77MIN: 2.65 / MAX: 3.76MIN: 2.61 / MAX: 3.751. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: mobilenet-v1-1.0CABDE246810Min: 2.65 / Avg: 2.68 / Max: 2.731. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16-INT8 - Device: CPUEADBC120240360480600SE +/- 0.37, N = 3569.25570.81572.19572.41572.841. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16-INT8 - Device: CPUEADBC100200300400500Min: 568.64 / Avg: 569.25 / Max: 569.911. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16-INT8 - Device: CPUEADBC714212835SE +/- 0.02, N = 328.0828.0127.9427.9227.91MIN: 27.25 / MAX: 44.19MIN: 27.19 / MAX: 33.76MIN: 27.08 / MAX: 44.05MIN: 27.09 / MAX: 44.07MIN: 27.1 / MAX: 34.121. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16-INT8 - Device: CPUEADBC612182430Min: 28.04 / Avg: 28.08 / Max: 28.111. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUEABDC3K6K9K12K15KSE +/- 1.59, N = 315751.2715767.8515826.0515834.9315841.701. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUEABDC3K6K9K12K15KMin: 15749.15 / Avg: 15751.27 / Max: 15754.391. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

GraphicsMagick

This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: EnhancedEBDAC80160240320400SE +/- 0.33, N = 33623633633643641. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: EnhancedEBDAC70140210280350Min: 361 / Avg: 361.67 / Max: 3621. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Face Detection FP16-INT8 - Device: CPUEADBC150300450600750SE +/- 0.32, N = 3696.40695.52693.83693.78692.60MIN: 688.69 / MAX: 721.06MIN: 691.3 / MAX: 708.41MIN: 684.96 / MAX: 701.04MIN: 686.88 / MAX: 717.7MIN: 686.95 / MAX: 699.351. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Face Detection FP16-INT8 - Device: CPUEADBC120240360480600Min: 695.8 / Avg: 696.4 / Max: 696.911. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Timed Wasmer Compilation

This test times how long it takes to compile Wasmer. Wasmer is written in the Rust programming language and is a WebAssembly runtime implementation that supports WASI and EmScripten. This test profile builds Wasmer with the Cranelift and Singlepast compiler features enabled. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Wasmer Compilation 2.3Time To CompileDEACB1326395265SE +/- 0.11, N = 357.6257.3557.3557.3257.321. (CC) gcc options: -m64 -ldl -lgcc_s -lutil -lrt -lpthread -lm -lc -pie -nodefaultlibs
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Wasmer Compilation 2.3Time To CompileDEACB1122334455Min: 57.24 / Avg: 57.35 / Max: 57.581. (CC) gcc options: -m64 -ldl -lgcc_s -lutil -lrt -lpthread -lm -lc -pie -nodefaultlibs

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Face Detection FP16-INT8 - Device: CPUAEBDC1.29832.59663.89495.19326.4915SE +/- 0.00, N = 35.745.745.765.765.771. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Face Detection FP16-INT8 - Device: CPUAEBDC246810Min: 5.74 / Avg: 5.74 / Max: 5.751. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16 - Device: CPUEBCAD70140210280350SE +/- 0.08, N = 3320.41321.18321.38321.78321.991. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16 - Device: CPUEBCAD60120180240300Min: 320.32 / Avg: 320.41 / Max: 320.571. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16 - Device: CPUEBCAD3691215SE +/- 0.00, N = 312.4612.4312.4212.4112.40MIN: 10.39 / MAX: 27.74MIN: 10.42 / MAX: 17.91MIN: 10.48 / MAX: 15.98MIN: 10.44 / MAX: 22.3MIN: 10.39 / MAX: 18.881. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16 - Device: CPUEBCAD48121620Min: 12.45 / Avg: 12.46 / Max: 12.461. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

7-Zip Compression

This is a test of 7-Zip compression/decompression with its integrated benchmark feature. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 22.01Test: Decompression RatingBEDCA16K32K48K64K80KSE +/- 23.02, N = 374832748987499475045751001. (CXX) g++ options: -lpthread -ldl -O2 -fPIC
OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 22.01Test: Decompression RatingBEDCA13K26K39K52K65KMin: 74855 / Avg: 74897.67 / Max: 749341. (CXX) g++ options: -lpthread -ldl -O2 -fPIC

GraphicsMagick

This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: Noise-GaussianBCDEA60120180240300SE +/- 0.33, N = 32892892892892901. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: Noise-GaussianBCDEA50100150200250Min: 288 / Avg: 288.67 / Max: 2891. (CC) gcc options: -fopenmp -O2 -ljbig -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread

Timed Node.js Compilation

This test profile times how long it takes to build/compile Node.js itself from source. Node.js is a JavaScript run-time built from the Chrome V8 JavaScript engine while itself is written in C/C++. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Node.js Compilation 18.8Time To CompileABEDC110220330440550SE +/- 0.12, N = 3497.19497.08496.37496.20495.67
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Node.js Compilation 18.8Time To CompileABEDC90180270360450Min: 496.13 / Avg: 496.37 / Max: 496.55

Timed CPython Compilation

This test times how long it takes to build the reference Python implementation, CPython, with optimizations and LTO enabled for a release build. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed CPython Compilation 3.10.6Build Configuration: DefaultBECDA51015202518.4118.3918.3818.3718.36

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Person Vehicle Bike Detection FP16 - Device: CPUEBCDA110220330440550SE +/- 0.63, N = 3503.15503.24504.03504.36504.461. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Person Vehicle Bike Detection FP16 - Device: CPUEBCDA90180270360450Min: 502.13 / Avg: 503.15 / Max: 504.31. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Person Vehicle Bike Detection FP16 - Device: CPUEBDCA246810SE +/- 0.01, N = 37.947.937.927.927.92MIN: 7.58 / MAX: 38.47MIN: 7.45 / MAX: 12.66MIN: 7.52 / MAX: 15.54MIN: 7.6 / MAX: 12.54MIN: 7.59 / MAX: 24.311. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared
OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Person Vehicle Bike Detection FP16 - Device: CPUEBDCA3691215Min: 7.92 / Avg: 7.94 / Max: 7.951. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Natron

Natron is an open-source, cross-platform compositing software for visual effects (VFX) and motion graphics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterNatron 2.4.3Input: SpaceshipACEBD0.7651.532.2953.063.825SE +/- 0.06, N = 123.33.33.33.43.4
OpenBenchmarking.orgFPS, More Is BetterNatron 2.4.3Input: SpaceshipACEBD246810Min: 2.8 / Avg: 3.32 / Max: 3.5

etcd

Etcd is a distributed, reliable key-value store intended for critical data of a distributed system. Etcd is written in Golang and part of the Cloud Native Computing Foundation (CNCF) and used by Kubernetes, Rook, CoreDNS, and other open-source software. This test profile uses Etcd's built-in benchmark to stress the PUT and RANGE performance of a single node / local system. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: RANGE - Connections: 100 - Clients: 1000 - Average LatencyEDBAC48121620SE +/- 0.46, N = 1518.111.711.711.711.6
OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: RANGE - Connections: 100 - Clients: 1000 - Average LatencyEDBAC510152025Min: 12 / Avg: 18.05 / Max: 19

OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: RANGE - Connections: 100 - Clients: 1000EBADC20K40K60K80K100KSE +/- 1989.42, N = 1555692.0184592.9184607.3484795.2584946.79
OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: RANGE - Connections: 100 - Clients: 1000EBADC15K30K45K60K75KMin: 52181.57 / Avg: 55692.01 / Max: 82561.73

OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: RANGE - Connections: 50 - Clients: 1000 - Average LatencyEDCBA48121620SE +/- 0.51, N = 1517.511.011.011.011.0
OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: RANGE - Connections: 50 - Clients: 1000 - Average LatencyEDCBA48121620Min: 11 / Avg: 17.47 / Max: 18.7

OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: RANGE - Connections: 50 - Clients: 1000EBCAD20K40K60K80K100KSE +/- 2395.30, N = 1557913.1889816.4189932.6990098.2090539.60
OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: RANGE - Connections: 50 - Clients: 1000EBCAD16K32K48K64K80KMin: 53176.27 / Avg: 57913.18 / Max: 89957.65

OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: PUT - Connections: 100 - Clients: 1000 - Average LatencyEDBCA48121620SE +/- 0.35, N = 1516.011.711.711.611.6
OpenBenchmarking.orgms, Fewer Is Betteretcd 3.5.4Test: PUT - Connections: 100 - Clients: 1000 - Average LatencyEDBCA48121620Min: 11.7 / Avg: 15.97 / Max: 16.8

OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: PUT - Connections: 100 - Clients: 1000EBDCA20K40K60K80K100KSE +/- 1719.86, N = 1562573.1584546.0684669.7984943.4385435.51
OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: PUT - Connections: 100 - Clients: 1000EBDCA15K30K45K60K75KMin: 58900.36 / Avg: 62573.15 / Max: 84531.19

OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: PUT - Connections: 50 - Clients: 1000EDBCA20K40K60K80K100KSE +/- 1306.71, N = 1572969.6589674.3689723.1389853.1790351.73
OpenBenchmarking.orgRequests/sec, More Is Betteretcd 3.5.4Test: PUT - Connections: 50 - Clients: 1000EDBCA16K32K48K64K80KMin: 68425.23 / Avg: 72969.65 / Max: 90034.51

75 Results Shown

etcd:
  RANGE - 50 - 100 - Average Latency
  RANGE - 50 - 100
  PUT - 500 - 100 - Average Latency
  PUT - 500 - 100
  RANGE - 100 - 100 - Average Latency
  PUT - 100 - 100 - Average Latency
  RANGE - 100 - 100
  PUT - 50 - 100 - Average Latency
  PUT - 100 - 100
  PUT - 50 - 100
  RANGE - 500 - 100
  RANGE - 500 - 100 - Average Latency
  RANGE - 500 - 1000 - Average Latency
  RANGE - 500 - 1000
  PUT - 500 - 1000 - Average Latency
  PUT - 500 - 1000
  PUT - 50 - 1000 - Average Latency
Mobile Neural Network
OpenVINO
Mobile Neural Network
OpenVINO
Mobile Neural Network
7-Zip Compression
Mobile Neural Network:
  mobilenetV3
  nasnet
OpenVINO:
  Person Detection FP16 - CPU
  Machine Translation EN To DE FP16 - CPU
  Machine Translation EN To DE FP16 - CPU
GraphicsMagick
OpenVINO:
  Face Detection FP16 - CPU
  Person Detection FP32 - CPU
  Person Detection FP16 - CPU
Mobile Neural Network
GraphicsMagick
OpenVINO
GraphicsMagick
OpenVINO
C-Blosc
Mobile Neural Network
Timed Erlang/OTP Compilation
OpenVINO
Timed PHP Compilation
OpenVINO
GraphicsMagick
Timed CPython Compilation
GraphicsMagick
OpenVINO
C-Blosc
OpenVINO:
  Vehicle Detection FP16-INT8 - CPU:
    ms
    FPS
Unpacking The Linux Kernel
Mobile Neural Network
OpenVINO:
  Weld Porosity Detection FP16-INT8 - CPU:
    FPS
    ms
  Age Gender Recognition Retail 0013 FP16-INT8 - CPU:
    FPS
GraphicsMagick
OpenVINO
Timed Wasmer Compilation
OpenVINO:
  Face Detection FP16-INT8 - CPU
  Weld Porosity Detection FP16 - CPU
  Weld Porosity Detection FP16 - CPU
7-Zip Compression
GraphicsMagick
Timed Node.js Compilation
Timed CPython Compilation
OpenVINO:
  Person Vehicle Bike Detection FP16 - CPU:
    FPS
    ms
Natron
etcd:
  RANGE - 100 - 1000 - Average Latency
  RANGE - 100 - 1000
  RANGE - 50 - 1000 - Average Latency
  RANGE - 50 - 1000
  PUT - 100 - 1000 - Average Latency
  PUT - 100 - 1000
  PUT - 50 - 1000