mnn ncnn xeon

2 x Intel Xeon Platinum 8380 testing with a Intel M50CYP2SB2U (SE5C6200.86B.0022.D08.2103221623 BIOS) and ASPEED on CentOS Stream 9 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2208133-NE-MNNNCNNXE63
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts

Limit displaying results to tests within:

HPC - High Performance Computing 2 Tests
Machine Learning 2 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
A
August 13 2022
  2 Hours, 54 Minutes
B
August 13 2022
  2 Hours, 53 Minutes
C
August 13 2022
  4 Hours, 30 Minutes
D
August 13 2022
  4 Hours, 39 Minutes
E
August 13 2022
  2 Hours, 55 Minutes
Invert Hiding All Results Option
  3 Hours, 34 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


mnn ncnn xeonOpenBenchmarking.orgPhoronix Test Suite2 x Intel Xeon Platinum 8380 @ 3.40GHz (80 Cores / 160 Threads)Intel M50CYP2SB2U (SE5C6200.86B.0022.D08.2103221623 BIOS)Intel Device 0998512GB7682GB INTEL SSDPF2KX076TZASPEEDVE2282 x Intel X710 for 10GBASE-T + 2 x Intel E810-C for QSFPCentOS Stream 95.14.0-142.el9.x86_64 (x86_64)GNOME Shell 40.10X ServerGCC 11.3.1 20220421xfs1920x1080ProcessorMotherboardChipsetMemoryDiskGraphicsMonitorNetworkOSKernelDesktopDisplay ServerCompilerFile-SystemScreen ResolutionMnn Ncnn Xeon BenchmarksSystem Logs- Transparent Huge Pages: always- --build=x86_64-redhat-linux --disable-libunwind-exceptions --enable-__cxa_atexit --enable-bootstrap --enable-cet --enable-checking=release --enable-gnu-indirect-function --enable-gnu-unique-object --enable-host-bind-now --enable-host-pie --enable-initfini-array --enable-languages=c,c++,fortran,lto --enable-link-serialization=1 --enable-multilib --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --mandir=/usr/share/man --with-arch_32=x86-64 --with-arch_64=x86-64-v2 --with-build-config=bootstrap-lto --with-gcc-major-version-only --with-linker-hash-style=gnu --with-tune=generic --without-cuda-driver --without-isl - Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0xd000363 - SELinux + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Mitigation of Clear buffers; SMT vulnerable + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected

ABCDEResult OverviewPhoronix Test Suite100%102%104%107%Mobile Neural NetworkNCNNMobile Neural NetworkNCNNNCNNNCNNNCNNNCNNMobile Neural NetworkMobile Neural NetworkNCNNNCNNNCNNNCNNNCNNNCNNMobile Neural NetworkMobile Neural NetworkNCNNMobile Neural NetworkNCNNNCNNNCNNNCNNMobileNetV2_224CPU - vgg16squeezenetv1.1CPU - efficientnet-b0CPU - alexnetCPU - mobilenetCPU - googlenetCPU - resnet18resnet-v2-50SqueezeNetV1.0CPU - squeezenet_ssdCPU - FastestDetCPU-v2-v2 - mobilenet-v2CPU - blazefaceCPU - yolov4-tinyCPU - resnet50mobilenetV3inception-v3CPU - regnety_400mmobilenet-v1-1.0CPU - vision_transformerCPU - mnasnetCPU - shufflenet-v2CPU-v3-v3 - mobilenet-v3

mnn ncnn xeonmnn: mobilenetV3mnn: squeezenetv1.1mnn: resnet-v2-50mnn: SqueezeNetV1.0mnn: MobileNetV2_224mnn: mobilenet-v1-1.0mnn: inception-v3ncnn: CPU - mobilenetncnn: CPU-v2-v2 - mobilenet-v2ncnn: CPU-v3-v3 - mobilenet-v3ncnn: CPU - shufflenet-v2ncnn: CPU - mnasnetncnn: CPU - efficientnet-b0ncnn: CPU - blazefacencnn: CPU - googlenetncnn: CPU - vgg16ncnn: CPU - resnet18ncnn: CPU - alexnetncnn: CPU - resnet50ncnn: CPU - yolov4-tinyncnn: CPU - squeezenet_ssdncnn: CPU - regnety_400mncnn: CPU - vision_transformerncnn: CPU - FastestDetABCDE1.8072.3178.7824.1213.4152.22820.69021.9112.9912.0313.4711.9216.897.1523.7430.9113.378.8924.7527.8026.3457.14151.7815.331.8372.3959.0814.2133.1512.19820.81221.6812.8112.2113.7212.0516.677.3422.5329.0513.188.5124.4927.2526.3657.76152.6314.981.8232.4258.7664.2953.2172.16720.85322.3512.7912.2513.4711.9117.757.2823.1129.6313.518.5325.1427.2827.1757.92152.7915.561.8652.4478.8864.2193.2292.21120.82922.8512.8112.1213.4511.7416.547.0623.5231.2013.879.1225.2727.3826.6656.90155.3114.921.8652.5209.1994.3213.1362.21721.34621.8513.3112.2113.6311.7916.507.2322.6428.6813.208.6224.3426.7425.9658.68155.9014.93OpenBenchmarking.org

Mobile Neural Network

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.0Model: mobilenetV3EDBCA0.41960.83921.25881.67842.098SE +/- 0.024, N = 3SE +/- 0.018, N = 15SE +/- 0.023, N = 3SE +/- 0.019, N = 14SE +/- 0.026, N = 31.8651.8651.8371.8231.807MIN: 1.79 / MAX: 2.12MIN: 1.77 / MAX: 4.16MIN: 1.78 / MAX: 1.98MIN: 1.67 / MAX: 2.18MIN: 1.72 / MAX: 4.251. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.0Model: squeezenetv1.1EDCBA0.5671.1341.7012.2682.835SE +/- 0.079, N = 3SE +/- 0.027, N = 15SE +/- 0.054, N = 14SE +/- 0.110, N = 3SE +/- 0.101, N = 32.5202.4472.4252.3952.317MIN: 2.34 / MAX: 6.7MIN: 2.3 / MAX: 3.94MIN: 2.11 / MAX: 3.4MIN: 2.17 / MAX: 4.44MIN: 2.17 / MAX: 3.611. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.0Model: resnet-v2-50EBDAC3691215SE +/- 0.027, N = 3SE +/- 0.142, N = 3SE +/- 0.044, N = 15SE +/- 0.149, N = 3SE +/- 0.057, N = 149.1999.0818.8868.7828.766MIN: 8.97 / MAX: 9.91MIN: 8.4 / MAX: 24.9MIN: 8.15 / MAX: 22.2MIN: 8.31 / MAX: 21.48MIN: 8.09 / MAX: 22.741. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.0Model: SqueezeNetV1.0ECDBA0.97221.94442.91663.88884.861SE +/- 0.116, N = 3SE +/- 0.058, N = 14SE +/- 0.039, N = 15SE +/- 0.097, N = 3SE +/- 0.164, N = 34.3214.2954.2194.2134.121MIN: 3.82 / MAX: 9.63MIN: 3.71 / MAX: 11.95MIN: 3.69 / MAX: 15.2MIN: 3.63 / MAX: 8.6MIN: 3.72 / MAX: 13.261. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.0Model: MobileNetV2_224ADCBE0.76841.53682.30523.07363.842SE +/- 0.143, N = 3SE +/- 0.047, N = 15SE +/- 0.049, N = 14SE +/- 0.076, N = 3SE +/- 0.105, N = 33.4153.2293.2173.1513.136MIN: 2.58 / MAX: 12.21MIN: 2.53 / MAX: 9.07MIN: 2.68 / MAX: 10.02MIN: 2.57 / MAX: 8.86MIN: 2.79 / MAX: 8.341. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.0Model: mobilenet-v1-1.0AEDBC0.50131.00261.50392.00522.5065SE +/- 0.021, N = 3SE +/- 0.021, N = 3SE +/- 0.013, N = 15SE +/- 0.007, N = 3SE +/- 0.028, N = 132.2282.2172.2112.1982.167MIN: 2.17 / MAX: 2.38MIN: 2.16 / MAX: 2.33MIN: 2.07 / MAX: 5.55MIN: 2.15 / MAX: 2.45MIN: 1.83 / MAX: 2.371. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.0Model: inception-v3ECDBA510152025SE +/- 0.10, N = 3SE +/- 0.12, N = 14SE +/- 0.09, N = 15SE +/- 0.18, N = 3SE +/- 0.44, N = 321.3520.8520.8320.8120.69MIN: 18.43 / MAX: 39.41MIN: 19.34 / MAX: 46.76MIN: 19.44 / MAX: 41.2MIN: 19.89 / MAX: 32.91MIN: 18.2 / MAX: 41.461. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: mobilenetDCAEB510152025SE +/- 0.10, N = 3SE +/- 0.16, N = 3SE +/- 0.29, N = 3SE +/- 0.16, N = 3SE +/- 0.02, N = 322.8522.3521.9121.8521.68MIN: 21.59 / MAX: 246.4MIN: 21.7 / MAX: 47.14MIN: 21.02 / MAX: 61.03MIN: 21.29 / MAX: 94.26MIN: 21.29 / MAX: 45.511. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU-v2-v2 - Model: mobilenet-v2EADBC3691215SE +/- 0.30, N = 3SE +/- 0.14, N = 3SE +/- 0.09, N = 3SE +/- 0.04, N = 3SE +/- 0.05, N = 313.3112.9912.8112.8112.79MIN: 12.14 / MAX: 236.38MIN: 12.1 / MAX: 146.34MIN: 12.34 / MAX: 89.44MIN: 12.31 / MAX: 37.43MIN: 12.42 / MAX: 18.331. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU-v3-v3 - Model: mobilenet-v3CEBDA3691215SE +/- 0.06, N = 3SE +/- 0.16, N = 3SE +/- 0.20, N = 3SE +/- 0.03, N = 3SE +/- 0.14, N = 312.2512.2112.2112.1212.03MIN: 11.87 / MAX: 36.44MIN: 11.67 / MAX: 35.66MIN: 11.64 / MAX: 151.41MIN: 11.84 / MAX: 34.59MIN: 11.55 / MAX: 35.771. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: shufflenet-v2BECAD48121620SE +/- 0.31, N = 3SE +/- 0.26, N = 3SE +/- 0.16, N = 3SE +/- 0.10, N = 3SE +/- 0.10, N = 313.7213.6313.4713.4713.45MIN: 13.1 / MAX: 82.58MIN: 12.59 / MAX: 105.4MIN: 12.86 / MAX: 37.29MIN: 12.89 / MAX: 36.76MIN: 12.63 / MAX: 79.261. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: mnasnetBACED3691215SE +/- 0.16, N = 3SE +/- 0.10, N = 3SE +/- 0.10, N = 3SE +/- 0.08, N = 3SE +/- 0.13, N = 312.0511.9211.9111.7911.74MIN: 11.5 / MAX: 157.89MIN: 11.57 / MAX: 36.77MIN: 11.32 / MAX: 35.73MIN: 11.19 / MAX: 63.29MIN: 11.27 / MAX: 17.511. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: efficientnet-b0CABDE48121620SE +/- 0.73, N = 3SE +/- 0.14, N = 3SE +/- 0.14, N = 3SE +/- 0.29, N = 3SE +/- 0.30, N = 317.7516.8916.6716.5416.50MIN: 15.75 / MAX: 590.87MIN: 16.1 / MAX: 73.41MIN: 15.93 / MAX: 60.85MIN: 15.61 / MAX: 86.38MIN: 15.68 / MAX: 40.871. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: blazefaceBCEAD246810SE +/- 0.20, N = 3SE +/- 0.11, N = 3SE +/- 0.05, N = 3SE +/- 0.08, N = 3SE +/- 0.12, N = 37.347.287.237.157.06MIN: 6.93 / MAX: 121.89MIN: 6.88 / MAX: 9.96MIN: 6.99 / MAX: 10.09MIN: 6.83 / MAX: 10.24MIN: 6.68 / MAX: 8.271. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: googlenetADCEB612182430SE +/- 0.56, N = 3SE +/- 0.45, N = 3SE +/- 0.48, N = 3SE +/- 0.29, N = 3SE +/- 0.21, N = 323.7423.5223.1122.6422.53MIN: 21.23 / MAX: 403.67MIN: 22.04 / MAX: 299.28MIN: 21.75 / MAX: 82.56MIN: 21.47 / MAX: 240.52MIN: 21.73 / MAX: 94.831. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: vgg16DACBE714212835SE +/- 0.60, N = 3SE +/- 1.59, N = 3SE +/- 0.23, N = 3SE +/- 0.07, N = 3SE +/- 0.04, N = 331.2030.9129.6329.0528.68MIN: 28.77 / MAX: 277.95MIN: 26.06 / MAX: 119.07MIN: 27.23 / MAX: 201.62MIN: 27.32 / MAX: 149.49MIN: 26.99 / MAX: 114.831. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: resnet18DCAEB48121620SE +/- 0.61, N = 3SE +/- 0.15, N = 3SE +/- 0.24, N = 3SE +/- 0.12, N = 3SE +/- 0.11, N = 313.8713.5113.3713.2013.18MIN: 12.86 / MAX: 18.89MIN: 12.95 / MAX: 158.24MIN: 12.55 / MAX: 19.07MIN: 12.71 / MAX: 82.75MIN: 12.67 / MAX: 65.171. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: alexnetDAECB3691215SE +/- 0.39, N = 3SE +/- 0.36, N = 3SE +/- 0.37, N = 3SE +/- 0.08, N = 3SE +/- 0.16, N = 39.128.898.628.538.51MIN: 8.11 / MAX: 130.97MIN: 7.96 / MAX: 119.99MIN: 7.91 / MAX: 231.7MIN: 8.13 / MAX: 49.08MIN: 8.08 / MAX: 100.341. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: resnet50DCABE612182430SE +/- 0.11, N = 3SE +/- 0.19, N = 3SE +/- 0.66, N = 3SE +/- 0.30, N = 3SE +/- 0.11, N = 325.2725.1424.7524.4924.34MIN: 23.98 / MAX: 190.68MIN: 23.76 / MAX: 106.76MIN: 23.11 / MAX: 123.68MIN: 23.44 / MAX: 111.14MIN: 23.27 / MAX: 121.621. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: yolov4-tinyADCBE714212835SE +/- 0.43, N = 3SE +/- 0.26, N = 3SE +/- 0.03, N = 3SE +/- 0.25, N = 3SE +/- 0.19, N = 327.8027.3827.2827.2526.74MIN: 25.29 / MAX: 281.03MIN: 25.63 / MAX: 358.69MIN: 26.44 / MAX: 51.63MIN: 25.94 / MAX: 352.26MIN: 25.94 / MAX: 113.751. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: squeezenet_ssdCDBAE612182430SE +/- 0.43, N = 3SE +/- 0.27, N = 3SE +/- 0.14, N = 3SE +/- 0.14, N = 3SE +/- 0.13, N = 327.1726.6626.3626.3425.96MIN: 25.76 / MAX: 481.76MIN: 24.96 / MAX: 198.39MIN: 25.39 / MAX: 197.69MIN: 24.99 / MAX: 249.3MIN: 25.3 / MAX: 49.751. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: regnety_400mECBAD1326395265SE +/- 0.57, N = 3SE +/- 0.98, N = 3SE +/- 0.73, N = 3SE +/- 1.60, N = 3SE +/- 1.22, N = 358.6857.9257.7657.1456.90MIN: 54.21 / MAX: 808.52MIN: 55.21 / MAX: 203.4MIN: 55.37 / MAX: 447.81MIN: 53.37 / MAX: 430.33MIN: 53.28 / MAX: 385.571. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: vision_transformerEDCBA306090120150SE +/- 0.33, N = 3SE +/- 0.39, N = 3SE +/- 0.64, N = 3SE +/- 0.18, N = 3SE +/- 1.55, N = 3155.90155.31152.79152.63151.78MIN: 146.82 / MAX: 812.77MIN: 145.23 / MAX: 1014.26MIN: 145.98 / MAX: 797.33MIN: 147.33 / MAX: 369.79MIN: 145.5 / MAX: 656.751. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: FastestDetCABED48121620SE +/- 0.45, N = 3SE +/- 0.14, N = 3SE +/- 0.20, N = 3SE +/- 0.09, N = 3SE +/- 0.22, N = 315.5615.3314.9814.9314.92MIN: 14.29 / MAX: 72.49MIN: 14.81 / MAX: 39.13MIN: 14.48 / MAX: 39.53MIN: 14.47 / MAX: 37.58MIN: 14.41 / MAX: 20.231. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread